Applied Mathematical Sciences, Vol. 8, 2014, no. 138, 6881 - 6888 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.47569 On Bounds of a Generalized ‘Useful’ Mean Codeword Length Sonali Saxena* *Post Doctoral Fellow of National Board of Higher Mathematics(NBHM)of Department of Atomic Energy(DAE),India Keerti Upadhyay Jaypee University of Engineering and Technology A.B. Road, Raghogarh – 473226 Dist. Guna – M.P., India D.S. Hooda Jaypee University of Engineering and Technology A.B. Road, Raghogarh – 473226 Dist. Guna – M.P., India Copyright © 2014 Sonali Saxena, Keerti Upadhyay and D. S. Hooda. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Abstract In the present paper a new generalized ‘useful’ mean codeword length is defined and characterized a new generalized information measure by obtaining bounds ofa new generalized ‘useful’ mean codeword length in terms of a new generalized information measure using Lagrange’s Multiplier method. The Shannon’s Noiseless coding theorem is verified by considering Huffman and Shannon -Fano coding schemes on taking empirical data.. Mathematics Subject Classification: 94A15, 94A17 Keywords: Kraft’s inequality, Utility distribution, Lagrange’s Multiplier method, Huffman Codes, Shannon- Fano Codes, Shannon’s Noiseless coding theorem 6882 Sonali Saxena, Keerti Upadhyay and D. S. Hooda 1. Introduction Information theory has ideas that are widely applicable to situations remote from its original inspiration. The applicability of ideas is not exact; however, they are very useful. One of the best applications of information measure is in noiseless coding theorem which gives the bounds for suitable encoding of information in terms of information measure. Let a finite set of n source symbols X x1 , x 2 , , x n be encoded using alphabet of D 2 symbols; then it has shown Feinstein [4] that there is a uniquely decipherable/ instantaneous code with lengths l1 , l 2 , , l n if and only if the following Kraft’s inequality [9] is satisfied: n D li 1 (1.1) i 1 n If L pi li be the average codeword length then for a code which satisfies (1.1), i 1 it has been shown that L H P , (1.2) with equality if and only if li log pi , for i 1,2,, n . This is Shannon’s noiseless coding theorem for a noiseless channel. Shannon-Fano coding is less efficient than Huffman coding, but we have an advantage that we can go directly from the probability pi to the codeword length li . Belis and Guiasu [2] observed that a source is not completely specified by the Probability distribution P over the source alphabet X, in the absence of qualitative character. So it can be assumed that the source alphabet letters are weighted according to their importance or utilities. Let U u1 , u 2 , , u n be the set of positive real numbers, where u i is the utility of the outcome x i having probability p i .The utility u i in general independent of p i , the probability of encoding of source symbol x i . The information source is thus given by x1 x2 xn S p1 p 2 p n , ui 0, 0 pi 1 for each i and u1 u 2 u n n p i 1 i 1 (1.3) Belis and Guiasu [2] introduced the “quantitative – qualitative” measure of infor- On bounds of a generalized ‘useful’ mean codeword length 6883 mation n H P; U u i p i log p i , (1.4) i 1 which can be taken as a measure for the average quantity of ‘valuable’ or ‘useful’ information provided by the information source (1.3). Guiasu and Picard [5] considered the problem of encoding the letter output by the source (1.3) by means of a single letter prefix code whose codeword w1 , w2 , , wn are of lengths l1 , l 2 , , l n , respectively and satisfy the Kraft’s inequality (1.1). They introduced the following ‘useful’ mean length of code by attaching utility distribution with probability distribution of a set of source symbols: n Lu u i 1 n i u j 1 pi li . j (1.5) pj Further they derived a lower bound for (1.5). However Longo [10] interpreted (1.5) as the average transmission cost of letters xi and derived the bounds for this cost function. Campbell [3] introduced the exponentiated mean codeword length given by 1 n LUD t log pi Dtli . (1.6) t i 1 Where LUD average codeword length for the uniquely decodable code, D is represents the size of the alphabet and li is the codeword length associated with xi of X . He proved the following noiseless coding theorem H X LUD t H X 1 , Under the condition n D li 1, (1.7) (1.8) i 1 where H X is the Renyi entropy of order 1 and li is the codeword 1 t length corresponding to source symbol xi . Taneja et al. [12] defined the ‘useful’ average code lengths of order t as given below: n tli u i pi D 1 , 1 or t 1 0 t . Lu t log D i 1 n (1.9) t t 1 u j p j j 1 6884 Sonali Saxena, Keerti Upadhyay and D. S. Hooda Evidently, when t 0 , (1.9) reduces to (1.5). They derived the bounds for the cost function (1.9) in terms of a generalized ‘useful’ information measure. Hooda and Bhaker [7] have studied one of its generalizations in terms of and given as below: 1 n li ui pi D . (1.10) L P;U log D i 1 n 1 uj pj j 1 They have studied the lower and upper bounds of codeword length (1.7) in term of generalized measure. Aczel and Daroczy [1] generalized entropy of order and type is given by: n 1 pi 1 i 1 . (1.11) H P log D 1 n pj j 1 It may be noted (1.11) was also characterized by Kapur [8] following different method. In the present paper we define a new generalized ‘useful’ mean codeword length andobtain the generalized ‘useful’ measure of information by using Lagrange’s multiplier method in section 2. In section3 we verify the Shannon’s noiseless coding theorem on average codeword lengthin cases of Huffman and Shannon-Fano coding schemes on taking empirical data. 2. A Generalized ‘Useful’ Mean Code Word Length and its Bounds In this section we define a new generalized ‘useful’ mean codeword length as follows: n li ui pi D , 0 1, 0 1, . (2.1) L P;U log D i 1 n ui pi i 1 where l i is the length of codeword xi and pi is the probability of occurrence of codeword xi . Theorem 2.1For all uniquely decipherable codes generalized ‘useful’mean codeword length L P;U defined in (2.1) satisfies the following relation On bounds of a generalized ‘useful’ mean codeword length H P;U L P;U H P;U 1, 6885 (2.2) 2 2 u p i i where H P;U log D , 0 1, 0 1, , (2.3) ui pi 2 under the generalized Kraft’s inequality given by 1 u i p i D li u i p i . (2.4) u i p i 1 D li , for each i 1,2, , n u i pi Substituting (2.5) in (2.1) we have 2 ui pi xi L P;U log D u p i i Thus we have to minimize (2.6) subject to the following constraint: u i pi 1 D li 1 x i u i pi Proof:Let us choose xi Since L P;U is a pseudo convex function for each (2.5) (2.6) (2.7) i 1,2,, n , therefore, we can obtain the minimum value of L P;U by applying the Lagrange’s multiplier method. Let us consider the corresponding Lagrangian as given below: 2 u i p i xi n L log xi 1 . u i pi i 1 Differentiating w.r.t. x i and equating to zero, we get dL u p x 1 i i i 0 u i pi dx i 1 It implies up 1 xi c i i , where c 0 u p i ii or xi cq i , where q i u i pi .(2.8) (2.8) together with (2.5) gives u i pi 6886 Sonali Saxena, Keerti Upadhyay and D. S. Hooda u i p i 1 D li u p i i u i pi u i pi It implies D li pi2 Taking log of both sides, we have l i log pi2 or li log D pi 2 (2.9) Multiplying both sides of (2.9) by 0 as we get 2 l i log D p i (2.10) From (2.1) and (2.10), we get the minimum value of L ( P;U ) as follows: 2 2 ui pi L ( P;U ) min log H ( P;U ) . u p i i again, since l i is always integral value in (2.9), so it must be equal to li ai i 2 where ai log D pi 2 and 0 i 1 . Putting (2.12) in (2.1), we have 2 i u p p D i i i L P;U log D ui pi 2 2 2 u i pi log i u i pi Since 0 i 1, therefore, (2.13) reduce to 2 2 u p i i 1 H P;U 1 . L P;U log ui pi Hence from (2.11) and (2.14), we get H P;U L P;U H P;U 1, which is (2.2). (2.11) (2.12) (2.13) 2 (2.14) On bounds of a generalized ‘useful’ mean codeword length 6887 Hence by using optimization technique we get the new generalized‘useful’ information measure given by (2.3). 3. Illustration In this section we illustrate the veracity of the theorem 2.1 by taking empirical data as given in table (3.1) and (3.2) on the lines of Hooda et al.[6]. Table- 3.1 Huffman codeword s li ui 0.3846 0.1795 0.1538 0 100 101 1 3 3 1 3 2 0.1538 110 3 2 0.1282 111 3 4 pi 0.2 0.3 H P;U H P;U L P;U L P;U 100% 6.0157 6.3764 94.34% Table -3.2 pi Shannon Fanocodew ords li ui 0.3846 00 2 1 0.2 0.3 0.1795 01 2 3 0.1538 10 2 2 0.1538 110 3 2 0.1282 111 3 4 H P;U 6.0157 L P;U 6.5516 H P;U L P;U 91.82% From table (3.1) and (3.2) we infer the following: 1. Theorem 2.1 holds in both cases of Shannon -Fano codes and Huffman codes. 2. Huffman mean codeword length is less than Shannon –Fano mean codeword length. 3. Coefficient of efficiency of Huffman Codes is greater than Shannon -Fano Codes i.e. it is concluded that Huffman coding Scheme is more efficient than Shannon -Fano coding scheme 100% 6888 Sonali Saxena, Keerti Upadhyay and D. S. Hooda Conclusion The various authors have characterized the generalized information measures by different methods, but we have introduced a new generalized ‘useful’ mean codeword length and studied its bounds in terms of the new generalized measure of information by optimization technique. Further, we have established the Shannon’s Noiseless Coding theorem with the help of two different coding techniques by taking experimental data and prove that Huffman coding scheme is more efficient than Shannon-Fano coding scheme. References [1] J.Aczel and Z. Daroczy, Uber Veallegemeineste Quasilinear Mittelwerte, Die Mitgrewincbtsfunktionen Gebilelet,Sind. Public Mathematics Debrecen,10(1963), 171-190. [2] M.Belis and S. Guiasu, A Quantitative-Qualitative Measure of Information in Cybernetic Systems,IEEE Transfer Information Theory,14(1968), 593-594. [3]L.L. Campbell, A coding theorem and Renyi’s Entropy, Information and control 8(1965) 423-429 [4] A.Feinstein, Foundation of Information Theory, Mc Grew-Hill, New York(1958). [5] S.Guiasu and C.F. Picard, Borne Inferieure De La Longueur De Certain Codes,C.R. Academy Science Paris, 273(1971), 248-251. [6]D.S.Hooda, Keerti Upadhyay and Sonali Saxena, Characterization of a Generalized Information Measure by Optimization Technique, International Journal of Engineering Research and Management Technology, I(2014), 302-311. [7] D.S.Hooda and U.S.Bhaker, A Profile on Noiseless Coding Theorem, International Journal of Management and Systems, 8(1992), 76-85. [8] J.Kapur, Generalized Entropy of Order and Type , Math.Seminar (Delhi),4(1967). [9] L.G.Kraft, A Device for Quantizing, Grouping and Coding Amplitude Modulated Pulses, M.S.Thesis, Electrical Engineering Department, MIT, (1949). [10] G.Longo,A Noiseless Coding Theorem for Sources having Utilities,SIAM Journal Applied Mathematics,30 (1972), 739-748. [11]C.E.Shannon, A Mathematical Theory of Communication, Bell System Technology Journal27 (1948), 379-423. [12]H.C.Taneja, D.S. Hooda and R.K.Tuteja,Coding Theorem on Generalized ‘useful’ Information, Soochow Journal of mathematics, 11 (1985), 123-131. Received: July 11, 2014
© Copyright 2025 Paperzz