**Commenced**in January 2007

**Frequency:**Monthly

**Edition:**International

**Paper Count:**30855

##### Codes and Formulation of Appropriate Constraints via Entropy Measures

**Authors:**
R. K. Tuli

**Abstract:**

**Keywords:**
Codeword,
Instantaneous code,
Prefix code,
Uniquely decipherable code,
Best one-one code,
Mean codewordlength

**Digital Object Identifier (DOI):**
doi.org/10.5281/zenodo.1329330

**References:**

[1] Campbell, L. L. (1965): "A coding theorem and Renyi's entropy", Information and Control, 8, 423-429.

[2] Cheng, J. and Huang, T. K. (2006): "New lower and upper bounds on the expected length of optimal one-to-one codes", Proceedings of the Data Compression Conference, 43-52.

[3] Cheng, J., Huang, T. K. and Weidmann, C. (2007): "New bounds on the expected length of optimal one-to-one codes", IEEE Trans. Inform. Theory, 53(5), 1884-1895.

[4] Feinstein, A. (1958): "Foundations of Information Theory", McGraw- Hill, New York.

[5] Kapur, J.N. (1995): "Measures of Information and Their Applications", Wiley Eastern, New York.

[6] Kraft, L. G. (1949): "A Device for Quantizing Grouping and Coding Amplitude Modulated Pulses", M.S. Thesis, Electrical Engineering Department, MIT.

[7] Leung-Yan-Cheong, S. K. and Cover, T. M. (1978): "Some equivalences between Shannon entropy and Kolmogrov complexity", IEEE Trans. Inform. Theory, 24, 331-338.

[8] Renyi, A. (1961): "On measures of entropy and information", Proceedings 4th Berkeley Symposium on Mathematical Statistics and Probability, 1, 547-561.

[9] Rissanen, J. (1992): "Tight lower bounds for optimal code length", IEEE Trans. Inform. Theory, 28(2), 348-349.

[10] Savari, S. A. and Naheta, A. (2004): "Bounds on the expected cost of one-to-one codes", Proc. IEEE Int. Symp. Information Theory, 94.

[11] Shannon, C. E. (1948): "A mathematical theory of communication", Bell System Tech J., 27, 379-423.