Monday, May 6, 2019

How is Entropy and Information Gain Theory used in Coding Theory Research Paper

How is Entropy and info Gain Theory used in Coding Theory - Research Paper ExampleIn a communication arrangement, two firsts of coding are evident, the first one is the source coding, which focuses on faculty readiness of digital presentation from source signal. Secondly, we nurture the channel coding mainly the error-control coding which is for the provision of reliable communications through noisy channels (Ashikhmin, Barg & Dimacs 47) In coding supposition, mho and info pip theory offers various roles. Firstly, it is concerned with data reduction, which happens mostly studying a given task and plan of attack across extra material called side information. This will result to the need of data reduction. Therefore, appliance of second and information will be of pronounced meaning because it will lessen the extra information. This theory displace also interpret the extra information to represent context or situation information (Roth 6). Secondly, entropy and information t heory concerns the determination of indecision that is allied with the given information. For example, if definite specific information is on conduction and this theory happens to known it before the transmission of that material, it will lead to the failure of that information passing through transmission. Entropy focuses on maximization during the equiprobable of meanings thus determining vagueness (Roth 7). Thirdly, this theory promotes intelligence and performance of secrecy to information. These concepts mostly apply to cryptography compounded with cryptanalysis (Roth 7). It focuses on the redundancy of plaintext by cock-a-hoop the least quantity of ciphertext ensuring exceptional decipherability. Here information theory makes us be certain that it is lumbering to keep any secret (Golomb, Peile & Scholtz 202). Fourthly, entropy and information gain theory deals with the gathering of unpolluted disciplines, which have already gone through investigation and transformed to en gineering practice. It deals with very broad applications thus the life force of coding theory. Fifthly, this theory is important when it comes to error-correcting codes in computers with high-speed memories (Cover & Thomas 13). These codes are of vital use when it comes to enhancing the dependability of computer memories. Here the computers contain unusual features that are rarely in communication applications. These errors are callable to encoding, decoding, and uncommon type of errors. When this occurs, the entropy and information gain theory are able to detect triplex error appearing at the same time, correcting the single errors (Cover & Thomas 13). Source coding theory is closely well-organized demonstration of given data that are from a certain information source (Gray 34). For example when it is an image-coding, achieving source coding can be through manipulating terminations of that image. To attain noiseless source coding, the measures of given information and its com plexity should be observed. Central to this, the entropy and information theory is of application when it comes to detection of total information. This theory also helps when it comes to arithmetic coding compounded with statistical modeling (Kannappan 174). Arithmetic coding avoids assigning certain bit of given patterns to the original source symbol. In entropy and information theory, a connotation with a code is given and it concludes the order of symbol. These code words have sub intervals showing the disparity

No comments:

Post a Comment

Note: Only a member of this blog may post a comment.