Basic Lecture Information Theory and Coding
Basic Lecture Information Theory and Coding
Dr. M. Arif Wahla EE Dept arif@mcs.edu.pk Military College of Signals National University of Sciences & Technology (NUST), Pakistan
Information Theory
Founded by Claude E. Shannon (1916-2001) The Mathematical Theory of Communication, 1948 Study fundamental limits in communications: transmission, storage,
etc
1:31 AM
Course Outline
1:31 AM
Course Outline
variables Information is digital: transmission should be 0s and 1s (bits) with no reference to what they represent
1:31 AM
Course Outline
fundamental limit in data compression (zip, MP3, JPEG, MPEG) Channel coding theorem fundamental limit for reliable communication through a noisy channel (telephone, cell phone, modem, data storage, etc)
1:31 AM
Course Outline
The fundamental problem of communication is that of reproducing at one point either exactly or approximately a message selected at another point.
1:31 AM
Course Outline
Course Outline
1:31 AM
Course Outline
classical information theory and coding theory. The main course objective is to introduce the students to wellknown information theoretic tools that can be used to solve engineering problems.
The course will begin by describing basic communication
systems problems where information theory may be applied. An explanation of information measurement and characterization will be given. Fundamentals of noiseless source coding and noisy channel coding will be taught next. Finally, some key information theory principles applied to communication security systems will be covered.
1:31 AM
Course Outline
Course Outline -I
Information theory is concerned with the fundamental limits of communication.
What is the ultimate limit to data compression? e.g. how many bits
channel, e.g. how many bits can be sent in one second over a telephone line.
1:31 AM
Course Outline
1:31 AM
Course Outline
10
Information Measurement
Definition and Properties of Entropy Uniqueness of the Entropy Measure Joint and Conditional Entropy Mutual Information and Conditional Mutual Information Information Divergence Measures
1:31 AM
Course Outline
11
1:31 AM
Course Outline
12
1:31 AM
Course Outline
13
A Mathematical Theory of Communication, Claude E. Shannon, Bell System Technical Journal, 1948 available for free on line
1:31 AM
Course Outline
14
Lecture Notes, Reading material and Assignments will be posted here. 1. Sessional marks and exam results will be uploaded 2. Students are encouraged to maintain a discussion blog and discuss the assignments and course topics
1:31 AM
Course Outline
15
A Mathematical Theory of Communication, Claude E. Shannon, Bell System Technical Journal, 1948 available for free on line
1:31 AM
Course Outline
16
Schedule
Class Meetings Wednesday (5pm-8pm) 3L
Consultancy Hours Wednesday (4pm-5pm), (8pm-8:30pm) Other times by appointment (phone or email)
1:31 AM
17
Teaching Methodology
Organized material will be presented in PPT slides. Concepts, mathematical expressions and examples will be performed on board
1:31 AM
18
will have their assignment and quiz marks cancelled OR marks would be shared by the group having similar solution
OHT 1&2 [30%] Exam during 7th & 11th week
Course Outline
19
1:31 AM
Course Outline
21
1:31 AM
22
famous)?
force is efficiency of expression, there is an interpolation of many feelings,
attitudes and perceptions; there is an efficient encoding of emotional and mental information.
1:31 AM
23
Information Theory
Two important questions in engineering: - What to do if information gets corrupted by errors?
1:31 AM
25
Problems in Communications
Speed Minimise length of transmitted data Accuracy Minimise and eliminate noise Security Ensure data is not changed or intercepted whilst in transit
1:31 AM
26
Solutions
Speed Minimise length of transmitted data Use Data Compression Accuracy Minimise and eliminate noise Use Error Detection / Correction Codes Security Ensure data is not changed or intercepted whilst in transit Use Data Encryption / Authentication
1:31 AM
27
Communications Model
Source data
Evesdropper
Destination data
signal Transmitter
received signal
Receiver
noise
1:31 AM
28
Data Compression
This is the study of encoding information so that it may be
1:31 AM
29
made due to noise or other impairments in the course of the transmission from the transmitter to the receiver.
Error correction has the additional feature that enables
Code, Reed-Solomon Code, Convolutional Codes, Turbo Codes and LDPC Codes
1:31 AM
30
Transmitter
Receiver
noise
Source Coding
Channel Coding
Compression
Encryption
Modulation
1:31 AM
31
What is information?
information: [m-w.org]
1: the communication or reception of knowledge or intelligence
1:31 AM
32
of messages to be communicated to a destination or receiver On a finer granularity, an information source produces symbols to be communicated to the destination In this lecture, we will focus on discrete sources
i.e., sources that produce discrete symbols from a predefined
alphabet
1:31 AM
33
should have more information For instance, consider a source, say S1, that wants to communicate its direction to a destination using the following symbols:
North (N), South (S), East (E), West (W)
using:
North (N), South (S), East (E), West (W), Northwest (NW),
34
1:31 AM
35
represent n symbols
2 symbols: 0, 1
4 symbols: 00, 01, 10, 11 8 symbols: 000, 001, 010, 011, 100, 101, 110, 111
1:31 AM
36
If 1000 symbols are generated by X, how many bits are required to transmit these 1000 symbols?
1:31 AM
37
If 1000 symbols are generated by X, how many bits are required to transmit these 1000 symbols?
1:31 AM
38
What is the minimum number of bits/symbol required to communicate an information source having n symbols? (n=4 in present example)
In fact, lets rephrase the question as:
Are 2 bits/symbol the minimum number of bits/symbol required to communicate an information source having n=4 symbols?
1:31 AM
39
1:31 AM
40
1:31 AM
41
Let us map the symbols to the following bit sequences: N: 0 S: 01 E: 011 W: 0111
1:31 AM
42
N: 0 S: 01 E: 011 W: 0111
0.3 0.05 N S E W X
Now if 1000 symbols are generated by X, how many bits are required to transmit these 1000 symbols?
1:31 AM
43
N: 0 S: 01 E: 011 W: 0111
0.3 0.05 N S E W X
Now if 1000 symbols are generated by X, how many bits are required to transmit these 1000 symbols?
On average, the 1000 symbols will have: 600 Ns, 300 Ss, 50 Es and 50 Ws
1:31 AM
44
Now if 1000 symbols are generated by X, how many bits are required to transmit these 1000 symbols? 600 Ns, 300 Ss, 50 Es and 50 Ws Total bits=6001+3002+503+504=1550
45
46
Are 1.55 bits/symbol the minimum number of bits/symbol required to communicate an information source having n=4 symbols?
1:31 AM
47
1:31 AM
48
communicate the symbols of a source is the information content of the source How to find a code that can provide the minimum information is a different question However, we can quantify the information of a source without knowing the code(s) that can achieve this minimum In this lecture, we will refer to the minimum number of bits/symbol of a source as the information content of the source
1:31 AM
49
1:31 AM
50
N: 0111 S: 011 E: 01 W: 0
0.3 0.05 N S E W X
Now if 1000 symbols are generated by X, how many bits are required to transmit these 1000 symbols?
Total bits=6004+3003+502+501=3450 3.45 bits/symbol
These are more bits than we would need if we assumed all symbols to be equally likely
1:31 AM
51
likely
1:31 AM
52
able to communicate all the symbols of the source using log2(n) bits/symbol
In other words, this is the maximum number of bits
1:31 AM
53
1:31 AM
54
likely (i.e., non-uniform source), log2(n) represents the maximum number of bits/symbol Among all discrete sources producing a given number of n symbols, a uniform source has the highest information content
Introduction to Information Theory
1:31 AM
55
1:31 AM
56
1:31 AM
57
source with a source having the symbols (North, South, East, West, Northwest, Northeast, Southeast, Southwest)
1:31 AM
58
symbols, the source with the maximum number of symbols has the maximum information content
In other words, for equally likely sources, a function H(.)
that quantifies information content of a source should be an increasing function of the number of symbols
Lets call this function H(n)
1:31 AM
59
H(n) = log2(n)
1:31 AM
60
English language
1:31 AM
61
language:
1:31 AM
62
likely ones, the total number of bits required by a code having the above properties will be less than log2(n)
1:31 AM
63
1:31 AM
64
increase as pX decreases
The following function has been proven to provide the right
1:31 AM
65
symbol is given by: H(pX=i)=log2(1/pX=i) So what is the expected or average value of the information content of all the symbols of pX?
1:31 AM
66
1:31 AM
67
This is called the entropy of the source and represents the minimum expected number of bits/symbol required to communicate this source
Introduction to Information Theory
68
1:31 AM
pX 1/n
1:31 AM
69
Note that this is the same function that we had deduced earlier
1:31 AM
70