DATA COMPRESSION AND ITS TYPES
Data Compression, also known as source coding, is the process of encoding or converting data in such a way that it consumes less memory space. Data compression reduces the number of resources required to store and transmit data.
It can be done in two ways lossless compression and lossy compression. Lossy compression reduces the size of data by removing unnecessary information, while there is no data loss in lossless compression.
WHAT IS SHANNON FANO CODING?
Shannon Fano Algorithm is an entropy encoding technique for lossless data compression of multimedia. Named after Claude Shannon and Robert Fano, it assigns a code to each symbol based on their probabilities of occurrence. It is a variable length encoding scheme, that is, the codes assigned to the symbols will be of varying length.
HOW DOES IT WORK?
The steps of the algorithm are as follows:
 Create a list of probabilities or frequency counts for the given set of symbols so that the relative frequency of occurrence of each symbol is known.
 Sort the list of symbols in decreasing order of probability, the most probable ones to the left and least probable to the right.
 Split the list into two parts, with the total probability of both the parts being as close to each other as possible.
 Assign the value 0 to the left part and 1 to the right part.
 Repeat the steps 3 and 4 for each part, until all the symbols are split into individual subgroups.
The Shannon codes are considered accurate if the code of each symbol is unique.
EXAMPLE:
Given task is to construct Shannon codes for the given set of symbols using the ShannonFano lossless compression technique.
Step:
Tree:
Solution:

Let P(x) be the probability of occurrence of symbol x:
 Upon arranging the symbols in decreasing order of probability:
P(D) + P(B) = 0.30 + 0.2 = 0.58
and,
P(A) + P(C) + P(E) = 0.22 + 0.15 + 0.05 = 0.42
And since thealmost equally split the table, the most is dividedit the blockquote table isblockquotento
{D, B} and {A, C, E}
and assign them the values 0 and 1 respectively.
Step:
Tree:
 Now, in {D, B} group,
P(D) = 0.30 and P(B) = 0.28
which means that P(D)~P(B), so divide {D, B} into {D} and {B} and assign 0 to D and 1 to B.
Step:
Tree:
 In {A, C, E} group,
P(A) = 0.22 and P(C) + P(E) = 0.20
So the group is divided into
{A} and {C, E}
and they are assigned values 0 and 1 respectively.
 In {C, E} group,
P(C) = 0.15 and P(E) = 0.05
So divide them into {C} and {E} and assign 0 to {C} and 1 to {E}
Step:
Tree:
Note: The splitting is now stopped as each symbol is separated now.
The Shannon codes for the set of symbols are:
As it can be seen, these are all unique and of varying lengths.
Below is the implementation of the above approach:
// C++ program for Shannon Fano Algorithm // include header files #include <bits/stdc++.h> using namespace std; // declare structure node struct node { // for storing symbol string sym; // for storing probability or frquency float pro; int arr[20]; int top; } p[20]; typedef struct node node; // function to find shannon code void shannon( int l, int h, node p[]) { float pack1 = 0, pack2 = 0, diff1 = 0, diff2 = 0; int i, d, k, j; if ((l + 1) == h  l == h  l > h) { if (l == h  l > h) return ; p[h].arr[++(p[h].top)] = 0; p[l].arr[++(p[l].top)] = 1; return ; } else { for (i = l; i <= h  1; i++) pack1 = pack1 + p[i].pro; pack2 = pack2 + p[h].pro; diff1 = pack1  pack2; if (diff1 < 0) diff1 = diff1 * 1; j = 2; while (j != h  l + 1) { k = h  j; pack1 = pack2 = 0; for (i = l; i <= k; i++) pack1 = pack1 + p[i].pro; for (i = h; i > k; i) pack2 = pack2 + p[i].pro; diff2 = pack1  pack2; if (diff2 < 0) diff2 = diff2 * 1; if (diff2 >= diff1) break ; diff1 = diff2; j++; } k++; for (i = l; i <= k; i++) p[i].arr[++(p[i].top)] = 1; for (i = k + 1; i <= h; i++) p[i].arr[++(p[i].top)] = 0; // Invoke shannon function shannon(l, k, p); shannon(k + 1, h, p); } } // Function to sort the symbols // based on their probability or frequency void sortByProbability( int n, node p[]) { int i, j; node temp; for (j = 1; j <= n  1; j++) { for (i = 0; i < n  1; i++) { if ((p[i].pro) > (p[i + 1].pro)) { temp.pro = p[i].pro; temp.sym = p[i].sym; p[i].pro = p[i + 1].pro; p[i].sym = p[i + 1].sym; p[i + 1].pro = temp.pro; p[i + 1].sym = temp.sym; } } } } // function to display shannon codes void display( int n, node p[]) { int i, j; cout << "\n\n\n\tSymbol\tProbability\tCode" ; for (i = n  1; i >= 0; i) { cout << "\n\t" << p[i].sym << "\t\t" << p[i].pro << "\t" ; for (j = 0; j <= p[i].top; j++) cout << p[i].arr[j]; } } // Driver code int main() { int n, i, j; float total = 0; string ch; node temp; // Input number of symbols cout << "Enter number of symbols\t: " ; n = 5; cout << n << endl; // Input symbols for (i = 0; i < n; i++) { cout << "Enter symbol " << i + 1 << " : " ; ch = ( char )(65 + i); cout << ch << endl; // Insert the symbol to node p[i].sym += ch; } // Input probability of symbols float x[] = { 0.22, 0.28, 0.15, 0.30, 0.05 }; for (i = 0; i < n; i++) { cout << "\nEnter probability of " << p[i].sym << " : " ; cout << x[i] << endl; // Insert the value to node p[i].pro = x[i]; total = total + p[i].pro; // checking max probability if (total > 1) { cout << "Invalid. Enter new values" ; total = total  p[i].pro; i; } } p[i].pro = 1  total; // Sorting the symbols based on // their probability or frequency sortByProbability(n, p); for (i = 0; i < n; i++) p[i].top = 1; // Find the shannon code shannon(0, n  1, p); // Display the codes display(n, p); return 0; } 
Enter number of symbols : 5 Enter symbol 1 : A Enter symbol 2 : B Enter symbol 3 : C Enter symbol 4 : D Enter symbol 5 : E Enter probability of A : 0.22 Enter probability of B : 0.28 Enter probability of C : 0.15 Enter probability of D : 0.3 Enter probability of E : 0.05 Symbol Probability Code D 0.3 00 B 0.28 01 A 0.22 10 C 0.15 110 E 0.05 111
Attention reader! Don’t stop learning now. Get hold of all the important DSA concepts with the DSA Self Paced Course at a studentfriendly price and become industry ready.
Recommended Posts:
 Difference between Lossy Compression and Lossless Compression
 Data Compression With Arithmetic Coding
 LZW (Lempel–Ziv–Welch) Compression technique
 PGP  Compression
 Compression of IPv6 address
 Difference between Inter and Intra Frame Compression
 DjVu Compression in Computer Network
 Difference between Data Scientist, Data Engineer, Data Analyst
 Move To Front Data Transform Algorithm
 Simplified International Data Encryption Algorithm (IDEA)
 Difference between Data Warehousing and Data Mining
 Introduction to Data Structures  10 most commonly used Data Structures
 Primitive data type vs. Object data type in Java with Examples
 Difference between Data Privacy and Data Security
 Difference between Data Privacy and Data Protection
 MO's Algorithm (Query Square Root Decomposition)  Set 1 (Introduction)
 kasai’s Algorithm for Construction of LCP array from Suffix Array
 EdgeRank Algorithm  Algo behind Facebook News Feed
 A* Search Algorithm
 DDA Line generation Algorithm in Computer Graphics
If you like GeeksforGeeks and would like to contribute, you can also write an article using contribute.geeksforgeeks.org or mail your article to contribute@geeksforgeeks.org. See your article appearing on the GeeksforGeeks main page and help other Geeks.
Please Improve this article if you find anything incorrect by clicking on the "Improve Article" button below.