# Difference between PCA VS t-SNE

**Principal Component analysis (PCA):** PCA is an unsupervised linear dimensionality reduction and data visualization technique for very high dimensional data. As having high dimensional data is very hard to gain insights from adding to that, it is very computationally intensive. The main idea behind this technique is to reduce the dimensionality of data that is highly correlated by transforming the original set of vectors to a new set which is known as **Principal component**.

PCA tries to preserve the Global Structure of data i.e when converting d-dimensional data to d’-dimensional data then it tries to map all the clusters as a whole due to which local structures might get lost. Application of this technique includes Noise filtering, feature extractions, stock market predictions, and gene data analysis.

**t-distributed stochastic neighbourhood embedding (t-SNE):** t-SNE is also a unsupervised non-linear dimensionality reduction and data visualization technique. The math behind t-SNE is quite complex but the idea is simple. It embeds the points from a higher dimension to a lower dimension trying to preserve the neighborhood of that point.

Unlike PCA it tries to preserve the Local structure of data by minimizing the **Kullback–Leibler divergence (KL divergence)** between the two distributions with respect to the locations of the points in the map. This technique finds application in computer security research, music analysis, cancer research, bioinformatics, and biomedical signal processing.

## Table of Difference between PCA and t-SNE

S.NO. | PCA | t-SNE |
---|---|---|

1. | It is a linear Dimensionality reduction technique. | It is a non-linear Dimensionality reduction technique. |

2. | It tries to preserve the global structure of the data. | It tries to preserve the local structure(cluster) of data. |

3. | It does not work well as compared to t-SNE. | It is one of the best dimensionality reduction technique. |

4. | It does not involve Hyperparameters. | It involves Hyperparameters such as perplexity, learning rate and number of steps. |

5. | It gets highly affected by outliers. | It can handle outliers. |

6. | PCA is a deterministic algorithm. | It is a non-deterministic or randomised algorithm. |

7. | It works by rotating the vectors for preserving variance. | It works by minimising the distance between the point in a guassian. |

8. | We can find decide on how much variance to preserve using eigen values. | We cannot preserve variance instead we can preserve distance using hyperparameters. |

## Recommended Posts:

- Difference between CPU and GPU
- Difference between SDN and NFV
- Difference Between SMO and SEO
- Web 1.0, Web 2.0 and Web 3.0 with their difference
- Difference between LAN and MAN
- Difference between H.323 and SIP
- Difference between DFD and ERD
- Difference between T-SQL and PL-SQL
- Difference between IoE and IoT
- Difference between Tor and VPN
- Difference between USB 2.0 and USB 3.0
- Difference between 4G and 5G
- Difference between BFS and DFS
- Difference between MP4 and MP3
- Difference between IoT and M2M
- Difference between 4NF and 5NF
- Difference between SQL and T-SQL
- Difference Between SEO and SEM
- Difference between CLI and GUI
- Difference between DTE and DCE

If you like GeeksforGeeks and would like to contribute, you can also write an article using contribute.geeksforgeeks.org or mail your article to contribute@geeksforgeeks.org. See your article appearing on the GeeksforGeeks main page and help other Geeks.

Please Improve this article if you find anything incorrect by clicking on the "Improve Article" button below.