Thanks to visit codestin.com
Credit goes to www.scribd.com

0% found this document useful (0 votes)
31 views3 pages

Pca 1

Principal Component Analysis (PCA) is an unsupervised learning technique aimed at reducing dataset dimensionality while preserving variance. It involves standardizing data, computing a covariance matrix, and performing eigenvalue decomposition to identify principal components. PCA is widely used for applications such as data visualization, noise reduction, and feature extraction, but it has limitations including information loss and sensitivity to outliers.

Uploaded by

luna luna
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
31 views3 pages

Pca 1

Principal Component Analysis (PCA) is an unsupervised learning technique aimed at reducing dataset dimensionality while preserving variance. It involves standardizing data, computing a covariance matrix, and performing eigenvalue decomposition to identify principal components. PCA is widely used for applications such as data visualization, noise reduction, and feature extraction, but it has limitations including information loss and sensitivity to outliers.

Uploaded by

luna luna
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 3

It seems like you're asking me to organize the detailed information I provided about Principal

Component Analysis (PCA) into a more structured format with clear titles. Here's a re-ordered
and titled presentation of the PCA information:

Principal Component Analysis (PCA): A Detailed


Overview
1. Introduction to PCA
Principal Component Analysis (PCA) is a fundamental unsupervised learning technique used
in machine learning and statistics. Its core objective is to reduce the dimensionality of a
dataset while retaining the maximum possible amount of original information (variance). PCA
achieves this by transforming a high-dimensional dataset into a lower-dimensional set of
uncorrelated features, called "principal components," which are linear combinations of the
original variables.

2. Key Concepts in PCA


●​ Dimensionality Reduction: The primary purpose of PCA, transforming data from a
higher to a lower-dimensional space.
●​ Principal Components (PCs): These are new variables created by PCA. They are
orthogonal (uncorrelated) linear combinations of the original variables.
○​ The first principal component captures the most variance in the data.
○​ Subsequent principal components capture the most remaining variance, orthogonal
to the previous ones.
●​ Variance: PCA's goal is to maximize the variance explained by the principal components,
identifying directions of greatest data spread.
●​ Orthogonal Transformation: PCA rotates the coordinate system so that new axes align
with the directions of maximum variance.

3. How PCA Works (Step-by-Step Process)


The process of PCA involves the following steps, underpinned by eigenvalue decomposition:
1.​ Standardize the Data:
○​ Crucial because PCA is sensitive to variable scales.
○​ Variables are transformed to have a mean of zero and a standard deviation of one,
ensuring equal contribution.
2.​ Compute the Covariance Matrix:
○​ This matrix quantifies the linear relationships (covariance) between all pairs of
standardized variables.
○​ It indicates how variables change together (positive, negative, or no correlation).
3.​ Calculate Eigenvalues and Eigenvectors:
○​ Eigenvectors: Represent the directions (axes) of maximum variance in the data;
these become the principal components.
○​ Eigenvalues: Quantify the amount of variance captured along each eigenvector
direction. Larger eigenvalues correspond to more significant variance.
4.​ Sort Eigenvalues and Select Principal Components:
○​ Eigenvalues are sorted in descending order, with the largest corresponding to the
first principal component.
○​ The number of components to retain is determined using methods like:
■​ Scree Plot: Visualizing eigenvalues to find an "elbow" where the drop-off
lessens.
■​ Cumulative Explained Variance: Selecting enough components to explain a
desired percentage (e.g., 90-95%) of the total variance.
5.​ Transform the Data:
○​ A "feature vector" is constructed using the selected eigenvectors.
○​ The original (standardized) data is then projected onto this new lower-dimensional
space by multiplying it with the feature vector, yielding the principal component
scores.

4. Mathematical Foundations of PCA


Given a dataset X (n observations, p variables):
1.​ Standardization: For each variable x_j, compute z_{ij} = \frac{x_{ij} - \mu_j}{\sigma_j} to
get the standardized matrix Z.
2.​ Covariance Matrix (\Sigma): Calculated from the standardized data: \Sigma =
\frac{1}{n-1} Z^T Z.
3.​ Eigenvalue Decomposition: Solve the equation \Sigma v = \lambda v to find
eigenvalues (\lambda) and eigenvectors (v). There will be p eigenvalues and p
corresponding eigenvectors.
4.​ Selecting Principal Components: Sort eigenvalues: \lambda_1 \ge \lambda_2 \ge ... \ge
\lambda_p. Select the top k eigenvectors (v_1, ..., v_k) to form the projection matrix W_k
= [v_1 | ... | v_k] (a p \times k matrix).
5.​ Transforming Data: The new dataset Y (principal component scores) is Y = Z W_k, an n
\times k matrix.

5. Applications of PCA
PCA is widely used across various domains for its ability to simplify and analyze complex data:
●​ Dimensionality Reduction: The primary use, simplifying datasets for analysis and
modeling.
●​ Data Visualization: Enabling 2D or 3D plots of high-dimensional data to reveal patterns.
●​ Feature Extraction: Deriving the most informative features for subsequent machine
learning algorithms.
●​ Noise Reduction: Filtering out less significant variance (often noise) by discarding lower
principal components.
●​ Data Compression: Representing data with fewer variables, reducing storage and
improving efficiency.
●​ Image Processing: Used in areas like image compression and facial recognition (e.g.,
Eigenfaces).
●​ Finance: Analyzing financial data and optimizing portfolios.
●​ Healthcare: Reducing dimensions in complex medical datasets for analysis.
6. Advantages of PCA
●​ Reduces Overfitting: Simplifies models by removing redundant features, improving
generalization.
●​ Speeds Up Computation: Faster training times for machine learning models on reduced
datasets.
●​ Enhances Data Visualization: Makes high-dimensional data plottable and
understandable.
●​ Removes Noise: Helps isolate and remove noise by focusing on high-variance
components.
●​ Unsupervised: Does not require labeled data, broadening its applicability.
●​ Creates Uncorrelated Features: The orthogonality of principal components can benefit
certain statistical models.

7. Disadvantages of PCA
●​ Loss of Information: Inherent to dimensionality reduction; some information is always
lost.
●​ Hard to Interpret Principal Components: The new components are abstract linear
combinations, making real-world interpretation challenging.
●​ Assumes Linearity: May not effectively capture non-linear relationships in the data.
●​ Requires Standardization: Sensitivity to scale necessitates pre-processing.
●​ Sensitive to Outliers: Outliers can significantly distort the computed variance and
principal component directions.
●​ Not Ideal for Categorical Data: Best suited for numerical data; categorical encoding can
lead to less meaningful results.

You might also like