PCA as a Last Resort


Introduction
Principal Component Analysis (PCA) is often the first dimensionality reduction technique that data scientists reach for when faced with high-dimensional data. While PCA is powerful and mathematically elegant, treating it as a default first step can lead to missed opportunities and suboptimal models. This post explores why feature engineering and feature removal should be your first considerations before applying PCA.
Understanding the Limitations of PCA
PCA transforms your original features into new components that capture maximum variance. However, this mathematical transformation comes with significant tradeoffs:
Loss of interpretability - Principal components are linear combinations of original features, making them difficult to explain to stakeholders
Domain knowledge is discarded - PCA is a purely statistical technique that ignores valuable domain expertise
Non-linear relationships are missed - Standard PCA only captures linear relationships between features
Feature Engineering: Creating Meaningful Representations
Before reducing dimensions, consider creating more informative features:
Ratio features that capture relationships between variables (e.g., debt-to-income ratio)
Interaction terms that represent how features work together
Domain-specific transformations based on expert knowledge
Polynomial features to capture non-linear relationships
These engineered features often provide more predictive power than abstract principal components while maintaining interpretability.
Feature Removal: The Simplest Form of Dimensionality Reduction
Feature removal should be your first dimensionality reduction approach because:
It preserves the original meaning of remaining features
It forces critical thinking about which variables truly matter
It simplifies your model and reduces overfitting
Methods for informed feature removal include:
Correlation analysis to identify redundant features
Feature importance rankings from tree-based models
Filter methods like variance thresholds and mutual information
Wrapper methods such as recursive feature elimination
When PCA Makes Sense
PCA becomes valuable after you've exhausted feature engineering and removal options, particularly when:
You still have high dimensionality after careful feature selection
Multicollinearity remains a significant issue
Computational efficiency is a critical concern
You're using specific algorithms that benefit from orthogonal features
Visualization of high-dimensional data is needed
A Better Workflow for Dimensionality Reduction
Instead of immediately applying PCA, follow this approach:
Start with domain knowledge to engineer meaningful features
Apply feature selection techniques to remove redundant or irrelevant variables
Use PCA only on the remaining features if dimensionality is still problematic
Consider non-linear dimensionality reduction techniques (t-SNE, UMAP) if linear PCA performs poorly
Conclusion
While PCA is a valuable tool in the data scientist's toolkit, it should rarely be your first choice for dimensionality reduction. By prioritizing feature engineering and thoughtful feature removal, you'll create models that are not only more accurate but also more interpretable and actionable. Save PCA for when you truly need it—as a last resort after you've leveraged your domain knowledge and simpler techniques.
Subscribe to my newsletter
Read articles from Harvey Ducay directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by
