dimensionality-reduction

aj-geddes/useful-ai-prompts · updated Apr 8, 2026

$npx skills add https://github.com/aj-geddes/useful-ai-prompts --skill dimensionality-reduction
0 commentsdiscussion
summary

Dimensionality reduction techniques reduce the number of features while preserving important information, improving model efficiency and enabling visualization of high-dimensional data.

skill.md

Dimensionality Reduction

Overview

Dimensionality reduction techniques reduce the number of features while preserving important information, improving model efficiency and enabling visualization of high-dimensional data.

When to Use

  • High-dimensional datasets with many features
  • Visualizing complex datasets in 2D or 3D
  • Reducing computational complexity and training time
  • Removing redundant or highly correlated features
  • Preventing overfitting in machine learning models
  • Preprocessing data before clustering or classification

Techniques

  • PCA: Principal Component Analysis
  • t-SNE: t-Distributed Stochastic Neighbor Embedding
  • UMAP: Uniform Manifold Approximation and Projection
  • Feature Selection: Selecting important features
  • Feature Extraction: Creating new features

Benefits

  • Reduce computational complexity
  • Remove noise and redundancy
  • Improve model generalization
  • Enable visualization
  • Prevent curse of dimensionality

Implementation with Python

import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
from sklearn.decomposition import PCA, TruncatedSVD, FactorAnalysis
from sklearn.manifold import TSNE, MDS
from sklearn.preprocessing import StandardScaler
from sklearn.datasets import load_iris
from sklearn.ensemble import RandomForestClassifier
from sklearn.feature_selection import SelectKBest, f_classif, mutual_info_classif
import seaborn as sns

# Load data
iris = load_iris()
X = iris.data
y = iris.target
feature_names = iris.feature_names

# Standardize
scaler = StandardScaler()
X_scaled = scaler.fit_transform(X)

# PCA
pca = PCA()
pca.fit(X_scaled)

# Explained variance
explained_variance = np.cumsum(pca.explained_variance_ratio_)
print("Explained Variance Ratio by Component:")
print(pca.explained_variance_ratio_)
print(f"Cumulative Variance (first 2): {explained_variance[1]:.4f}")

# Scree plot
fig, axes = plt.subplots(1, 2, figsize=(14, 4))

axes[0].plot(range(1, len(pca.explained_variance_ratio_) + 1),
             pca.explained_variance_ratio_, 'bo-')
axes[0].set_xlabel('Principal Component')
axes[0].set_ylabel('Explained Variance Ratio')
axes[0].set_title('Scree Plot')
axes[0].grid(True, alpha=0.3)

axes[1].plot(range(1, len(explained_variance) + 1),
             explained_variance, 'go-')
axes[1].axhline(y=0.95, color='r', linestyle='--', label='95% Variance')
axes[1].set_xlabel('Number of Components')
axes[1].set_ylabel('Cumulative Explained Variance')
axes[1].set_title('Cumulative Explained Variance')
axes[1].legend()
axes[1].grid(True, alpha=0.3)

plt.tight_layout()
plt.show()

# PCA with 2 components
pca_2d = PCA(n_components=2)
X_pca_2d = pca_2d.fit_transform(X_scaled)

# PCA with 3 components
pca_3d = PCA(n_components=3)
X_pca_3d = pca_3d.fit_transform(X_scaled)

# PCA visualization
fig = plt.figure(figsize=(14, 5))

# 2D PCA
ax1 = fig.add_subplot(131)
scatter = ax1.scatter(X_pca_2d[:, 0], X_pca_2d[:, 1], c=y, cmap='viridis', alpha=0.6)
ax1.set_xlabel(f'PC1 ({pca_2d.explained_variance_ratio_[0]:.2%})')
ax1.set_ylabel(f'PC2 ({pca_2d.explained_variance_ratio_[1]:.2%})')
ax1.set_title('PCA 2D')
plt.colorbar(scatter, ax=ax1)

# 3D PCA
ax2 = fig.add_subplot(132, projection='3d')
scatter = ax2.scatter(X_pca_3d[:, 0], X_pca_3d[:, 1], X_pca_3d[:, 2],
                      c=y, cmap='viridis', alpha=0.6)
ax2.set_xlabel(f'PC1 ({pca_3d.explained_variance_ratio_[0]:.2%})')
ax2.set_ylabel(f'PC2 ({pca_3d.explained_variance_ratio_[1]:.2%})')
ax2.set_zlabel(f'PC3 ({pca_3d.explained_variance_ratio_[2]:.2%})')
ax2.set_title('PCA 3D')

# Loading plot
ax3 = fig.add_subplot(133)
loadings = pca_2d.components_.T
for i, feature in enumerate(feature_names):
    ax3.arrow(0, 0, loadings[i, 0], loadings[i, 1],
             head_width=0.05, head_length=0.05, fc=

Discussion

Product Hunt–style comments (not star reviews)
  • No comments yet — start the thread.
general reviews

Ratings

4.529 reviews
  • Chaitanya Patil· Dec 20, 2024

    Registry listing for dimensionality-reduction matched our evaluation — installs cleanly and behaves as described in the markdown.

  • Isabella Sharma· Dec 16, 2024

    We added dimensionality-reduction from the explainx registry; install was straightforward and the SKILL.md answered most questions upfront.

  • James Harris· Dec 4, 2024

    dimensionality-reduction reduced setup friction for our internal harness; good balance of opinion and flexibility.

  • James Smith· Nov 23, 2024

    Registry listing for dimensionality-reduction matched our evaluation — installs cleanly and behaves as described in the markdown.

  • Amina Choi· Nov 15, 2024

    dimensionality-reduction is among the better-maintained entries we tried; worth keeping pinned for repeat workflows.

  • Piyush G· Nov 11, 2024

    dimensionality-reduction reduced setup friction for our internal harness; good balance of opinion and flexibility.

  • Neel Ndlovu· Nov 7, 2024

    Keeps context tight: dimensionality-reduction is the kind of skill you can hand to a new teammate without a long onboarding doc.

  • Isabella Yang· Oct 26, 2024

    dimensionality-reduction has been reliable in day-to-day use. Documentation quality is above average for community skills.

  • Ira Rahman· Oct 14, 2024

    dimensionality-reduction fits our agent workflows well — practical, well scoped, and easy to wire into existing repos.

  • Tariq Abebe· Oct 6, 2024

    Useful defaults in dimensionality-reduction — fewer surprises than typical one-off scripts, and it plays nicely with `npx skills` flows.

showing 1-10 of 29

1 / 3