Link Copied.
System and Method for Fast Processing Singular Value Decomposition (SVD) of Extremely Large-Scale Low-Rank Matrices

中文版本

Opportunity

Singular Value Decomposition (SVD) is a fundamental tool for dimensionality reduction, data compression, and feature extraction. However, computing SVD for large-scale matrices is prohibitively expensive. A dense 1,000,000 × 1,000,000 matrix requires 7.27 TB of memory—far beyond typical desktop computers. Traditional SVD algorithms still require storing the entire matrix, performing large matrix multiplications, and have very high time complexity. Even when the matrix has low-rank structure (common in practical applications like image sets or time-series data), existing methods cannot scale to million-scale matrices under typical memory constraints. There is an urgent need for an SVD algorithm that processes extremely large matrices using only partial data, without loading the full matrix into RAM.

Technology  

This patent presents CUR-SVD, a method for computing Singular Value Decomposition (SVD) of extremely large matrices using only a small subset of rows and columns. Its core innovation is enabling SVD on matrices that are too large to fit in a computer's memory, solving a fundamental limitation of traditional approaches which require loading the entire matrix into RAM. The method first randomly samples a small number of columns to form matrix C and a small number of rows to form matrix R. Instead of relying on random selection alone, it then applies the Discrete Empirical Interpolation Method (DEIM) to intelligently pick the most informative rows and columns from these samples, ensuring high approximation accuracy. With these compact C and R matrices, CUR-SVD computes their SVD and uses a lightweight intermediate matrix to recover the full SVD result—singular values and singular vectors—without ever processing the original giant matrix directly. A key breakthrough is its ability to handle unknown matrix ranks automatically: the method starts with a small block of rows and columns, checks whether the singular values have stabilized, and incrementally adds more blocks using Incremental SVD until convergence, all without recomputing from scratch. Throughout the entire process, the method never loads the full matrix into RAM, only reading the selected columns and rows from disk. This memory efficiency is transformative: the patent demonstrates computing SVD of a one-million-by-one-million matrix using approximately three gigabytes of memory, whereas traditional methods would require over seven terabytes, making large-scale matrix decomposition feasible on ordinary desktop computers for applications like image processing, PCA, and machine learning.

Advantages  

  • Extreme Scalability: Computes SVD of 1,000,000 × 1,000,000 matrix (7.27 TB) using only 3 GB memory—impossible for traditional methods.
  • Fast Runtime: Top 200 singular values/vectors of a million-by-million matrix computed in about 82 seconds (fixed-rank) or 15 minutes (fixed-precision).
  • Low Memory Footprint: Requires memory proportional only to the number of selected rows/columns, not the full matrix dimensions.
  • Comparable Accuracy: Achieves relative error of about 1 part in 1 trillion for low-rank matrices, matching standard SVD accuracy.
  • Works with Unknown Rank: Fixed-precision version automatically estimates the rank using a convergence stopping criterion.
  • Proven on Real Data: 12 times faster than standard SVD for Principal Component Analysis on a large face image dataset; reconstructed faces are visually indistinguishable from the original.

Applications  

  • Principal Component Analysis (PCA): Dimensionality reduction for large image datasets (facial recognition, medical imaging) where full data cannot fit in memory.
  • Recommendation Systems: SVD for collaborative filtering on user-item matrices with millions of users and items.
  • Time-Series Analysis: Extracting dominant patterns from fluid dynamics, climate data, or large sensor networks.
  • Compressed Sensing & Data Compression: Low-rank approximation for large scientific computing matrices.
  • Machine Learning Preprocessing: Reducing feature space for large-scale text or genomic data before model training.
Remarks
CIMDA: P00128
IP Status
Patent filed
Technology Readiness Level (TRL)
4
Questions about this Technology?
Contact Our Tech Manager
Contact Our Tech Manager
System and Method for Fast Processing Singular Value Decomposition (SVD) of Extremely Large-Scale Low-Rank Matrices

Personal Information

(ReCaptcha V3 Hidden Field)

We use cookies to ensure you get the best experience on our website.

More Information