From: https://stats.stackexchange.com/questions/107533/how-to-use-svd-for-dimensionality-reduction-to-reduce-the-number-of-columns-fea
My original data has many more columns (features) than rows (users). I am trying to reduce the features of my SVD (I need all of the rows). I found one method of doing so in a book called “Machine Learning in Action” but I don’t think it will work for the data I am using.
The method is as follows. Define SVD as

Set an optimization threshold (i.e., 90%). Calculate the total sum of the squares of the diagonal 𝑆� matrix. Calculate how many 𝑆� values it takes to reach 90% of the total sum of squares. So if that turns out to be 100 𝑆� values, then I would take the first 100 columns of the 𝑈� matrix, first 100 rows of the 𝑉^⊤ matrix, and a 100×100100×100 square matrix out of the 𝑆� matrix. I would then calculate 𝐴=𝑈𝑆𝑉^⊤ using the reduced matrices.
However, this method does not target the columns of my original data, since the dimensions of the resulting 𝐴� matrix are the same as before. How would I target the columns of my original matrix?