|
You are here |
thomvolker.github.io | ||
| | | | |
www.jeremykun.com
|
|
| | | | | The singular value decomposition (SVD) of a matrix is a fundamental tool in computer science, data analysis, and statistics. It's used for all kinds of applications from regression to prediction, to finding approximate solutions to optimization problems. In this series of two posts we'll motivate, define, compute, and use the singular value decomposition to analyze some data. (Jump to the second post) I want to spend the first post entirely on motivation and background. | |
| | | | |
www.aleksandrhovhannisyan.com
|
|
| | | | | Some systems of equations do not have a unique solution, but we can find an approximate solution using the method of least squares. Applications of this method include linear and polynomial regression. | |
| | | | |
peterbloem.nl
|
|
| | | | | [AI summary] The pseudo-inverse is a powerful tool for solving matrix equations, especially when the inverse does not exist. It provides exact solutions when they exist and least squares solutions otherwise. If multiple solutions exist, it selects the one with the smallest norm. The pseudo-inverse can be computed using the singular value decomposition (SVD), which is numerically stable and handles cases where the matrix does not have full column rank. The SVD approach involves computing the SVD of the matrix, inverting the non-zero singular values, and then reconstructing the pseudo-inverse using the modified SVD components. This method is preferred due to its stability and ability to handle noisy data effectively. | |
| | | | |
weisser-zwerg.dev
|
|
| | | A series about Monte Carlo methods and generating samples from probability distributions. | ||