|
You are here |
teddykoker.com | ||
| | | | |
francisbach.com
|
|
| | | | | [AI summary] The blog post discusses the spectral properties of kernel matrices, focusing on the analysis of eigenvalues and their estimation using tools like the matrix Bernstein inequality. It also covers the estimation of the number of integer vectors with a given L1 norm and the relationship between these counts and combinatorial structures. The post includes a detailed derivation of bounds for the difference between true and estimated eigenvalues, highlighting the role of the degrees of freedom and the impact of regularization in kernel methods. Additionally, it touches on the importance of spectral analysis in machine learning and its applications in various domains. | |
| | | | |
jaketae.github.io
|
|
| | | | | In this post, we will take a look at Nyström approximation, a technique that I came across in Nyströmformer: A Nyström-based Algorithm for Approximating Self-Attention by Xiong et al. This is yet another interesting paper that seeks to make the self-attention algorithm more efficient down to linear runtime. While there are many intricacies to the Nyström method, the goal of this post is to provide a high level intuition of how the method can be used to approximate large matrices, and how this method was used in the aforementioned paper. | |
| | | | |
d2l.ai
|
|
| | | | | [AI summary] This chapter provides an in-depth exploration of recommender systems, covering fundamental concepts and advanced techniques. It begins with an overview of collaborative filtering and the distinction between explicit and implicit feedback. The chapter then delves into various recommendation tasks and their evaluation methods. It introduces the MovieLens dataset as a practical example for building recommendation models. Subsequent sections discuss matrix factorization, AutoRec using autoencoders, personalized ranking with Bayesian personalized ranking and hinge loss, neural collaborative filtering, sequence-aware recommenders, feature-rich models, and deep factorization machines like DeepFM. The chapter concludes with implementation details and ev... | |
| | | | |
poissonisfish.com
|
|
| | | Someof the most fundamental functions in R, in my opinion, are those that deal with probability distributions. Whenever you compute a P-value you relyon a probability distribution, and there are many types out there. In this exercise I will cover four: Bernoulli, Binomial, Poisson, and Normal distributions. Let me begin with some theory first: Bernoulli... | ||