Explore >> Select a destination


You are here

jeremykun.wordpress.com
| | www.jeremykun.com
0.0 parsecs away

Travel
| | This post is a sequel to Formulating the Support Vector Machine Optimization Problem. The Karush-Kuhn-Tucker theorem Generic optimization problems are hard to solve efficiently. However, optimization problems whose objective and constraints have special structure often succumb to analytic simplifications. For example, if you want to optimize a linear function subject to linear equality constraints, one can compute the Lagrangian of the system and find the zeros of its gradient. More generally, optimizing...
| | francisbach.com
2.8 parsecs away

Travel
| | [AI summary] The blog post discusses non-convex quadratic optimization problems and their solutions, including the use of strong duality, semidefinite programming (SDP) relaxations, and efficient algorithms. It highlights the importance of these problems in machine learning and optimization, particularly for non-convex problems where strong duality holds. The post also mentions the equivalence between certain non-convex problems and their convex relaxations, such as SDP, and provides examples of when these relaxations are tight or not. Key concepts include the role of eigenvalues in quadratic optimization, the use of Lagrange multipliers, and the application of methods like Newton-Raphson for solving these problems. The author also acknowledges contributions...
| | thenumb.at
3.7 parsecs away

Travel
| | [AI summary] This text provides a comprehensive overview of differentiable programming, focusing on its application in machine learning and image processing. It explains the fundamentals of automatic differentiation, including forward and backward passes, and demonstrates how to implement these concepts in a custom framework. The text also discusses higher-order differentiation and its implementation in frameworks like JAX and PyTorch. A practical example is given using differentiable programming to de-blur an image, showcasing how optimization techniques like gradient descent can be applied to solve real-world problems. The text emphasizes the importance of differentiable programming in enabling efficient and flexible computation for various domains, includ...
| | extremal010101.wordpress.com
27.8 parsecs away

Travel
| With Alexandros Eskenazis we posted a paper on arxiv "Learning low-degree functions from a logarithmic number of random queries" exponentially improving randomized query complexity for low degree functions. Perhaps a very basic question one asks in learning theory is as follows: there is an unknown function $latex f : \{-1,1\}^{n} \to \mathbb{R}$, and we are...