|
You are here |
11011110.github.io | ||
| | | | |
windowsontheory.org
|
|
| | | | | Previous post: ML theory with bad drawings Next post: What do neural networks learn and when do they learn it, see also all seminar posts and course webpage. Lecture video (starts in slide 2 since I hit record button 30 seconds too late - sorry!) - slides (pdf) - slides (Powerpoint with ink and animation)... | |
| | | | |
a3nm.net
|
|
| | | | | ||
| | | | |
anuragbishnoi.wordpress.com
|
|
| | | | | The Ramsey number $latex R(s, t)$ is the smallest $latex n$ such that every graph on $latex \geq n$ vertices either contains a clique of size $latex s$ or an independent set of size $latex t$. Ramsey's theorem implies that these numbers always exist, and determining them (precisely or asymptotically) has been a major challenge... | |
| | | | |
iclr-blog-track.github.io
|
|
| | | [AI summary] The provided text is an extensive blog post discussing the implementation and reproduction of the Proximal Policy Optimization (PPO) algorithm in various environments, including Atari, Procgen, and others. It highlights key implementation details, such as MultiDiscrete action spaces, vectorized environments, and accelerated training techniques like Envpool. The post also compares PPO with other algorithms like IMPALA and APPO, and emphasizes the importance of documentation and efficient code for reproducibility and research. | ||