A Clojure port of XinJingHao’s PPO implementation using libpython-clj2, PyTorch, and Quil. PPO is a reinforcement learning method which has become popular because it addresses the problem of stability. The PPO implementation is tested using the inverted pendulum problem.
My Windows gaming graphics performance dropped by 40% after some update. I already was using Linux for everything else except gaming for many years. So I tried out Steam on Linux and I was quite amazed how many games run on Linux via Proton. Just check out protondb.com for compatibility reports.
I did a bit of scripting trying to automate the TDD cycle given a task description. The problem is, that the LLM tends to jump forward and submit a full implementation instead of a minimal change. I guess the problem is, that LLMs are trained on complete solutions instead of minimal steps.
In the past I have done some rigid body physics in GNU Guile (see https://www.youtube.com/watch?v=zBq3kW2jVxs for example). Of course if you need to simulate many objects, you will hit performance problems sooner if you don't use C/C++/Rust. Also the developer of Jolt has solved quite difficult problems, so I was quite happy to use it instead of rolling my own.
I use a few macros for creating contexts (i.e. with-texture, with-stencil, with-scissors, with-tar). Also I have macros for rendering (onscreen-render, offscreen-render). However I try not to overuse macros.
reply