Skip to content Skip to sidebar Skip to footer

Reinforcement Learning for Physics: ODEs and Hyperparameter Tuning | by Robert Etter | Oct, 2024

Working with ODEs Physical systems can typically be modeled through differential equations, or equations including derivatives. Forces, hence Newton’s Laws, can be expressed as derivatives, as can Maxwell’s Equations, so differential equations can describe most physics problems. A differential equation describes how a system changes based on the system’s current state, in effect defining state…

Read More

LLM vs LLM: Codenames Tournament. A mini multi-agent competition among 3… | by Yennie Jun | Oct, 2024

A mini multi-agent competition among 3 different LLM agents Generated using ChatGPT 4o.This article was originally posted on Art Fish Intelligence. LLMs are good at many things, and one of those things is playing games. People have used LLMs to play all sorts of games such as Minecraft, Chess, murder mystery games, Werewolf, and the…

Read More

Supercharge Your LLM Apps Using DSPy and Langfuse | by Raghav Bali | Oct, 2024

As illustrated in figure 1, DSPy is a pytorch-like/lego-like framework for building LLM-based apps. Out of the box, it comes with: Signatures: These are specifications to define input and output behaviour of a DSPy program. These can be defined using short-hand notation (like “question -> answer” where the framework automatically understands question is the input…

Read More

Transformer? Diffusion? Transfusion!

A gentle introduction to the latest multi-modal transfusion model Recently, Meta and Waymo released their latest paper — Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model, which integrates the popular transformer model with the diffusion model for multi-modal training and prediction purposes. Like Meta’s previous work, the Transfusion model is based on the…

Read More

Training AI Models on CPU. Revisiting CPU for ML in an Era of GPU… | by Chaim Rand | Sep, 2024

Revisiting CPU for ML in an Era of GPU Scarcity Photo by Quino Al on UnsplashThe recent successes in AI are often attributed to the emergence and evolutions of the GPU. The GPU’s architecture, which typically includes thousands of multi-processors, high-speed memory, dedicated tensor cores, and more, is particularly well-suited to meet the intensive demands…

Read More