Open RL infra for materials R&D at Supermatter Agent
Technical intelligence at Manuel AI — MAS infra, HVAC · energy · marine

I believe AI is the ultimate compressor for hardest problems and a superpower for productivity and I'm dedicated to accelerating its progress. Through the lens of multi-agent systems, reinforcement learning (RL) infrastructure for LLMs; I'm working on high-compute RL, megakernels, large-scale retrieval and memory components.
Working on Supermatter Agent AI agents, materials science, physics simulation, multiscale modeling, scientific computing, agent systems
Building an agent system for the physical sciences — materials science, simulation and rapid prototyping. Multiscale materials design powered by AI agents running natively on local compute.
Feb 19, 2026

Teaching agents GPU/TPU/QPU compute: An open skill catalog GPU programming, TPU, QPU, quantum computing, HPC, CUDA, agent skills, open source, accelerated computing
An open-source, growing catalog of structured agent skills for GPU/TPU/QPU-accelerated frameworks and simulators. The goal: cover every mature and experimental HPC compute workload so any agent can pick up working knowledge at inference time.
Feb 15, 2026

Claude Code-Time Skill Acquisition with Agent Teams AI agents, Claude Code, multi-agent systems, skill acquisition, agent coordination, knowledge base, React Native
A team of agents researched, synthesized, and integrated a production-grade React Native skill into a shared knowledge base in under 15 minutes — just through coordination at Claude Code-time.
Feb 7, 2026

On Compression, Computation and the Space Between Kolmogorov complexity, information theory, computation theory, neural networks, Wolfram, compression, mathematics
Kolmogorov complexity, neural networks as program search and Wolfram's ruliology seem to be looking at the same thing from different rooms.
Feb 1, 2026

Defeating Nondeterminism in LLM Inference: Reproducing Batch-Invariant Ops (RMSNorm & Tiled Matrix Multiplication) in JAX JAX, GPU kernels, RMSNorm, matrix multiplication, LLM inference, nondeterminism, batch invariance, deep learning
This learning log is my beginning of a series exploring various kernel-related topics. As a starting point, I will reproduce the implementation of batch-invariant NN operations in JAX, drawing from Thinking Machines Lab's seminal collaborative work, \"Defeating Nondeterminism in LLM Inference.\"
Nov 25, 2025

Streaming deepagents and task delegation with real-time output LLM agents, streaming, deep agents, multi-agent systems, task delegation, Python, AI engineering
This post demonstrates how to implement streaming capabilities on top of DeepAgents' package with multi-agent setup, with practical code examples and architectural patterns you can apply to your own projects.
Oct 20, 2025