
Jonathan co-founded LLM Stats to build independent, reproducible measurement infrastructure for AI models. He leads the platform's benchmark evaluation methodology, arena design, and data pipeline architecture. His work focuses on eliminating bias from AI model evaluation — designing blind voting systems, standardizing benchmark collection across providers, and publishing transparent ranking methodologies that frontier labs and Fortune 500 teams rely on for model selection decisions.
Expertise
Articles (5)
Claude Mythos Preview: Benchmarks, Pricing & Project Glasswing
Apr 7, 2026
How to Calculate Hardware Requirements for Running LLMs Locally
Apr 3, 2026
Post-Training in 2026: GRPO, DAPO, RLVR & Beyond
Mar 11, 2026
Nemotron 3 Super: Pricing, Benchmarks, Architecture & API
Mar 11, 2026
Model Quantization Across Providers
Nov 28, 2024