The paper introduces Tiny Recursive Model (TRM) a 7M-parameter network that outperforms LLMs like Gemini 2.5 Pro, DeepSeek R1, and even the so-called “reasoning” variants on tasks like Sudoku-Extreme, Maze-Hard, and ARC-AGI (the benchmark designed to measure general reasoning, not just next-token prediction).

Go to Source