This started with Addition Under Pressure, where I gave Claude Code and Codex the same prompt: train the smallest possible transformer that can do 10-digit addition with at least 99% accuracy. Claude Code came back with 6,080 parameters and Codex came back with 1,644. The community has since pushed this dramatically lower.
If you have any questions ask in the comments section
,推荐阅读服务器推荐获取更多信息
FT Videos & Podcasts
时隔五年,MacBook Pro 大升级