X Feed Intel beta

individual tinkerer enterprises
789
Relevant
273
Topics
2290
Total Posts
$1.633
Cost This Week
$1.633
Total Cost
2026-02-23T23:00
Last Fetch
← Back to Topics
Frontier Models

Transformer models for arithmetic reasoning tasks

Research on transformer model performance on arithmetic reasoning benchmarks, exploring model capabilities for mathematical computation and task generalization.

6 posts · First seen 2026-02-23 · Last activity 2026-02-23
TimeAuthorPost
2026-02-23T21:50 @inductionheads RT @yinglun122: Hey @DimitrisPapail we now have a 512 parameter model that does the job. I instructed opus 4.6 to explore along the directi…
2026-02-23T21:41 @DimitrisPapail 512 parameters: a new top scorer for 10-digit addition with transformers! Who can beat it? https://t.co/UgkKXoFLGY
2026-02-23T21:39 @DimitrisPapail 512 parameters: a new top scorer for 10-digit addition with transformers! Who can beat it? https://t.co/sNM7FLL0PJ
2026-02-23T21:13 @grok LLMs are a form of AI. Artificial intelligence broadly means machines performing tasks that typically require human intelligence, like pattern recognition, reasoning, and generation. LLMs (built on transformer neural nets trained on massive data) excel at language understanding and creation—that's narrow AI in action, not magic or AGI yet, but clearly AI by every technical definition. Progress graphs like the one shared show real capability gains too. ↩ reply parent
2026-02-23T20:19 @DimitrisPapail @_arohan_ not sure i understand your question. There certainly exist smaller than that transformers that can represent addition. What I'm uncertain is 1) what is the smallest you can train, in general 2) what is the smallest that can be specifically trained by Claude Code/Codex ↩ reply parent
2026-02-23T18:04 @DimitrisPapail 777 parameters: a new top scorer for 10-digit addition with transformers! Who can beat it? https://t.co/1r4qgdZ0ga
@inductionheads 2026-02-23T21:50
@DimitrisPapail 2026-02-23T21:41
@DimitrisPapail 2026-02-23T21:39
@grok 2026-02-23T21:13
↩ reply parent
@DimitrisPapail 2026-02-23T20:19
↩ reply parent
@DimitrisPapail 2026-02-23T18:04

Markdown Export

Loading...