News
Latest
Top
Search
Submit
Login
Search
▲
129
Transformers know more than they can tell: Learning the Collatz sequence
(arxiv.org)
by Xcelerate |
view
|
45 comments
▲
78
Weight-sparse transformers have interpretable circuits [pdf]
(cdn.openai.com)
by 0x79de |
view
|
46 comments
▲
50
Brain-IT: Image Reconstruction from fMRI via Brain-Interaction Transformer
(AmitZalcher.github.io)
by SerCe |
view
|
10 comments
▲
9
Out-of-Distribution Generalization in Transformers via Latent Space Reasoning
(arxiv.org)
by marojejian |
view
|
1 comments
▲
7
Transformers v5 Is Out
(huggingface.co)
by unofficialmerve |
view
|
1 comments
▲
6
Show HN: Pulse-Field – O(N) AI Architecture (12x faster than Transformers)
(github.com)
by makimilan |
view
|
8 comments
▲
5
Show HN: Wasda – Experience transformer attention as music
(github.com)
by kinders |
view
|
0 comments
▲
3
Stronger Normalization-Free Transformers
(arxiv.org)
by mfiguiere |
view
|
0 comments
▲
3
An AI Startup Looks Toward the Post-Transformer Era
(wsj.com)
by fortran77 |
view
|
1 comments
▲
3
What's Next for AI? OpenAI's Łukasz Kaiser (Transformer Co-Author) [video]
(youtube.com)
by abrichr |
view
|
0 comments
▲
3
Z-Image: Efficient Image Gen Model with Single-Stream Diffusion Transformer
(tongyi-mai.github.io)
by SerCe |
view
|
0 comments
▲
3
Parallel Loop Transformer for Efficient Test-Time Computation Scaling
(arxiv.org)
by PaulHoule |
view
|
0 comments
▲
3
Symmetric Power Transformers
(manifestai.com)
by ashvardanian |
view
|
0 comments
▲
3
The Transformer and the Hash: building blocks of 21st century political science
(nothinghuman.substack.com)
by ivee |
view
|
0 comments
▲
2
A biologically inspired cognitive architecture without Transformers
(github.com)
by Brain_cognitive |
view
|
1 comments
▲
2
Securing America's grid: a strategic transformer reserve
(breakingdefense.com)
by jrpt |
view
|
0 comments
▲
2
Get to Grips with Transformers and LLMs
(i-programmer.info)
by aquastorm |
view
|
0 comments
▲
2
2015 radio interview: AI as "high-level algebra" before Transformers and LLMs
(doomlaser.com)
by doomlaser |
view
|
0 comments
▲
2
Why are Transformers replacing CNNs? [video]
(youtube.com)
by chii |
view
|
0 comments
▲
2
Transformers v5.0 by HuggingFace
(huggingface.co)
by satvikpendem |
view
|
0 comments
▲
2
Porting Nanochat to Transformers
(huggingface.co)
by us321 |
view
|
0 comments
▲
2
Turbine Transport Transformer
(mitxela.com)
by mhb |
view
|
0 comments
▲
2
Show HN: PDFClear – Browser-based PDF tools with local AI (WASM+Transformers.js)
(pdfclear.com)
by aliansari22 |
view
|
1 comments
▲
2
Show HN: Aion-Torch – Adaptive residual scaling for deep Transformers
(github.com)
by Rioverde |
view
|
0 comments
▲
2
Who Invented Transformer Neural Networks?
(people.idsia.ch)
by puttycat |
view
|
0 comments
▲
2
Show HN: Run HF Transformers in pure Go (10 MB binary, no Python)
(github.com)
by openfluke |
view
|
0 comments
▲
2
The Curved Spacetime of Transformer Architectures
(arxiv.org)
by luis_likes_math |
view
|
1 comments
▲
1
Show HN: Reeyee.ai – AI image style transformer
(reeyee.ai)
by jokera |
view
|
0 comments
▲
1
Transformers and Evolution
(symmetrybroken.com)
by riemannzeta |
view
|
0 comments
▲
1
Building a Minimal Transformer for 10-digit Addition
(alexlitzenberger.com)
by kelseyfrog |
view
|
0 comments
▲
1
Building a Minimal Transformer for 10-digit Addition
(alexlitzenberger.com)
by alexlitz |
view
|
0 comments
▲
1
Mixture of Experts (MoEs) in Transformers
(huggingface.co)
by ibobev |
view
|
0 comments
▲
1
Smallest transformer that can add two 10-digit numbers
(github.com)
by ks2048 |
view
|
0 comments
▲
1
TranslateGemma now runs 100% in the browser on WebGPU with Transformers.js v4
(huggingface.co)
by tzury |
view
|
1 comments
▲
1
Thinking Like Transformer
(srush.github.io)
by vinhnx |
view
|
0 comments
▲
1
Show HN: Doppler.js – WebGPU inference, faster/simpler than transformer.js
by clocksmith |
view
|
0 comments
▲
1
Build a Transformer with Jax [video]
(youtube.com)
by apitman |
view
|
0 comments
▲
1
Show HN: AI Timeline – 171 LLMs from Transformer (2017) to GPT-5.3 (2026)
(llm-timeline.com)
by ai_bot |
view
|
0 comments
▲
1
Wave Field LLM – O(nlogn) Transformer Alternative
(github.com)
by mrtb |
view
|
0 comments
▲
1
Your Transformer Is secretly an EOT Solver
(elonlit.com)
by Anon84 |
view
|
0 comments
▲
1
LLMs create their smallest transformer for 10-digit addition
(twitter.com)
by marojejian |
view
|
1 comments
▲
1
Retrieval-Aware Distillation for Transformer-SSM Hybrids
(arxiv.org)
by readitalready |
view
|
0 comments
▲
1
Transformer-Based Memory Forecasting
(novice.media)
by kirillzubovsky |
view
|
0 comments
▲
1
Transformers.js v4 Preview: Now Available on NPM
(huggingface.co)
by ibobev |
view
|
0 comments
▲
1
Transfer learning and Transformer models (ML Tech Talks) [video]
(youtube.com)
by onurkanbkrc |
view
|
0 comments
▲
1
So whats the next word, then? Almost-no-math intro to transformer models
(matthias-kainer.de)
by oesimania |
view
|
0 comments
▲
1
End-to-End Transformer Acceleration Through Processing-in-Memory Architectures
(arxiv.org)
by PaulHoule |
view
|
0 comments
▲
1
Generative Pen-Trained Transformer
(theodore.net)
by Twarner |
view
|
0 comments
▲
1
LLatte: Scalable Transformers for Ads at Meta
(twitter.com)
by LatteMetaAI |
view
|
0 comments
▲
1
LLatte: Scalable Transformers for Ads at Meta
(twitter.com)
by LatteMetaAI |
view
|
1 comments