The eighth-generation TPU: An architecture deep dive
meetpateltech
67 points
10 comments
April 22, 2026
Related Discussions
Found 5 related stories in 64.9ms across 5,335 title embeddings via pgvector HNSW
- Our eighth generation TPUs: two chips for the agentic era xnx · 427 pts · April 22, 2026 · 77% similar
- Google unveils chips for AI training and inference in latest shot at Nvidia wslh · 12 pts · April 22, 2026 · 52% similar
- TurboQuant: Redefining AI efficiency with extreme compression ray__ · 509 pts · March 25, 2026 · 50% similar
- NanoGPT Slowrun: 10x Data Efficiency with Infinite Compute sdpmas · 122 pts · March 19, 2026 · 50% similar
- Intel's make-or-break 18A process node debuts for data center with 288-core Xeon vanburen · 270 pts · March 03, 2026 · 49% similar
Discussion Highlights (5 comments)
zshn25
Splitting TPUs into dedicated training vs inference chips feels like an admission that the bottleneck has shifted from FLOPs to memory bandwidth + latency. Are future gains to come more from memory/system design than raw compute scaling? What’s that saying about Scaling laws?
ricardo81
dupe https://news.ycombinator.com/item?id=47862497
ttul
No matter how smart your large language model is, if you can’t find the energy to power it, it won’t run. I could imagine Google winning merely because their chips are more efficient. Of course, the other labs are capable of making chips, but Google has been doing it for years.
speedping
2.764 petabytes of HBM per 8i? So that's where all the RAM went.
juancn
Super interesting but it's so damn hard to find any detail. I would love to see an instruction set reference for one of these, all you have is hardware architectural diagrams or high level APIs.