CERN uses ultra-compact AI models on FPGAs for real-time LHC data filtering
TORcicada
306 points
138 comments
March 28, 2026
Related Discussions
Found 5 related stories in 138.7ms across 3,471 title embeddings via pgvector HNSW
- The Inference Shift – How Cheap Chips Could Put Frontier AI in Everyone's Hands arcanemachiner · 11 pts · March 31, 2026 · 50% similar
- Switzerland hosts 'CERN of semiconductor research' teleforce · 32 pts · April 03, 2026 · 48% similar
- TurboQuant: Redefining AI efficiency with extreme compression ray__ · 509 pts · March 25, 2026 · 48% similar
- FFmpeg at Meta: Media Processing at Scale sudhakaran88 · 239 pts · March 09, 2026 · 47% similar
- Executing programs inside transformers with exponentially faster inference u1hcw9nx · 17 pts · March 12, 2026 · 47% similar
Discussion Highlights (20 comments)
rakel_rakel
Hey Siri, show me an example of an oxymoron! > CERN is using extremely small, custom large language models physically burned into silicon chips to perform real-time filtering of the enormous data generated by the Large Hadron Collider (LHC).
100721
Does anyone know why they are using language models instead of a more purpose-built statistical model? My intuition is that a language model would either be overfit, or its training data would have a lot of noise unrelated to the application and significantly drive up costs.
serendipty01
Might be related: https://www.youtube.com/watch?v=T8HT_XBGQUI (Big Data and AI at the CERN LHC by Dr. Thea Klaeboe Aarrestad) https://www.youtube.com/watch?v=8IZwhbsjhvE (From Zettabytes to a Few Precious Events: Nanosecond AI at the Large Hadron Collider by Thea Aarrestad) Page: https://www.scylladb.com/tech-talk/from-zettabytes-to-a-few-...
randomNumber7
Does string theory finally make sense when we ad AI hallucinations?
quijoteuniv
A bit of hype in the AI wording here. This could be called a chip with hardcoded logic obtained with machine learning
seydor
cern has been using neural networks for decades
intoXbox
They used a custom neural net with autoencoders, which contain convolutional layers. They trained it on previous experiment data. https://arxiv.org/html/2411.19506v1 Why is it so hard to elaborate what AI algorithm / technique they integrate? Would have made this article much better
WhyNotHugo
Intuitively, I’ve always had an impression that using an analogue circuit would be feasible for neural networks (they just matrix multiplication!). These should provide instantaneous output. Isn’t this kind of approach feasible for something so purpose-built?
mentalgear
That's what Groq did as well: burning the Transformer right onto a chip (I have to say I was impressed by the simplicity, but afterwards less so by their controversial Kushner/Saudi investment) .
v9v
Do they actually have ASICs or just FPGAs? The article seems a bit unclear.
Janicc
I think chips having a single LLM directly on them will be very common once LLMs have matured/reached a ceiling.
nerolawa
the fact that 99% of LHC data is just gone forever is insane
amelius
When is the price of fabbing silicon coming down, so every SMB can do it?
Surac
Very important! This is not a LLM like the ones so often called AI these days. Its a neural network in a FPGA.
armcat
Not on the same extreme level, but I know that some coffee machines use a tiny CNN based model locally/embedded. There is a small super cheap camera integrated in the coffee machine, and the model does three things: (1) classifies the container type in order to select type of coffee, (2) image segmentation - to determine where the cup/hole is placed, (3) regression - to determine the volume and regulate how much coffee to pour.
konradha
How are FPGAs "bruned into silicon"? Would be news to me that there are ASICs being taped out at CERN
quantum_state
CERN has been doing HEP experiments for decades. What did it use before the current incarnation of AI? The AI label seems to be more marketing and superficial than substantial. It’s a bit sad that a place like CERN feels the need to make it public that it is on the bandwagon.
TORcicada
Thanks for the thoughtful comments and links really appreciated the high-signal feedback. We've updated the article to better reflect the actual VAE-based AXOL1TL architecture (variational autoencoder for anomaly detection). Added the arXiv paper and Thea Aarrestad's talks to the Primary Sources.
jurschreuder
I've got news for you, everybody with a modern cpu uses this, which use a perceptron for branch prediction.
logicallee
I hope they have good results and keep all the data they need, and identify all the interesting data they're looking for. I do have a cautionary tale about mini neural networks in new experiments. We recently spent a large amount of time training a mini neural network (200k parameters) to make new predictions in a very difficult domain (predicting specific trails for further round collisions in a hash function than anyone did before.) We put up a spiffy internal dashboard[1] where we could tune parameters and see how well the neural network learns the existing results. We got to r^2 of 0.85 (that is very good correlation) on the data that already existed, from other people's records and from the data we solved for previously. It showed such a nicely dropping loss function as it trained, brings tears to the eye, we were pumped to see how it performs on data it didn't see before, data that was too far out to solve for. So many parameters to tune! We thought we could beat the world record by 1 round with it (40 instead of 39 rounds), and then let the community play with it to see if they can train it even better, to predict the inputs that let us brute force 42 round collisions, or even more. We could put up a leaderboard. The possiblities were endless, all it had to do was do extrapolate some input values by one round. We'd take the rest from there with the rest of our solving instrastructure. After training it fully, we moved on to the inference stage, trying it on the round counts we didn't have data for! It turned out ... to have zero predictive ability on data it didn't see before. This is on well-structured, sensible extrapolations for what worked at lower round counts, and what could be selected based on real algabraic correlations. This mini neural network isn't part of our pipeline now. [1] screenshot: https://taonexus.com/publicfiles/mar2026/neural-network.png