LoGeR – 3D reconstruction from extremely long videos (DeepMind, UC Berkeley)
helloplanets
137 points
27 comments
March 10, 2026
Related Discussions
Found 5 related stories in 49.2ms across 3,471 title embeddings via pgvector HNSW
- Helios: Real real-time long video generation model tzury · 13 pts · March 09, 2026 · 50% similar
- Learning athletic humanoid tennis skills from imperfect human motion data danielmorozoff · 137 pts · March 15, 2026 · 47% similar
- Datasets for Reconstructing Visual Perception from Brain Data katsee · 55 pts · March 05, 2026 · 45% similar
- Mamba-3 WarmWash · 41 pts · March 18, 2026 · 44% similar
- Show HN: A 3D Body Scan for Nine Cents – Without SMPL arkadiuss · 11 pts · March 30, 2026 · 43% similar
Discussion Highlights (6 comments)
msuniverse2026
Truly don't understand what is happening in the heads of these researchers. Can't they see how the main use of this is going to be mass surveillance?
IshKebab
Very cool. Doesn't seem like they've actually released the code: > This is a reimplementation of LoGeR; complete code and models will be released upon approval. I don't understand why it's a re implementation either? I would guess it's "research" code anyway so not really usable unless you are an expert.
Dead_Lemon
What is the actual objective of this, is it solving an issue or creating a solution to a problem, that is still to be determined? It seems like a lot of energy to replicate a lidar mapping system. It's not like you can expect accurate dimensions from this approximate guess work, excluding the expected hallucinations adding to inaccuracy.
tmilard
Very interesting paper. I can see street-view using it to perfect the 3D analysing of the photo-video they catch with there google-car. What a wonderfull time we are living in ! Specificaly in the Video to 3D reconstruction. Every month, a new brick is put in place.Super
_fw
This is like something straight out of Cyberpunk 2077 - the braindances investigation scenes.
priowise
very interesting direction. One thing I’m curious about with extremely long videos is how you handle temporal drift over time. Do you periodically re-anchor the reconstruction or rely purely on accumulated frame consistency?