Discussion about this post

User's avatar
Nir Ronen's avatar

Great overview — I really appreciated the thoughtful comparison between scaling in LLMs and recent trends in robotics / physical AI.

This resonates strongly with patterns we’ve seen repeatedly in semiconductors over the last 20+ years. “One size fits all” compute tends to break down once you hit hard real-time loops, tight power envelopes, and latency/jitter constraints — especially in embedded systems.

GPUs (and NVIDIA’s ecosystem in particular) will clearly dominate for a long time due to tooling, legacy code, and scale, but historically we’ve seen specialized solutions emerge around the edges where general-purpose architectures struggle. Curious how you see this playing out as perception and control loops get tighter.

Bharath Suresh's avatar

This is a really detailed post, Avik. I'm quite new to this, so this might be a naive question: are world models being built as foundation models + some domain/environment specific post training/fine tuning? I'm trying to draw a comparison between language models which have one foundational model like GPT, but many domain specific implementations that use the foundation model.

2 more comments...

No posts

Ready for more?