Zane Chen
Zane Chen @chenzeling4 ·
World model for video-to-video inference. InSpatio-World: 14B and 1.3B models. Transform video with AI - depth estimation, diffusion, text-to-video capabilities. HuggingFace ready. ⭐ 596 stars #AI #VideoGeneration #WorldModel
1
5
Quant Signals
Quant Signals @QuantSignalsXYZ ·
18 days. The World Model is coming. 5,000 AI agents simulating the social battlefield of trading. April 15, 2026. QS 1-year anniversary. Be ready. #QuantSignals #WorldModel
16
Yu-Cheng Chou
Yu-Cheng Chou @johnson111788 ·
CVPR 2026🎥We built a model that lets you fly through a generated video world. Not just generating frames — but maintaining a consistent 3D world under complex camera motion. Code, ckpt, and even the data pipeline are all open-sourced ↓ #AI #worldmodel #videogen #cvpr #drone9
13
47
21.8K
PlanX
PlanX @PlanX_DEX ·
Deeper understanding leads to better execution PlanX-Execution Beyond Human #AI #LLM #WorldModel #Xgent
Lex Lex @PlanX_Lex ·
LLMs vs World Models — what’s the real difference? Most people treat them as the same thing. They are not. They solve fundamentally different problems. 1. LLMs (Large Language Models) Core function: pattern completion in token space (1) Learn statistical relationships between next token given context (3) Operate on language, not reality They are excellent at: (1) Natural language understanding & generation (2) Code synthesis (3) Structured output (JSON, workflows, APIs) (4) Interface between humans and systems LLMs are best understood as: a universal interface layer for cognition 2. World Models Core function: modeling state transitions of a system (1) Learn how environments evolve over time (2) Capture dynamics, causality, and feedback loops (3) Operate on states, not tokens They are used for: (1) Simulation and planning (2) Control systems and robotics (3) Financial market dynamics (4) Risk and environment modeling World Models are: predictive representations of how the world changes 3. Key Difference LLMs answer: → “Given this context, what should be said next?” World Models answer: → “Given this state, what will happen next?” 4. Where each fits Use LLMs when: (1) The problem is semantic (2) You need interpretation, structuring, or communication (3) Output is consumed by humans or systems Use World Models when: (1) The problem is dynamic (2) You need prediction under uncertainty (3) Decisions depend on system evolution 5. In advanced systems, they are not competitors — they are complements. A powerful stack looks like: (1) LLM → understands intent & constructs strategy (2) World Model → simulates outcomes & evaluates risk (3) Execution system → acts on validated decisions Example (Trading Systems) (1) LLM: translates intent → strategy structure (2) World Model: evaluates regime, risk, expected behavior (3) Engine: executes under constraints 6. Take away LLMs model language. World Models model reality. Confusing the two leads to fragile systems. Combining them correctly leads to intelligent ones. #LLM #worldmodel #AI #Xgent
181
Quant Signals
Quant Signals @QuantSignalsXYZ ·
The future of trading isn't a signal. It's a simulation. 5000 AI agents modeling the entire market battlefield in real-time. QS V5 World Model. April 15. 🌍⚡ #QuantSignals #WorldModel
22
Hod Lipson
Hod Lipson @hodlipson ·
We found evidence of an emergent "Self" in a robot learning in a nonstationary environment. There is a lot of talk about robot #WorldModel learning, but the #SelfModel is where the magic really happens. See the paper here: lnkd.in/eycu5GEs
2
35
9.7K
Repandre.com
Repandre.com @repandrecom ·
Intéressant @ylecun #worldmodel
Alex Ruben Alex Ruben @rubenxela ·
Recent PNAS paper from Coimbra & Carnegie Mellon: the brain builds actions from recombinable kinematic synergies, like words from letters Seems our gyrus supramarginal has been doing compositional world models all along Ça va dans le bon sens @ylecun ? pnas.org/doi/10.1073/pn…
Object-directed action representations are componentially built in parietal cortex | PNAS

The inferior parietal lobule supports action representations that are necessary to grasp and use objects in a functionally appropriate manner [S. H...

From pnas.org
4
Marouene Chaibi
Marouene Chaibi @Marouenechaibi ·
#AI #WorldModel
alphaXiv alphaXiv @askalphaxiv ·
Yann LeCun and his team can't stop cooking "LeWorldModel: Stable End-to-End Joint-Embedding Predictive Architecture from Pixels" One of the biggest bottlenecks of JEPA is they are hard to train, and this new research changes that. They propose LeWorldModel, which shows that a usable world model directly from raw pixels end-to-end. Sitting at 15M parameters, they made it without needing heuristics and avoiding anti-collapse hacks while staying competitive and planning up to 48x faster. Making JEPA based modeling much more accessible, cheaper, and stabler.
40