431
OpenAI introduces Sora, its text-to-video AI model
(www.theverge.com)
This is a most excellent place for technology news and articles.
I think we're talking about different kinds of implementations.
One being an ai generated 'video' that is interactive, generating new frames continuously to simulate a 3d space that you can move around in. That seems pretty hard to accomplish for the reasons you're describing. These models are not particularly stable or consistent between frames. The software does not have an understanding of the physical rules, just how a scene might look based on it's training data.
Another and probably more plausible approach is likely to come from the same frame generation technology in use today with things like DLSS and FSR. I'm imagining a sort of post-processing that can draw details on top of traditional 3d geometry. You could classically render a simple scene and allow ai to draw on top of the geometry in realtime to sort of fake higher levels of detail. This is already possible, but it seems reasonable to imagine that these tools could get more creative and turn a simple blocky undetailed 3d model into a photo-realistic object. Still insanely computationally expensive but grounding the AI with classic rendering to stabilize it's output could be really interesting.