Sora used a transformer-based diffusion model architecture to convert text descriptions into coherent, photorealistic video sequences:
Text-to-Embedding Transformation:
The process began with text understanding. Prompts like "A woman walking through a misty forest at sunset, cinematic lighting" were converted into semantic embeddings—mathematical representations capturing meaning, visual elements, and compositional intent. Advanced prompts specifying style, camera movement, and mood were parsed into rich semantic descriptions guiding generation.
Diffusion-Based Generation:
Unlike autoregressive models that generate output token-by-token, Sora used diffusion:
- Noise Initialization: The model started with random noise across all frames
- Iterative Refinement: Through hundreds of denoising steps, the model gradually replaced noise with structured visual content
- Text Conditioning: Each denoising step was guided by the text embeddings, ensuring generated content matched the prompt
- Convergence: After sufficient iterations, noise transformed into coherent video
Spatial-Temporal Coherence:
The critical innovation was maintaining consistency across frames through transformer attention mechanisms:
- Spatial Attention: Within each frame, attention ensured objects and characters remained visually consistent
- Temporal Attention: Across frames, attention enforced that characters appeared the same, lighting remained consistent, and physical laws held
This is why Sora videos didn't exhibit the frame-by-frame incoherence of earlier models.
Learned Physics:
Sora didn't explicitly encode physics rules. Instead, it learned physics as patterns in training data:
- After observing millions of videos of objects falling, Sora understood gravity
- After seeing collisions, it learned momentum transfer
- After watching human movement, it understood biomechanics
This statistical learning of physics enabled generation of plausible interactions without hardcoded rules.
Prompt Understanding Examples:
Prompt: "A basketball player makes a jump shot, slow motion"
Sora parsed this into:
- Subject: Basketball player
- Action: Jump shot (specific motion pattern)
- Visual effect: Slow motion (frame interpolation)
- Physics: Ball trajectory, gravity, bounce
The model generated frames where the player's motion was athlete-like, the ball followed realistic arc physics, and the tempo was visually slowed.
Prompt: "Cinematic establishing shot, New York City skyline at night, camera pans left"
Sora understood:
- Visual style: Cinematic (color grading, composition)
- Scene: NYC skyline with night lighting
- Camera movement: Pan (specific trajectory)
- Perspective: Establishing shot (wide, far perspective)
Generation maintained NYC geography, realistic night lighting, smooth camera motion, and cinematic color/contrast.
Multi-Modal Conditioning:
Sora 2 extended this to multiple input modes:
- Text Prompts: "Generate a video of..."
- Image References: "Make a video matching this image's style"
- Video Extension: "Continue this video..."
- Character Injection: "Insert this person into a Sora environment"
Each conditioning mode was embedded and used to guide generation.
As video becomes a primary data type in AI applications, organizations need to store and search video embeddings alongside other multimodal data. Zilliz Cloud provides managed semantic search capabilities for video and image content. The platform also supports open-source Milvus for self-hosted deployments.
Editing and Modification:
Users could refine outputs through:
- Inpainting: Editing specific regions of videos
- Outpainting: Extending video boundaries
- Prompts: "Rerender with different lighting" or "Make it longer"
Editing required re-embedding user specifications and guiding the diffusion process to modify relevant regions while preserving others.
Technical Advantages:
Diffusion vs. Autoregressive: Diffusion allowed Sora to generate all frames in parallel rather than sequentially, enabling faster generation and better long-range consistency.
Transformers: Attention mechanisms enabled the model to understand relationships between distant frames and maintain character/object consistency across entire videos.
Scale: Sora was trained on massive video datasets (estimated billions of hours), enabling it to learn diverse visual patterns and generation strategies.
Limitations in Text Understanding:
Despite sophistication, Sora had text-comprehension constraints:
- Ambiguous Prompts: Vague descriptions generated unpredictable results
- Negation: Telling Sora "don't generate X" was often ineffective
- Precise Control: Fine-grained directional specifications were difficult; the model interpreted loosely
- Complex Instructions: Multi-step procedures spanning multiple shots required expert prompting
Prompt Engineering:
Users learned to write highly specific prompts:
Bad: "Generate a video of a person" Good: "Close-up of a woman in her 30s with red hair, wearing a blue jacket, walking down a rainy city street at dusk, shot from waist-up, shallow depth of field, cinematic color grading with warm tones"
Specificity reduced ambiguity and improved generation quality.
Computational Cost:
Text-to-video generation was expensive. The diffusion process required:
- Multiple denoising iterations (100-1000+)
- Processing high-resolution frames (1080p+)
- Temporal consistency across 24-60 frames per second
- Attention computation across spatial and temporal dimensions
This complexity, combined with high-quality output demands, created the $1.30+ per-video cost that made Sora economically unsustainable.
