Back to Blog
EngineeringMarch 22, 2026·8 min read

How Velora's New Engine Reduces Render Times by 40%

V

Team Velora

Engineering

40%

Render Speed Increase

0.8s

Average Frame Time

10×

Concurrent Jobs

The Problem We Set Out to Solve

When we launched Velora, our rendering pipeline was built on a conventional sequential processing model — each scene was processed one element at a time: script → voiceover → imagery → compositing → export. For short-form content, this felt fast enough. But as creators began requesting 15- and 30-minute productions, cracks appeared. What should have been a 12-minute render was routinely taking 35–45 minutes per job. Creator satisfaction scores dropped. Compute costs ballooned.

We knew we needed a fundamental re-architecture — not a patch, but a ground-up rethink of how Velora's engine parallelizes work.

Introducing the Velora Parallel Scene Graph (PSG)

The core breakthrough is what we internally call the Parallel Scene Graph. Instead of processing scenes sequentially, PSG constructs a dependency graph of every scene in a project at the moment you submit a job. Scenes with no interdependencies — which typically accounts for 70–80% of all scenes in a video — are immediately dispatched to separate GPU workers simultaneously.

A 30-minute video that previously had 90 scenes processed one-by-one now fans out across our cluster in waves. The first wave processes ~60 independent scenes in parallel. The second wave handles the remaining scenes that depend on earlier output. Total wall-clock render time collapses dramatically.

Voiceover Pre-Synthesis Pipeline

A second major bottleneck was voiceover generation. Previously, TTS synthesis happened after imagery was composited — a sequential dependency that stalled the entire pipeline waiting on ElevenLabs API calls.

In the new engine, voiceover synthesis is kicked off in parallel with scene imagery generation the moment a job is submitted. By the time the final visual frame is composited, the audio track is already fully rendered and waiting. This alone shaved an average of 4.2 minutes off every job regardless of length.

Smart Frame Caching

The third pillar of the new engine is a perceptual frame cache. If two scenes share very similar visual elements — same background, same avatar model, same lighting — Velora now reuses pre-computed frame data instead of re-rendering from scratch. Our perceptual hashing algorithm compares scenes at the semantic level, not pixel-by-pixel, which means meaningful re-use even when scenes aren't identical.

In real-world creator projects, average cache hit rates are sitting at 22–31%, which translates directly into skipped compute cycles and faster delivery.

Real-World Results

After rolling out to 100% of production traffic over a two-week canary period, the numbers confirm what our internal testing predicted:

  • Average render time for 5-min videos: 3.1 min → 1.9 min (−39%)
  • Average render time for 30-min videos: 41 min → 24.4 min (−40%)
  • P99 tail latency reduced from 68 min → 38 min
  • GPU utilization efficiency increased from 61% → 89%
  • Compute cost per job reduced by ~28% despite higher parallelism

What's Next

PSG v2 is already in development. The next iteration will introduce predictive pre-warming — spinning up GPU workers before a creator even hits "Generate" based on behavioral signals. We're also exploring streaming delivery, where completed scenes begin downloading to the creator's browser while later scenes are still rendering. Our target is sub-60-second delivery for 5-minute videos by Q3 2026.

Engineering blogs like this one are part of our commitment to radical transparency with the creator community. If you're a developer or infrastructure engineer who's passionate about these challenges — we're hiring.

🎬
Velora Support
Online
✨ Upgrade to Advanced AI for real-time web research
Hi! 👋 How can I help you today? Ask me anything about Velora, pricing, or your account.