Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Congrats on the launch. The client-side rendering on WebCodecs/WebGL2 is impressive — that alone is a hard technical problem most teams avoid.

One thing I've been thinking about building in this space: there's a fundamental split between understanding what to edit (where VLMs/agents shine) and executing the edit precisely (where you need deterministic operations, not model inference).

Most "AI video editors" blur these two together — they use the same probabilistic approach for both understanding and execution. But when a user says "cut the first 3 seconds and add a 0.5s crossfade," that shouldn't go through a model. That should be a precise, repeatable operation.

The Cursor analogy in your roadmap is apt — Cursor works because it predicts intent but executes through deterministic code transforms, not by asking an LLM to write the whole file. Same principle applies to video.

Curious how you handle the boundary between agent-proposed edits and deterministic timeline operations under the hood?

 help



- models are good at precise edits, we have an internal benchmark for the same

- plan mode / agent mode is something that'd be helpful in deciding / executing




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: