Chris Messina

X-Pilot - Explain anything accurately, from document to video course

X-Pilot turns docs into video courses for people who explain anything and can't risk hallucinations. Every visual is rendered programmatically via Remotion in isolated sandboxes — deterministic, not generative. Formulas, diagrams, and code stay accurate.

Add a comment

Replies

Best
Bian Heshan

Hi Product Hunt — I’m Heshan, founder of X-Pilot.

After leaving Baidu Apollo, I built 3 edtech companies (1M+ users total) and kept seeing the same issue: the people with the deepest knowledge are often the least equipped to turn it into video. Hand a professor a video editor and everything slows down. I started calling it the “Expert Paradox.”

X-Pilot is our attempt to solve it: upload a document, and X-Pilot generates an accurate, multi-module video course—complete with a syllabus, learning objectives, and animated visuals (diagrams, rendered formulas, code walkthroughs) you can publish.

A key difference vs. HeyGen/Synthesia: those are talking-head/avatar script readers. X-Pilot focuses on knowledge visualization. Every visual is rendered programmatically via Remotion in isolated sandboxes—deterministic code, not generative visuals—so if your doc says 2+2=4, the video shows 2+2=4.

Free to start (no credit card).

I’d love feedback from anyone who’s tried to turn a document into a course: what broke for you—structuring, visuals, editing time, accuracy, or distribution?

zijiaWang

Hey Product Hunt 👋 I’m one of the devs behind X-Pilot.

We built this because we kept hitting the same issue: AI video tools look great, but they hallucinate—especially on code, charts, and formulas. For educators and trainers, that’s a dealbreaker.

So we went a different direction. X-Pilot takes your PDF / PPT / Markdown and renders everything deterministically in a sandbox (via Remotion), instead of “imagining” the video.

That means what you write is exactly what shows up—no hallucinations, especially for technical content.

It's been a massive engineering challenge to build this rendering engine from scratch, but seeing it save course creators hours of manual editing has been incredibly rewarding for our team.

We’ve added free credits so you can try the real workflow. I’ll be around in the comments—would love feedback, ideas, or any questions.

Mykyta Semenov 🇺🇦🇳🇱

Great idea! Is Ukrainian supported? Or only English?

Bian Heshan

@mykyta_semenov_ Currently, the video supports 9 languages. If you require support for Ukrainian, we will immediately arrange for our technical team to add it.

Klara Minarikova

Curious what the actual workflow looks like for a non-technical creator. Upload a doc — and then what? How many decisions do I need to make before I have something I'd actually want to publish?

Lizi Xiao

@klara_minarikova 
Thanks for the question — here’s what it actually looks like for a non‑technical creator.

1.Before you upload

You can set a few “defaults” up front so the first draft matches how you publish: output language, visual style, a custom Brand Kit (colors/typography/logo rules), and voice (voice model).

2.After you upload a document

The agent reads your document end‑to‑end—including images, equations, formulas, and chart/table data—and turns what’s in the doc into visual elements on the timeline (graphics, on‑screen math, charts, and other visuals) so the video reflects the source material, not just a plain narration.

3.If it’s not quite right

You don’t need to “operate software.” Just talk to it in natural language to request changes (tone, pacing, emphasis, a segment rewrite, etc.) and iterate until you’re comfortable publishing.

4.How many decisions before publish‑ready?

If your document and goal are clear, most creators reach something they’re happy to publish in about 1–3 rounds of natural‑language conversation—for example, small follow‑up tweaks to tone, pacing, emphasis, or a specific segment. You’re giving plain‑language feedback, not working through a long technical checklist.

Lizzy Lee

AI video tools usually mess up charts. Does this fully avoid that?

Lizi Xiao

@lizzy_leeeee We build visuals as structured components in Remotion (code‑driven timelines), not by “guessing” a chart from a blurry screenshot. That means charts and diagrams (including things like flowcharts) are treated as first‑class UI elements: layout, labels, axes, and connections stay consistent across frames, and they’re much less likely to warp, smear, or drift the way purely generative video pipelines often do.

So it doesn’t rely on “redrawing the chart from pixels”, which is usually what breaks charts in typical AI video tools. That said, no system can promise perfection for every edge case—if the underlying data or instructions are ambiguous, you may still want a quick conversational tweak—but the approach is designed to avoid the common ‘messed‑up chart’ failure mode by keeping charts in a stable, component‑based representation end‑to‑end.

Antler Kaku

I'm a bit curious about the generation speed and cost. If it's stable, there should be a market for educational content.

Lizi Xiao

@antler_kaku The generation speed largely depends on the user's video requirements. Current video pre-settings allow for settings of 1-3 minutes, 3-10 minutes, and 10-20 minutes. As the selected duration increases, the depth of the agent's planning and explanation also increases, along with the number of visual components, leading to increased generation time and cost.

I must admit that our product still has much room for improvement. We need user feedback to guide our upgrades, and I thank x-pilot users for their support.

Samir Asadov

The "Expert Paradox" is real — I see it constantly in finance. The best modelers often produce the worst training materials because they skip steps they've internalized. X-Pilot solving this for video is compelling. I teach Excel for financial modelling on Udemy and structuring course content that works for both beginners and intermediate practitioners is genuinely hard. A tool that can take a document and render accurate multi-module content around it would be a game changer for technical finance education. Congrats on the launch!

Bian Heshan

@samir_asadov thx,I would really appreciate receiving your feedback on your experience using it.

lumm

does this actually prevent hallucinations in the formulas and diagrams, or just reduce them? been burned before by AI-generated math visuals that looked right but werent. the Remotion approach sounds promising tho

Bian Heshan

@lumm To ensure absolute accuracy, we have implemented a four-layer verification process—spanning from the data source to code generation, animation rendering, and finally, the export stage.

Daulen Zhangabylov

I wanted to try first but could not find a free plan

Bian Heshan

@daulen_zhangabylov Currently, you are allotted two free video generations. You can simply log in via Google to generate your video directly. The first two generations are free; however, these free videos are limited to the default voiceover and will include a watermark.

Charlene Chen

Congrats! Does it support file input like image and .docx files, besides text?

Lizi Xiao

@charlenechen_123 Thank you! Yes, we currently support document uploads in .docx, .doc, .pdf, .ppt, and .pptx formats. However, we do not support image files at this moment.

123
Next
Last