Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
Show HN: Script Snap – Extract code from videos (script-snap.com)
14 points by liumw1203 3 days ago | hide | past | favorite | 7 comments
Hi HN, I'm lmw-lab, the builder behind Script Snap.

The Backstory: I built this out of pure frustration. A while ago, I was trying to figure out a specific configuration for a project, and the only good resource I could find was a 25-minute YouTube video. I had to scrub through endless "smash the like button" intros and sponsor reads just to find a single 5-line JSON payload.

I realized I didn't want an "AI summary" of the video; I just wanted the raw code hidden inside it.

What's different: There are dozens of "YouTube to Text" summarizers out there. Script Snap is different because it is explicitly designed as a technical extraction engine.

It doesn't give you bullet points about how the YouTuber feels. It scans the transcript and on-screen visuals to extract specifically:

Code snippets

Terminal commands

API payloads (JSON/YAML)

Security warnings (like flagging sketchy npm installs)

It strips out the "vibe" and outputs raw, formatted Markdown that you can copy straight into your IDE.

Full disclosure on the launch: Our payment processor (Stripe) flagged us on day one (banks seem to hate AI tools), so I've pivoted to a manual "Concierge Alpha" for onboarding. The extraction engine is fully operational, just doing things the hard way for now.

I'd love to hear your thoughts or harsh feedback on the extraction quality!

 help



Very cool!

It would be a very interesting project to extract text from a bunch of youtubers’ videos and put it in a (private) github repo to let github’s secret scanner run over it: https://docs.github.com/en/code-security/concepts/secret-sec...

Could detect API keys they forgot to redact (hence the suggestion to use a private repo.)


Not sure if this is just decoration for your demo but this fake console output shown on the landing page and every video transcribed feels weird.

04:03:26> [STATUS] Deepgram Nova-3 Engine docked. Ready. 04:03:26> [LOAD] Mounting 'Tech Stack Dictionary v4.0'... 04:03:26> [INFO] Loaded modules: CUDA, Ada Lovelace, M4 Pro. 04:03:26> [STREAM] Ingesting audio stream (48kHz)... 04:03:26> [WARN] Ambiguity detected at 04:12: 'Pie Torch'

The video I submitted (https://youtu.be/0q6Ujn_zNH8) definitely doesn’t mention “Pie Torch”.

And I also haven’t seen any actual output except for “ ◧ Processing...”, 10 minutes after submitting the video. A progress bar of some sort or maybe timestamp currently being analyzed would be nice.

Also would this work in situations where the video creator is scrolling through a long code file in their IDE? Would the engine show it as one file multiple separate sections?

Apart from that, cool project!


Hi beans42, thanks for the stress test! You completely caught me red-handed on the UI, but I actually have the raw output for your video now.

The fake console: You are 100% right. The terminal is currently a hardcoded React animation to set the "vibe" while the actual worker chugs in the background. Seeing a "Pie Torch" warning pop up during a hardcore Ben Eater video about ca65/ld65 and 6502 assembly is objectively hilarious. I am ripping that fake logger out today and piping in real status events.

The 10-minute hang: A 21-minute dense technical video is exactly what breaks my current queue. The background worker actually did finish extracting the ld65 memory map configs and the BIOS segments, but the frontend WebSocket connection silently dropped, leaving you stuck on "Processing...". A real progress indicator is my P1 task right now.

To prove the engine actually works: I pulled the completed Markdown extraction for your Ben Eater video from the database. It successfully pulled the exact linker config and the BIOS segment code. I've hosted the raw output here so you can see what it should have returned after 10 minutes: https://gist.github.com/lmw-dev/d9f276cc0d90c05b7bab5ec0758d...

Scrolling files: You hit on the hardest problem. Right now, it extracts them as separate code blocks based on timestamps. Stitching scrolling frames into a single, deduplicated file without the AI hallucinating is my next major research hurdle.

Seriously, thank you for the roast. This is exactly the kind of feedback I needed to stop focusing on the "vibe" and fix the actual plumbing.


You know this reply is transparently LLM generated, right?

It's what happens when you paste the HN feedback into your vibe session as if you wrote it, then copy the response back here.


Guilty as charged.

English isn't my native language (I'm based in China), so I use an LLM to translate my messy notes and polish my replies. I get nervous about messing up the technical nuances when talking to guys like you and beans42 on here.

But to be clear: the actions are 100% human and real. The raw extraction in that Gist is actual backend output I queried from the DB, and the fake UI logger really is getting ripped out today.

Appreciate the callout—I clearly need to tune my prompt to sound less like a PR bot.


"Use clear plain English for grade level 10 avoiding jargon and buzzwords." goes a long way.

Might also toss "bullets" in the things to avoid; the endless bullets are out of hand. :-)


That is actually a great prompt tip. Adding "under penalty of death, no bullet points" to my system prompt right now. Appreciate the advice and the understanding! :-)



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: