by divyaprakash
3 subcomments
- I built this because I was tired of "AI tools" that were just wrappers around expensive APIs with high latency. As a developer who lives in the terminal (Arch/Nushell), I wanted something that felt like a CLI tool and respected my hardware.
The Tech:
GPU Heavy: It uses decord and PyTorch for scene analysis. I’m calculating action density and spectral flux locally to find hooks before hitting an LLM.
Local Audio: I’m using ChatterBox locally for TTS to avoid recurring costs and privacy leaks.
Rendering: Final assembly is offloaded to NVENC.
Looking for Collaborators: I’m currently looking for PRs specifically around: Intelligent Auto-Zoom: Using YOLO/RT-DETR to follow the action in a 9:16 crop.
Voice Engine Upgrades: Moving toward ChatterBoxTurbo or NVIDIA's latest TTS.
It's fully dockerized, and also has a makefile. Would love some feedback on the pipeline architecture!
by HeartofCPU
2 subcomments
- It looks like it’s written by a LLM
- Can I use this for other use cases instead of game videos?
I want to create film-style scenes, cinematic elements, and smooth motion effects.
I’m also thinking of deploying it as a SaaS and using it for video creation features in my app: https://picxstudio.com/
by wasmainiac
2 subcomments
- This does not seem local first. Misleading.
Regardless, we need more tools like this to speed social media towards death.
- Wow, great job.
I did smth similar 4 years ago with YOLO ultralytics.
Back then I used chat messsges spike as one of several variables to detect highs and fails moments. It needed a lot a human validation but was so fun.
Keep going
by 8organicbits
1 subcomments
- What's the intended use case for this? It seems like you'd create slop videos for social media. I'd love to see more AI use cases that aren't: uninteresting content people would prefer to avoid.
by Huston1992
0 subcomment
- big fan of the 'respects my hardware' philosophy. i feel like 90% of ai tools right now are just expensive middleware for openai, so seeing something that actually leverages local compute (and doesn't leak data) is refreshing
- How much memory do you need locally? Is a rtx 3090 with 24gb enough?