A tool that converts a slide deck into a video, complete with your voice narration. Support multiple languages.
Tested on Ubuntu 20.04.
ffmpeg:
sudo apt-get install ffmpegpip if you haven't already.git clone [email protected]:Changochen/slide-to-video.git
cd slide-to-video
pip install .slide-to-videoNEWSLIDE.slide-to-video --model MODEL_NAME --slide slide --script script --output-dir OUTPUT_PATH --config ADDITIONAL_CONFIG.yamlTo use a local voice model:
slide-to-video --model local --slide example/slide.pdf --script example/script.txt --voice example/sample.mp3 --output-dir outputA final video will be generated in the OUTPUT_PATH directory as output.mp4.
For more options, including adjusting speech speed, run:
slide-to-video --helpCurrently Supported Model:
Currently Supported Languages: 'en', 'es', 'fr', 'de', 'it', 'pt', 'pl', 'tr', 'ru', 'nl', 'cs', 'ar', 'zh-cn', 'hu', 'ko', 'ja', 'hi'
After generating the video, the output directory will contain a project.yaml file, which helps skip the generation of unchanged content. If inputs remain the same, the tool skips the video generation process.
If you modify the slide, script, or settings (like speech speed), the tool regenerates the affected content. To force regeneration of specific parts, set the force_reset field of the corresponding item in project.yaml in the output directory.
To support a new voice model, you need to implement a new class in src/slide_to_video/tts_engine and register the class by calling register_engine (See an example at here).
> You must confirm the following:
| > "I have purchased a commercial license from Coqui: [email protected]"
| > "Otherwise, I agree to the terms of the non-commercial CPML: https://coqui.ai/cpml" - [y/n]
| | >
y.