Create high-resolution visually dubbed videos with DINet
-
Updated
Jun 27, 2024 - Python
Create high-resolution visually dubbed videos with DINet
Personalized Virtual Webcam for WebRTC
A multi-language virtual assistant
This is tremendous approach for implementing super light weight real-time lip-sync AI engine. This engine can be forked for the purpose of building real-time consistent character generation system and other purposes. I am open to live discussion with AI engineers and fans.
Tool designed to transform research papers from arXiv into engaging presentations and videos, ready for upload to YouTube.
Multi-angle Lip Multimodal Video Data
Generative AI | Automated Video Translation and Content Generation using Talking Heads, Leveraging Wav2Lip Model
A mouth operated gaming joystick based on the LipSync.
Presenting SIPHI, a revolutionary assistive technology designed to empower the deaf community with the ability to speak English. SIPHI utilizes cutting-edge lip-sync animation and graphical voice representation to provide an immersive and effective learning experience.
Interactive news summarizer system that leverages avatar narration and text to speech conversion techniques.
I have collected my own dataset of lip movements to develop and train the model.
DoyenTalker uses deep learning techniques to generate personalized avatar videos that speak user-provided text in a specified voice. The system utilizes Coqui TTS for text-to-speech generation, along with various face rendering and animation techniques to create a video where the given avatar articulates the speech.
Installing VSeeFaceSDK on Unity is such a hassle. So here, just download this project and you're good to go.
Add a description, image, and links to the lipsync topic page so that developers can more easily learn about it.
To associate your repository with the lipsync topic, visit your repo's landing page and select "manage topics."