canvas-based talking head model using viseme data
-
Updated
Sep 4, 2023 - JavaScript
canvas-based talking head model using viseme data
canvas-based talking head model using viseme data
Open-source example for integrating ElevenLabs conversational AI with animated avatars using Mascotbot SDK. Features real-time lip sync and natural voice interactions.
Playing Audio with lipsync using different Avathar expressions
FastAPI backend for a multilingual AI avatar system with text-to-speech and voice-to-voice translation. Integrates AWS Bedrock, Polly, Transcribe, and S3 for speech synthesis, transcription, and viseme mapping to enable real-time avatar lip-sync across multiple languages.
Frontend implementation of Conflicta. Submitted to ScienceHack 2025, Munich.
LipGANs is a text-to-viseme GAN framework that generates realistic mouth movements directly from text, without requiring audio. It maps phonemes → visemes, predicts phoneme durations, and uses per-viseme 3D GANs to synthesize photorealistic frames that can be exported as PNG sequences, GIFs, or MP4 videos.
Add a description, image, and links to the viseme topic page so that developers can more easily learn about it.
To associate your repository with the viseme topic, visit your repo's landing page and select "manage topics."