Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

fullstackpeng/LiveTalking

Repository files navigation

Real-time interactive streaming digital human enables synchronous audio and video dialogue. It can basically achieve commercial effects.

Effect of wav2lip | Effect of ernerf | Effect of musetalk

中文版

News

  • December 8, 2024: Improved multi-concurrency, and the video memory does not increase with the number of concurrent connections.
  • December 21, 2024: Added model warm-up for wav2lip and musetalk to solve the problem of stuttering during the first inference. Thanks to @heimaojinzhangyz
  • December 28, 2024: Added the digital human model Ultralight-Digital-Human. Thanks to @lijihua2017
  • February 7, 2025: Added fish-speech tts
  • February 21, 2025: Added the open-source model wav2lip256. Thanks to @不蠢不蠢
  • March 2, 2025: Added Tencent's speech synthesis service
  • March 16, 2025: Supports mac gpu inference. Thanks to @GcsSloop

Features

  1. Supports multiple digital human models: ernerf, musetalk, wav2lip, Ultralight-Digital-Human
  2. Supports voice cloning
  3. Supports interrupting the digital human while it is speaking
  4. Supports full-body video stitching
  5. Supports rtmp and webrtc
  6. Supports video arrangement: Play custom videos when not speaking
  7. Supports multi-concurrency

1. Installation

Tested on Ubuntu 20.04, Python 3.10, Pytorch 1.12 and CUDA 11.3

1.1 Install dependency

conda create -n nerfstream python=3.10
conda activate nerfstream
# If the cuda version is not 11.3 (confirm the version by running nvidia-smi), install the corresponding version of pytorch according to <https://pytorch.org/get-started/previous-versions/> 
conda install pytorch==1.12.1 torchvision==0.13.1 cudatoolkit=11.3 -c pytorch
pip install -r requirements.txt
# If you need to train the ernerf model, install the following libraries
# pip install "git+https://github.com/facebookresearch/pytorch3d.git"
# pip install tensorflow-gpu==2.8.0
# pip install --upgrade "protobuf<=3.20.1"

Common installation issues FAQ
For setting up the linux cuda environment, you can refer to this article https://zhuanlan.zhihu.com/p/674972886

2. Quick Start

If you can't access huggingface, before running

export HF_ENDPOINT=https://hf-mirror.com

3. More Usage

Usage instructions: https://livetalking-doc.readthedocs.io/en/latest

4. Docker Run

No need for the previous installation, just run directly.

docker run --gpus all -it --network=host --rm registry.cn-beijing.aliyuncs.com/codewithgpu2/lipku-metahuman-stream:2K9qaMBu8v

The code is in /root/metahuman-stream. First, git pull to get the latest code, and then execute the commands as in steps 2 and 3.

The following images are provided:

5. TODO

  • Added chatgpt to enable digital human dialogue
  • Voice cloning
  • Replace the digital human with a video when it is silent
  • MuseTalk
  • Wav2Lip
  • Ultralight-Digital-Human

If this project is helpful to you, please give it a star. Friends who are interested are also welcome to join in and improve this project together.

  • Knowledge Planet: https://t.zsxq.com/7NMyO, where high-quality common problems, best practice experiences, and problem solutions are accumulated.
  • WeChat Official Account: Digital Human Technology

About

Real time interactive streaming digital human

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 70.3%
  • Cuda 12.0%
  • JavaScript 9.4%
  • HTML 6.9%
  • C 0.8%
  • Dockerfile 0.3%
  • Other 0.3%

AltStyle によって変換されたページ (->オリジナル) /