Skip to content

Latest commit

History

History
197 lines (143 loc) 路 6.93 KB

README.md

File metadata and controls

197 lines (143 loc) 路 6.93 KB
logo logo logo

AI Rapper

Talking Head videos of your favorite rapper rapping about anything. Using open-source NLP and TTS libraries.

Github


Table of Contents

    馃摑 About
    馃捇 How to build
    馃敡 Tools used
    馃懁 Contact

馃摑About

Overview

  • Input a prompt, reference audio, reference photo
  • Output auto-generated rap lyrics in style of rapper, synthesized audio using cloned voice, and Talking Head video.

Features

  • Intelligent text (lyrics): input a prompt and harness state-of-the-art LLMs to craft creative and engaging rap verses.

  • Synthetic audio (voice): a text-to-speech (TTS) system to clone a voice based on audio sample and feed it generated lyrics.

  • Talking Head (video): input a reference image and cobine with generated audio to create a realistic, engaging talking head.

馃捇 How to build

Prerequisites

  • Clone MakeItTalk (for video generation) https://github.com/adobe-research/MakeItTalk/ into root directory of ai-rapper
  • Add a strictly 256 x 256 image of rapper in MakeItTalk/examples. Face should be clear and un-obstructed. Ex: MakeItTalk/examples/eminem.png
  • Add an audio .wav file ( ~ 10-30 sec) of rapper in a separate directory of audio_samples i.e.audio_samples/eminem_00.wav

Install dependencies and run

pip install -r requirements.txt
python src/app.py

Output

Look for generated video in MakeItTalk/examples:

/tmp/tmpx_swo6p1eminem_00.wav
/tmp/tmp7zx0u65zem.png
Audio-----> tmpx_swo6p1eminem_00.wav
Parameters===== tmpx_swo6p1eminem_00.wav 48000 [-29 -36 -43 ... 120 125 124]
Loaded the voice encoder model on cuda in 0.04 seconds.
Processing audio file tmpx_swo6p1eminem_00.wav
Loaded the voice encoder model on cuda in 0.03 seconds.
source shape: torch.Size([1, 576, 80]) torch.Size([1, 256]) torch.Size([1, 256]) torch.Size([1, 576, 257])
converted shape: torch.Size([1, 576, 80]) torch.Size([1, 1152])
Run on device: cuda
======== LOAD PRETRAINED FACE ID MODEL examples/ckpt/ckpt_speaker_branch.pth =========
....
....
....
====================================
z = torch.tensor(torch.zeros(aus.shape[0], 128), requires_grad=False, dtype=torch.float).to(device)
OpenCV: FFMPEG: fallback to use tag 0x7634706d/'mp4v'
examples/tmpx_swo6p1eminem_00.wav
ffmpeg version 4.4.2-0ubuntu0.22.04.1 Copyright (c) 2000-2021 the FFmpeg developers
....
....
....
OpenCV: FFMPEG: tag 0x67706a6d/'mjpg' is not supported with codec id 7 and format 'mp4 / MP4 (MPEG-4 Part 14)'
OpenCV: FFMPEG: fallback to use tag 0x7634706d/'mp4v'
Time - ffmpeg add audio: 15.704241514205933
finish image2image gen
examples/test_pred_fls_tmpx_swo6p1eminem_00_audio_embed.mp4

video preview

video

馃敡Tools Used

Python Hugging Face Transformers library pyTorch audio CUDA Toolkit Tortoise TTS FFMPEG OpenCV

NLP

HuggingFace Transformers libary

  • Harnesses fine-tuned and pre-trained language models for rap lyric generation
  • AutoModelForCausalLM generates text by predicting the next word based on previous ones, not on the ones that follow. Useful for speciifc creative tasks such as generating rap lyrics, which rely on stylistic model outputs that have been trained on vast amounts of diverse text data (thus enabling it to generate coherent and contextually relevant text based on a given user prompt)
  • AutoTokenizer efficiently tokenizes input prompts, enabling seamless integration with LLMs. DistilGPT2 (a distilled, more efficient version of GPT-2) efficiently handles this. See usage in src/text_generation/text_generator.py

TTS

Tortoise TTS

  • Used for synthesizing audio from text
  • Supports custom voice models to mimic specific rappers' voices

CUDA Toolkit

  • Trained Eminem's voice (as in the example) on a custom TTS model.
  • NVIDIA's CUDA Toolkit used to accelerate GPU training.

PyTorch Audio

  • torchaudio library handles audio data, saving synthesized rap audio in *.wav format

Talking Head generation

MakeItTalk

  • Open-source Github repo used for video synthesis, harnessing OpenCV and FFMPEG
  • Demo: https://github.com/yzhou359/MakeItTalk/blob/main/quick_demo_tdlr.ipynb

OpenCV

  • used to segment facial features in input image and lip-sync to audio

FFMPEG

  • Used to handle smooth, compatible audio + video synthesis

馃懁Contact

Email Twitter