Skip to content

vdutts7/ai-rapper

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

14 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

logo logo logo

AI Rapper

Talking Head videos of your favorite rapper rapping about anything. Using open-source NLP and TTS libraries.

Github


Table of Contents

    📝 About
    💻 How to build
    🔧 Tools used
    👤 Contact

📝About

Overview

  • Input a prompt, reference audio, reference photo
  • Output auto-generated rap lyrics in style of rapper, synthesized audio using cloned voice, and Talking Head video.

Features

  • Intelligent text (lyrics): input a prompt and harness state-of-the-art LLMs to craft creative and engaging rap verses.

  • Synthetic audio (voice): a text-to-speech (TTS) system to clone a voice based on audio sample and feed it generated lyrics.

  • Talking Head (video): input a reference image and cobine with generated audio to create a realistic, engaging talking head.

💻 How to build

Prerequisites

  • Clone MakeItTalk (for video generation) https://github.com/adobe-research/MakeItTalk/ into root directory of ai-rapper
  • Add a strictly 256 x 256 image of rapper in MakeItTalk/examples. Face should be clear and un-obstructed. Ex: MakeItTalk/examples/eminem.png
  • Add an audio .wav file ( ~ 10-30 sec) of rapper in a separate directory of audio_samples i.e.audio_samples/eminem_00.wav

Install dependencies and run

pip install -r requirements.txt
python src/app.py

Output

Look for generated video in MakeItTalk/examples:

/tmp/tmpx_swo6p1eminem_00.wav
/tmp/tmp7zx0u65zem.png
Audio-----> tmpx_swo6p1eminem_00.wav
Parameters===== tmpx_swo6p1eminem_00.wav 48000 [-29 -36 -43 ... 120 125 124]
Loaded the voice encoder model on cuda in 0.04 seconds.
Processing audio file tmpx_swo6p1eminem_00.wav
Loaded the voice encoder model on cuda in 0.03 seconds.
source shape: torch.Size([1, 576, 80]) torch.Size([1, 256]) torch.Size([1, 256]) torch.Size([1, 576, 257])
converted shape: torch.Size([1, 576, 80]) torch.Size([1, 1152])
Run on device: cuda
======== LOAD PRETRAINED FACE ID MODEL examples/ckpt/ckpt_speaker_branch.pth =========
....
....
....
====================================
z = torch.tensor(torch.zeros(aus.shape[0], 128), requires_grad=False, dtype=torch.float).to(device)
OpenCV: FFMPEG: fallback to use tag 0x7634706d/'mp4v'
examples/tmpx_swo6p1eminem_00.wav
ffmpeg version 4.4.2-0ubuntu0.22.04.1 Copyright (c) 2000-2021 the FFmpeg developers
....
....
....
OpenCV: FFMPEG: tag 0x67706a6d/'mjpg' is not supported with codec id 7 and format 'mp4 / MP4 (MPEG-4 Part 14)'
OpenCV: FFMPEG: fallback to use tag 0x7634706d/'mp4v'
Time - ffmpeg add audio: 15.704241514205933
finish image2image gen
examples/test_pred_fls_tmpx_swo6p1eminem_00_audio_embed.mp4

video preview

video

🔧Tools Used

Python Hugging Face Transformers library pyTorch audio CUDA Toolkit Tortoise TTS FFMPEG OpenCV

NLP

HuggingFace Transformers libary

  • Harnesses fine-tuned and pre-trained language models for rap lyric generation
  • AutoModelForCausalLM generates text by predicting the next word based on previous ones, not on the ones that follow. Useful for speciifc creative tasks such as generating rap lyrics, which rely on stylistic model outputs that have been trained on vast amounts of diverse text data (thus enabling it to generate coherent and contextually relevant text based on a given user prompt)
  • AutoTokenizer efficiently tokenizes input prompts, enabling seamless integration with LLMs. DistilGPT2 (a distilled, more efficient version of GPT-2) efficiently handles this. See usage in src/text_generation/text_generator.py

TTS

Tortoise TTS

  • Used for synthesizing audio from text
  • Supports custom voice models to mimic specific rappers' voices

CUDA Toolkit

  • Trained Eminem's voice (as in the example) on a custom TTS model.
  • NVIDIA's CUDA Toolkit used to accelerate GPU training.

PyTorch Audio

  • torchaudio library handles audio data, saving synthesized rap audio in *.wav format

Talking Head generation

MakeItTalk

  • Open-source Github repo used for video synthesis, harnessing OpenCV and FFMPEG
  • Demo: https://github.com/yzhou359/MakeItTalk/blob/main/quick_demo_tdlr.ipynb

OpenCV

  • used to segment facial features in input image and lip-sync to audio

FFMPEG

  • Used to handle smooth, compatible audio + video synthesis

👤Contact

Email Twitter

About

Talking Head of your favorite rapper using Transformers, PyTorch, Tortoise TTS, and OpenCV 🎵

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages