Skip to content

Official implementation of "PhonMatchNet: Phoneme-Guided Zero-Shot Keyword Spotting for User-Defined Keywords" (INTERSPEECH 2023)

License

Notifications You must be signed in to change notification settings

ncsoft/PhonMatchNet

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

🔍 PhonMatchNet: Phoneme-Guided Zero-Shot Keyword Spotting for User-Defined Keywords

Official implementation of PhonMatchNet: Phoneme-Guided Zero-Shot Keyword Spotting for User-Defined Keywords.

PyTorch version: https://github.com/ncsoft/PhonMatchNet/tree/pytorch

Requirements

Datasets

Download the dataset and prepare it according to each guide.

Getting started

Environment

cd ./docker
docker build --tag udkws .

Training

docker run -it --rm --gpus '"device=0,1"' \
    -v /path/to/this/repo:/home/ \
    -v /path/to/prepared/dataset:/home/DB \
    ukws \
    /bin/bash -c \
    "python train.py \
        --epoch 100 \
        --lr 1e-3 \
        --loss_weight 1.0 1.0 \
        --audio_input both \
        --text_input g2p_embed \
        --stack_extractor \
        --comment 'user comments for each experiment'"

Monitoring

tensorboard --logdir ./log/ --bind_all

CONTRIBUTING

Please post bug reports and new feature suggestions to the Issues and Pull requests tabs of this repo.

About

Official implementation of "PhonMatchNet: Phoneme-Guided Zero-Shot Keyword Spotting for User-Defined Keywords" (INTERSPEECH 2023)

Resources

License

Stars

Watchers

Forks

Languages