🔚 Digital Sound lecture module 2: Speaker Recognition
This project is a module #2 of Digital Sound lecture: Speaker Recognition
Implementing voice recognition that is used in AI speakers
Recognizing speaker’s voice
Distinguishing Speaker’s voice patterns in that every person have different voice feature
We want to provide personal service for each family member by AI speaker using Speaker Recognition
All data that have used for training models will NOT be OPENED to PUBLIC due to privacy issues
I have looked up the code from this blog and applied to my project
Image down below is a flow chart how the code works in total
train_model.py
test_model.py
show_all_graphs.py
Mel Spectogram, MFCC, GMM graphs.
issues | solutions |
---|---|
IO library problems | Not very often, io library has some problems with reading wav files. Before inputting the wav, you have to get rid of implicit metadata in wav files. Using ffmpeg is one way. |
Unicode problems | Try adding “utf-8” when reading wav files |
Mainly used sklearn and speakerfeatures with Python