--- library_name: keras tags: - SpeakerRecognition - Fast Fourier Transform (FFT) - Convnet - speech-recordings - SpeechClassification --- ## Model description This model helps to classify speakers from the frequency domain representation of speech recordings, obtained via Fast Fourier Transform (FFT). The model is created by a 1D convolutional network with residual connections for audio classification. This repo contains the model for the notebook [**Speaker Recognition**](https://keras.io/examples/audio/speaker_recognition_using_cnn/). Full credits go to [**Fadi Badine**](https://twitter.com/fadibadine) ## Dataset Used This model uses a [**speaker recognition dataset**](https://www.kaggle.com/kongaevans/speaker-recognition-dataset) of Kaggle ## Intended uses & limitations This should be run with `TensorFlow 2.3` or higher, or `tf-nightly`. Also, The noise samples in the dataset need to be resampled to a sampling rate of 16000 Hz before using for this model so, In order to do this, you will need to have installed `ffmpg`. ## Training and evaluation data During dataset preparation, the speech samples & background noise samples were sorted and categorized into 2 folders - audio & noise, and then noise samples were resampled to 16000Hz & then the background noise was added to the speech samples to augment the data. After that, the FFT of these samples was given to the model for the training & evaluation part. ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: | name | learning_rate | decay | beta_1 | beta_2 | epsilon | amsgrad | training_precision | |----|-------------|-----|------|------|-------|-------|------------------| |Adam|0.0010000000474974513|0.0|0.8999999761581421|0.9990000128746033|1e-07|False|float32| ## Training Metrics Model history needed ## Model Plot
View Model Plot ![Model Image](./model.png)
Model By : Kavya Bisht