asahi417 commited on
Commit
7b838b6
1 Parent(s): d9a2565
experiment_cache/cluster/w2v_bert_se.expresso.speaker_id.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a7b8722c152981c4e514dfaf90eb17dbefbc653dbee44c59bee558aa2a9cb52
3
- size 162416
 
 
 
 
experiment_cache/cluster/w2v_bert_se.expresso.style.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:201cafca59ca308d8882d3434dd2637fe49a0c5b11b6cb8bccec7396bfc170be
3
- size 181626
 
 
 
 
experiment_cache/cluster/w2v_bert_se.voxceleb1-test-split.speaker_id.csv DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0a9a82594fdef76a933e5a21534a8af37487f7ec1b6b2e4662be6950c52ce76
3
- size 123308
 
 
 
 
experiment_cache/embeddings/w2v_bert_se.expresso.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:6adea5c716277b2525cf78b38234da223ff9af3722635aeae19ac26e78226ef5
3
- size 268823097
 
 
 
 
experiment_cache/embeddings/w2v_bert_se.voxceleb1-test-split.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:03324e439f856b30e136b3b13fe9d841919b792aeaca8a6f7d65787977dca1dc
3
- size 112682448
 
 
 
 
experiment_cache/figure/2d.latent_space.w2v_bert_se.expresso.speaker_id.png DELETED

Git LFS Details

  • SHA256: bdeecba8d662eec6e355636c8866e88c00a26d10839fe4ef4d8a0c08b0836344
  • Pointer size: 131 Bytes
  • Size of remote file: 886 kB
experiment_cache/figure/2d.latent_space.w2v_bert_se.expresso.style.png DELETED

Git LFS Details

  • SHA256: 5f83ea0b753f6acc579b7f70d9d49cc1b1d3dae62ee5f0a888a6114c83cac65c
  • Pointer size: 132 Bytes
  • Size of remote file: 1.48 MB
experiment_cache/figure/2d.latent_space.w2v_bert_se.voxceleb1-test-split.speaker_id.png DELETED

Git LFS Details

  • SHA256: 9c33c1d7432a36e2796728e4eb573eb7db121cbbb76845cf3f1c6aabefc32b74
  • Pointer size: 132 Bytes
  • Size of remote file: 1.53 MB
experiment_cache/tsne/w2v_bert_se.expresso.speaker_id.npy DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e186ef5932df77876df41323523fdef80e17d9ae2db2e1f6452a66cd5c4d0a61
3
- size 93048
 
 
 
 
experiment_cache/tsne/w2v_bert_se.expresso.style.npy DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:600c2be510bb6044ab2d38a87a88ac64224d832f622b0bc8647e5c07ff2c03de
3
- size 93048
 
 
 
 
experiment_cache/tsne/w2v_bert_se.voxceleb1-test-split.speaker_id.npy DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d09ce160684b68806d052b6c699819e41410df0a47fe5f1bb51bb4039b1d9c6
3
- size 39120
 
 
 
 
experiment_speaker_verification.py CHANGED
@@ -16,10 +16,11 @@ from datasets import load_dataset
16
 
17
  from model_meta_voice import MetaVoiceEmbedding
18
  from model_pyannote_embedding import PyannoteEmbedding
19
- from model_w2v_bert import W2VBERTEmbedding
20
  from model_clap import CLAPEmbedding, CLAPGeneralEmbedding
21
- from model_xls import Wav2VecEmbedding, XLSR300MEmbedding, XLSR1BEmbedding, XLSR2BEmbedding
22
- from model_hubert import HuBERTBaseEmbedding, HuBERTLargeEmbedding, HuBERTXLEmbedding
 
 
23
 
24
 
25
  def get_embedding(model_class, model_name: str, dataset_name: str, data_split: str):
@@ -118,68 +119,68 @@ def analyze_embedding(model_name: str, dataset_name: str, n_shot: int = 5, n_cro
118
  if __name__ == '__main__':
119
  # get_embedding(MetaVoiceEmbedding, "meta_voice_se", "asahi417/voxceleb1-test-split", "test")
120
  # get_embedding(PyannoteEmbedding, "pyannote_se", "asahi417/voxceleb1-test-split", "test")
121
- # get_embedding(W2VBERTEmbedding, "w2v_bert_se", "asahi417/voxceleb1-test-split", "test")
122
  # get_embedding(CLAPEmbedding, "clap_se", "asahi417/voxceleb1-test-split", "test")
123
  # get_embedding(CLAPGeneralEmbedding, "clap_general_se", "asahi417/voxceleb1-test-split", "test")
124
  # get_embedding(HuBERTBaseEmbedding, "hubert_base_se", "asahi417/voxceleb1-test-split", "test")
125
  # get_embedding(HuBERTLargeEmbedding, "hubert_large_se", "asahi417/voxceleb1-test-split", "test")
126
  # get_embedding(HuBERTXLEmbedding, "hubert_xl_se", "asahi417/voxceleb1-test-split", "test")
127
- get_embedding(Wav2VecEmbedding, "wav2vec_se", "asahi417/voxceleb1-test-split", "test")
128
- get_embedding(XLSR300MEmbedding, "xlsr_300m_se", "asahi417/voxceleb1-test-split", "test")
129
- get_embedding(XLSR1BEmbedding, "xlsr_1b_se", "asahi417/voxceleb1-test-split", "test")
130
- get_embedding(XLSR2BEmbedding, "xlsr_2b_se", "asahi417/voxceleb1-test-split", "test")
 
131
 
132
  # get_embedding(MetaVoiceEmbedding, "meta_voice_se", "ylacombe/expresso", "train")
133
  # get_embedding(PyannoteEmbedding, "pyannote_se", "ylacombe/expresso", "train")
134
- # get_embedding(W2VBERTEmbedding, "w2v_bert_se", "ylacombe/expresso", "train")
135
  # get_embedding(CLAPEmbedding, "clap_se", "ylacombe/expresso", "train")
136
  # get_embedding(CLAPGeneralEmbedding, "clap_general_se", "ylacombe/expresso", "train")
137
  # get_embedding(HuBERTBaseEmbedding, "hubert_base_se", "ylacombe/expresso", "train")
138
  # get_embedding(HuBERTLargeEmbedding, "hubert_large_se", "ylacombe/expresso", "train")
139
  # get_embedding(HuBERTXLEmbedding, "hubert_xl_se", "ylacombe/expresso", "train")
140
- get_embedding(Wav2VecEmbedding, "wav2vec_se", "ylacombe/expresso", "train")
141
- get_embedding(XLSR300MEmbedding, "xlsr_300m_se", "ylacombe/expresso", "train")
142
- get_embedding(XLSR1BEmbedding, "xlsr_1b_se", "ylacombe/expresso", "train")
143
- get_embedding(XLSR2BEmbedding, "xlsr_2b_se", "ylacombe/expresso", "train")
 
144
 
145
  # cluster_embedding("meta_voice_se", "asahi417/voxceleb1-test-split", "speaker_id")
146
  # cluster_embedding("pyannote_se", "asahi417/voxceleb1-test-split", "speaker_id")
147
- # cluster_embedding("w2v_bert_se", "asahi417/voxceleb1-test-split", "speaker_id")
148
  # cluster_embedding("clap_se", "asahi417/voxceleb1-test-split", "speaker_id")
149
  # cluster_embedding("clap_general_se", "asahi417/voxceleb1-test-split", "speaker_id")
150
  # cluster_embedding("hubert_base_se", "asahi417/voxceleb1-test-split", "speaker_id")
151
  # cluster_embedding("hubert_large_se", "asahi417/voxceleb1-test-split", "speaker_id")
152
  # cluster_embedding("hubert_xl_se", "asahi417/voxceleb1-test-split", "speaker_id")
153
- cluster_embedding("wav2vec_se", "asahi417/voxceleb1-test-split", "speaker_id")
154
- cluster_embedding("xlsr_300m_se", "asahi417/voxceleb1-test-split", "speaker_id")
155
- cluster_embedding("xlsr_1b_se", "asahi417/voxceleb1-test-split", "speaker_id")
156
- cluster_embedding("xlsr_2b_se", "asahi417/voxceleb1-test-split", "speaker_id")
 
157
 
158
  # cluster_embedding("meta_voice_se", "ylacombe/expresso", "speaker_id")
159
  # cluster_embedding("pyannote_se", "ylacombe/expresso", "speaker_id")
160
- # cluster_embedding("w2v_bert_se", "ylacombe/expresso", "speaker_id")
161
  # cluster_embedding("clap_se", "ylacombe/expresso", "speaker_id")
162
  # cluster_embedding("clap_general_se", "ylacombe/expresso", "speaker_id")
163
  # cluster_embedding("hubert_base_se", "ylacombe/expresso", "speaker_id")
164
  # cluster_embedding("hubert_large_se", "ylacombe/expresso", "speaker_id")
165
  # cluster_embedding("hubert_xl_se", "ylacombe/expresso", "speaker_id")
166
- cluster_embedding("wav2vec_se", "ylacombe/expresso", "speaker_id")
167
- cluster_embedding("xlsr_300m_se", "ylacombe/expresso", "speaker_id")
168
- cluster_embedding("xlsr_1b_se", "ylacombe/expresso", "speaker_id")
169
- cluster_embedding("xlsr_2b_se", "ylacombe/expresso", "speaker_id")
 
170
 
171
  # cluster_embedding("meta_voice_se", "ylacombe/expresso", "style")
172
  # cluster_embedding("pyannote_se", "ylacombe/expresso", "style")
173
- # cluster_embedding("w2v_bert_se", "ylacombe/expresso", "style")
174
  # cluster_embedding("clap_se", "ylacombe/expresso", "style")
175
  # cluster_embedding("clap_general_se", "ylacombe/expresso", "style")
176
  # cluster_embedding("hubert_base_se", "ylacombe/expresso", "style")
177
  # cluster_embedding("hubert_large_se", "ylacombe/expresso", "style")
178
  # cluster_embedding("hubert_xl_se", "ylacombe/expresso", "style")
179
- cluster_embedding("wav2vec_se", "ylacombe/expresso", "style")
180
- cluster_embedding("xlsr_300m_se", "ylacombe/expresso", "style")
181
- cluster_embedding("xlsr_1b_se", "ylacombe/expresso", "style")
182
- cluster_embedding("xlsr_2b_se", "ylacombe/expresso", "style")
 
183
 
184
 
185
 
 
16
 
17
  from model_meta_voice import MetaVoiceEmbedding
18
  from model_pyannote_embedding import PyannoteEmbedding
 
19
  from model_clap import CLAPEmbedding, CLAPGeneralEmbedding
20
+ from model_speaker_embedding import (
21
+ W2VBERTEmbedding, Wav2VecEmbedding, XLSR300MEmbedding, XLSR1BEmbedding, XLSR2BEmbedding,
22
+ HuBERTBaseEmbedding, HuBERTLargeEmbedding, HuBERTXLEmbedding
23
+ )
24
 
25
 
26
  def get_embedding(model_class, model_name: str, dataset_name: str, data_split: str):
 
119
  if __name__ == '__main__':
120
  # get_embedding(MetaVoiceEmbedding, "meta_voice_se", "asahi417/voxceleb1-test-split", "test")
121
  # get_embedding(PyannoteEmbedding, "pyannote_se", "asahi417/voxceleb1-test-split", "test")
 
122
  # get_embedding(CLAPEmbedding, "clap_se", "asahi417/voxceleb1-test-split", "test")
123
  # get_embedding(CLAPGeneralEmbedding, "clap_general_se", "asahi417/voxceleb1-test-split", "test")
124
  # get_embedding(HuBERTBaseEmbedding, "hubert_base_se", "asahi417/voxceleb1-test-split", "test")
125
  # get_embedding(HuBERTLargeEmbedding, "hubert_large_se", "asahi417/voxceleb1-test-split", "test")
126
  # get_embedding(HuBERTXLEmbedding, "hubert_xl_se", "asahi417/voxceleb1-test-split", "test")
127
+ get_embedding(W2VBERTEmbedding, "w2v_bert_se", "asahi417/voxceleb1-test-split", "test")
128
+ # get_embedding(Wav2VecEmbedding, "wav2vec_se", "asahi417/voxceleb1-test-split", "test")
129
+ # get_embedding(XLSR300MEmbedding, "xlsr_300m_se", "asahi417/voxceleb1-test-split", "test")
130
+ # get_embedding(XLSR1BEmbedding, "xlsr_1b_se", "asahi417/voxceleb1-test-split", "test")
131
+ # get_embedding(XLSR2BEmbedding, "xlsr_2b_se", "asahi417/voxceleb1-test-split", "test")
132
 
133
  # get_embedding(MetaVoiceEmbedding, "meta_voice_se", "ylacombe/expresso", "train")
134
  # get_embedding(PyannoteEmbedding, "pyannote_se", "ylacombe/expresso", "train")
 
135
  # get_embedding(CLAPEmbedding, "clap_se", "ylacombe/expresso", "train")
136
  # get_embedding(CLAPGeneralEmbedding, "clap_general_se", "ylacombe/expresso", "train")
137
  # get_embedding(HuBERTBaseEmbedding, "hubert_base_se", "ylacombe/expresso", "train")
138
  # get_embedding(HuBERTLargeEmbedding, "hubert_large_se", "ylacombe/expresso", "train")
139
  # get_embedding(HuBERTXLEmbedding, "hubert_xl_se", "ylacombe/expresso", "train")
140
+ get_embedding(W2VBERTEmbedding, "w2v_bert_se", "ylacombe/expresso", "train")
141
+ # get_embedding(Wav2VecEmbedding, "wav2vec_se", "ylacombe/expresso", "train")
142
+ # get_embedding(XLSR300MEmbedding, "xlsr_300m_se", "ylacombe/expresso", "train")
143
+ # get_embedding(XLSR1BEmbedding, "xlsr_1b_se", "ylacombe/expresso", "train")
144
+ # get_embedding(XLSR2BEmbedding, "xlsr_2b_se", "ylacombe/expresso", "train")
145
 
146
  # cluster_embedding("meta_voice_se", "asahi417/voxceleb1-test-split", "speaker_id")
147
  # cluster_embedding("pyannote_se", "asahi417/voxceleb1-test-split", "speaker_id")
 
148
  # cluster_embedding("clap_se", "asahi417/voxceleb1-test-split", "speaker_id")
149
  # cluster_embedding("clap_general_se", "asahi417/voxceleb1-test-split", "speaker_id")
150
  # cluster_embedding("hubert_base_se", "asahi417/voxceleb1-test-split", "speaker_id")
151
  # cluster_embedding("hubert_large_se", "asahi417/voxceleb1-test-split", "speaker_id")
152
  # cluster_embedding("hubert_xl_se", "asahi417/voxceleb1-test-split", "speaker_id")
153
+ cluster_embedding("w2v_bert_se", "asahi417/voxceleb1-test-split", "speaker_id")
154
+ # cluster_embedding("wav2vec_se", "asahi417/voxceleb1-test-split", "speaker_id")
155
+ # cluster_embedding("xlsr_300m_se", "asahi417/voxceleb1-test-split", "speaker_id")
156
+ # cluster_embedding("xlsr_1b_se", "asahi417/voxceleb1-test-split", "speaker_id")
157
+ # cluster_embedding("xlsr_2b_se", "asahi417/voxceleb1-test-split", "speaker_id")
158
 
159
  # cluster_embedding("meta_voice_se", "ylacombe/expresso", "speaker_id")
160
  # cluster_embedding("pyannote_se", "ylacombe/expresso", "speaker_id")
 
161
  # cluster_embedding("clap_se", "ylacombe/expresso", "speaker_id")
162
  # cluster_embedding("clap_general_se", "ylacombe/expresso", "speaker_id")
163
  # cluster_embedding("hubert_base_se", "ylacombe/expresso", "speaker_id")
164
  # cluster_embedding("hubert_large_se", "ylacombe/expresso", "speaker_id")
165
  # cluster_embedding("hubert_xl_se", "ylacombe/expresso", "speaker_id")
166
+ cluster_embedding("w2v_bert_se", "ylacombe/expresso", "speaker_id")
167
+ # cluster_embedding("wav2vec_se", "ylacombe/expresso", "speaker_id")
168
+ # cluster_embedding("xlsr_300m_se", "ylacombe/expresso", "speaker_id")
169
+ # cluster_embedding("xlsr_1b_se", "ylacombe/expresso", "speaker_id")
170
+ # cluster_embedding("xlsr_2b_se", "ylacombe/expresso", "speaker_id")
171
 
172
  # cluster_embedding("meta_voice_se", "ylacombe/expresso", "style")
173
  # cluster_embedding("pyannote_se", "ylacombe/expresso", "style")
 
174
  # cluster_embedding("clap_se", "ylacombe/expresso", "style")
175
  # cluster_embedding("clap_general_se", "ylacombe/expresso", "style")
176
  # cluster_embedding("hubert_base_se", "ylacombe/expresso", "style")
177
  # cluster_embedding("hubert_large_se", "ylacombe/expresso", "style")
178
  # cluster_embedding("hubert_xl_se", "ylacombe/expresso", "style")
179
+ cluster_embedding("w2v_bert_se", "ylacombe/expresso", "style")
180
+ # cluster_embedding("wav2vec_se", "ylacombe/expresso", "style")
181
+ # cluster_embedding("xlsr_300m_se", "ylacombe/expresso", "style")
182
+ # cluster_embedding("xlsr_1b_se", "ylacombe/expresso", "style")
183
+ # cluster_embedding("xlsr_2b_se", "ylacombe/expresso", "style")
184
 
185
 
186
 
model_speaker_embedding.py CHANGED
@@ -1,7 +1,4 @@
1
- """Meta's w2vBERT based speaker embedding.
2
- - feature dimension: 1024
3
- - source: https://huggingface.co/facebook/w2v-bert-2.0
4
- """
5
  from typing import Optional
6
 
7
  import torch
 
1
+ """Meta's w2vBERT based speaker embedding."""
 
 
 
2
  from typing import Optional
3
 
4
  import torch
test.py CHANGED
@@ -33,6 +33,5 @@ def test():
33
  print(v.shape)
34
 
35
 
36
-
37
  if __name__ == '__main__':
38
  test()
 
33
  print(v.shape)
34
 
35
 
 
36
  if __name__ == '__main__':
37
  test()