Oysiyl commited on
Commit
36d4baa
1 Parent(s): 8f7b0d3

Training in progress, step 525

Browse files
config.json CHANGED
@@ -47,7 +47,7 @@
47
  "num_hidden_layers": 24,
48
  "num_negatives": 100,
49
  "output_hidden_size": 1024,
50
- "pad_token_id": 30,
51
  "position_embeddings_type": "relative_key",
52
  "proj_codevector_dim": 768,
53
  "right_max_position_embeddings": 8,
@@ -77,6 +77,6 @@
77
  "transformers_version": "4.37.0.dev0",
78
  "use_intermediate_ffn_before_adapter": false,
79
  "use_weighted_layer_sum": false,
80
- "vocab_size": 33,
81
  "xvector_output_dim": 512
82
  }
 
47
  "num_hidden_layers": 24,
48
  "num_negatives": 100,
49
  "output_hidden_size": 1024,
50
+ "pad_token_id": 36,
51
  "position_embeddings_type": "relative_key",
52
  "proj_codevector_dim": 768,
53
  "right_max_position_embeddings": 8,
 
77
  "transformers_version": "4.37.0.dev0",
78
  "use_intermediate_ffn_before_adapter": false,
79
  "use_weighted_layer_sum": false,
80
+ "vocab_size": 39,
81
  "xvector_output_dim": 512
82
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:283c290b7083d1a5c116e170f33df3bdb7b49e22b701fe370590b758090e1c11
3
- size 2422949860
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4c03e6240ac5e64845577b616a6741d9a93db026cfcef0d8dd552d528250765
3
+ size 2422974460
runs/Feb05_09-16-01_nbbl5s368u/events.out.tfevents.1707124967.nbbl5s368u.66.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f17f2ac5c8844a7c311f0e57ec76816adbd718ffa70a9f8140388613f815d8b
3
+ size 6134
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:510ddda6de5e5bd5ce42bd855b0eafd449a3663742a7869411bd88250beb330b
3
  size 4271
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eebdf22ece0cffa29a59ccc3904fe6354603445b353c20b51902c2bb9552b9cc
3
  size 4271