mllm-dev commited on
Commit
b752b49
1 Parent(s): b9aeffd

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -49,7 +49,7 @@
49
  }
50
  },
51
  "torch_dtype": "float32",
52
- "transformers_version": "4.38.2",
53
  "use_cache": true,
54
  "vocab_size": 50257
55
  }
 
49
  }
50
  },
51
  "torch_dtype": "float32",
52
+ "transformers_version": "4.37.1",
53
  "use_cache": true,
54
  "vocab_size": 50257
55
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4090092241945e27e2abddb1c370884d514e406fd988c3308a6db168642a07ae
3
  size 497789648
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65cdea4e57693fb129b4257c58233ae86f77f67e2e1113eb745c4c397760a4ba
3
  size 497789648
runs/Apr09_16-34-03_lambda-hyperplane05/events.out.tfevents.1712698445.lambda-hyperplane05.3431601.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ed10756cbe142c5695acce25933059ba132038c44f6b5c8e3e6cc99ddacdd52
3
+ size 13748
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 1024,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:202a5542d25d5172e8161c0589eaf5ec460cc8e907a908b0e1388d20c67b502b
3
- size 4539
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b94a3b72022afedcd8a0a62e1e0c9badd127614dfbeeb61f54d6d68580c0058
3
+ size 4347