abgoswam commited on
Commit
4ed42e0
1 Parent(s): eda7d11

Training in progress, step 100

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/mnt/phyagimodelsdev/phi7b_higherlr_phase2_hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/mnt/phyagimodelsdev/phi7b_lowerlr_phase2_hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
runs/Mar05_18-01-48_node-0/events.out.tfevents.1709690564.node-0.123102.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be90df59e33f82d267a89d7e1a0d69e55cb609eeab62e235b978065c3270da91
3
+ size 12956
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a07684d5040b30181cbaab7d56e82bbbffd35a0263cf4190de70c7d69651e661
3
  size 6200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c64276fa240ff0eb70dcd5ff05e57bd4927eeba68b6be48a0acf69c3e5553e9c
3
  size 6200