alexandrualexandru commited on
Commit
56df300
1 Parent(s): b33403b

End of training

Browse files
Files changed (4) hide show
  1. README.md +23 -23
  2. adapter_config.json +1 -1
  3. adapter_model.bin +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: llama2
3
- base_model: codellama/CodeLlama-7b-hf
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,9 +13,9 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # code-llama-text-to-sparql
15
 
16
- This model is a fine-tuned version of [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.1969
19
 
20
  ## Model description
21
 
@@ -50,26 +50,26 @@ The following hyperparameters were used during training:
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:------:|:----:|:---------------:|
53
- | 1.1337 | 0.0710 | 20 | 1.1247 |
54
- | 0.7318 | 0.1421 | 40 | 0.4634 |
55
- | 0.321 | 0.2131 | 60 | 0.3095 |
56
- | 0.2883 | 0.2842 | 80 | 0.2670 |
57
- | 0.2226 | 0.3552 | 100 | 0.2881 |
58
- | 0.2483 | 0.4263 | 120 | 0.2445 |
59
- | 0.2368 | 0.4973 | 140 | 0.2339 |
60
- | 0.2355 | 0.5684 | 160 | 0.2262 |
61
- | 0.2216 | 0.6394 | 180 | 0.2223 |
62
- | 0.2076 | 0.7105 | 200 | 0.2345 |
63
- | 0.2094 | 0.7815 | 220 | 0.2145 |
64
- | 0.2192 | 0.8526 | 240 | 0.2095 |
65
- | 0.2145 | 0.9236 | 260 | 0.2080 |
66
- | 0.2042 | 0.9947 | 280 | 0.2047 |
67
- | 0.2027 | 1.0657 | 300 | 0.2028 |
68
- | 0.2078 | 1.1368 | 320 | 0.2002 |
69
- | 0.1988 | 1.2078 | 340 | 0.2008 |
70
- | 0.2028 | 1.2789 | 360 | 0.1979 |
71
- | 0.1986 | 1.3499 | 380 | 0.1972 |
72
- | 0.2018 | 1.4210 | 400 | 0.1969 |
73
 
74
 
75
  ### Framework versions
 
1
  ---
2
  license: llama2
3
+ base_model: codellama/CodeLlama-13b-hf
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # code-llama-text-to-sparql
15
 
16
+ This model is a fine-tuned version of [codellama/CodeLlama-13b-hf](https://huggingface.co/codellama/CodeLlama-13b-hf) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.1870
19
 
20
  ## Model description
21
 
 
50
 
51
  | Training Loss | Epoch | Step | Validation Loss |
52
  |:-------------:|:------:|:----:|:---------------:|
53
+ | 1.0976 | 0.0710 | 20 | 1.0941 |
54
+ | 0.5959 | 0.1421 | 40 | 0.3399 |
55
+ | 0.3037 | 0.2131 | 60 | 0.2744 |
56
+ | 0.2385 | 0.2842 | 80 | 0.2441 |
57
+ | 0.2178 | 0.3552 | 100 | 0.2767 |
58
+ | 0.2271 | 0.4263 | 120 | 0.2303 |
59
+ | 0.2302 | 0.4973 | 140 | 0.2219 |
60
+ | 0.2433 | 0.5684 | 160 | 0.2162 |
61
+ | 0.2115 | 0.6394 | 180 | 0.2108 |
62
+ | 0.199 | 0.7105 | 200 | 0.2151 |
63
+ | 0.2117 | 0.7815 | 220 | 0.2071 |
64
+ | 0.2107 | 0.8526 | 240 | 0.2004 |
65
+ | 0.2008 | 0.9236 | 260 | 0.1990 |
66
+ | 0.2083 | 0.9947 | 280 | 0.1953 |
67
+ | 0.1905 | 1.0657 | 300 | 0.1933 |
68
+ | 0.2001 | 1.1368 | 320 | 0.1905 |
69
+ | 0.1887 | 1.2078 | 340 | 0.1905 |
70
+ | 0.1921 | 1.2789 | 360 | 0.1880 |
71
+ | 0.1837 | 1.3499 | 380 | 0.1870 |
72
+ | 0.1951 | 1.4210 | 400 | 0.1870 |
73
 
74
 
75
  ### Framework versions
adapter_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "base_model_name_or_path": "codellama/CodeLlama-7b-hf",
3
  "bias": "none",
4
  "enable_lora": null,
5
  "fan_in_fan_out": false,
 
1
  {
2
+ "base_model_name_or_path": "codellama/CodeLlama-13b-hf",
3
  "bias": "none",
4
  "enable_lora": null,
5
  "fan_in_fan_out": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94ad8700b9a4a5daf3437d5f88d2fe69cfc1fe29b79232d9212bdb310819c1aa
3
- size 67201802
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c45481da4dbc3bf130d955da7fe7c8ac262f9f093647695c0d0d083f1573b45
3
+ size 104973834
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9484c2af318f1062491673b0333ffde7684776871e763cb0eb8c600c43677fc5
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee61becd0124455401f98e9232ca9efe80fc5210deb6f63ad99d546146021fcf
3
  size 5176