pszemraj commited on
Commit
68b4958
1 Parent(s): cf208af

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -18
README.md CHANGED
@@ -1,35 +1,30 @@
1
  ---
2
- base_model: silu-griffin-1024-c3t-8layer
3
  tags:
4
  - generated_from_trainer
5
  metrics:
6
  - accuracy
7
- model-index:
8
- - name: silu-griffin-1024-c3t-8layer-simple_wikipedia_LM-vN
9
- results: []
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- # silu-griffin-1024-c3t-8layer-simple_wikipedia_LM-vN
16
 
17
- This model is a fine-tuned version of [silu-griffin-1024-c3t-8layer](https://huggingface.co/silu-griffin-1024-c3t-8layer) on the pszemraj/simple_wikipedia_LM dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 4.0476
20
- - Accuracy: 0.4224
21
 
22
  ## Model description
23
 
24
- More information needed
25
-
26
- ## Intended uses & limitations
27
-
28
- More information needed
29
 
30
- ## Training and evaluation data
31
-
32
- More information needed
33
 
34
  ## Training procedure
35
 
@@ -65,4 +60,4 @@ The following hyperparameters were used during training:
65
  - Transformers 4.40.1
66
  - Pytorch 2.2.0+cu121
67
  - Datasets 2.19.0
68
- - Tokenizers 0.19.1
 
1
  ---
 
2
  tags:
3
  - generated_from_trainer
4
  metrics:
5
  - accuracy
6
+ license: apache-2.0
7
+ datasets:
8
+ - pszemraj/simple_wikipedia_LM
9
+ language:
10
+ - en
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ # griffin-v0.01-c3t-8layer-simplewiki-silu
17
 
18
+ - griffin/recurrent_gemma arch
19
+ - claude3 tokenizer (as an HF gpt2 tokenizer)
 
 
20
 
21
  ## Model description
22
 
23
+ pretrain experiment on the pszemraj/simple_wikipedia_LM dataset.
 
 
 
 
24
 
25
+ It achieves the following results on the evaluation set:
26
+ - Loss: 4.0476
27
+ - Accuracy: 0.4224
28
 
29
  ## Training procedure
30
 
 
60
  - Transformers 4.40.1
61
  - Pytorch 2.2.0+cu121
62
  - Datasets 2.19.0
63
+ - Tokenizers 0.19.1