Muennighoff commited on
Commit
83b89f8
1 Parent(s): 9a00d10

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -25
README.md CHANGED
@@ -2,6 +2,8 @@
2
  pipeline_tag: text-generation
3
  inference: true
4
  license: apache-2.0
 
 
5
  ---
6
 
7
  # Table of Contents
@@ -13,39 +15,22 @@ license: apache-2.0
13
 
14
  # Model Summary
15
 
16
- > GritLM is a generative-representational instruction-tuned language model. It performs well at both text representation and text generation.
17
 
18
  - **Repository:** [ContextualAI/gritlm](https://github.com/ContextualAI/gritlm)
19
  - **Paper:** [TODO](https://arxiv.org/abs/2308.07124)
20
 
21
- # Use
22
-
23
- The models usage is documented [here](TODO). It supports GritLM, Transformers, Sentence Transformers.
24
-
25
- # Training
26
-
27
- ## Model
28
 
29
- - **Architecture:** Mistral-8x7B
30
- - **Steps:** 250k pretraining & 30 instruction tuning
31
- - **Pretraining tokens:** ? pretraining & 2M instruction tuning
32
- - **Precision:** bfloat16
33
-
34
- ## Hardware
35
-
36
- - **Pretraining:**
37
- - **GPUs:** 512 Tesla A100
38
- - **Training time:** 1 day
39
- - **Instruction tuning:**
40
- - **GPUs:** 8 Tesla A100
41
- - **Training time:** 4 hours
42
-
43
- ## Software
44
 
45
- https://github.com/ContextualAI/gritlm
46
 
47
  # Citation
48
 
49
  ```bibtex
50
  TODO
51
- ```
 
2
  pipeline_tag: text-generation
3
  inference: true
4
  license: apache-2.0
5
+ datasets:
6
+ - GritLM/tulu2
7
  ---
8
 
9
  # Table of Contents
 
15
 
16
  # Model Summary
17
 
18
+ > GritLM is a generative representational instruction tuned language model. It unifies text representation (embedding) and text generation into a single model achieving state-of-the-art performance on both types of tasks.
19
 
20
  - **Repository:** [ContextualAI/gritlm](https://github.com/ContextualAI/gritlm)
21
  - **Paper:** [TODO](https://arxiv.org/abs/2308.07124)
22
 
23
+ | Model | Description |
24
+ |-------|-------------|
25
+ | [GritLM 7B](https://hf.co/GritLM/GritLM-7B) | Mistral 7B finetuned using GRIT |
26
+ | [GritLM 8x7B](https://hf.co/GritLM/GritLM-8x7B) | Mixtral 8x7B finetuned using GRIT |
 
 
 
27
 
28
+ # Use
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
+ The model usage is documented [here](TODO). It supports GritLM, Transformers, Sentence Transformers.
31
 
32
  # Citation
33
 
34
  ```bibtex
35
  TODO
36
+ ```