alpindale commited on
Commit
547c8f6
1 Parent(s): 35febfc

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +22 -26
README.md CHANGED
@@ -16,6 +16,8 @@ language:
16
 
17
  # Model Card for C4AI Command R+
18
 
 
 
19
 
20
  ## Model Summary
21
 
@@ -92,32 +94,7 @@ print(gen_text)
92
 
93
  **Quantized model through bitsandbytes, 4-bit precision**
94
 
95
- ```python
96
- # pip install 'git+https://github.com/huggingface/transformers.git' bitsandbytes accelerate
97
- from transformers import AutoTokenizer, AutoModelForCausalLM, BitsAndBytesConfig
98
-
99
- bnb_config = BitsAndBytesConfig(load_in_4bit=True)
100
-
101
- model_id = "CohereForAI/c4ai-command-r-plus"
102
- tokenizer = AutoTokenizer.from_pretrained(model_id)
103
- model = AutoModelForCausalLM.from_pretrained(model_id, quantization_config=bnb_config)
104
-
105
- # Format message with the command-r-plus chat template
106
- messages = [{"role": "user", "content": "Hello, how are you?"}]
107
- input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
108
- ## <BOS_TOKEN><|START_OF_TURN_TOKEN|><|USER_TOKEN|>Hello, how are you?<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>
109
-
110
- gen_tokens = model.generate(
111
- input_ids,
112
- max_new_tokens=100,
113
- do_sample=True,
114
- temperature=0.3,
115
- )
116
-
117
- gen_text = tokenizer.decode(gen_tokens[0])
118
- print(gen_text)
119
- ```
120
-
121
 
122
  ## Model Details
123
 
@@ -133,6 +110,25 @@ Pre-training data additionally included the following 13 languages: Russian, Pol
133
 
134
  **Context length**: Command R+ supports a context length of 128K.
135
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
136
  ### Tool use & multihop capabilities:
137
 
138
  Command R+ has been specifically trained with conversational tool use capabilities. These have been trained into the model via a mixture of supervised fine-tuning and preference fine-tuning, using a specific prompt template. Deviating from this prompt template will likely reduce performance, but we encourage experimentation.
 
16
 
17
  # Model Card for C4AI Command R+
18
 
19
+ 🚨 **This model is non-quantized version of C4AI Command R+. You can find the quantized version of C4AI Command R+ using bitsandbytes [here](https://huggingface.co/CohereForAI/c4ai-command-r-plus-4bit)**.
20
+
21
 
22
  ## Model Summary
23
 
 
94
 
95
  **Quantized model through bitsandbytes, 4-bit precision**
96
 
97
+ This model is non-quantized version of C4AI Command R+. You can find the quantized version of C4AI Command R+ using bitsandbytes [here](https://huggingface.co/CohereForAI/c4ai-command-r-plus-4bit).
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
98
 
99
  ## Model Details
100
 
 
110
 
111
  **Context length**: Command R+ supports a context length of 128K.
112
 
113
+ ## Evaluations
114
+
115
+ Command R+ has been submitted to the [Open LLM leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). We include the results below, along with a direct comparison to the strongest state-of-art open weights models currently available on Hugging Face. We note that these results are only useful to compare when evaluations are implemented for all models in a [standardized way](https://github.com/EleutherAI/lm-evaluation-harness) using publically available code, and hence shouldn't be used for comparison outside of models submitted to the leaderboard or compared to self-reported numbers which can't be replicated in the same way.
116
+
117
+ | Model | Average | Arc (Challenge) | Hella Swag | MMLU | Truthful QA | Winogrande | GSM8k |
118
+ |:--------------------------------|----------:|------------------:|-------------:|-------:|--------------:|-------------:|--------:|
119
+ | **CohereForAI/c4ai-command-r-plus** | 74.6 | 70.99 | 88.6 | 75.7 | 56.3 | 85.4 | 70.7 |
120
+ | [DBRX Instruct](https://huggingface.co/databricks/dbrx-instruct) | 74.5 | 68.9 | 89 | 73.7 | 66.9 | 81.8 | 66.9 |
121
+ | [Mixtral 8x7B-Instruct](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) | 72.7 | 70.1 | 87.6 | 71.4 | 65 | 81.1 | 61.1 |
122
+ | [Mixtral 8x7B Chat](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) | 72.6 | 70.2 | 87.6 | 71.2 | 64.6 | 81.4 | 60.7 |
123
+ | [CohereForAI/c4ai-command-r-v01](https://huggingface.co/CohereForAI/c4ai-command-r-v01) | 68.5 | 65.5 | 87 | 68.2 | 52.3 | 81.5 | 56.6 |
124
+ | [Llama 2 70B](https://huggingface.co/meta-llama/Llama-2-70b-hf) | 67.9 | 67.3 | 87.3 | 69.8 | 44.9 | 83.7 | 54.1 |
125
+ | [Yi-34B-Chat](https://huggingface.co/01-ai/Yi-34B-Chat) | 65.3 | 65.4 | 84.2 | 74.9 | 55.4 | 80.1 | 31.9 |
126
+ | [Gemma-7B](https://huggingface.co/google/gemma-7b) | 63.8 | 61.1 | 82.2 | 64.6 | 44.8 | 79 | 50.9 |
127
+ | [LLama 2 70B Chat](https://huggingface.co/meta-llama/Llama-2-70b-chat-hf) | 62.4 | 64.6 | 85.9 | 63.9 | 52.8 | 80.5 | 26.7 |
128
+ | [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) | 61 | 60 | 83.3 | 64.2 | 42.2 | 78.4 | 37.8 |
129
+
130
+ We include these metrics here because they are frequently requested, but note that these metrics do not capture RAG, multilingual, tooling performance or the evaluation of open ended generations which we believe Command R+ to be state-of-art at. For evaluations of RAG, multilingual and tooling read more [here](https://txt.cohere.com/command-r-plus-microsoft-azure/). For evaluation of open ended generation, Command R+ is currently being evaluated on the [chatbot arena](https://chat.lmsys.org/).
131
+
132
  ### Tool use & multihop capabilities:
133
 
134
  Command R+ has been specifically trained with conversational tool use capabilities. These have been trained into the model via a mixture of supervised fine-tuning and preference fine-tuning, using a specific prompt template. Deviating from this prompt template will likely reduce performance, but we encourage experimentation.