Update README.md
Browse files
README.md
CHANGED
@@ -25,9 +25,9 @@ _Fast Inference with Customization:_ As with our previous version, once trained,
|
|
25 |
- **HuggingFace** (access ELM Turbo Models in HF): π [here](https://huggingface.co/collections/slicexai/elm-turbo-66945032f3626024aa066fde)
|
26 |
|
27 |
## ELM Turbo Model Release
|
28 |
-
In this version, we employed our new, improved decomposable ELM techniques on a widely used open-source LLM, `meta-llama/Meta-Llama-3.1-8B-Instruct` (8B params) (check [Llama-license](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B/blob/main/LICENSE) for usage). After training, we generated three smaller slices with parameter counts ranging from 3B billion to 6B billion.
|
29 |
|
30 |
-
- [Section 1.](https://huggingface.co/slicexai/Llama3.1-elm-turbo-
|
31 |
|
32 |
**NOTE**: The open-source datasets from the HuggingFace hub used for instruction fine-tuning ELM Turbo include, but are not limited to: `allenai/tulu-v2-sft-mixture`, `microsoft/orca-math-word-problems-200k`, `mlabonne/WizardLM_evol_instruct_70k-ShareGPT`, and `mlabonne/WizardLM_evol_instruct_v2_196K-ShareGPT`. We advise users to exercise caution when utilizing ELM Turbo, as these datasets may contain factually incorrect information, unintended biases, inappropriate content, and other potential issues. It is recommended to thoroughly evaluate the model's outputs and implement appropriate safeguards for your specific use case.
|
33 |
|
@@ -39,12 +39,12 @@ There are three ELM Turbo slices derived from the `Meta-Llama-3.1-8B-Instruct` m
|
|
39 |
|
40 |
Make sure to update your transformers installation via pip install --upgrade transformers.
|
41 |
|
42 |
-
Example - To run the `slicexai/Llama3.1-elm-turbo-
|
43 |
```python
|
44 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
45 |
import torch
|
46 |
|
47 |
-
elm_turbo_model = "slicexai/Llama3.1-elm-turbo-
|
48 |
model = AutoModelForCausalLM.from_pretrained(
|
49 |
elm_turbo_model,
|
50 |
device_map="cuda",
|
|
|
25 |
- **HuggingFace** (access ELM Turbo Models in HF): π [here](https://huggingface.co/collections/slicexai/elm-turbo-66945032f3626024aa066fde)
|
26 |
|
27 |
## ELM Turbo Model Release
|
28 |
+
In this version, we employed our new, improved decomposable ELM techniques on a widely used open-source LLM, `meta-llama/Meta-Llama-3.1-8B-Instruct` (8B params) (check [Llama-license](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B/blob/main/LICENSE) for usage). After training, we generated three smaller slices with parameter counts ranging from 3B billion to 6B billion.
|
29 |
|
30 |
+
- [Section 1.](https://huggingface.co/slicexai/Llama3.1-elm-turbo-4B-instruct#1-run-elm-turbo-models-with-huggingface-transformers-library) π instructions to run ELM-Turbo with the Huggingface Transformers library.
|
31 |
|
32 |
**NOTE**: The open-source datasets from the HuggingFace hub used for instruction fine-tuning ELM Turbo include, but are not limited to: `allenai/tulu-v2-sft-mixture`, `microsoft/orca-math-word-problems-200k`, `mlabonne/WizardLM_evol_instruct_70k-ShareGPT`, and `mlabonne/WizardLM_evol_instruct_v2_196K-ShareGPT`. We advise users to exercise caution when utilizing ELM Turbo, as these datasets may contain factually incorrect information, unintended biases, inappropriate content, and other potential issues. It is recommended to thoroughly evaluate the model's outputs and implement appropriate safeguards for your specific use case.
|
33 |
|
|
|
39 |
|
40 |
Make sure to update your transformers installation via pip install --upgrade transformers.
|
41 |
|
42 |
+
Example - To run the `slicexai/Llama3.1-elm-turbo-4B-instruct`
|
43 |
```python
|
44 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
45 |
import torch
|
46 |
|
47 |
+
elm_turbo_model = "slicexai/Llama3.1-elm-turbo-4B-instruct"
|
48 |
model = AutoModelForCausalLM.from_pretrained(
|
49 |
elm_turbo_model,
|
50 |
device_map="cuda",
|