Update README.md
Browse files
README.md
CHANGED
@@ -27,7 +27,7 @@ _Fast Inference with Customization:_ As with our previous version, once trained,
|
|
27 |
## ELM Turbo Model Release
|
28 |
In this version, we employed our new, improved decomposable ELM techniques on a widely used open-source LLM, `microsoft/Phi-3-mini-128k-instruct` (3.82B params) (check [phi3-license](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct/resolve/main/LICENSE) for usage). After training, we generated three smaller slices with parameter counts ranging from 1.33 billion to 2.01 billion. Furthermore, we seamlessly integrated these slices into NVIDIA's [TensoRT-LLM](https://github.com/NVIDIA/TensorRT-LLM), providing trtllm engines compatible with A100 and H100 GPUs, respectively.
|
29 |
|
30 |
-
- [Section 1.](https://huggingface.co/slicexai/elm-turbo-0.25-instruct#1-run-elm-turbo-models-with-huggingface-transformers-library) π instructions to run ELM-Turbo with the Huggingface Transformers library
|
31 |
- [Section 2.](https://huggingface.co/slicexai/elm-turbo-0.25-instruct#2-running-elm-turbo-via-nvidias-tensorrt-llm) π instructions to run ELM-Turbo engines powered by NVIDIA's TensoRT-LLM.
|
32 |
|
33 |
**NOTE**: The open-source datasets from the HuggingFace hub used for instruction fine-tuning ELM Turbo include, but are not limited to: `allenai/tulu-v2-sft-mixture`, `microsoft/orca-math-word-problems-200k`, `mlabonne/WizardLM_evol_instruct_70k-ShareGPT`, and `mlabonne/WizardLM_evol_instruct_v2_196K-ShareGPT`. We advise users to exercise caution when utilizing ELM Turbo, as these datasets may contain factually incorrect information, unintended biases, inappropriate content, and other potential issues. It is recommended to thoroughly evaluate the model's outputs and implement appropriate safeguards for your specific use case.
|
|
|
27 |
## ELM Turbo Model Release
|
28 |
In this version, we employed our new, improved decomposable ELM techniques on a widely used open-source LLM, `microsoft/Phi-3-mini-128k-instruct` (3.82B params) (check [phi3-license](https://huggingface.co/microsoft/Phi-3-mini-128k-instruct/resolve/main/LICENSE) for usage). After training, we generated three smaller slices with parameter counts ranging from 1.33 billion to 2.01 billion. Furthermore, we seamlessly integrated these slices into NVIDIA's [TensoRT-LLM](https://github.com/NVIDIA/TensorRT-LLM), providing trtllm engines compatible with A100 and H100 GPUs, respectively.
|
29 |
|
30 |
+
- [Section 1.](https://huggingface.co/slicexai/elm-turbo-0.25-instruct#1-run-elm-turbo-models-with-huggingface-transformers-library) π instructions to run ELM-Turbo with the Huggingface Transformers library.
|
31 |
- [Section 2.](https://huggingface.co/slicexai/elm-turbo-0.25-instruct#2-running-elm-turbo-via-nvidias-tensorrt-llm) π instructions to run ELM-Turbo engines powered by NVIDIA's TensoRT-LLM.
|
32 |
|
33 |
**NOTE**: The open-source datasets from the HuggingFace hub used for instruction fine-tuning ELM Turbo include, but are not limited to: `allenai/tulu-v2-sft-mixture`, `microsoft/orca-math-word-problems-200k`, `mlabonne/WizardLM_evol_instruct_70k-ShareGPT`, and `mlabonne/WizardLM_evol_instruct_v2_196K-ShareGPT`. We advise users to exercise caution when utilizing ELM Turbo, as these datasets may contain factually incorrect information, unintended biases, inappropriate content, and other potential issues. It is recommended to thoroughly evaluate the model's outputs and implement appropriate safeguards for your specific use case.
|