seungrok81
commited on
Commit
•
fed8070
1
Parent(s):
0d53a29
Update README.md
Browse files
README.md
CHANGED
@@ -48,7 +48,7 @@ Select 'Stable + Linux + Pip + Python + ROCm' to get the specific pip installati
|
|
48 |
|
49 |
An example command line (note the versioning of the whl file):
|
50 |
|
51 |
-
> `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/
|
52 |
|
53 |
### TensorFlow
|
54 |
|
@@ -99,7 +99,7 @@ with Hugging Face models are available on the [Optimum page](https://huggingface
|
|
99 |
# Serving a model with TGI
|
100 |
|
101 |
Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
|
102 |
-
TGI is already usable in production on AMD Instinct™ GPUs through the docker image `ghcr.io/huggingface/text-generation-inference:
|
103 |
Make sure to refer to the [documentation](https://huggingface.co/docs/text-generation-inference/supported_models#supported-hardware)
|
104 |
concerning the support and any limitations.
|
105 |
|
@@ -110,6 +110,7 @@ across normal and distributed settings, with various supported optimizations and
|
|
110 |
|
111 |
# Useful Links and Blogs
|
112 |
|
|
|
113 |
- Detailed Llama-2 results show casing the [Optimum benchmark on AMD Instinct MI250](https://huggingface.co/blog/huggingface-and-optimum-amd)
|
114 |
- Check out our blog titled [Run a Chatgpt-like Chatbot on a Single GPU with ROCm](https://huggingface.co/blog/chatbot-amd-gpu)
|
115 |
- Complete ROCm [Documentation](https://rocm.docs.amd.com/en/latest/) for installation and usage
|
|
|
48 |
|
49 |
An example command line (note the versioning of the whl file):
|
50 |
|
51 |
+
> `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.2`
|
52 |
|
53 |
### TensorFlow
|
54 |
|
|
|
99 |
# Serving a model with TGI
|
100 |
|
101 |
Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
|
102 |
+
TGI is already usable in production on AMD Instinct™ GPUs through the docker image `ghcr.io/huggingface/text-generation-inference:latest-rocm`.
|
103 |
Make sure to refer to the [documentation](https://huggingface.co/docs/text-generation-inference/supported_models#supported-hardware)
|
104 |
concerning the support and any limitations.
|
105 |
|
|
|
110 |
|
111 |
# Useful Links and Blogs
|
112 |
|
113 |
+
- Detailed Llama-3 results [Run TGI on AMD Instinct MI300X](https://huggingface.co/blog/huggingface-amd-mi300)
|
114 |
- Detailed Llama-2 results show casing the [Optimum benchmark on AMD Instinct MI250](https://huggingface.co/blog/huggingface-and-optimum-amd)
|
115 |
- Check out our blog titled [Run a Chatgpt-like Chatbot on a Single GPU with ROCm](https://huggingface.co/blog/chatbot-amd-gpu)
|
116 |
- Complete ROCm [Documentation](https://rocm.docs.amd.com/en/latest/) for installation and usage
|