seungrok81 commited on
Commit
fed8070
1 Parent(s): 0d53a29

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -48,7 +48,7 @@ Select 'Stable + Linux + Pip + Python + ROCm' to get the specific pip installati
48
 
49
  An example command line (note the versioning of the whl file):
50
 
51
- > `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.4.2`
52
 
53
  ### TensorFlow
54
 
@@ -99,7 +99,7 @@ with Hugging Face models are available on the [Optimum page](https://huggingface
99
  # Serving a model with TGI
100
 
101
  Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
102
- TGI is already usable in production on AMD Instinct™ GPUs through the docker image `ghcr.io/huggingface/text-generation-inference:1.2-rocm`.
103
  Make sure to refer to the [documentation](https://huggingface.co/docs/text-generation-inference/supported_models#supported-hardware)
104
  concerning the support and any limitations.
105
 
@@ -110,6 +110,7 @@ across normal and distributed settings, with various supported optimizations and
110
 
111
  # Useful Links and Blogs
112
 
 
113
  - Detailed Llama-2 results show casing the [Optimum benchmark on AMD Instinct MI250](https://huggingface.co/blog/huggingface-and-optimum-amd)
114
  - Check out our blog titled [Run a Chatgpt-like Chatbot on a Single GPU with ROCm](https://huggingface.co/blog/chatbot-amd-gpu)
115
  - Complete ROCm [Documentation](https://rocm.docs.amd.com/en/latest/) for installation and usage
 
48
 
49
  An example command line (note the versioning of the whl file):
50
 
51
+ > `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm6.2`
52
 
53
  ### TensorFlow
54
 
 
99
  # Serving a model with TGI
100
 
101
  Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
102
+ TGI is already usable in production on AMD Instinct™ GPUs through the docker image `ghcr.io/huggingface/text-generation-inference:latest-rocm`.
103
  Make sure to refer to the [documentation](https://huggingface.co/docs/text-generation-inference/supported_models#supported-hardware)
104
  concerning the support and any limitations.
105
 
 
110
 
111
  # Useful Links and Blogs
112
 
113
+ - Detailed Llama-3 results [Run TGI on AMD Instinct MI300X](https://huggingface.co/blog/huggingface-amd-mi300)
114
  - Detailed Llama-2 results show casing the [Optimum benchmark on AMD Instinct MI250](https://huggingface.co/blog/huggingface-and-optimum-amd)
115
  - Check out our blog titled [Run a Chatgpt-like Chatbot on a Single GPU with ROCm](https://huggingface.co/blog/chatbot-amd-gpu)
116
  - Complete ROCm [Documentation](https://rocm.docs.amd.com/en/latest/) for installation and usage