Update README.md
Browse files
README.md
CHANGED
@@ -18,11 +18,10 @@ enabling high performance and high efficiency to make the world smarter.
|
|
18 |
|
19 |
# Getting Started with Hugging Face Transformers
|
20 |
|
21 |
-
Details on getting started
|
22 |
-
with Hugging Face models are available on the [Optimum page](https://huggingface.co/docs/optimum/main/en/amd/index)
|
23 |
|
24 |
-
|
25 |
-
for
|
|
|
26 |
This base knowledge can be leveraged to start fine-tuning from a base model or even start developing your own model.
|
27 |
General Linux and ML experience is a required pre-requisite.
|
28 |
|
@@ -94,6 +93,9 @@ Click on the 'Use in Transformers' button to see the exact code to import a spec
|
|
94 |
For a deeper dive into using Hugging Face libraries on AMD GPUs, check out the [Optimum](https://huggingface.co/docs/optimum/main/en/amd/amdgpu/overview) page
|
95 |
describing details on Flash Attention 2, GPTQ Quantization and ONNX Runtime integration.
|
96 |
|
|
|
|
|
|
|
97 |
# Serving a model with TGI
|
98 |
|
99 |
Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
|
|
|
18 |
|
19 |
# Getting Started with Hugging Face Transformers
|
20 |
|
|
|
|
|
21 |
|
22 |
+
|
23 |
+
Looking for how to use the most common transformers on Hugging Face
|
24 |
+
for inference workloads on select AMD Instinct™ accelerators and AMD Radeon™ GPUs using the AMD ROCm software ecosystem?
|
25 |
This base knowledge can be leveraged to start fine-tuning from a base model or even start developing your own model.
|
26 |
General Linux and ML experience is a required pre-requisite.
|
27 |
|
|
|
93 |
For a deeper dive into using Hugging Face libraries on AMD GPUs, check out the [Optimum](https://huggingface.co/docs/optimum/main/en/amd/amdgpu/overview) page
|
94 |
describing details on Flash Attention 2, GPTQ Quantization and ONNX Runtime integration.
|
95 |
|
96 |
+
Details on getting started
|
97 |
+
with Hugging Face models are available on the [Optimum page](https://huggingface.co/docs/optimum/main/en/amd/index)
|
98 |
+
|
99 |
# Serving a model with TGI
|
100 |
|
101 |
Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
|