lindseyb commited on
Commit
ae3454c
·
verified ·
1 Parent(s): a278645

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -18,11 +18,10 @@ enabling high performance and high efficiency to make the world smarter.
18
 
19
  # Getting Started with Hugging Face Transformers
20
 
21
- Details on getting started
22
- with Hugging Face models are available on the [Optimum page](https://huggingface.co/docs/optimum/main/en/amd/index)
23
 
24
- The following section describes how to use the most common transformers on Hugging Face
25
- for inference workloads on select AMD Instinct™ accelerators and AMD Radeon™ GPUs using the AMD ROCm software ecosystem.
 
26
  This base knowledge can be leveraged to start fine-tuning from a base model or even start developing your own model.
27
  General Linux and ML experience is a required pre-requisite.
28
 
@@ -94,6 +93,9 @@ Click on the 'Use in Transformers' button to see the exact code to import a spec
94
  For a deeper dive into using Hugging Face libraries on AMD GPUs, check out the [Optimum](https://huggingface.co/docs/optimum/main/en/amd/amdgpu/overview) page
95
  describing details on Flash Attention 2, GPTQ Quantization and ONNX Runtime integration.
96
 
 
 
 
97
  # Serving a model with TGI
98
 
99
  Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.
 
18
 
19
  # Getting Started with Hugging Face Transformers
20
 
 
 
21
 
22
+
23
+ Looking for how to use the most common transformers on Hugging Face
24
+ for inference workloads on select AMD Instinct™ accelerators and AMD Radeon™ GPUs using the AMD ROCm software ecosystem?
25
  This base knowledge can be leveraged to start fine-tuning from a base model or even start developing your own model.
26
  General Linux and ML experience is a required pre-requisite.
27
 
 
93
  For a deeper dive into using Hugging Face libraries on AMD GPUs, check out the [Optimum](https://huggingface.co/docs/optimum/main/en/amd/amdgpu/overview) page
94
  describing details on Flash Attention 2, GPTQ Quantization and ONNX Runtime integration.
95
 
96
+ Details on getting started
97
+ with Hugging Face models are available on the [Optimum page](https://huggingface.co/docs/optimum/main/en/amd/index)
98
+
99
  # Serving a model with TGI
100
 
101
  Text Generation Inference (a.k.a “TGI”) provides an end-to-end solution to deploy large language models for inference at scale.