Spaces:
Running
Update README.md
Browse filesHugging Face is working with Amazon Web Services to make it easier than ever for startups and enterprises to train and deploy Hugging Face models to [AWS Inferentia](https://aws.amazon.com/machine-learning/inferentia/) and [AWS Trainium] (https://aws.amazon.com/machine-learning/trainium/) based instances available on Amazon EC2 or through Amazon SageMaker.
AWS Inferentia accelerators are designed by AWS to deliver high performance at the lowest cost for your deep learning (DL) inference applications. Inferentia2-based Amazon EC2 Inf2 instances are designed to deliver high performance at the lowest cost in Amazon EC2 for your DL inference and generative artificial intelligence (AI) applications. They are optimized to deploy increasingly complex models, such as large language models (LLM) and vision transformers, at scale. Inf2 instances are the first inference-optimized instances in Amazon EC2 to support scale-out distributed inference with ultra-high-speed connectivity between accelerators.
AWS Trainium is the second-generation machine learning (ML) accelerator that AWS purpose built for deep learning training of 100B+ parameter models. Each Amazon Elastic Compute Cloud (EC2) Trn1 instance deploys up to 16 AWS Trainium accelerators to deliver a high-performance, low-cost solution for deep learning (DL) training in the cloud.
** Optimum Neuron **
🤗 Optimum Neuron is the interface between the 🤗 Transformers library and AWS Accelerators including AWS Trainium and AWS Inferentia. It provides a set of tools enabling easy model loading, training and inference on single- and multi-Accelerator settings for different downstream tasks. The list of officially validated models and tasks is available [here](https://huggingface.co/docs/optimum-neuron/package_reference/configuration#supported-architectures).
** Learn More**
[Tutorials](https://huggingface.co/docs/optimum-neuron/tutorials/overview)
[How to Get started](https://huggingface.co/docs/optimum-neuron/guides/overview)
[References](https://huggingface.co/docs/optimum-neuron/package_reference/trainer)
[AWS Neuron Documentation](https://awsdocs-neuron.readthedocs-hosted.com/en/latest/)
**Blogs and Videos**
[Deploy Llama 2 7B on AWS inferentia2 with Amazon SageMaker](https://www.philschmid.de/inferentia2-llama-7b)
[Fine-tune Llama 7B on AWS Trainium](https://www.philschmid.de/fine-tune-llama-7b-trainium)
[Deploy Embedding Models on AWS inferentia2 with Amazon SageMaker](https://www.philschmid.de/inferentia2-embeddings)
[Deploy Stable Diffusion XL on AWS inferentia2 with Amazon SageMaker](https://www.philschmid.de/inferentia2-stable-diffusion-xl)
[Accelerating Transformers with Optimum Neuron, AWS Trainium and AWS Inferentia2](https://www.youtube.com/watch?v=FmjTWags__Q)
[Training Transformers with AWS Trainium and the Hugging Face Neuron AMI](https://www.youtube.com/watch?v=0Y5E8RI_D2E)
Join the conversation on the Hugging Face Optimum Neuron Forum (here](https://discuss.huggingface.co/c/aws-inferentia-trainium/66)
@@ -1,10 +1,10 @@
|
|
1 |
---
|
2 |
title: README
|
3 |
emoji: 🐢
|
4 |
-
colorFrom:
|
5 |
-
colorTo:
|
6 |
sdk: static
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
-
Edit this `README.md` markdown file to author your organization card.
|
|
|
1 |
---
|
2 |
title: README
|
3 |
emoji: 🐢
|
4 |
+
colorFrom: pink
|
5 |
+
colorTo: purple
|
6 |
sdk: static
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
+
Edit this `README.md` markdown file to author your organization card.
|