Update README.md
Browse files
README.md
CHANGED
@@ -1,152 +1,89 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
border-collapse: collapse;
|
10 |
-
font-family: Arial, sans-serif;
|
11 |
-
}
|
12 |
-
th, td {
|
13 |
-
border: 1px solid #ddd;
|
14 |
-
padding: 8px;
|
15 |
-
}
|
16 |
-
th {
|
17 |
-
background-color: #f4f4f4;
|
18 |
-
text-align: left;
|
19 |
-
}
|
20 |
-
tr:nth-child(even) {
|
21 |
-
background-color: #f9f9f9;
|
22 |
-
}
|
23 |
-
tr:hover {
|
24 |
-
background-color: #f1f1f1;
|
25 |
-
}
|
26 |
-
h2, p {
|
27 |
-
font-family: Arial, sans-serif;
|
28 |
-
}
|
29 |
-
</style>
|
30 |
-
</head>
|
31 |
-
<body>
|
32 |
-
|
33 |
-
<h2>Explore Arm-Optimized Learning Paths on Hugging Face</h2>
|
34 |
-
<p>
|
35 |
-
Discover curated <strong>Learning Paths</strong> that showcase AI models optimized for Arm platforms across key market applications.
|
36 |
-
Each learning path highlights specific models featured within our dedicated Hugging Face <strong>Model Collections</strong>,
|
37 |
-
simplifying your journey from learning to deployment on Arm technologies.
|
38 |
-
</p>
|
39 |
-
|
40 |
-
<table>
|
41 |
-
<thead>
|
42 |
-
<tr>
|
43 |
-
<th>Learning Path</th>
|
44 |
-
<th>Market Application</th>
|
45 |
-
<th>Model(s) Featured</th>
|
46 |
-
<th>Arm Platform</th>
|
47 |
-
</tr>
|
48 |
-
</thead>
|
49 |
-
<tbody>
|
50 |
-
<tr>
|
51 |
-
<td>Build a RAG application using Zilliz Cloud on Arm servers</td>
|
52 |
-
<td>Cloud, Datacenter</td>
|
53 |
-
<td>all-MiniLM-L6-v2</td>
|
54 |
-
<td>AWS Graviton3 (c7g.2xlarge)</td>
|
55 |
-
</tr>
|
56 |
-
<tr>
|
57 |
-
<td>Accelerate NLP models from Hugging Face on Arm servers</td>
|
58 |
-
<td>Cloud, Datacenter</td>
|
59 |
-
<td>DistilBERT base uncased finetuned SST-2</td>
|
60 |
-
<td>AWS Graviton3</td>
|
61 |
-
</tr>
|
62 |
-
<tr>
|
63 |
-
<td>Deploy a Large Language Model (LLM) chatbot with llama.cpp using KleidiAI on Arm servers</td>
|
64 |
-
<td>Cloud, Datacenter</td>
|
65 |
-
<td>Dolphin 2.9.4 Llama 3.1 8b</td>
|
66 |
-
<td>AWS Graviton4 (r8g.16xlarge)</td>
|
67 |
-
</tr>
|
68 |
-
<tr>
|
69 |
-
<td>Run a Large Language Model (LLM) chatbot with PyTorch using KleidiAI on Arm servers</td>
|
70 |
-
<td>Cloud, Datacenter</td>
|
71 |
-
<td>Llama-3.1-8B-Instruct</td>
|
72 |
-
<td>Arm-based instance with at least 16 CPUs</td>
|
73 |
-
</tr>
|
74 |
-
<tr>
|
75 |
-
<td>Deploy a RAG-based Chatbot with llama-cpp-python using KleidiAI on Google Axion processors</td>
|
76 |
-
<td>Cloud, Datacenter</td>
|
77 |
-
<td>Llama3.1 8b</td>
|
78 |
-
<td>Google Cloud Axion</td>
|
79 |
-
</tr>
|
80 |
-
<tr>
|
81 |
-
<td>Build an Android chat app with Llama, KleidiAI, ExecuTorch, and XNNPACK</td>
|
82 |
-
<td>Edge Device (Smartphone)</td>
|
83 |
-
<td>Llama-3.2-1B-Instruct</td>
|
84 |
-
<td>Smartphone based on Arm CPU (with i8mm support)</td>
|
85 |
-
</tr>
|
86 |
-
<tr>
|
87 |
-
<td>Run a local LLM chatbot on a Raspberry Pi 5</td>
|
88 |
-
<td>Edge Device (Raspberry Pi)</td>
|
89 |
-
<td>Orca-Mini-3B</td>
|
90 |
-
<td>Raspberry Pi 5</td>
|
91 |
-
</tr>
|
92 |
-
<tr>
|
93 |
-
<td>Build an Android chat application with ONNX Runtime API</td>
|
94 |
-
<td>Edge Device (Smartphone)</td>
|
95 |
-
<td>Phi-3-vision-128k-instruct-onnx-cuda</td>
|
96 |
-
<td>Samsung Galaxy S24 Android smartphone</td>
|
97 |
-
</tr>
|
98 |
-
<tr>
|
99 |
-
<td>Run an LLM chatbot with rtp-llm on Arm-based servers</td>
|
100 |
-
<td>Cloud, Datacenter</td>
|
101 |
-
<td>Qwen2 0.5B-Instruct</td>
|
102 |
-
<td>Arm server based on Neoverse N2 or Arm Neoverse V2</td>
|
103 |
-
</tr>
|
104 |
-
<tr>
|
105 |
-
<td>Build and Run a Virtual Large Language Model (vLLM) on Arm Servers</td>
|
106 |
-
<td>Cloud, Datacenter</td>
|
107 |
-
<td>Qwen2.5-0.5B-Instruct</td>
|
108 |
-
<td>Arm-based instance with at least 8 CPUs and 16 GB RAM</td>
|
109 |
-
</tr>
|
110 |
-
<tr>
|
111 |
-
<td>Run a Natural Language Processing (NLP) model from Hugging Face on Arm servers</td>
|
112 |
-
<td>Cloud, Datacenter</td>
|
113 |
-
<td>twitter-roberta-base-sentiment-latest</td>
|
114 |
-
<td>Arm AArch64 CPU based server</td>
|
115 |
-
</tr>
|
116 |
-
<tr>
|
117 |
-
<td>Profile the Performance of AI and ML Mobile Applications on Arm</td>
|
118 |
-
<td>Edge Device (Smartphone)</td>
|
119 |
-
<td>MobileNet V2</td>
|
120 |
-
<td>Arm-based Android smartphone</td>
|
121 |
-
</tr>
|
122 |
-
<tr>
|
123 |
-
<td>Get started with object detection using a Jetson Orin Nano</td>
|
124 |
-
<td>Edge Device</td>
|
125 |
-
<td>MobileNet V2</td>
|
126 |
-
<td>NVIDIA Jetson Oron Nano</td>
|
127 |
-
</tr>
|
128 |
-
<tr>
|
129 |
-
<td>Create a ChatGPT voice bot on a Raspberry Pi</td>
|
130 |
-
<td>Edge Device (Raspberry Pi)</td>
|
131 |
-
<td>gpt-4-turbo-preview</td>
|
132 |
-
<td>Raspberry Pi 5</td>
|
133 |
-
</tr>
|
134 |
-
<tr>
|
135 |
-
<td>LLM inference on Android with KleidiAI, MediaPipe, and XNNPACK</td>
|
136 |
-
<td>Edge Device (Smartphone)</td>
|
137 |
-
<td>Gemma 2B</td>
|
138 |
-
<td>Google Pixel 8 Pro (Android phone with support for i8mm)</td>
|
139 |
-
</tr>
|
140 |
-
<tr>
|
141 |
-
<td>Run Llama 3 on a Raspberry Pi 5 using ExecuTorch</td>
|
142 |
-
<td>Edge Device (Raspberry Pi)</td>
|
143 |
-
<td>Llama 3.1 8B</td>
|
144 |
-
<td>Raspberry Pi 5</td>
|
145 |
-
</tr>
|
146 |
-
</tbody>
|
147 |
-
</table>
|
148 |
-
|
149 |
-
<p><i><small>Note: The data collated here is sourced from Arm and third parties. While Arm uses reasonable efforts to keep this information accurate, Arm does not warrant (express or implied) or provide any guarantee of data correctness due to the ever-evolving AI and software landscape. Any links to third party sites and resources are provided for ease and convenience. Your use of such third-party sites and resources is subject to the third party’s terms of use, and use is at your own risk.</small></i></p>
|
150 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
151 |
</body>
|
152 |
-
</html>
|
|
|
1 |
+
---
|
2 |
+
title: README
|
3 |
+
emoji: 🦀
|
4 |
+
colorFrom: indigo
|
5 |
+
colorTo: purple
|
6 |
+
sdk: static
|
7 |
+
pinned: false
|
8 |
+
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
|
10 |
+
<p>Arm’s AI development resources ensure you can deploy at pace, achieving best performance on Arm by default. Our aim is to make your AI development easier, ensuring integration with all major operating systems and AI frameworks, enabling portability for deploying AI on Arm at scale.</p>
|
11 |
+
<p>Discover below some key resources and content from Arm, including our software libraries and tools, that enable you to optimize for Arm architectures and pass-on significant performance uplift for models – from traditional ML and computer vision workloads to small and large language models - running on Arm-based devices.</p>
|
12 |
+
<br>
|
13 |
+
<strong>Arm and Meta: <a href="https://huggingface.co/collections/meta-llama/llama-32-66f448ffc8c32f949b04c8cf" target="_blank">Llama 3.2</a><br>Accelerated cloud to edge AI performance</strong>
|
14 |
+
<p>The availability of smaller LLMs that enable fundamental text-based generative AI workloads, such as Llama 3.2 1B and 3B, are critical to enabling AI inference at scale. Running the new Llama 3.2 3B LLM on Arm-powered mobile devices through the Arm CPU optimized kernel leads to a 5x improvement in prompt processing and 3x improvement in token generation, achieving 19.92 tokens per second in the generation phase. This means less latency when processing AI workloads on the device and a far faster overall user experience. Also, the more AI processed at the edge, the more power that is saved from data traveling to and from the cloud, leading to energy and cost savings.</p>
|
15 |
+
<p>Alongside running small models at the edge, we are also able to run larger models, such as Llama 3.2 11B and 90B, in the cloud. The 11B and 90B models are a great fit for CPU based inference workloads in the cloud that generate text and image, as our data on Arm Neoverse V2 shows. When we run the 11B image and text model on the Arm-based AWS Graviton4, we can achieve 29.3 tokens per second in the generation phase. When you consider that the human reading speed is around 5 tokens per second, it’s far outpacing that.</p>
|
16 |
+
<ul><p>
|
17 |
+
<li><a href="https://newsroom.arm.com/news/ai-inference-everywhere-with-new-llama-llms-on-arm" target="_blank">Accelerating and Scaling AI Inference Everywhere with New Llama 3.2 LLMs on Arm</a></li>
|
18 |
+
<li><a href="https://ai.meta.com/blog/llama-3-2-connect-2024-vision-edge-mobile-devices/" target="_blank">Meta Llama 3.2 Blog</a></li>
|
19 |
+
<li><a href=" https://www.llama.com/docs/getting-the-models/1b3b-partners" target="_blank">Meta Llama 3.2 1b/3/b Partner Guide</a></li>
|
20 |
+
<li><a href="https://www.youtube.com/watch?v=AVqm7SfNQrw" target="_blank">How Arm and Meta are Transforming AI Software Development</a></li>
|
21 |
+
<li><a href="https://www.arm.com/markets/artificial-intelligence/software" target="_blank">Arm AI Software Page</a></li>
|
22 |
+
</p></ul>
|
23 |
+
<br>
|
24 |
+
<strong>Arm Kleidi: Unleashing Mass-Market AI Performance on Arm</strong>
|
25 |
+
<p>Arm Kleidi is a targeted software suite, expediting optimizations for any framework and enabling accelerations for billions of AI workloads across Arm-based devices everywhere. Application developers achieve top performance by default, with no additional work or investment in new skills or tools training required.</p>
|
26 |
+
<p><b>Useful Resources on Arm Kleidi:</b></p>
|
27 |
+
<ul><p>
|
28 |
+
<li><a href="https://newsroom.arm.com/blog/kleidiai-integration-mediapipe" target="_blank">KleidiAI integration with Google's MediaPipe framework</a></li>
|
29 |
+
<li>Arm KleidiAI for optimizing any AI framework: <a href="https://gitlab.arm.com/kleidi/kleidiai" target="_blank">Gitlab repo</a> and <a href="https://community.arm.com/arm-community-blogs/b/ai-and-ml-blog/posts/kleidiai" target="_blank"> blog</a></li>
|
30 |
+
<li>Arm KleidiCV for optimizing any computer vision framework: <a href="https://gitlab.arm.com/kleidi/kleidicv" target="_blank">Gitlab repo</a> and <a href="https://community.arm.com/arm-community-blogs/b/ai-and-ml-blog/posts/kleidicv" target="_blank"> blog</a></li>
|
31 |
+
<li><a href="https://github.com/ARM-software/ComputeLibrary" target="_blank">Arm Compute Library for all AI software</a></li>
|
32 |
+
</p></ul>
|
33 |
+
<br>
|
34 |
+
<strong>Running LLMs on Mobile</strong>
|
35 |
+
<br>
|
36 |
+
<p>Our foundation of pervasiveness, flexible performance and energy efficiency mean that Arm CPUs are already the hardware of choice for a variety of AI workloads. Alongside Arm-based servers excelling with LLM workloads, the Arm Kleidi software suite, optimizations to our software libraries, combined with the <a href="https://github.com/ggerganov/llama.cpp" target="_blank">open-source llama.cpp project</a> enable generative AI to run efficiently on mobile devices.</p>
|
37 |
+
<p>Our work includes a virtual assistant demo which at first utilized Meta’s Llama2-7B LLM on mobile via a chat-based application, and has since expanded to include the Llama3 model and Phi-3 3.8B. <a href="https://community.arm.com/arm-community-blogs/b/ai-and-ml-blog/posts/generative-ai-on-mobile-on-arm-cpu" target="_blank">You can learn more about the technical implementation of the demos here. </a></p>
|
38 |
+
<p><b>Find out more about the community contributions that make this happen:</b></p>
|
39 |
+
<ul><p>
|
40 |
+
<li><a href="https://huggingface.co/TheBloke/Llama-2-7B-GGUF" target="_blank">Llama-2-7B-GGUF</a></li>
|
41 |
+
<li><a href="https://huggingface.co/TheBloke/phi-2-GGUF" target="_blank">Phi-2-GGUF</a></li>
|
42 |
+
<li><a href="https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF" target="_blank">Meta-Llama-3-8B-Instruct-GGUF</a></li>
|
43 |
+
<li><a href="https://huggingface.co/SanctumAI/Phi-3-mini-4k-instruct-GGUF" target="_blank">Phi-3-mini-4k-instruct-GGUF</a></li>
|
44 |
+
<!-- Add more links here -->
|
45 |
+
</p></ul>
|
46 |
+
<p>These advancements are also highlighted in our Learning Paths below.</p>
|
47 |
+
<br>
|
48 |
+
<strong>AI on Arm in the Cloud</strong>
|
49 |
+
<br>
|
50 |
+
<p>Arm Neoverse platforms give our infrastructure partners access to leading performance, efficiency and unparalleled flexibility to innovate in pursuit of the optimal solutions for emerging AI workloads. The flexibility of the Neoverse platform enables our innovative hardware partners to closely integrate additional compute acceleration into their designs, creating a new generation of built-for-AI custom data center silicon.</p>
|
51 |
+
<p><b>Read the latest on AI-on-Neoverse:</b></p>
|
52 |
+
<ul><p><li>
|
53 |
+
<a href="https://www.arm.com/developer-hub/servers-and-cloud-computing/ai-cloud" target="_blank">Accelerate Your GenAI, AI and ML Workloads on Arm CPUs</a>
|
54 |
+
</li>
|
55 |
+
<li>
|
56 |
+
<a href="https://community.arm.com/arm-community-blogs/b/infrastructure-solutions-blog/posts/accelerating-sentiment-analysis-on-arm-neoverse-cpus" target="_blank">Accelerating Popular Hugging Face Models using Arm Neoverse</a>
|
57 |
+
</li>
|
58 |
+
<li>
|
59 |
+
<a href="https://newsroom.arm.com/blog/small-language-models-on-arm" target="_blank">Small Language Models: Efficient Arm Computing Enables a Custom AI Future</a>
|
60 |
+
</li>
|
61 |
+
<li>
|
62 |
+
<a href="https://community.arm.com/arm-community-blogs/b/infrastructure-solutions-blog/posts/best-in-class-llm-performance-on-arm-neoverse-v1-based-aws-graviton3-servers" target="_blank">Best-in-class LLM Performance on Arm Neoverse V1 based AWS Graviton3 CPUs</a>
|
63 |
+
</li></p>
|
64 |
+
</ul>
|
65 |
+
<!-- Add relevant links or content here -->
|
66 |
+
<br>
|
67 |
+
<strong>Arm Learning Paths</strong>
|
68 |
+
<br><p>Tutorials designed to help you develop quality Arm software faster.</p>
|
69 |
+
<ul><p>
|
70 |
+
<li>
|
71 |
+
<a href="https://learn.arm.com/learning-paths/smartphones-and-mobile/kleidiai-on-android-with-mediapipe-and-xnnpack/" class="underline" target="_blank">LLMs on Android with KleidiAI, MediaPipe and XNNPACK</a>
|
72 |
+
</li>
|
73 |
+
<li>
|
74 |
+
<a href="https://learn.arm.com/learning-paths/servers-and-cloud-computing/llama-cpu/llama-chatbot/" class="underline" target="_blank">Run a Large Language model (LLM) chatbot on Arm servers</a>
|
75 |
+
</li>
|
76 |
+
<li>
|
77 |
+
<a href="https://learn.arm.com/learning-paths/servers-and-cloud-computing/nlp-hugging-face/pytorch-nlp-hf/" target="_blank">Deploy an NLP model using PyTorch on Arm-based device</a>
|
78 |
+
</li>
|
79 |
+
<li>
|
80 |
+
<a href="https://learn.arm.com/learning-paths/embedded-systems/llama-python-cpu/" target="_blank">Run a local LLM chatbot on a Raspberry Pi 5</a>
|
81 |
+
</li>
|
82 |
+
<li>
|
83 |
+
<a href="https://learn.arm.com/learning-paths/servers-and-cloud-computing/benchmark-nlp/" target="_blank">Accelerate Natural Language Processing (NLP) models from Hugging Face on Arm servers</a>
|
84 |
+
</li>
|
85 |
+
</p></ul>
|
86 |
+
<p>Contribute to our Learning Paths: <a href="https://github.com/ArmDeveloperEcosystem/arm-learning-paths/discussions/categories/ideas" target="_blank">suggest a new Learning Path</a> or <a href="https://learn.arm.com/learning-paths/cross-platform/_example-learning-path/" target="_blank">create one yourself </a>with support from the Arm community.</p>
|
87 |
+
<br>
|
88 |
+
<p><i><small>Note: The data collated here is sourced from Arm and third parties. While Arm uses reasonable efforts to keep this information accurate, Arm does not warrant (express or implied) or provide any guarantee of data correctness due to the ever-evolving AI and software landscape. Any links to third party sites and resources are provided for ease and convenience. Your use of such third-party sites and resources is subject to the third party’s terms of use, and use is at your own risk.</small></i></p>
|
89 |
</body>
|
|