RyzenAI-1.5_LLM_Hybrid_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid Updated about 1 month ago • 139 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
Gumiho Official Model Parameters for "Gumiho: A Hybrid Architecture to Prioritize Early Tokens in Speculative Decoding" Gumiho: A Hybrid Architecture to Prioritize Early Tokens in Speculative Decoding Paper • 2503.10135 • Published Mar 13
Gumiho: A Hybrid Architecture to Prioritize Early Tokens in Speculative Decoding Paper • 2503.10135 • Published Mar 13
OGA CPU LLM Collection This collection contains AMD-Quark quantized OGA exported models for CPU execution amd/Phi-3-mini-4k-instruct_int4_float16_onnx_cpu Updated Apr 12 amd/Qwen1.5-7B-Chat_uint4_asym_g128_float16_onnx_cpu Updated Apr 12 amd/Llama-3.2-1B-Instruct-awq-uint4-float16-cpu-onnx Updated Apr 28 amd/Llama-3.2-3B-Instruct-awq-uint4-float16-cpu-onnx Updated Apr 28
RyzenAI-1.4_LLM_NPU_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
AMD-RyzenAI-Deepseek-R1-Distill-Hybrid amd/DeepSeek-R1-Distill-Llama-8B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 267 • 1 amd/DeepSeek-R1-Distill-Qwen-1.5B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 16 • 1 amd/DeepSeek-R1-Distill-Qwen-7B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 229 • 1
amd/DeepSeek-R1-Distill-Llama-8B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 267 • 1
amd/DeepSeek-R1-Distill-Qwen-1.5B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 16 • 1
amd/DeepSeek-R1-Distill-Qwen-7B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 229 • 1
RyzenAI-1.3_LLM_NPU_Models Models quantized by Quark and prepared for the OGA-based NPU-only execution flow (Ryzen AI 1.3) amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
Nitro Diffusion 💥 Nitro Diffusion is a series of efficient text-to-image diffusion models built on AMD Instinct™ GPUs. amd/Nitro-1-SD Text-to-Image • Updated 29 days ago • 38 • 7 amd/Nitro-1-PixArt Text-to-Image • Updated 29 days ago • 28 • 3 amd/Nitro-T-0.6B Text-to-Image • Updated 15 days ago • 56 • 2 amd/Nitro-T-1.2B Text-to-Image • Updated 15 days ago • 141 • 3
Quark Quantized ONNX LLMs for Ryzen AI 1.3 EA ONNX Runtime generate() API based models quantized by Quark and optimized for Ryzen AI Strix Point NPU amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Llama-2-7b-hf-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 30 amd/Llama-3-8B-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 14 • 1
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Llama-2-7b-hf-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 30
amd/Llama-3-8B-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 14 • 1
RyzenAI-1.5_LLM_NPU_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
PARD Official Model Parameters for "PARD: Accelerating LLM Inference with Low-Cost PARallel Draft Model Adaptation" amd/PARD-Llama-3.2-1B Text Generation • 1B • Updated May 19 • 1.27k • 2 amd/PARD-DeepSeek-R1-Distill-Qwen-1.5B Text Generation • 2B • Updated May 19 • 165 • 1 amd/PARD-Qwen2.5-0.5B Text Generation • 0.6B • Updated May 19 • 1.24k amd/PARD-Qwen3-0.6B Text Generation • 0.8B • Updated 15 days ago • 31
Quark Quantized MXFP4 models amd/DeepSeek-R1-MXFP4-Preview 357B • Updated 3 days ago • 523 amd/Llama-3.1-405B-Instruct-MXFP4-Preview 218B • Updated 27 days ago • 112
AMDGPU OnnxGenAI Collection ONNX GenAI compatible Language Models to run on AMD Ryzen(TM) GPUs and Radeon Discrete GPUs amd/Llama-2-7b-chat-hf-awq-g128-int4-onnx-directml Updated Apr 8 amd/Llama-2-7b-hf-awq-g128-int4-onnx-directml Updated Apr 10 amd/Llama-3.1-8B-awq-g128-int4-onnx-directml Updated Apr 9 amd/Llama-3.1-8B-Instruct-awq-g128-int4-onnx-directml Updated Apr 9
RyzenAI-1.4_LLM_Hybrid_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid Updated about 1 month ago • 139 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
Instella ✨ Announcing Instella, a series of 3 billion parameter language models developed by AMD, trained from scratch on 128 Instinct MI300X GPUs. amd/Instella-3B-Stage1 Text Generation • 3B • Updated Mar 6 • 37 • 13 amd/Instella-3B Text Generation • 3B • Updated 29 days ago • 93 • 36 amd/Instella-3B-SFT Text Generation • 3B • Updated Mar 6 • 6 • 9 amd/Instella-3B-Instruct Text Generation • 3B • Updated Mar 28 • 154 • 51
AMDGPU onnx optimized image generation ONNX models for AMD Ryzen (TM) AI GPUs and Radeon Discrete GPUs TensorStack/Flux_schnell-f16-onnx Text-to-Image • Updated Nov 27, 2024 • 3 TensorStack/Juggernaut-XL-v11-onnx Updated Oct 9, 2024 • 1 amd/stable-diffusion-xl-1.0_io32_amdgpu Text-to-Image • Updated Apr 3 • 11 amd/stable-diffusion-1.5_io32_amdgpu Text-to-Image • Updated 14 days ago • 26
RyzenAI-1.3_LLM_Hybrid_Models Models quantized by Quark and prepared for the OGA-based hybrid execution flow (Ryzen AI 1.3) amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid Updated about 1 month ago • 139 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
AMD-OLMo AMD-OLMo are a series of 1 billion parameter language models trained by AMD on AMD Instinct™ MI250 GPUs based on OLMo. amd/AMD-OLMo Text Generation • Updated Nov 3, 2024 • 79 amd/AMD-OLMo-1B Text Generation • 1B • Updated Nov 3, 2024 • 2.2k • 25 amd/AMD-OLMo-1B-SFT Text Generation • 1B • Updated Nov 7, 2024 • 7.3k • 20 amd/AMD-OLMo-1B-SFT-DPO Text Generation • 1B • Updated Nov 7, 2024 • 1.21k • 22
Quark Quantized OCP FP8 Models amd/Llama-3.1-8B-Instruct-FP8-KV 8B • Updated Dec 19, 2024 • 13.6k • 6 amd/Llama-3.1-70B-Instruct-FP8-KV 71B • Updated Dec 19, 2024 • 5.66k • 3 amd/Llama-3.1-405B-Instruct-FP8-KV 406B • Updated Dec 19, 2024 • 2.96k • 4 amd/Mixtral-8x7B-Instruct-v0.1-FP8-KV 3B • Updated Dec 19, 2024 • 12k • 3
RyzenAI-1.5_LLM_Hybrid_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid Updated about 1 month ago • 139 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
RyzenAI-1.5_LLM_NPU_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
Gumiho Official Model Parameters for "Gumiho: A Hybrid Architecture to Prioritize Early Tokens in Speculative Decoding" Gumiho: A Hybrid Architecture to Prioritize Early Tokens in Speculative Decoding Paper • 2503.10135 • Published Mar 13
Gumiho: A Hybrid Architecture to Prioritize Early Tokens in Speculative Decoding Paper • 2503.10135 • Published Mar 13
PARD Official Model Parameters for "PARD: Accelerating LLM Inference with Low-Cost PARallel Draft Model Adaptation" amd/PARD-Llama-3.2-1B Text Generation • 1B • Updated May 19 • 1.27k • 2 amd/PARD-DeepSeek-R1-Distill-Qwen-1.5B Text Generation • 2B • Updated May 19 • 165 • 1 amd/PARD-Qwen2.5-0.5B Text Generation • 0.6B • Updated May 19 • 1.24k amd/PARD-Qwen3-0.6B Text Generation • 0.8B • Updated 15 days ago • 31
OGA CPU LLM Collection This collection contains AMD-Quark quantized OGA exported models for CPU execution amd/Phi-3-mini-4k-instruct_int4_float16_onnx_cpu Updated Apr 12 amd/Qwen1.5-7B-Chat_uint4_asym_g128_float16_onnx_cpu Updated Apr 12 amd/Llama-3.2-1B-Instruct-awq-uint4-float16-cpu-onnx Updated Apr 28 amd/Llama-3.2-3B-Instruct-awq-uint4-float16-cpu-onnx Updated Apr 28
Quark Quantized MXFP4 models amd/DeepSeek-R1-MXFP4-Preview 357B • Updated 3 days ago • 523 amd/Llama-3.1-405B-Instruct-MXFP4-Preview 218B • Updated 27 days ago • 112
AMDGPU OnnxGenAI Collection ONNX GenAI compatible Language Models to run on AMD Ryzen(TM) GPUs and Radeon Discrete GPUs amd/Llama-2-7b-chat-hf-awq-g128-int4-onnx-directml Updated Apr 8 amd/Llama-2-7b-hf-awq-g128-int4-onnx-directml Updated Apr 10 amd/Llama-3.1-8B-awq-g128-int4-onnx-directml Updated Apr 9 amd/Llama-3.1-8B-Instruct-awq-g128-int4-onnx-directml Updated Apr 9
RyzenAI-1.4_LLM_NPU_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
RyzenAI-1.4_LLM_Hybrid_Models amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid Updated about 1 month ago • 139 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
Instella ✨ Announcing Instella, a series of 3 billion parameter language models developed by AMD, trained from scratch on 128 Instinct MI300X GPUs. amd/Instella-3B-Stage1 Text Generation • 3B • Updated Mar 6 • 37 • 13 amd/Instella-3B Text Generation • 3B • Updated 29 days ago • 93 • 36 amd/Instella-3B-SFT Text Generation • 3B • Updated Mar 6 • 6 • 9 amd/Instella-3B-Instruct Text Generation • 3B • Updated Mar 28 • 154 • 51
AMD-RyzenAI-Deepseek-R1-Distill-Hybrid amd/DeepSeek-R1-Distill-Llama-8B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 267 • 1 amd/DeepSeek-R1-Distill-Qwen-1.5B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 16 • 1 amd/DeepSeek-R1-Distill-Qwen-7B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 229 • 1
amd/DeepSeek-R1-Distill-Llama-8B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 267 • 1
amd/DeepSeek-R1-Distill-Qwen-1.5B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 16 • 1
amd/DeepSeek-R1-Distill-Qwen-7B-awq-asym-uint4-g128-lmhead-onnx-hybrid Updated about 1 month ago • 229 • 1
AMDGPU onnx optimized image generation ONNX models for AMD Ryzen (TM) AI GPUs and Radeon Discrete GPUs TensorStack/Flux_schnell-f16-onnx Text-to-Image • Updated Nov 27, 2024 • 3 TensorStack/Juggernaut-XL-v11-onnx Updated Oct 9, 2024 • 1 amd/stable-diffusion-xl-1.0_io32_amdgpu Text-to-Image • Updated Apr 3 • 11 amd/stable-diffusion-1.5_io32_amdgpu Text-to-Image • Updated 14 days ago • 26
RyzenAI-1.3_LLM_NPU_Models Models quantized by Quark and prepared for the OGA-based NPU-only execution flow (Ryzen AI 1.3) amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 2
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 2.53k • 2
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
RyzenAI-1.3_LLM_Hybrid_Models Models quantized by Quark and prepared for the OGA-based hybrid execution flow (Ryzen AI 1.3) amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176 amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-fp16-onnx-hybrid Updated about 1 month ago • 139 amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
amd/Phi-3-mini-4k-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 176
amd/Phi-3.5-mini-instruct-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 41
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-fp16-onnx-hybrid Text Generation • Updated about 1 month ago • 179
Nitro Diffusion 💥 Nitro Diffusion is a series of efficient text-to-image diffusion models built on AMD Instinct™ GPUs. amd/Nitro-1-SD Text-to-Image • Updated 29 days ago • 38 • 7 amd/Nitro-1-PixArt Text-to-Image • Updated 29 days ago • 28 • 3 amd/Nitro-T-0.6B Text-to-Image • Updated 15 days ago • 56 • 2 amd/Nitro-T-1.2B Text-to-Image • Updated 15 days ago • 141 • 3
AMD-OLMo AMD-OLMo are a series of 1 billion parameter language models trained by AMD on AMD Instinct™ MI250 GPUs based on OLMo. amd/AMD-OLMo Text Generation • Updated Nov 3, 2024 • 79 amd/AMD-OLMo-1B Text Generation • 1B • Updated Nov 3, 2024 • 2.2k • 25 amd/AMD-OLMo-1B-SFT Text Generation • 1B • Updated Nov 7, 2024 • 7.3k • 20 amd/AMD-OLMo-1B-SFT-DPO Text Generation • 1B • Updated Nov 7, 2024 • 1.21k • 22
Quark Quantized ONNX LLMs for Ryzen AI 1.3 EA ONNX Runtime generate() API based models quantized by Quark and optimized for Ryzen AI Strix Point NPU amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1 amd/Mistral-7B-Instruct-v0.3-awq-g128-int4-asym-bf16-onnx-ryzen-strix Updated 26 days ago • 25 amd/Llama-2-7b-hf-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 30 amd/Llama-3-8B-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 14 • 1
amd/Qwen1.5-7B-Chat-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 24 • 1
amd/Llama-2-7b-hf-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 30
amd/Llama-3-8B-awq-g128-int4-asym-bf16-onnx-ryzen-strix Text Generation • Updated 26 days ago • 14 • 1
Quark Quantized OCP FP8 Models amd/Llama-3.1-8B-Instruct-FP8-KV 8B • Updated Dec 19, 2024 • 13.6k • 6 amd/Llama-3.1-70B-Instruct-FP8-KV 71B • Updated Dec 19, 2024 • 5.66k • 3 amd/Llama-3.1-405B-Instruct-FP8-KV 406B • Updated Dec 19, 2024 • 2.96k • 4 amd/Mixtral-8x7B-Instruct-v0.1-FP8-KV 3B • Updated Dec 19, 2024 • 12k • 3