A collection of models that are able to be run using onnxruntime-genai and can be served through embeddedllm library.
AI & ML interests
None defined yet.
Recent Activity
View all activity
Model Powered by Onnxruntime CPU GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 8 -
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 5 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 6 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 8
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 334k • • 1.24k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 603k • • 1.66k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 14.5k • 223 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.2k • 382
A collection of model that can be hosted using OpenVINO Model Server
Model Powered by Onnxruntime DirectML GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-directml
Text Generation • Updated • 7 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-directml
Text Generation • Updated • 6 -
EmbeddedLLM/Phi-3-medium-4k-instruct-onnx-directml
Text Generation • Updated • 5 -
EmbeddedLLM/Phi-3-medium-128k-instruct-onnx-directml
Text Generation • Updated • 6
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 334k • • 1.24k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 603k • • 1.66k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 14.5k • 223 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.2k • 382
A collection of models that are able to be run using onnxruntime-genai and can be served through embeddedllm library.
Model Powered by Onnxruntime DirectML GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-directml
Text Generation • Updated • 7 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-directml
Text Generation • Updated • 6 -
EmbeddedLLM/Phi-3-medium-4k-instruct-onnx-directml
Text Generation • Updated • 5 -
EmbeddedLLM/Phi-3-medium-128k-instruct-onnx-directml
Text Generation • Updated • 6
Model Powered by Onnxruntime CPU GenAI
-
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 8 -
EmbeddedLLM/Phi-3-mini-4k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 5 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32
Text Generation • Updated • 6 -
EmbeddedLLM/Phi-3-mini-128k-instruct-onnx-cpu-int4-rtn-block-32-acc-level-4
Text Generation • Updated • 8
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 334k • • 1.24k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 603k • • 1.66k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 14.5k • 223 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.2k • 382
-
microsoft/Phi-3-mini-4k-instruct
Text Generation • 4B • Updated • 334k • • 1.24k -
microsoft/Phi-3-mini-128k-instruct
Text Generation • 4B • Updated • 603k • • 1.66k -
microsoft/Phi-3-medium-4k-instruct
Text Generation • 14B • Updated • 14.5k • 223 -
microsoft/Phi-3-medium-128k-instruct
Text Generation • 14B • Updated • 11.2k • 382
A collection of model that can be hosted using OpenVINO Model Server