
AmpereComputing/llama-3.3-70b-instruct-gguf
71B
•
Updated
•
214
AI inference, AI in the cloud, AI on edge, software acceleration of AI workloads on hardware, efficient AI deployments, GPU-Free AI inference, AI model optimization.