AmpereComputing/llama-3.1-8b-instruct-gguf
8B
•
Updated
•
86
AI inference, AI in the cloud, AI on edge, software acceleration of AI workloads on hardware, efficient AI deployments, GPU-Free AI inference, AI model optimization.