[experimental] My GGUF-IQ-ARM-Imatrix quants for CohereForAI/c4ai-command-r7b-12-2024.

Quants uploaded.

model-image

Downloads last month
746
GGUF
Model size
8.03B params
Architecture
cohere2

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference API
Inference API (serverless) has been turned off for this model.

Model tree for Lewdiculous/c4ai-command-r7b-12-2024-GGUF-IQ-ARM-Imatrix

Quantized
(10)
this model

Collection including Lewdiculous/c4ai-command-r7b-12-2024-GGUF-IQ-ARM-Imatrix