Meta-Llama-3.1-8B-Instruct-f32.gguf | f32 | 32.13GB | false | Full F32 weights. |
Meta-Llama-3.1-8B-Instruct-Q8_0.gguf | Q8_0 | 8.54GB | false | Extremely high quality, generally unneeded but max available quant. |
Meta-Llama-3.1-8B-Instruct-Q6_K_L.gguf | Q6_K_L | 6.85GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, recommended. |
Meta-Llama-3.1-8B-Instruct-Q6_K.gguf | Q6_K | 6.60GB | false | Very high quality, near perfect, recommended. |
Meta-Llama-3.1-8B-Instruct-Q5_K_L.gguf | Q5_K_L | 6.06GB | false | Uses Q8_0 for embed and output weights. High quality, recommended. |
Meta-Llama-3.1-8B-Instruct-Q5_K_M.gguf | Q5_K_M | 5.73GB | false | High quality, recommended. |
Meta-Llama-3.1-8B-Instruct-Q5_K_S.gguf | Q5_K_S | 5.60GB | false | High quality, recommended. |
Meta-Llama-3.1-8B-Instruct-Q4_K_L.gguf | Q4_K_L | 5.31GB | false | Uses Q8_0 for embed and output weights. Good quality, recommended. |
Meta-Llama-3.1-8B-Instruct-Q4_K_M.gguf | Q4_K_M | 4.92GB | false | Good quality, default size for must use cases, recommended. |
Meta-Llama-3.1-8B-Instruct-Q3_K_XL.gguf | Q3_K_XL | 4.78GB | false | Uses Q8_0 for embed and output weights. Lower quality but usable, good for low RAM availability. |
Meta-Llama-3.1-8B-Instruct-Q4_K_S.gguf | Q4_K_S | 4.69GB | false | Slightly lower quality with more space savings, recommended. |
Meta-Llama-3.1-8B-Instruct-IQ4_XS.gguf | IQ4_XS | 4.45GB | false | Decent quality, smaller than Q4_K_S with similar performance, recommended. |
Meta-Llama-3.1-8B-Instruct-Q3_K_L.gguf | Q3_K_L | 4.32GB | false | Lower quality but usable, good for low RAM availability. |
Meta-Llama-3.1-8B-Instruct-Q3_K_M.gguf | Q3_K_M | 4.02GB | false | Low quality. |
Meta-Llama-3.1-8B-Instruct-IQ3_M.gguf | IQ3_M | 3.78GB | false | Medium-low quality, new method with decent performance comparable to Q3_K_M. |
Meta-Llama-3.1-8B-Instruct-Q2_K_L.gguf | Q2_K_L | 3.69GB | false | Uses Q8_0 for embed and output weights. Very low quality but surprisingly usable. |
Meta-Llama-3.1-8B-Instruct-Q3_K_S.gguf | Q3_K_S | 3.66GB | false | Low quality, not recommended. |
Meta-Llama-3.1-8B-Instruct-IQ3_XS.gguf | IQ3_XS | 3.52GB | false | Lower quality, new method with decent performance, slightly better than Q3_K_S. |
Meta-Llama-3.1-8B-Instruct-Q2_K.gguf | Q2_K | 3.18GB | false | Very low quality but surprisingly usable. |
Meta-Llama-3.1-8B-Instruct-IQ2_M.gguf | IQ2_M | 2.95GB | false | Relatively low quality, uses SOTA techniques to be surprisingly usable. |