All Our Models
See the table below for all GGUF, 16-bit and 4-bit bnb uploaded models on Hugging Face.
GGUFs can be used to run in your favorite places like Ollama, Open WebUI and llama.cpp.
4-bit and 16-bit models can be used for inference serving or fine-tuning.
Here's a table of all our GGUF + 4-bit model uploads:
Model
GGUF
Instruct (4-bit)
Base (4-bit)
Mistral
Gemma 2
Phi-3.5
Phi-3
Llama 3
Llava
Qwen2 VL
Llama 2
SmolLM2
TinyLlama
Qwen2
Zephyr SFT
CodeLlama
Yi
Last updated
Was this helpful?