matrixportal/X-Ray_Alpha-GGUF
This model was converted to GGUF format from SicariusSicariiStuff/X-Ray_Alpha using llama.cpp via the ggml.ai's all-gguf-same-where space.
Refer to the original model card for more details on the model.
β Quantized Models Download List
π Recommended Quantizations
- β¨ General CPU Use:
Q4_K_M(Best balance of speed/quality) - π± ARM Devices:
Q4_0(Optimized for ARM CPUs) - π Maximum Quality:
Q8_0(Near-original quality)
π¦ Full Quantization Options
| π Download | π’ Type | π Notes |
|---|---|---|
| Download | Basic quantization | |
| Download | Small size | |
| Download | Balanced quality | |
| Download | Better quality | |
| Download | Fast on ARM | |
| Download | Fast, recommended | |
| Download | Best balance | |
| Download | Good quality | |
| Download | Balanced | |
| Download | High quality | |
| Download | Very good quality | |
| Download | Fast, best quality | |
| Download | Maximum accuracy | |
| Download | Multimodal projection file for image processing |
π‘ Pro Tip: Start with Q4_K_M for most use cases, only use F16 if you need maximum precision.
- Downloads last month
- 106
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for matrixportalx/X-Ray_Alpha-GGUF
Base model
google/gemma-3-4b-pt
Finetuned
google/gemma-3-4b-it
Finetuned
SicariusSicariiStuff/X-Ray_Alpha