๐Ÿง  Jesteban247/brats_medgemma-GGUF

This repository provides GGUF-format variants of the fine-tuned model
Jesteban247/brats_medgemma,
which originates from the BraTS (Brain Tumor Segmentation and TextBraTS) domain adaptation experiments detailed in
Jesteban247/medgemma-brats-experiments.


๐Ÿงฉ Model Description

These files are direct GGUF conversions of the brats_medgemma model, compatible with llama.cpp and the GGUF-my-repo conversion space.

The GGUF format allows efficient inference across CPU and GPU environments, offering a lightweight deployment option for medical visual-language applications such as MRI interpretation and radiology text reasoning.


๐Ÿงฎ Available Files

File Precision Size Description
model_f16.gguf FP16 7.77 GB Full-precision float16 version
model_q8.gguf Q8_0 4.13 GB Balanced quantization for accuracy and speed
brats_medgemma-q5_k_m.gguf Q5_K_M 2.83 GB Medium quantization for balanced size and performance
mmproj_model_f16.gguf FP16 851 MB Multimodal projection weights (image-text alignment)

๐Ÿ”— Related Resources

Downloads last month
119
GGUF
Model size
4B params
Architecture
gemma3
Hardware compatibility
Log In to view the estimation

5-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Jesteban247/brats_medgemma-GGUF

Quantized
(1)
this model

Dataset used to train Jesteban247/brats_medgemma-GGUF