Medical Gemma-3N GGUF: Offline Medical Assistant πŸ₯πŸš€

GGUF quantized version of Medical Gemma-3N for ultra-fast local inference with Ollama. Optimized for offline emergency medical assistance on consumer hardware without requiring Python dependencies. Base Model: unsloth/gemma-3n-E4B-it

πŸ“¦ Model File

File Size Description Quality
medical-gemma-3n-4b.gguf 13.7GB Full precision GGUF Maximum

Note: This is a high-precision GGUF conversion. For smaller sizes, you can quantize further using llama.cpp tools.

πŸš€ Quick Start with Ollama

βœ… 1. Download the GGUF model from Hugging Face

Download it:

wget https://huggingface.co/ericrisco/medical-gemma-3n-4b-gguf/resolve/main/medical-gemma-3n-4b.gguf

πŸ—‚οΈ 2. Create a Modelfile for Ollama

Create a new directory:

mkdir -p ~/ollama/models/medical-gemma-3n-4b
cd ~/ollama/models/medical-gemma-3n-4b

And add a file called Modelfile with this content:

FROM ./medical-gemma-3n-4b.gguf

# You can adjust these parameters if needed
PARAMETER temperature 0.7
PARAMETER top_k 40
PARAMETER top_p 0.9
PARAMETER repeat_penalty 1.1

πŸ—οΈ 3. Create the model with ollama

Now build the model:

ollama create medical-gemma-3n-4b -f Modelfile

🧠 4. Run the model

Once created, simply run:

ollama run medical-gemma-3n-4b

This will give you an interactive interface to talk to your model.

Interactive Medical Consultation

# Start interactive session
ollama run medical-gemma-3n-4b

# Example conversation:
>>> What are the signs of a stroke?
The main signs of a stroke can be ...

⚠️ Important Disclaimers

  • 🚨 NOT A SUBSTITUTE for professional medical advice
  • πŸ₯ ALWAYS consult qualified healthcare professionals
  • πŸ“ž CALL EMERGENCY SERVICES (911/112) for life-threatening situations
  • πŸ”¬ FOR EDUCATIONAL AND RESEARCH PURPOSES ONLY
  • βš–οΈ Users assume FULL RESPONSIBILITY for model usage
  • πŸ”’ Ensure compliance with medical privacy regulations (HIPAA, GDPR)

πŸ“– Citation

@misc{medical_gemma_gguf,
  title={Medical Gemma-3N GGUF: Quantized Emergency Medical Assistant},
  author={Eric Risco},
  year={2025},
  url={https://huggingface.co/ericrisco/medical-gemma-3n-lora-gguf},
  note={GGUF quantized model for offline medical assistance}
}

πŸ”— Related Models

πŸ“œ License

This model is released under the Gemma License. See LICENSE for details.


πŸš€ Ultra-fast offline medical assistance - Deploy anywhere with zero dependencies

Downloads last month
31
GGUF
Model size
6.87B params
Architecture
gemma3n
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for ericrisco/medical-gemma-3n-4b-gguf

Quantized
(3)
this model

Dataset used to train ericrisco/medical-gemma-3n-4b-gguf

Collection including ericrisco/medical-gemma-3n-4b-gguf