althayr's picture
Updating README and TOS
4a0e792
---
library_name: ggml
tags:
- gguf
- quantized
- conversational
- brazilian-portuguese
- portuguese
- instruction-tuned
base_model: CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it
license: gemma
language:
- pt
pipeline_tag: text-generation
model_type: gemma3
quantized_by: Althayr
---
# Gemma-3-Gaia-PT-BR-4b-it-GGUF
This model was converted to GGUF format from [CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it](https://huggingface.co/CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it) using llama.cpp.
## About GAIA
**GAIA** is an open, state-of-the-art language model for Brazilian Portuguese. It was developed by continuously pre-training the google/gemma-3-4b-pt model on an extensive, high-quality corpus of Portuguese data. The goal of GAIA is to democratize access to cutting-edge AI technology in Brazil.
### Original Model Developed by
- Brazilian Association of AI (ABRIA)
- Center of Excellence in Artificial Intelligence (CEIA-UFG)
- Nama
- Amadeus AI
- Google DeepMind
## Model Details
- **Base Model**: [CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it](https://huggingface.co/CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it)
- **Original Model**: [google/gemma-3-4b-pt](https://huggingface.co/google/gemma-3-4b-pt)
- **Quantization**: BF16 (default)
- **Format**: GGUF
- **Size**: ~8.5GB
- **Parameters**: 4.3B
- **Architecture**: Gemma3
- **Context Window**: 128K tokens
- **Language**: Brazilian Portuguese
- **Converted by**: Althayr Nazaret [![GitHub](https://img.shields.io/badge/GitHub-100000?style=flat&logo=github&logoColor=white)](https://github.com/althayr) [![LinkedIn](https://img.shields.io/badge/LinkedIn-0077B5?style=flat&logo=linkedin&logoColor=white)](https://www.linkedin.com/in/althayr-santos/) [![HuggingFace](https://img.shields.io/badge/🤗_Hugging_Face-FFD21E?style=flat)](https://huggingface.co/Althayr)
## Usage
### Ollama
```bash
ollama pull https://huggingface.co/Althayr/Gemma-3-Gaia-PT-BR-4b-it-GGUF && \
ollama cp huggingface.co/Althayr/Gemma-3-Gaia-PT-BR-4b-it-GGUF Gemma-3-Gaia-PT-BR-4b-it-GGUF && \
ollama run Gemma-3-Gaia-PT-BR-4b-it-GGUF
```
### llama.cpp
#### Installation
Follow the OS specific instructions at [llama.cpp](https://github.com/ggml-org/llama.cpp?tab=readme-ov-file#quick-start)
#### CLI Execution
```bash
llama-cli --hf-repo Althayr/Gemma-3-Gaia-PT-BR-4b-it-GGUF --hf-file gemma-3-gaia-pt-br-4b-it.gguf -p "Me explique brevemente o que é inteligência artificial"
```
#### Server Execution
```bash
llama-server --hf-repo Althayr/Gemma-3-Gaia-PT-BR-4b-it-GGUF --hf-file gemma-3-gaia-pt-br-4b-it.gguf -c 2048
```
## Capabilities
The model is optimized for Brazilian Portuguese tasks, including:
- 💬 **Conversation** - Chatbots and virtual assistants
-**Question Answering** - Factual question responses
- 📝 **Summarization** - Summarizing long texts
- ✍️ **Text Generation** - Creative content creation
- 🎯 **Sentiment Analysis** - Emotion analysis in text
- 🔍 **Text Understanding** - Document interpretation
## License and Terms
This model is provided under and subject to the **Gemma Terms of Use**.
By downloading or using this model, you agree to be bound by these terms.
**Key obligations include:**
* Compliance with the [Gemma Prohibited Use Policy](https://ai.google.dev/gemma/prohibited_use_policy).
* Providing a copy of the [Gemma Terms of Use](https://ai.google.dev/gemma/terms) to any third-party recipients.
* Prominent notice that this is a modified (quantized) version.
## Citation
If you use this model in research or applications, please cite the original GAIA paper:
```bibtex
@misc{gaia-gemma-3-4b-2025,
title={GAIA: An Open Language Model for Brazilian Portuguese},
author={CAMILO-JUNIOR, C. G.; OLIVEIRA, S. S. T.; PEREIRA, L. A.; AMADEUS, M.; FAZZIONI, D.; NOVAIS, A. M. A.; JORDÃO, S. A. A.},
year={2025},
publisher={Hugging Face},
journal={Hugging Face repository},
howpublished={\url{https://huggingface.co/CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it}}
}
```
## Acknowledgments
If you use this specific GGUF version, please acknowledge:
- Original model: CEIA-UFG/Gemma-3-Gaia-PT-BR-4b-it
- GGUF conversion: Althayr/Gemma-3-Gaia-PT-BR-4b-it-GGUF