--- library_name: transformers tags: - mlx base_model: - Qwen/Qwen2.5-VL-3B-Instruct --- # 📸 Adasah - Qwen 2.5 3B (4-bit) Fine-tuned on Arabic Photo Q&A and Descriptions ## Demo Video: ## Adasah - IOS App [App Store](https://apps.apple.com/us/app/adasah/id6745417467) ### Warning - The app downloads a 2GB model, so it takes some time for the first time. ## Forked from Huggingsnap https://github.com/huggingface/HuggingSnap **Model Name**: `Adasah` **Base Model**: [Qwen/Qwen2.5-VL-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct) **Quantization**: 4-bit (GGUF) **Platform**: iOS (mobile-compatible) **Language**: Arabic (translated from English) **Use Case**: Arabic Visual Q&A and Photo Description Understanding --- ## 🧠 Model Overview **Adasah** is a fine-tuned variant of the Qwen 2.5 3B base model, optimized for Arabic language understanding in visual contexts. The model was trained on a custom dataset consisting of **English visual question-answer pairs and photo descriptions translated into Arabic**, allowing it to: - Answer Arabic questions about images - Generate Arabic descriptions of visual content - Serve as a mobile assistant for Arabic-speaking users The model is quantized to **4-bit** to ensure smooth on-device performance on **iOS apps**. ## 📱 Mobile Optimization The model is quantized using **4-bit precision** to make it lightweight and suitable for **on-device inference** in: - **iOS apps** - **Offline-first mobile experiences** - **Arabic language educational or accessibility tools** --- ## Use with mlx ```bash pip install -U mlx-vlm ``` ```bash python -m mlx_vlm.generate --model NAMAA-Space/Adasah-QA-0.1-3B-Instruct-merged-4bits --max-tokens 100 --temperature 0.0 --prompt "Describe this image." --image ```