ArunKr commited on
Commit
52b01a0
·
verified ·
1 Parent(s): 4937cc9

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +42 -13
README.md CHANGED
@@ -1,21 +1,50 @@
 
1
  ---
2
- base_model: unsloth/gemma-3-270m-it
3
- tags:
4
- - text-generation-inference
5
- - transformers
6
- - unsloth
7
- - gemma3_text
8
- license: apache-2.0
9
  language:
10
  - en
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
- # Uploaded finetuned model
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
 
15
- - **Developed by:** ArunKr
16
- - **License:** apache-2.0
17
- - **Finetuned from model :** unsloth/gemma-3-270m-it
18
 
19
- This gemma3_text model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
20
 
21
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
+
2
  ---
 
 
 
 
 
 
 
3
  language:
4
  - en
5
+ license: apache-2.0
6
+ tags:
7
+ - text-generation
8
+ - instruct
9
+ - manim
10
+ - lora
11
+ - gguf
12
+ datasets:
13
+ - ArunKr/verified-data-manim
14
+ base_model: HuggingFaceTB/SmolLM-135M-Instruct
15
+ library_name: transformers
16
+ pipeline_tag: text-generation
17
  ---
18
 
19
+ # gemma-3-270m-it-web-agent - Fine-tuned
20
+
21
+ This repository contains three variants of the model:
22
+
23
+ - **LoRA adapters** → [ArunKr/gemma-3-270m-it-web-agent-lora](https://huggingface.co/ArunKr/gemma-3-270m-it-web-agent-lora)
24
+ - **Merged FP16 weights** → [ArunKr/gemma-3-270m-it-web-agent-16bit](https://huggingface.co/ArunKr/gemma-3-270m-it-web-agent-16bit)
25
+ - **GGUF quantizations** → [ArunKr/gemma-3-270m-it-web-agent-gguf](https://huggingface.co/ArunKr/gemma-3-270m-it-web-agent-gguf)
26
+
27
+ ### Training
28
+ - Base model: `unsloth/gemma-3-270m-it`
29
+ - Dataset: `ArunKr/gui_grounding_dataset-100`
30
+ - Method: LoRA fine-tuning with [Unsloth](https://github.com/unslothai/unsloth)
31
+
32
+ ### Quantizations
33
+ We provide `f16`, `bf16`, `f32`, and `q8_0` GGUF files for llama.cpp / Ollama.
34
+
35
+ ### Usage Example
36
+ ```python
37
+ from transformers import AutoModelForCausalLM, AutoTokenizer
38
+
39
+ tok = AutoTokenizer.from_pretrained("ArunKr/gemma-3-270m-it-web-agent-16bit")
40
+ model = AutoModelForCausalLM.from_pretrained("ArunKr/gemma-3-270m-it-web-agent-16bit")
41
+ print(model.generate(**tok("Hello", return_tensors="pt")))
42
+ ```
43
 
44
+ ### Ollama Example
 
 
45
 
46
+ ```bash
47
+ ollama run ArunKr/SmolLM-135M-Instruct-manim-gguf:<file_name>.gguf
48
+ ```
49
 
50
+ www.ollama.com