chhatramani commited on
Commit
3c5df4c
·
verified ·
1 Parent(s): 62cc5b8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -5
README.md CHANGED
@@ -1,5 +1,39 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - unsloth
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Gemma3-4b-it-finetune-test
2
+
3
+ This is a fine-tuned version of the **Gemma 3 4B** model, specifically adapted for learning purposes using the sample dataset from Hugging Face: [`mlabonne/FineTome-100k`](https://huggingface.co/datasets/mlabonne/FineTome-100k ). The fine-tuning process was aimed at enhancing the model's capabilities in specific tasks and domains based on this dataset.
4
+
5
+ ## Model Description
6
+
7
+ **Gemma 3** is a lightweight, state-of-the-art open model built from the same research and technology as Gemini 2.0, offering high-quality results across various tasks. This particular model has been further refined through fine-tuning to better understand and generate content relevant to the themes present in the `FineTome-100k` dataset.
8
+
9
+ ## Training Details
10
+
11
+ The model was fine-tuned using a subset of the `mlabonne/FineTome-100k` dataset, which provides a diverse collection of text samples designed for training and evaluation in natural language processing tasks. The process involved adjusting the pre-trained model's parameters to specialize in the patterns and nuances found within this dataset.
12
+
13
+ ### Training Setup
14
+
15
+ - **Base Model**: [google/gemma-3-4b-it](https://huggingface.co/google/gemma-3-4b-it )
16
+ - **Dataset**: [mlabonne/FineTome-100k](https://huggingface.co/datasets/mlabonne/FineTome-100k )
17
+ - **Training Framework**: [Hugging Face Transformers](https://github.com/huggingface/transformers )
18
+ - **Additional Tools**: QloRA (for efficient fine-tuning with quantization)
19
+
20
+ ## How to Use
21
+
22
+ To utilize this model, you can load it via the Hugging Face Transformers library. Here’s a simple example:
23
+
24
+ ```python
25
+ from transformers import AutoTokenizer, AutoModelForCausalLM
26
+
27
+ tokenizer = AutoTokenizer.from_pretrained("chhatramani/gemma3-4b-it-finetune-test")
28
+ model = AutoModelForCausalLM.from_pretrained("chhatramani/gemma3-4b-it-finetune-test")
29
+
30
+ input_text = "Your input text here."
31
+ inputs = tokenizer(input_text, return_tensors="pt")
32
+ outputs = model.generate(**inputs)
33
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
34
+
35
+ ---
36
+ license: apache-2.0
37
+ tags:
38
+ - unsloth
39
+ ---