Gemma-3-270M - Fine-tuned for Financial Instructions

This is a fine-tuned version of Google's gemma-3-270m-it model, adapted for financial instruction-following tasks.

Model Description

This model was fine-tuned using the Apple MLX framework. The goal was to specialize the base model for financial reporting summary and decision-making assistance. It was trained on the Josephgflowers/Finance-Instruct-500k dataset.

Intended Use

This model is intended for tasks related to the financial domain, such as:

  • Answering questions about financial concepts.
  • Summarizing financial reports.
  • Following instructions based on financial data.

How to Use

You can use this model with the transformers library just like any other standard Hugging Face model.

from transformers import AutoTokenizer, AutoModelForCausalLM

model_name = "tlgoa/tmr-ai-nano"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name)

# Note: Gemma 3 uses a specific chat template.
# For single-turn inference, you can format it like this:
prompt = "What is the difference between revenue and profit?"
formatted_prompt = f"### User:\n{prompt}\n\n### Assistant:"

inputs = tokenizer(formatted_prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=200)

response = tokenizer.decode(outputs[0], skip_special_tokens=True)
# Clean up the response to only show the assistant's part
assistant_response = response.split("### Assistant:")[1].strip()

print(assistant_response)

Training Procedure

Dataset

The model was fine-tuned on the Josephgflowers/Finance-Instruct-500k dataset. The data was preprocessed to fit the following format:

### User:
{user_prompt}

### Assistant:
{assistant_response}

Fine-tuning

The model was fine-tuned directly (full parameter tuning) using an Adam optimizer. Due to challenges with LoRA implementation in the available MLX version, a full fine-tuning approach was chosen. The fine-tuned weights were originally saved in MLX's .npz format and subsequently converted back to Hugging Face safetensors format for distribution.

Licenses

  • Base Model: This model is based on Google's Gemma-3-270M, which is subject to the Gemma Terms of Use.
  • Dataset: The training data from Josephgflowers/Finance-Instruct-500k is available under the Apache 2.0 License.
Downloads last month
19
Safetensors
Model size
268M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Tlgoa/tmr-ai-nano

Finetuned
(497)
this model
Quantizations
1 model

Dataset used to train Tlgoa/tmr-ai-nano