Model Card for Model ID

Model Details

A base model OuteAI/Lite-Oute-1-300M-Instruct was fine-tuned on a tweet sentiment dataset cardiffnlp/tweet_eval in order to determine tweets tonality by positive, neutral or negative.

Model Description

We used a system prompt to instruct the model:

SYSTEM PROMPT:

You are a tweet sentiment classifier. For each tweet input, analyze its sentiment and output exactly one word: "negative", "neutral", or "positive". Do not include any extra text.

But the model is not trained to return only the sentiment name.

Different from the previous task, the PEFT method changed to DoRA, which not simply learn a low rank matrix, it trains the magnitute vector and the directional matrix separately. By replacing the k_proj and v_proj layers as we did in the previos task to modify the model.

Training Details

batch_size=16 rank = 8 alpha = 16 lr = 3e-5

The model achieved 0.34 macro f1-score on the test dataset(initial model 0.06)

Downloads last month
6
Safetensors
Model size
300M params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for xinyuema/llm-course-hw3-dora

Finetuned
(32)
this model

Dataset used to train xinyuema/llm-course-hw3-dora

Collection including xinyuema/llm-course-hw3-dora