ChavyvAkvar's picture
Update README.md
89de6a2 verified
---
library_name: transformers
license: other
license_name: lfm1.0
license_link: LICENSE
tags:
- unsloth
- trl
- sft
datasets:
- HelpingAI/Intermediate-Thinking-130k
language:
- en
base_model:
- LiquidAI/LFM2-1.2B
pipeline_tag: text-generation
---
<center>
<div style="text-align: center;">
<img
src="https://cdn-uploads.huggingface.co/production/uploads/63b6f2e752c02ae8acbaa4d8/Qn87jdxhfCaqQUiAqD_60.png"
alt="Liquid with Thinking"
style="width: 100%; max-width: 66%; height: auto; display: inline-block; margin-bottom: 0.5em; margin-top: 0.5em;"
/>
</div>
</center>
---
# Liquid-Thinking-Preview
This repository contains a full-parameter fine-tuned version of `LiquidAI/LFM2-1.2B` on the `HelpingAI/Intermediate-Thinking-130k` dataset. The goal of this fine-tune is to enhance the model's ability to perform step-by-step reasoning and solve problems that require intermediate thought processes.
This model was trained with high efficiency using the [Unsloth](https://github.com/unslothai/unsloth) library.
- **GGUF Quantizations:** For CPU inference, GGUF versions are available in the [GGUF repository here](https://huggingface.co/kreasof-ai/Liquid-Thinking-Preview-GGUF).
## Model Details
* **Base Model:** [LiquidAI/LFM2-1.2B](https://huggingface.co/LiquidAI/LFM2-1.2B)
* **Dataset:** [HelpingAI/Intermediate-Thinking-130k](https://huggingface.co/datasets/HelpingAI/Intermediate-Thinking-130k)
* **Fine-tuning Method:** Full Parameter Fine-tuning.
* **Frameworks:** Unsloth, Transformers, PyTorch, TRL.
## How to Use with Unsloth
You can easily run this model for fast inference using the `unsloth` library.
First, install the necessary packages:
```bash
pip install "unsloth[colab-new] @ git+https://github.com/unslothai/unsloth.git"
pip install --no-deps git+https://github.com/huggingface/transformers.git
pip install --no-deps causal-conv1d==1.5.0.post8
```
Next, use the following Python code to load the model and generate text:
```python
from unsloth import FastModel
import torch
from transformers import TextStreamer
model_repo_id = "kreasof-ai/Liquid-Thinking-Preview"
model, tokenizer = FastModel.from_pretrained(
model_name = model_repo_id,
dtype = None, # None for auto detection
load_in_4bit = True, # Use 4bit quantization for faster inference
)
# Define the conversation
messages = [{
"role": "user",
"content": "Explain the step-by-step process of photosynthesis in simple terms.",
}]
# Apply the chat template
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt = True, # Must add for generation
return_tensors = "pt",
).to("cuda")
# Set up the streamer for real-time output
streamer = TextStreamer(tokenizer, skip_prompt = True)
# Generate the response
_ = model.generate(
**inputs,
max_new_tokens = 4096,
# Recommended generation parameters from LiquidAI
temperature = 0.3,
min_p = 0.15,
repetition_penalty = 1.05,
use_cache = True,
streamer = streamer,
)
```
## Training Procedure
### Data Preprocessing
The `HelpingAI/Intermediate-Thinking-130k` dataset, consisting of `instruction` and `output` pairs, was formatted into a conversational format using the LFM2's chat template. The template follows this structure:
```
<|im_start|>user
{instruction}<|im_end|>
<|im_start|>assistant
{output}<|im_end|>
```
The training loss was calculated **only on the assistant's responses** (`train_on_responses_only`), encouraging the model to learn the desired output format and reasoning style without being penalized for the user's prompt.
### Hyperparameters
The model was trained with the following hyperparameters using `trl.SFTTrainer`:
| Hyperparameter | Value |
| ----------------------------- | --------------- |
| `per_device_train_batch_size` | 8 |
| `gradient_accumulation_steps` | 4 |
| `Effective Batch Size` | 32 |
| `num_train_epochs` | 3 |
| `learning_rate` | 4e-5 |
| `lr_scheduler_type` | `cosine` |
| `warmup_ratio` | 0.1 |
| `optim` | `adamw_torch` |
| `weight_decay` | 0.01 |
| `max_seq_length` | 8192 |
| `seed` | 3407 |
### Training Results
The training loss curve showed a steady decrease, indicating successful learning.
![image/png](https://cdn-uploads.huggingface.co/production/uploads/63b6f2e752c02ae8acbaa4d8/6Xn_p2rJbP2qBtaywhlvK.png)
## Intended Use & Limitations
This model is intended for research and experimentation in multi-step reasoning, complex problem-solving, and chain-of-thought generation.
* **Limitations:** Like all language models, it is prone to hallucination and may produce factually incorrect or biased information.
* **Out-of-Scope:** This model is not designed for production use without extensive testing and safety evaluations. It should not be used for critical applications where incorrect information could cause harm.