Smily-ultra-1 / README.md
boning123's picture
Update README.md
234fb21 verified
---
tags:
- text-generation
- transformer
- chatbot
license: mit
library_name: transformers
language: en
datasets:
- custom
inference: true
---
# Smily-ultra-1
Smily-ultra-1 is a custom fine-tuned language model designed for reasoning and chatbot-like interactions.
# NOTE!!!: As this model has reasoning capacities, it is considerably slower than SAM-flash-mini-v1
However it is also more powerful and smarter than the SAM-flash-mini-v1 model at the expense of speed and size.
# Smily-ultra-1
**Smily-ultra-1** is a fine-tuned language model optimized for chatbot-style conversations and basic logical reasoning. It was created by **Smilyai-labs** using a small dataset of synthetic examples and trained in Google Colab. The model is small and lightweight, making it suitable for experimentation, education, and simple chatbot tasks.
## try it yourself!
Try it with this space: [Try it here!] (https://huggingface.co/spaces/Smilyai-labs/smily-ultra-chatbot)
## Model Details
- **Base model**: GPT-Neo 125M
- **Fine-tuned by**: Smilyai-labs
- **Parameter count**: ~125 million
- **Training examples**: ~1000 inline synthetic reasoning and dialogue samples
- **Framework**: Hugging Face Transformers
- **Trained in**: Google Colab
- **Stored in**: Google Drive
- **Uploaded to**: Hugging Face Hub
## Intended Uses
This model can be used for:
- Learning how transformers work
- Building experimental chatbots
- Simple reasoning demos
- Generating creative or silly responses
## Example Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("Smilyai-labs/Smily-ultra-1")
model = AutoModelForCausalLM.from_pretrained("Smilyai-labs/Smily-ultra-1")
prompt = "What is 2 + 2?"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=30)
print(tokenizer.decode(outputs[0]))
```
## Limitations
- Not accurate for factual tasks
- Reasoning is simple and inconsistent
- Can repeat or produce nonsensical outputs
- Not safe for critical systems or real-world advice
- Small training data limits its knowledge
## Training
- Trained for 3 epochs on ~1000 examples
- Used Hugging Face `Trainer` API
- Mixed reasoning and chatbot-style prompts
- Stored in Google Drive and uploaded via `HfApi`
## License
MIT License or similar open-source license
## Citation
```
@misc{smilyultra1,
author = {Smilyai-labs},
title = {Smily-ultra-1: Chatbot + Reasoning Toy Model},
year = 2025,
url = {https://huggingface.co/Smilyai-labs/Smily-ultra-1}
}
```