|
--- |
|
tags: |
|
- text-generation |
|
- transformer |
|
- chatbot |
|
license: mit |
|
library_name: transformers |
|
language: en |
|
datasets: |
|
- custom |
|
inference: true |
|
--- |
|
|
|
# Smily-ultra-1 |
|
|
|
Smily-ultra-1 is a custom fine-tuned language model designed for reasoning and chatbot-like interactions. |
|
|
|
# NOTE!!!: As this model has reasoning capacities, it is considerably slower than SAM-flash-mini-v1 |
|
However it is also more powerful and smarter than the SAM-flash-mini-v1 model at the expense of speed and size. |
|
|
|
|
|
|
|
# Smily-ultra-1 |
|
|
|
**Smily-ultra-1** is a fine-tuned language model optimized for chatbot-style conversations and basic logical reasoning. It was created by **Smilyai-labs** using a small dataset of synthetic examples and trained in Google Colab. The model is small and lightweight, making it suitable for experimentation, education, and simple chatbot tasks. |
|
|
|
|
|
## try it yourself! |
|
|
|
Try it with this space: [Try it here!] (https://huggingface.co/spaces/Smilyai-labs/smily-ultra-chatbot) |
|
## Model Details |
|
|
|
- **Base model**: GPT-Neo 125M |
|
- **Fine-tuned by**: Smilyai-labs |
|
- **Parameter count**: ~125 million |
|
- **Training examples**: ~1000 inline synthetic reasoning and dialogue samples |
|
- **Framework**: Hugging Face Transformers |
|
- **Trained in**: Google Colab |
|
- **Stored in**: Google Drive |
|
- **Uploaded to**: Hugging Face Hub |
|
|
|
## Intended Uses |
|
|
|
This model can be used for: |
|
- Learning how transformers work |
|
- Building experimental chatbots |
|
- Simple reasoning demos |
|
- Generating creative or silly responses |
|
|
|
## Example Usage |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("Smilyai-labs/Smily-ultra-1") |
|
model = AutoModelForCausalLM.from_pretrained("Smilyai-labs/Smily-ultra-1") |
|
|
|
prompt = "What is 2 + 2?" |
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
outputs = model.generate(**inputs, max_new_tokens=30) |
|
print(tokenizer.decode(outputs[0])) |
|
``` |
|
|
|
## Limitations |
|
|
|
- Not accurate for factual tasks |
|
- Reasoning is simple and inconsistent |
|
- Can repeat or produce nonsensical outputs |
|
- Not safe for critical systems or real-world advice |
|
- Small training data limits its knowledge |
|
|
|
## Training |
|
|
|
- Trained for 3 epochs on ~1000 examples |
|
- Used Hugging Face `Trainer` API |
|
- Mixed reasoning and chatbot-style prompts |
|
- Stored in Google Drive and uploaded via `HfApi` |
|
|
|
## License |
|
|
|
MIT License or similar open-source license |
|
|
|
## Citation |
|
|
|
``` |
|
@misc{smilyultra1, |
|
author = {Smilyai-labs}, |
|
title = {Smily-ultra-1: Chatbot + Reasoning Toy Model}, |
|
year = 2025, |
|
url = {https://huggingface.co/Smilyai-labs/Smily-ultra-1} |
|
} |
|
``` |
|
|