ProseFlow-v1-360M-Instruct

ProseFlow-v1-360M-Instruct is a lightweight, experimental instruction-tuned model created for the ProseFlow desktop application. This model is a fine-tune of HuggingFace's SmolLM-360M-Instruct and was created to explore the capabilities of smaller language models on a diverse set of text-processing tasks.

The model was fine-tuned on the ProseFlow-Actions-v1 dataset.

Note: This model is provided for research and experimental purposes and low-resource devices. For the best user experience in the ProseFlow application, the larger and more capable ProseFlow-v1-1.5B-Instruct model is strongly recommended.

Model Description

ProseFlow is a universal AI text processor that allows users to create and execute custom AI "Actions" on text in any application. This model was an experiment to see if a ~360M parameter model could reliably perform the wide range of tasks defined in the training dataset.

Performance and Capabilities

Evaluations show that while this model is extremely fast and has very low resource requirements, its capabilities are limited.

Strengths:

  • Extremely Lightweight: Can run on devices with very limited RAM and computational power.
  • Strict Formatting Adherence (sometimes): In some cases where it understands the task, it can follow rigid formatting instructions (like creating a bulleted list) more strictly than its larger counterpart.
  • Simple Data Extraction: It shows some capability in basic data extraction and formatting tasks, such as creating Markdown tables or extracting contact information.

Provided Files & Quantization Details

This repository provides multiple versions of the model, allowing users to choose the best balance of performance and resource usage for their specific hardware. All quantized versions are provided in the GGUF format for broad compatibility.

File Name (Quantization) VRAM Usage (Approx.) Performance Recommended Use Case
Q8_0 ~1 GB Best Overall. Nearly identical to FP16. The recommended default for most users.
Q4_K_M ~900 MB Low Quality. Noticeable degradation in nuance. For maximum speed on low-power devices.

Note on Quantization: To maintain the highest possible quality, the token embeddings and the final output layer were kept at F16 precision. Additionally, an importance matrix was used for calibration during the quantization process. This is why the quantized files are larger than what might typically be expected, as this method significantly improves their performance and coherence.

Weaknesses & Limitations:

  • Poor Reasoning: The model struggles significantly with tasks that require logical reasoning, inference, or multi-step problem-solving. It often fails on word problems and logical puzzles.
  • Limited Creativity: It is not effective at creative writing tasks like continuing a story or generating novel content. Its outputs are often repetitive or nonsensical.
  • Instructional Failures: The model frequently violates the "no extra text" rule by adding conversational chatter. In many cases, it fails the task entirely and repeats the input verbatim.
  • Hallucination: On some tasks (e.g., To Paragraph), the model hallucinates content completely unrelated to the input.
  • Unreliable for Complex Tasks: It is not suitable for complex tasks like code refactoring, bug finding, or drafting professional business correspondence.

Intended Use

This model is intended for experimental use and for users on extremely resource-constrained systems who are willing to accept a significant trade-off in performance and reliability. It may be suitable for a very limited subset of simple, repetitive text-formatting tasks.

It is designed to be used within the ProseFlow desktop application, but it is not the recommended model for general use.

How to Use in ProseFlow

  1. Download and install the ProseFlow application.
  2. Navigate to the Providers -> Local Provider tab.
  3. Click "Manage Models..." and select the desired version of ProseFlow-v1-360M-Instruct from the "Available for Download" list. We recommend starting with Q8_0.
  4. Once downloaded, select it from the "My Models" list.
  5. Set your "Primary Service Type" in ProseFlow to Local.
  6. Be aware of the limitations described above when executing actions.

Training Details

License

This model is licensed under the Apache License, Version 2.0.

Downloads last month
36
GGUF
Model size
362M params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for LSXPrime/ProseFlow-v1-360M-Instruct-GGUF

Dataset used to train LSXPrime/ProseFlow-v1-360M-Instruct-GGUF