ProseFlow-v1-1.5B-Instruct

ProseFlow-v1-1.5B-Instruct is a versatile instruction-tuned model designed to be the local AI engine for the ProseFlow desktop application. This model excels at a wide variety of text-processing and code-related tasks, making it an ideal choice for users who want a high-performance, private, and offline-capable AI assistant integrated into their daily workflow.

This model was fine-tuned from Qwen/Qwen2.5-Coder-1.5B-Instruct, inheriting its strong coding and logical reasoning capabilities, and has been further specialized to follow the specific, structured prompts used by ProseFlow "Actions".

The model was fine-tuned on the ProseFlow-Actions-v1 dataset.

Model Description

ProseFlow is a universal AI text processor that works via global hotkeys. Users create "Actions"—reusable instructions for the AI—to perform tasks like proofreading, summarizing, refactoring code, or changing the tone of a text. This model is the brain that executes those instructions.

ProseFlow-v1-1.5B-Instruct strikes an excellent balance between performance, resource requirements, and capability, making it the recommended primary local model for the application.

Key Strengths

Based on comprehensive evaluations, this model demonstrates:

  • Excellent Task Comprehension: Consistently understands the intent behind a wide variety of instructions, from simple text manipulation to complex business and logical tasks.
  • Superior Code Intelligence: Thanks to its Qwen2.5-Coder base, the model is exceptionally proficient at code-related tasks like explaining code snippets, finding bugs, refactoring for efficiency, and adding comments.
  • High-Quality Text Generation: Produces coherent, high-quality output for summarization, expansion, and creative writing prompts.
  • Strong Reasoning Capabilities: Successfully solves multi-step word problems and performs logical deductions required by more advanced actions.
  • Versatility: Performs reliably across dozens of distinct tasks, including sentiment analysis, data extraction (JSON conversion), and professional email drafting.

Provided Files & Quantization Details

This repository provides multiple versions of the model, allowing users to choose the best balance of performance and resource usage for their specific hardware. All quantized versions are provided in the GGUF format for broad compatibility.

File Name (Quantization) VRAM Usage (Approx.) Performance Recommended Use Case
Q8_0 ~2.5 GB Best Overall. Nearly identical to FP16, superior at logic. The recommended default for most users.
Q6_K ~1.9 GB Good Quality. A minor step down from Q8_0. Users with very limited VRAM (<2GB).
Q4_K_M ~1.75 GB Acceptable Quality. Noticeable degradation in nuance. For maximum speed on low-power devices.

Key Finding: Comprehensive A/B testing revealed that the Q8_0 version is the optimal choice for most users. It retains nearly 100% of the FP16 model's capability—and even outperforms it on logical reasoning tasks—while offering a significant reduction in VRAM usage and an increase in inference speed.

Note on Quantization: To maintain the highest possible quality, the token embeddings and the final output layer were kept at F16 precision. Additionally, an importance matrix was used for calibration during the quantization process. This is why the quantized files are larger than what might typically be expected, as this method significantly improves their performance and coherence.

Intended Use

This model is primarily intended to be used within the ProseFlow desktop application. Its prompt format and output style are specifically tuned to work seamlessly with the app's "Action" system. When used in ProseFlow, it provides a powerful, private, and offline alternative to cloud-based AI services.

Primary Use Cases:

  • Code refactoring, debugging, and documentation.
  • Drafting and improving professional emails and documents.
  • Summarizing long articles or meeting notes.
  • Proofreading and enhancing creative or technical writing.
  • Brainstorming ideas and generating structured content.

Limitations and Considerations

While highly capable, this model has a few known behaviors:

  • Instruction Following vs. Helpfulness: The model is so well-aligned to be a helpful assistant that it sometimes adds conversational headers or brief explanations to its output (e.g., "Here is the refactored code:"). This violates the strict "output only" constraint in some of the training prompts. In the context of the ProseFlow application, this is generally a minor issue, but it's a deviation from the prompt instructions.
  • Complex Logic: While it can handle multi-step logic, it may fail on more abstract or tricky logical puzzles (e.g., identifying an anomaly in a nuanced list).

How to Use in ProseFlow

  1. Download and install the ProseFlow application.
  2. Navigate to the Providers -> Local Provider tab.
  3. Click "Manage Models..." and select the desired version of ProseFlow-v1-1.5B-Instruct from the "Available for Download" list. We recommend starting with Q8_0.
  4. Once downloaded, select it from the "My Models" list.
  5. Set your "Primary Service Type" in ProseFlow to Local.
  6. You're all set! The application will now use this model for all AI actions.

Training Details

License

This model is licensed under the Apache License, Version 2.0.

Downloads last month
84
GGUF
Model size
1.54B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

4-bit

6-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for LSXPrime/ProseFlow-v1-1.5B-Instruct-GGUF

Dataset used to train LSXPrime/ProseFlow-v1-1.5B-Instruct-GGUF