LSXPrime commited on
Commit
7e2a195
·
verified ·
1 Parent(s): 2bce5f4

Initial upload of ProseFlow-v1-360M-Instruct-GGUF

Browse files
.gitattributes CHANGED
@@ -33,3 +33,6 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ ProseFlow-v1-360M-Instruct-imatrix.gguf filter=lfs diff=lfs merge=lfs -text
37
+ ProseFlow-v1-360M-Instruct-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
38
+ ProseFlow-v1-360M-Instruct-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
ProseFlow-v1-360M-Instruct-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e2cd2d6f724a562e6fff2fe19fa53d12a83ec91837bcb5ef287c85130ffff05
3
+ size 314829504
ProseFlow-v1-360M-Instruct-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35e793fb78e0b805855b5d38a48868880ae83d1b8f91a1c777373a007532cf93
3
+ size 430643904
ProseFlow-v1-360M-Instruct-imatrix.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7df578bc333fe3c460aa41f09a69de270da99c7566b501e0af1cea1e264097aa
3
+ size 1099456
README.md ADDED
@@ -0,0 +1,111 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: LSXPrime/ProseFlow-v1-360M-Instruct
3
+ base_model_relation: quantized
4
+ language:
5
+ - en
6
+ library_name: gguf
7
+ pipeline_tag: text-generation
8
+ license: apache-2.0
9
+ datasets:
10
+ - LSXPrime/ProseFlow-Actions-v1
11
+ tags:
12
+ - text-generation
13
+ - instruction
14
+ - proseflow
15
+ - unsloth
16
+ - smollm
17
+ - writing-assistant
18
+ ---
19
+
20
+ # ProseFlow-v1-360M-Instruct
21
+
22
+ **ProseFlow-v1-360M-Instruct** is a lightweight, experimental instruction-tuned model created for
23
+ the [ProseFlow desktop application](https://github.com/LSXPrime/ProseFlow). This model is a fine-tune of HuggingFace's [
24
+ **SmolLM-360M-Instruct**](https://huggingface.co/HuggingFaceTB/SmolLM-360M-Instruct) and was created to explore the
25
+ capabilities of smaller language models on a diverse set of text-processing tasks.
26
+
27
+ The model was fine-tuned on the **[ProseFlow-Actions-v1
28
+ ](https://huggingface.co/datasets/LSXPrime/ProseFlow-Actions-v1)** dataset.
29
+
30
+ **Note:** This model is provided for research and experimental purposes and low-resource devices. For the best user
31
+ experience in the ProseFlow application, the larger and more capable [
32
+ `ProseFlow-v1-1.5B-Instruct`](https://huggingface.co/LSXPrime/ProseFlow-v1-1.5B-Instruct) model is strongly recommended.
33
+
34
+ ## Model Description
35
+
36
+ ProseFlow is a universal AI text processor that allows users to create and execute custom AI "Actions" on text in any
37
+ application. This model was an experiment to see if a ~360M parameter model could reliably perform the wide range of
38
+ tasks defined in the training dataset.
39
+
40
+ ### Performance and Capabilities
41
+
42
+ Evaluations show that while this model is extremely fast and has very low resource requirements, its capabilities are
43
+ limited.
44
+
45
+ #### Strengths:
46
+
47
+ * **Extremely Lightweight:** Can run on devices with very limited RAM and computational power.
48
+ * **Strict Formatting Adherence (sometimes):** In some cases where it understands the task, it can follow rigid
49
+ formatting instructions (like creating a bulleted list) more strictly than its larger counterpart.
50
+ * **Simple Data Extraction:** It shows some capability in basic data extraction and formatting tasks, such as creating
51
+ Markdown tables or extracting contact information.
52
+
53
+ ## Provided Files & Quantization Details
54
+
55
+ This repository provides multiple versions of the model, allowing users to choose the best balance of performance and
56
+ resource usage for their specific hardware. All quantized versions are provided in the GGUF format for broad
57
+ compatibility.
58
+
59
+ | File Name (Quantization) | VRAM Usage (Approx.) | Performance | Recommended Use Case |
60
+ |:-------------------------|:---------------------|:---------------------------------------------------|:--------------------------------------------|
61
+ | `Q8_0` | ~1 GB | **Best Overall.** Nearly identical to FP16. | **The recommended default for most users.** |
62
+ | `Q4_K_M` | ~900 MB | **Low Quality.** Noticeable degradation in nuance. | For maximum speed on low-power devices. |
63
+
64
+ **Note on Quantization:** To maintain the highest possible quality, the token embeddings and the final output layer were
65
+ kept at F16 precision. Additionally, an importance matrix was used for calibration during the quantization process. This
66
+ is why the quantized files are larger than what might typically be expected, as this method significantly improves their
67
+ performance and coherence.
68
+
69
+ #### Weaknesses & Limitations:
70
+
71
+ * **Poor Reasoning:** The model struggles significantly with tasks that require logical reasoning, inference, or
72
+ multi-step problem-solving. It often fails on word problems and logical puzzles.
73
+ * **Limited Creativity:** It is not effective at creative writing tasks like continuing a story or generating novel
74
+ content. Its outputs are often repetitive or nonsensical.
75
+ * **Instructional Failures:** The model frequently violates the "no extra text" rule by adding conversational chatter.
76
+ In many cases, it fails the task entirely and repeats the input verbatim.
77
+ * **Hallucination:** On some tasks (e.g., `To Paragraph`), the model hallucinates content completely unrelated to the
78
+ input.
79
+ * **Unreliable for Complex Tasks:** It is not suitable for complex tasks like code refactoring, bug finding, or drafting
80
+ professional business correspondence.
81
+
82
+ ### Intended Use
83
+
84
+ This model is intended for **experimental use** and for users on **extremely resource-constrained systems** who are
85
+ willing to accept a significant trade-off in performance and reliability. It may be suitable for a very limited subset
86
+ of simple, repetitive text-formatting tasks.
87
+
88
+ It is designed to be used within the **ProseFlow desktop application**, but it is **not the recommended model for
89
+ general use**.
90
+
91
+ ## How to Use in ProseFlow
92
+
93
+ 1. [Download and install the ProseFlow application](https://github.com/LSXPrime/ProseFlow/releases).
94
+ 2. Navigate to the **Providers -> Local Provider** tab.
95
+ 3. Click "Manage Models..." and select the desired version of `ProseFlow-v1-360M-Instruct` from the "Available for
96
+ Download" list. **We recommend starting with `Q8_0`.**
97
+ 4. Once downloaded, select it from the "My Models" list.
98
+ 5. Set your "Primary Service Type" in ProseFlow to **Local**.
99
+ 6. Be aware of the limitations described above when executing actions.
100
+
101
+ ## Training Details
102
+
103
+ * **Base Model:** [HuggingFaceTB/SmolLM-360M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM-360M-Instruct)
104
+ * **Dataset:** [LSXPrime/ProseFlow-Actions-v1](https://huggingface.co/datasets/LSXPrime/ProseFlow-Actions-v1)
105
+ * **Fine-tuning Library:** [Unsloth](https://github.com/unslothai/unsloth)
106
+ * **Fine-tuning Method:** Supervised fine-tuning using LoRA on a dataset of structured instruction-input-output
107
+ triplets.
108
+
109
+ ## License
110
+
111
+ This model is licensed under the [Apache License, Version 2.0](https://www.apache.org/licenses/LICENSE-2.0).