Files changed (1) hide show
  1. README.md +56 -37
README.md CHANGED
@@ -1,37 +1,56 @@
1
- ---
2
- license: apache-2.0
3
- inference: false
4
- base_model: Qwen/Qwen2.5-14B-Instruct
5
- base_model_relation: quantized
6
- tags: [green, llmware-chat, p14, ov,emerald]
7
- ---
8
-
9
- # qwen2.5-14b-instruct-ov
10
-
11
- **qwen2.5-14b-instruct-ov** is an OpenVino int4 quantized version of [Qwen2.5-14B-Instruct](https://www.huggingface.co/Qwen/Qwen2.5-14B-Instruct), providing a very fast inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
12
-
13
- This is from the latest release series from Qwen, and is one of the largest models in the collection.
14
-
15
- This model will run on an AI PC, with GPU acceleration and 32 GB of memory. Please note that the loading can take a little bit of time, but inference is still quite fast.
16
-
17
-
18
- ### Model Description
19
-
20
- - **Developed by:** Qwen
21
- - **Quantized by:** llmware
22
- - **Model type:** qwen2.5
23
- - **Parameters:** 14 billion
24
- - **Model Parent:** Qwen/Qwen2.5-14B-Instruct
25
- - **Language(s) (NLP):** English
26
- - **License:** Apache 2.0
27
- - **Uses:** Chat, general-purpose LLM
28
- - **Quantization:** int4
29
-
30
-
31
- ## Model Card Contact
32
-
33
- [llmware on github](https://www.github.com/llmware-ai/llmware)
34
-
35
- [llmware on hf](https://www.huggingface.co/llmware)
36
-
37
- [llmware website](https://www.llmware.ai)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ inference: false
4
+ base_model: Qwen/Qwen2.5-14B-Instruct
5
+ base_model_relation: quantized
6
+ tags:
7
+ - green
8
+ - llmware-chat
9
+ - p14
10
+ - ov
11
+ - emerald
12
+ language:
13
+ - zho
14
+ - eng
15
+ - fra
16
+ - spa
17
+ - por
18
+ - deu
19
+ - ita
20
+ - rus
21
+ - jpn
22
+ - kor
23
+ - vie
24
+ - tha
25
+ - ara
26
+ ---
27
+
28
+ # qwen2.5-14b-instruct-ov
29
+
30
+ **qwen2.5-14b-instruct-ov** is an OpenVino int4 quantized version of [Qwen2.5-14B-Instruct](https://www.huggingface.co/Qwen/Qwen2.5-14B-Instruct), providing a very fast inference implementation, optimized for AI PCs using Intel GPU, CPU and NPU.
31
+
32
+ This is from the latest release series from Qwen, and is one of the largest models in the collection.
33
+
34
+ This model will run on an AI PC, with GPU acceleration and 32 GB of memory. Please note that the loading can take a little bit of time, but inference is still quite fast.
35
+
36
+
37
+ ### Model Description
38
+
39
+ - **Developed by:** Qwen
40
+ - **Quantized by:** llmware
41
+ - **Model type:** qwen2.5
42
+ - **Parameters:** 14 billion
43
+ - **Model Parent:** Qwen/Qwen2.5-14B-Instruct
44
+ - **Language(s) (NLP):** English
45
+ - **License:** Apache 2.0
46
+ - **Uses:** Chat, general-purpose LLM
47
+ - **Quantization:** int4
48
+
49
+
50
+ ## Model Card Contact
51
+
52
+ [llmware on github](https://www.github.com/llmware-ai/llmware)
53
+
54
+ [llmware on hf](https://www.huggingface.co/llmware)
55
+
56
+ [llmware website](https://www.llmware.ai)