Question Answering
Safetensors
moellama
custom_code
davzoku commited on
Commit
9cd01f9
verified
1 Parent(s): 97cb482

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +15 -4
README.md CHANGED
@@ -16,8 +16,18 @@ pipeline_tag: question-answering
16
 
17
  ## Model Details
18
 
19
- This model is a mixture of experts (MoE) using the [RhuiDih/moetify](https://github.com/RhuiDih/moetify) library. It combines multiple domain-specific experts, LoRA adapters, and datasets, all available at [Moecule Ingredients]
20
- (https://huggingface.co/collections/davzoku/moecule-ingredients-67dac0e6210eb1d95abc6411).
 
 
 
 
 
 
 
 
 
 
21
 
22
  ## MoE Creation
23
 
@@ -39,7 +49,7 @@ To reproduce this model, run the following command:
39
  davzoku/stock_market_expert_3b
40
  ```
41
 
42
- ### Model Parameters
43
 
44
  ```shell
45
  INFO:root:Stem parameters: 1228581888
@@ -50,7 +60,7 @@ INFO:root:MOE total parameters : 8363609088
50
  INFO:root:MOE active parameters: 5985438720
51
  ```
52
 
53
- ## Quick Start
54
 
55
  ```python
56
  # git clone moetify fork that fixes dependency issue
@@ -102,3 +112,4 @@ print(generated_text)
102
 
103
  - [Flexible and Effective Mixing of Large Language Models into a Mixture of Domain Experts](https://arxiv.org/abs/2408.17280v2)
104
  - [RhuiDih/moetify](https://github.com/RhuiDih/moetify)
 
 
16
 
17
  ## Model Details
18
 
19
+ This model is a mixture of experts (MoE) using the [RhuiDih/moetify](https://github.com/RhuiDih/moetify) library. It combines multiple domain-specific experts, LoRA adapters, and datasets, all available at [Moecule Ingredients](https://huggingface.co/collections/davzoku/moecule-ingredients-67dac0e6210eb1d95abc6411).
20
+
21
+ ## Key Features
22
+
23
+ - **Zero Additional Training:** Combine existing domain-specific / task-specific experts into a powerful MoE model without additional training!
24
+
25
+ ## System Requirements
26
+
27
+ | Steps | System Requirements |
28
+ | ---------------- | ---------------------- |
29
+ | MoE Creation | > 54.2 GB System RAM |
30
+ | Inference (fp16) | GPU with > 15.5GB VRAM |
31
 
32
  ## MoE Creation
33
 
 
49
  davzoku/stock_market_expert_3b
50
  ```
51
 
52
+ ## Model Parameters
53
 
54
  ```shell
55
  INFO:root:Stem parameters: 1228581888
 
60
  INFO:root:MOE active parameters: 5985438720
61
  ```
62
 
63
+ ## Inference
64
 
65
  ```python
66
  # git clone moetify fork that fixes dependency issue
 
112
 
113
  - [Flexible and Effective Mixing of Large Language Models into a Mixture of Domain Experts](https://arxiv.org/abs/2408.17280v2)
114
  - [RhuiDih/moetify](https://github.com/RhuiDih/moetify)
115
+