lbourdois commited on
Commit
ff2c993
·
verified ·
1 Parent(s): b45d894

Improve language tag

Browse files

Hi! As the model is multilingual, this is a PR to add other languages than English to the language tag to improve the referencing. Note that 29 languages are announced in the README, but only 13 are explicitly listed. I was therefore only able to add these 13 languages.

Files changed (1) hide show
  1. README.md +109 -97
README.md CHANGED
@@ -1,98 +1,110 @@
1
- ---
2
- license: creativeml-openrail-m
3
- datasets:
4
- - AI-MO/NuminaMath-CoT
5
- language:
6
- - en
7
- base_model:
8
- - Qwen/Qwen2.5-7B-Instruct
9
- pipeline_tag: text-generation
10
- library_name: transformers
11
- tags:
12
- - Qwen2.5
13
- - Ollama
14
- - Neumind
15
- - Math
16
- - Instruct
17
- - safetensors
18
- - pytorch
19
- - trl
20
- ---
21
-
22
- ### Neumind-Math-7B-Instruct Model Files
23
-
24
- The **Neumind-Math-7B-Instruct** is a fine-tuned model based on **Qwen2.5-7B-Instruct**, optimized for mathematical reasoning, step-by-step problem-solving, and instruction-based tasks in the mathematics domain. The model is designed for applications requiring structured reasoning, numerical computations, and mathematical proof generation.
25
-
26
- | File Name | Size | Description | Upload Status |
27
- |------------------------------------|------------|------------------------------------------|----------------|
28
- | `.gitattributes` | 1.57 kB | Git attributes configuration file | Uploaded |
29
- | `README.md` | 265 Bytes | ReadMe file with basic information | Updated |
30
- | `added_tokens.json` | 657 Bytes | Additional token definitions | Uploaded |
31
- | `config.json` | 860 Bytes | Model configuration settings | Uploaded |
32
- | `generation_config.json` | 281 Bytes | Generation settings | Uploaded |
33
- | `merges.txt` | 1.82 MB | Tokenizer merge rules | Uploaded |
34
- | `pytorch_model-00001-of-00004.bin` | 4.88 GB | Model shard 1 of 4 | Uploaded (LFS) |
35
- | `pytorch_model-00002-of-00004.bin` | 4.93 GB | Model shard 2 of 4 | Uploaded (LFS) |
36
- | `pytorch_model-00003-of-00004.bin` | 4.33 GB | Model shard 3 of 4 | Uploaded (LFS) |
37
- | `pytorch_model-00004-of-00004.bin` | 1.09 GB | Model shard 4 of 4 | Uploaded (LFS) |
38
- | `pytorch_model.bin.index.json` | 28.1 kB | Model index JSON | Uploaded |
39
- | `special_tokens_map.json` | 644 Bytes | Mapping of special tokens | Uploaded |
40
- | `tokenizer.json` | 11.4 MB | Tokenizer configuration | Uploaded (LFS) |
41
- | `tokenizer_config.json` | 7.73 kB | Additional tokenizer settings | Uploaded |
42
- | `vocab.json` | 2.78 MB | Vocabulary for tokenization | Uploaded |
43
-
44
- ---
45
-
46
- ### **Key Features:**
47
-
48
- 1. **Mathematical Reasoning:**
49
- Specifically fine-tuned for solving mathematical problems, including arithmetic, algebra, calculus, and geometry.
50
-
51
- 2. **Step-by-Step Problem Solving:**
52
- Provides detailed, logical solutions for complex mathematical tasks and demonstrates problem-solving methodologies.
53
-
54
- 3. **Instructional Applications:**
55
- Tailored for use in educational settings, such as tutoring systems, math content creation, and interactive learning tools.
56
-
57
- ---
58
-
59
- ### **Training Details:**
60
- - **Base Model:** [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B)
61
- - **Dataset:** Trained on **AI-MO/NuminaMath-CoT**, a large dataset of mathematical problems and chain-of-thought (CoT) reasoning. The dataset contains **860k problems** across various difficulty levels, enabling the model to tackle a wide spectrum of mathematical tasks.
62
-
63
- ---
64
-
65
- ### **Capabilities:**
66
-
67
- - **Complex Problem Solving:**
68
- Solves a wide range of mathematical problems, from basic arithmetic to advanced calculus and algebraic equations.
69
-
70
- - **Chain-of-Thought Reasoning:**
71
- Excels in step-by-step logical reasoning, making it suitable for tasks requiring detailed explanations.
72
-
73
- - **Instruction-Based Generation:**
74
- Ideal for generating educational content, such as worked examples, quizzes, and tutorials.
75
-
76
- ---
77
-
78
- ### **Usage Instructions:**
79
-
80
- 1. **Model Setup:**
81
- Download all model shards and the associated configuration files. Ensure the files are correctly placed for seamless loading.
82
-
83
- 2. **Inference:**
84
- Load the model using frameworks like PyTorch and Hugging Face Transformers. Ensure the `pytorch_model.bin.index.json` file is in the same directory for shard-based loading.
85
-
86
- 3. **Customization:**
87
- Adjust generation parameters using `generation_config.json` to optimize outputs for your specific application.
88
- ---
89
-
90
- ### **Applications:**
91
-
92
- - **Education:**
93
- Interactive math tutoring, content creation, and step-by-step problem-solving tools.
94
- - **Research:**
95
- Automated theorem proving and symbolic mathematics.
96
- - **General Use:**
97
- Solving everyday mathematical queries and generating numerical datasets.
 
 
 
 
 
 
 
 
 
 
 
 
98
  ---
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ datasets:
4
+ - AI-MO/NuminaMath-CoT
5
+ language:
6
+ - zho
7
+ - eng
8
+ - fra
9
+ - spa
10
+ - por
11
+ - deu
12
+ - ita
13
+ - rus
14
+ - jpn
15
+ - kor
16
+ - vie
17
+ - tha
18
+ - ara
19
+ base_model:
20
+ - Qwen/Qwen2.5-7B-Instruct
21
+ pipeline_tag: text-generation
22
+ library_name: transformers
23
+ tags:
24
+ - Qwen2.5
25
+ - Ollama
26
+ - Neumind
27
+ - Math
28
+ - Instruct
29
+ - safetensors
30
+ - pytorch
31
+ - trl
32
+ ---
33
+
34
+ ### Neumind-Math-7B-Instruct Model Files
35
+
36
+ The **Neumind-Math-7B-Instruct** is a fine-tuned model based on **Qwen2.5-7B-Instruct**, optimized for mathematical reasoning, step-by-step problem-solving, and instruction-based tasks in the mathematics domain. The model is designed for applications requiring structured reasoning, numerical computations, and mathematical proof generation.
37
+
38
+ | File Name | Size | Description | Upload Status |
39
+ |------------------------------------|------------|------------------------------------------|----------------|
40
+ | `.gitattributes` | 1.57 kB | Git attributes configuration file | Uploaded |
41
+ | `README.md` | 265 Bytes | ReadMe file with basic information | Updated |
42
+ | `added_tokens.json` | 657 Bytes | Additional token definitions | Uploaded |
43
+ | `config.json` | 860 Bytes | Model configuration settings | Uploaded |
44
+ | `generation_config.json` | 281 Bytes | Generation settings | Uploaded |
45
+ | `merges.txt` | 1.82 MB | Tokenizer merge rules | Uploaded |
46
+ | `pytorch_model-00001-of-00004.bin` | 4.88 GB | Model shard 1 of 4 | Uploaded (LFS) |
47
+ | `pytorch_model-00002-of-00004.bin` | 4.93 GB | Model shard 2 of 4 | Uploaded (LFS) |
48
+ | `pytorch_model-00003-of-00004.bin` | 4.33 GB | Model shard 3 of 4 | Uploaded (LFS) |
49
+ | `pytorch_model-00004-of-00004.bin` | 1.09 GB | Model shard 4 of 4 | Uploaded (LFS) |
50
+ | `pytorch_model.bin.index.json` | 28.1 kB | Model index JSON | Uploaded |
51
+ | `special_tokens_map.json` | 644 Bytes | Mapping of special tokens | Uploaded |
52
+ | `tokenizer.json` | 11.4 MB | Tokenizer configuration | Uploaded (LFS) |
53
+ | `tokenizer_config.json` | 7.73 kB | Additional tokenizer settings | Uploaded |
54
+ | `vocab.json` | 2.78 MB | Vocabulary for tokenization | Uploaded |
55
+
56
+ ---
57
+
58
+ ### **Key Features:**
59
+
60
+ 1. **Mathematical Reasoning:**
61
+ Specifically fine-tuned for solving mathematical problems, including arithmetic, algebra, calculus, and geometry.
62
+
63
+ 2. **Step-by-Step Problem Solving:**
64
+ Provides detailed, logical solutions for complex mathematical tasks and demonstrates problem-solving methodologies.
65
+
66
+ 3. **Instructional Applications:**
67
+ Tailored for use in educational settings, such as tutoring systems, math content creation, and interactive learning tools.
68
+
69
+ ---
70
+
71
+ ### **Training Details:**
72
+ - **Base Model:** [Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B)
73
+ - **Dataset:** Trained on **AI-MO/NuminaMath-CoT**, a large dataset of mathematical problems and chain-of-thought (CoT) reasoning. The dataset contains **860k problems** across various difficulty levels, enabling the model to tackle a wide spectrum of mathematical tasks.
74
+
75
+ ---
76
+
77
+ ### **Capabilities:**
78
+
79
+ - **Complex Problem Solving:**
80
+ Solves a wide range of mathematical problems, from basic arithmetic to advanced calculus and algebraic equations.
81
+
82
+ - **Chain-of-Thought Reasoning:**
83
+ Excels in step-by-step logical reasoning, making it suitable for tasks requiring detailed explanations.
84
+
85
+ - **Instruction-Based Generation:**
86
+ Ideal for generating educational content, such as worked examples, quizzes, and tutorials.
87
+
88
+ ---
89
+
90
+ ### **Usage Instructions:**
91
+
92
+ 1. **Model Setup:**
93
+ Download all model shards and the associated configuration files. Ensure the files are correctly placed for seamless loading.
94
+
95
+ 2. **Inference:**
96
+ Load the model using frameworks like PyTorch and Hugging Face Transformers. Ensure the `pytorch_model.bin.index.json` file is in the same directory for shard-based loading.
97
+
98
+ 3. **Customization:**
99
+ Adjust generation parameters using `generation_config.json` to optimize outputs for your specific application.
100
+ ---
101
+
102
+ ### **Applications:**
103
+
104
+ - **Education:**
105
+ Interactive math tutoring, content creation, and step-by-step problem-solving tools.
106
+ - **Research:**
107
+ Automated theorem proving and symbolic mathematics.
108
+ - **General Use:**
109
+ Solving everyday mathematical queries and generating numerical datasets.
110
  ---