Update README.md
Browse files
README.md
CHANGED
@@ -48,8 +48,31 @@ tags:
|
|
48 |
| **Llama-3.2 (11B vision)** | [▶️ Start on Colab](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Llama3.2_(11B)-Vision.ipynb) | 2x faster | 60% less |
|
49 |
| **Qwen2.5 (7B)** | [▶️ Start on Colab](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Qwen2.5_(7B)-Alpaca.ipynb) | 2x faster | 60% less |
|
50 |
| **Phi-4 (14B)** | [▶️ Start on Colab](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Phi_4-Conversational.ipynb) | 2x faster | 50% less |
|
51 |
-
# Qwen3-30B-A3B
|
52 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
53 |
## Qwen3 Highlights
|
54 |
|
55 |
Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support, with the following key features:
|
|
|
48 |
| **Llama-3.2 (11B vision)** | [▶️ Start on Colab](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Llama3.2_(11B)-Vision.ipynb) | 2x faster | 60% less |
|
49 |
| **Qwen2.5 (7B)** | [▶️ Start on Colab](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Qwen2.5_(7B)-Alpaca.ipynb) | 2x faster | 60% less |
|
50 |
| **Phi-4 (14B)** | [▶️ Start on Colab](https://colab.research.google.com/github/unslothai/notebooks/blob/main/nb/Phi_4-Conversational.ipynb) | 2x faster | 50% less |
|
|
|
51 |
|
52 |
+
# To Switch Between Thinking and Non-Thinking
|
53 |
+
If you are using llama.cpp, Ollama, Open WebUI etc., you can add `/think` and `/no_think` to user prompts or system messages to switch the model's thinking mode from turn to turn. The model will follow the most recent instruction in multi-turn conversations.
|
54 |
+
|
55 |
+
Here is an example of multi-turn conversation:
|
56 |
+
|
57 |
+
```
|
58 |
+
> Who are you /no_think
|
59 |
+
|
60 |
+
<think>
|
61 |
+
|
62 |
+
</think>
|
63 |
+
|
64 |
+
I am Qwen, a large-scale language model developed by Alibaba Cloud. [...]
|
65 |
+
|
66 |
+
> How many 'r's are in 'strawberries'? /think
|
67 |
+
|
68 |
+
<think>
|
69 |
+
Okay, let's see. The user is asking how many times the letter 'r' appears in the word "strawberries". [...]
|
70 |
+
</think>
|
71 |
+
|
72 |
+
The word strawberries contains 3 instances of the letter r. [...]
|
73 |
+
```
|
74 |
+
|
75 |
+
# Qwen3-30B-A3B
|
76 |
## Qwen3 Highlights
|
77 |
|
78 |
Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support, with the following key features:
|