Update README.md
Browse files
README.md
CHANGED
@@ -48,7 +48,7 @@ The following contains a code snippet illustrating how to use the model generate
|
|
48 |
from mlx_lm import load, generate
|
49 |
|
50 |
model, tokenizer = load("Qwen/Qwen3-4B-MLX-6bit")
|
51 |
-
prompt = "
|
52 |
|
53 |
if tokenizer.chat_template is not None:
|
54 |
messages = [{"role": "user", "content": prompt}]
|
|
|
48 |
from mlx_lm import load, generate
|
49 |
|
50 |
model, tokenizer = load("Qwen/Qwen3-4B-MLX-6bit")
|
51 |
+
prompt = "Hello, please introduce yourself and tell me what you can do."
|
52 |
|
53 |
if tokenizer.chat_template is not None:
|
54 |
messages = [{"role": "user", "content": prompt}]
|