Quantized Qwen/Qwen3-0.6B-Base
This is a quantized version of Qwen/Qwen3-0.6B-Base.
Model Details
- Base Model: Qwen/Qwen3-0.6B-Base
- Model Type: Quantized Causal Language Model
- Language(s): English, Chinese
- License: Apache 2.0
Usage
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("your-username/your-repo-name")
model = AutoModelForCausalLM.from_pretrained("your-username/your-repo-name")
# Generate text
inputs = tokenizer("Hello, how are you?", return_tensors="pt")
outputs = model.generate(**inputs, max_length=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Training Details
This model has been quantized to reduce memory usage while maintaining performance.
Ethical Considerations
Please refer to the original model's documentation for ethical considerations and limitations.
- Downloads last month
- 41
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
馃檵
Ask for provider support
Model tree for MWeste/MNLP_M2_quantized_model
Base model
Qwen/Qwen3-0.6B-Base