Code World Model (CWM)
This repository contains the MLX conversion of CWM with a 4bit quantization.
mlx-lm mlx_lm.generate --model . --prompt "hello"
==========
Hello! It's nice to meet you. Is there something I can help you with or would you like to chat?
</think>
==========
Prompt: 14 tokens, 20.107 tokens-per-sec
Generation: 28 tokens, 18.347 tokens-per-sec
Peak memory: 18.408 GB
- Downloads last month
- -
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support