Code World Model (CWM)

This repository contains the MLX conversion of CWM with a 4bit quantization.

mlx-lm mlx_lm.generate --model . --prompt "hello"                                 
==========
Hello! It's nice to meet you. Is there something I can help you with or would you like to chat?
</think>
==========
Prompt: 14 tokens, 20.107 tokens-per-sec
Generation: 28 tokens, 18.347 tokens-per-sec
Peak memory: 18.408 GB
Downloads last month
-
Safetensors
Model size
32.6B params
Tensor type
BF16
·
U32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support