Cotype Nano GGUF🤖

Cotype-Nano-GGUF– это легковесная LLM, которая может запускаться даже на мобильных девайсах

Cotype-Nano-GGUF is a lightweight LLM that can run even on mobile devices

Inference

from llama_cpp import Llama

llm = Llama.from_pretrained(
    repo_id="MTSAIR/Cotype-Nano-GGUF",
    filename="cotype_nano_8bit.gguf",
)

llm.create_chat_completion(
    messages = [
        {
            "role": "user",
            "content": "What is the capital of France?"
        }
    ]
)
Downloads last month
161
GGUF
Model size
1.54B params
Architecture
qwen2

16-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Collection including MTSAIR/Cotype-Nano-GGUF