llama.cpp๋ฅผ ์ฌ์ฉํด gguf๋ก ๋ณํํ์ต๋๋ค.
from llama_cpp import Llama
llm = Llama(
model_path="HyperCLOVAX-SEED-Text-Instruct-1.5B-gguf-q8_0.gguf",
n_gpu_layers=-1,
main_gpu=0,
n_ctx=2048
)
output = llm(
"์ฌ๋ฏธ์๋ ์ด์ผ๊ธฐ ํ๋ ๋ง๋ค์ด์ค. 1000์ ์ด์์ด์ด์ผ ํด. ์์:", # Prompt
max_tokens=2048,
echo=True,
)
print(output)
geforce 3070 RTX๋ก ํ ์คํธํ์ผ๋ฉฐ, ์ฑ๋ฅ์ ๋ค์๊ณผ ๊ฐ์ต๋๋ค.
q8_0, peak: 1.8G
llama_perf_context_print: load time = 186.23 ms
llama_perf_context_print: prompt eval time = 186.15 ms / 19 tokens ( 9.80 ms per token, 102.07 tokens per second)
llama_perf_context_print: eval time = 3141.36 ms / 536 runs ( 5.86 ms per token, 170.63 tokens per second)
llama_perf_context_print: total time = 3758.06 ms / 555 tokens
- Downloads last month
- 12
Hardware compatibility
Log In
to view the estimation
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support