Spaces:
Configuration error
Configuration error
name: codellama-7b | |
backend: transformers | |
type: AutoModelForCausalLM | |
parameters: | |
model: codellama/CodeLlama-7b-hf | |
temperature: 0.2 | |
top_k: 40 | |
top_p: 0.95 | |
usage: | | |
curl http://localhost:8080/v1/completions -H "Content-Type: application/json" -d '{ | |
"model": "codellama-7b", | |
"prompt": "import socket\n\ndef ping_exponential_backoff(host: str):" | |
}' | |