Update vLLM serve with parameters
Browse files
    	
        README.md
    CHANGED
    
    | @@ -101,7 +101,12 @@ messages = [ | |
| 101 | 
             
            ]
         | 
| 102 |  | 
| 103 | 
             
            response = client.chat.completions.create(model="ilsp/Llama-Krikri-8B-Instruct",
         | 
| 104 | 
            -
                                                      messages=messages | 
|  | |
|  | |
|  | |
|  | |
|  | |
| 105 | 
             
            print(response.choices[0].message.content)
         | 
| 106 | 
             
            ```
         | 
| 107 |  | 
|  | |
| 101 | 
             
            ]
         | 
| 102 |  | 
| 103 | 
             
            response = client.chat.completions.create(model="ilsp/Llama-Krikri-8B-Instruct",
         | 
| 104 | 
            +
                                                      messages=messages,
         | 
| 105 | 
            +
                                                      temperature=0.1,
         | 
| 106 | 
            +
                                                      top_p=0.95,
         | 
| 107 | 
            +
                                                      max_tokens=8192,
         | 
| 108 | 
            +
                                                      stream=False)
         | 
| 109 | 
            +
             | 
| 110 | 
             
            print(response.choices[0].message.content)
         | 
| 111 | 
             
            ```
         | 
| 112 |  | 

