Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -7,7 +7,7 @@ from threading import Thread
|
|
7 |
# Remove GPU decorator since we are CPU-only
|
8 |
def predict(message, history):
|
9 |
# Load model and tokenizer on CPU
|
10 |
-
model_id = "kurakurai/Luth-LFM2-
|
11 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
12 |
model = AutoModelForCausalLM.from_pretrained(
|
13 |
model_id,
|
@@ -59,9 +59,9 @@ def predict(message, history):
|
|
59 |
gr.ChatInterface(
|
60 |
predict,
|
61 |
description="""
|
62 |
-
<center><h2>Kurakura AI Luth-LFM2-
|
63 |
|
64 |
-
Chat with [Luth-LFM2-
|
65 |
""",
|
66 |
examples=[
|
67 |
"Peux-tu résoudre l'équation 3x - 7 = 11 pour x ?",
|
|
|
7 |
# Remove GPU decorator since we are CPU-only
|
8 |
def predict(message, history):
|
9 |
# Load model and tokenizer on CPU
|
10 |
+
model_id = "kurakurai/Luth-LFM2-350M"
|
11 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
12 |
model = AutoModelForCausalLM.from_pretrained(
|
13 |
model_id,
|
|
|
59 |
gr.ChatInterface(
|
60 |
predict,
|
61 |
description="""
|
62 |
+
<center><h2>Kurakura AI Luth-LFM2-350M Chat</h2></center>
|
63 |
|
64 |
+
Chat with [Luth-LFM2-350M](https://huggingface.co/kurakurai/Luth-LFM2-350M), a French-tuned version of LFM2-350M.
|
65 |
""",
|
66 |
examples=[
|
67 |
"Peux-tu résoudre l'équation 3x - 7 = 11 pour x ?",
|