Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -17,7 +17,7 @@ bi_encoder.max_seq_length = 256
|
|
| 17 |
top_k = 3
|
| 18 |
|
| 19 |
# Load your fine-tuned model and tokenizer
|
| 20 |
-
model_name = "legacy107/flan-t5-large-ia3-
|
| 21 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 22 |
model = T5ForConditionalGeneration.from_pretrained(model_name)
|
| 23 |
max_length = 512
|
|
@@ -98,7 +98,7 @@ def generate_answer(question, context, ground):
|
|
| 98 |
# Decode and return the generated answer
|
| 99 |
generated_answer = tokenizer.decode(generated_ids[0], skip_special_tokens=True)
|
| 100 |
|
| 101 |
-
return generated_answer, context
|
| 102 |
|
| 103 |
|
| 104 |
# Define a function to list examples from the dataset
|
|
@@ -122,8 +122,7 @@ iface = gr.Interface(
|
|
| 122 |
],
|
| 123 |
outputs=[
|
| 124 |
Textbox(label="Generated Answer"),
|
| 125 |
-
Textbox(label="Retrieved Context")
|
| 126 |
-
Textbox(label="Ground Truth")
|
| 127 |
],
|
| 128 |
examples=list_examples()
|
| 129 |
)
|
|
|
|
| 17 |
top_k = 3
|
| 18 |
|
| 19 |
# Load your fine-tuned model and tokenizer
|
| 20 |
+
model_name = "legacy107/flan-t5-large-ia3-wiki2-100-merged"
|
| 21 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
| 22 |
model = T5ForConditionalGeneration.from_pretrained(model_name)
|
| 23 |
max_length = 512
|
|
|
|
| 98 |
# Decode and return the generated answer
|
| 99 |
generated_answer = tokenizer.decode(generated_ids[0], skip_special_tokens=True)
|
| 100 |
|
| 101 |
+
return generated_answer, context
|
| 102 |
|
| 103 |
|
| 104 |
# Define a function to list examples from the dataset
|
|
|
|
| 122 |
],
|
| 123 |
outputs=[
|
| 124 |
Textbox(label="Generated Answer"),
|
| 125 |
+
Textbox(label="Retrieved Context")
|
|
|
|
| 126 |
],
|
| 127 |
examples=list_examples()
|
| 128 |
)
|