tajuarAkash commited on
Commit
350f453
·
verified ·
1 Parent(s): d661bad

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +26 -0
app.py ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import streamlit as st
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
+
4
+ # Load the locally saved fine-tuned model inside your space
5
+ MODEL_DIR = "./laptop-tinyllama"
6
+
7
+ @st.cache_resource
8
+ def load_pipeline():
9
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_DIR)
10
+ model = AutoModelForCausalLM.from_pretrained(MODEL_DIR)
11
+ return pipeline("text-generation", model=model, tokenizer=tokenizer)
12
+
13
+ # Load model pipeline
14
+ generator = load_pipeline()
15
+
16
+ # Streamlit UI
17
+ st.title("💻 Laptop Recommendation with TinyLlama")
18
+ st.write("Enter a question like: *Suggest a laptop for gaming under 1 lakh BDT.*")
19
+
20
+ # Prompt input
21
+ prompt = st.text_area("Enter your query", value="Suggest a laptop for programming under 70000 BDT.")
22
+
23
+ if st.button("Generate Response"):
24
+ with st.spinner("Generating..."):
25
+ result = generator(prompt, max_new_tokens=100, temperature=0.7)
26
+ st.success(result[0]["generated_text"])