Spaces:
Running
Running
Abid Ali Awan
commited on
Commit
·
0d6d0e1
1
Parent(s):
baa0fe2
Update requirements.txt to maintain mem0ai dependency and adjust UIHandler to streamline message processing by removing the system prompt. Simplify header formatting in the UI for improved clarity.
Browse files- agents/ui_handler.py +3 -5
- requirements.txt +1 -4
agents/ui_handler.py
CHANGED
@@ -43,12 +43,10 @@ class UIHandler:
|
|
43 |
|
44 |
# Clear processing message and stream response
|
45 |
history.pop()
|
46 |
-
|
47 |
-
chat_prompt = f"You are a friendly AI assistant. Respond conversationally to: {message}"
|
48 |
streaming_content = ""
|
49 |
history.append(ChatMessage(role="assistant", content=""))
|
50 |
|
51 |
-
for chunk in stream_llm(
|
52 |
streaming_content += chunk
|
53 |
history[-1] = ChatMessage(role="assistant", content=streaming_content)
|
54 |
yield history, "", gr.update(interactive=False)
|
@@ -204,8 +202,8 @@ Found {len(memory_results)} similar past queries in memory.
|
|
204 |
# Header
|
205 |
gr.HTML("""
|
206 |
<center>
|
207 |
-
<h1
|
208 |
-
|
209 |
</center>
|
210 |
""")
|
211 |
|
|
|
43 |
|
44 |
# Clear processing message and stream response
|
45 |
history.pop()
|
|
|
|
|
46 |
streaming_content = ""
|
47 |
history.append(ChatMessage(role="assistant", content=""))
|
48 |
|
49 |
+
for chunk in stream_llm(message):
|
50 |
streaming_content += chunk
|
51 |
history[-1] = ChatMessage(role="assistant", content=streaming_content)
|
52 |
yield history, "", gr.update(interactive=False)
|
|
|
202 |
# Header
|
203 |
gr.HTML("""
|
204 |
<center>
|
205 |
+
<h1>🛡️ RegRadar</h1>
|
206 |
+
AI-powered regulatory compliance assistant that monitors global regulations
|
207 |
</center>
|
208 |
""")
|
209 |
|
requirements.txt
CHANGED
@@ -1,6 +1,3 @@
|
|
1 |
-
langgraph==0.4.8
|
2 |
openai==1.88.0
|
3 |
tavily-python==0.7.6
|
4 |
-
mem0ai==0.1.108
|
5 |
-
openai-agents==0.0.19
|
6 |
-
gradio==5.34.0
|
|
|
|
|
1 |
openai==1.88.0
|
2 |
tavily-python==0.7.6
|
3 |
+
mem0ai==0.1.108
|
|
|
|