Update app.py
Browse files
app.py
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
import os
|
2 |
import json
|
3 |
import re
|
4 |
-
import openai
|
5 |
import pandas as pd
|
6 |
from datetime import datetime
|
7 |
from dotenv import load_dotenv
|
@@ -9,6 +8,7 @@ from fpdf import FPDF
|
|
9 |
from gtts import gTTS
|
10 |
import gradio as gr
|
11 |
|
|
|
12 |
from langchain.prompts import ChatPromptTemplate, SystemMessagePromptTemplate, HumanMessagePromptTemplate
|
13 |
from langchain.memory import ConversationBufferMemory
|
14 |
from llama_index.core import SimpleDirectoryReader, VectorStoreIndex
|
@@ -47,8 +47,9 @@ llm = ChatOpenAI(temperature=0.6, model_name="gpt-3.5-turbo", openai_api_key=ope
|
|
47 |
teen_memory = ConversationBufferMemory()
|
48 |
embed_model = HuggingFaceEmbedding(model_name="sentence-transformers/all-MiniLM-L6-v2")
|
49 |
|
|
|
50 |
def moderate_output(text):
|
51 |
-
|
52 |
try:
|
53 |
response = openai.Moderation.create(input=text)
|
54 |
flagged = response["results"][0]["flagged"]
|
|
|
1 |
import os
|
2 |
import json
|
3 |
import re
|
|
|
4 |
import pandas as pd
|
5 |
from datetime import datetime
|
6 |
from dotenv import load_dotenv
|
|
|
8 |
from gtts import gTTS
|
9 |
import gradio as gr
|
10 |
|
11 |
+
from openai import OpenAI
|
12 |
from langchain.prompts import ChatPromptTemplate, SystemMessagePromptTemplate, HumanMessagePromptTemplate
|
13 |
from langchain.memory import ConversationBufferMemory
|
14 |
from llama_index.core import SimpleDirectoryReader, VectorStoreIndex
|
|
|
47 |
teen_memory = ConversationBufferMemory()
|
48 |
embed_model = HuggingFaceEmbedding(model_name="sentence-transformers/all-MiniLM-L6-v2")
|
49 |
|
50 |
+
client = OpenAI(api_key=openai_key)
|
51 |
def moderate_output(text):
|
52 |
+
# already loaded from env
|
53 |
try:
|
54 |
response = openai.Moderation.create(input=text)
|
55 |
flagged = response["results"][0]["flagged"]
|