Update handler.py
Browse files- handler.py +4 -4
handler.py
CHANGED
@@ -6,12 +6,12 @@ class EndpointHandler():
|
|
6 |
# init
|
7 |
# load the model
|
8 |
tokenizer = AutoTokenizer.from_pretrained("verseAI/vai-GPT-NeoXT-Chat-Base-20B")
|
9 |
-
|
10 |
# THROWS ERROR model = AutoModelForQuestionAnswering.from_pretrained("verseAI/vai-GPT-NeoXT-Chat-Base-20B", device_map="auto", load_in_8bit=True)
|
11 |
-
model = AutoModel.from_pretrained("verseAI/vai-GPT-NeoXT-Chat-Base-20B", device_map="auto", load_in_8bit=True)
|
12 |
# create inference pipeline
|
13 |
-
|
14 |
-
self.pipeline = pipeline("question-answering", model=model, tokenizer=tokenizer)
|
15 |
|
16 |
def __call__(self, data: Dict[str, Any]) -> List[List[Dict[str, float]]]:
|
17 |
"""
|
|
|
6 |
# init
|
7 |
# load the model
|
8 |
tokenizer = AutoTokenizer.from_pretrained("verseAI/vai-GPT-NeoXT-Chat-Base-20B")
|
9 |
+
model = AutoModelForCausalLM.from_pretrained("verseAI/vai-GPT-NeoXT-Chat-Base-20B", device_map="auto", load_in_8bit=True)
|
10 |
# THROWS ERROR model = AutoModelForQuestionAnswering.from_pretrained("verseAI/vai-GPT-NeoXT-Chat-Base-20B", device_map="auto", load_in_8bit=True)
|
11 |
+
# model = AutoModel.from_pretrained("verseAI/vai-GPT-NeoXT-Chat-Base-20B", device_map="auto", load_in_8bit=True)
|
12 |
# create inference pipeline
|
13 |
+
self.pipeline = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
14 |
+
#self.pipeline = pipeline("question-answering", model=model, tokenizer=tokenizer)
|
15 |
|
16 |
def __call__(self, data: Dict[str, Any]) -> List[List[Dict[str, float]]]:
|
17 |
"""
|