GilbertAkham commited on
Commit
da90641
·
verified ·
1 Parent(s): 337042d

Upload app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -5
app.py CHANGED
@@ -1,18 +1,26 @@
1
 
 
2
  import gradio as gr
 
3
  from transformers import AutoTokenizer, AutoModelForCausalLM
4
  from peft import PeftModel
5
 
6
- BASE = "openlm-research/open_llama_3b" # base of your LoRA
7
- LORA = "GilbertAkham/openlm-llama-lora-codetrans" # your LoRA on HF
 
 
 
 
 
 
8
 
9
- tokenizer = AutoTokenizer.from_pretrained(BASE)
10
  model = AutoModelForCausalLM.from_pretrained(
11
  BASE,
12
  load_in_8bit=True,
13
- use_fast=False, # fits on T4/low VRAM
14
  device_map="auto"
15
  )
 
 
16
  model = PeftModel.from_pretrained(model, LORA)
17
  model.eval()
18
 
@@ -35,5 +43,5 @@ demo = gr.Interface(
35
  title="openlm-llama-LoRA codetrans",
36
  )
37
 
38
- demo.launch(share=True) # share=True gives you a free public link
39
 
 
1
 
2
+ import os
3
  import gradio as gr
4
+ import torch
5
  from transformers import AutoTokenizer, AutoModelForCausalLM
6
  from peft import PeftModel
7
 
8
+ BASE = "openlm-research/open_llama_3b"
9
+ LORA = "GilbertAkham/openlm-llama-lora-codetrans"
10
+
11
+ # ---- FIX HERE ----
12
+ tokenizer = AutoTokenizer.from_pretrained(
13
+ BASE,
14
+ use_fast=False # MUST be here, not on model
15
+ )
16
 
 
17
  model = AutoModelForCausalLM.from_pretrained(
18
  BASE,
19
  load_in_8bit=True,
 
20
  device_map="auto"
21
  )
22
+ # ------------------
23
+
24
  model = PeftModel.from_pretrained(model, LORA)
25
  model.eval()
26
 
 
43
  title="openlm-llama-LoRA codetrans",
44
  )
45
 
46
+ demo.launch(share=True)
47