ricklegac commited on
Commit
04b3975
·
1 Parent(s): b516e8a

sin tensor

Browse files
Files changed (2) hide show
  1. app.py +5 -1
  2. requirements.txt +1 -1
app.py CHANGED
@@ -1,18 +1,22 @@
1
  import gradio as gr
2
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
 
4
- model_id = "HuggingFaceH4/zephyr-7b-alpha" # sin restricciones
5
 
 
6
  tokenizer = AutoTokenizer.from_pretrained(model_id)
7
  model = AutoModelForCausalLM.from_pretrained(model_id)
8
 
 
9
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
10
 
 
11
  def generar_commit(diff):
12
  prompt = f"Escribí un mensaje de commit claro, corto y técnico basado en los siguientes cambios:\n{diff}"
13
  output = pipe(prompt, max_new_tokens=60, do_sample=True, temperature=0.7)
14
  return [output[0]["generated_text"]]
15
 
 
16
  demo = gr.Interface(
17
  fn=generar_commit,
18
  inputs=gr.Textbox(lines=10, label="Código diff"),
 
1
  import gradio as gr
2
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
 
4
+ model_id = "HuggingFaceH4/zephyr-7b-alpha" # modelo 100% libre
5
 
6
+ # Carga del modelo sin device_map ni dtype
7
  tokenizer = AutoTokenizer.from_pretrained(model_id)
8
  model = AutoModelForCausalLM.from_pretrained(model_id)
9
 
10
+ # Crear pipeline
11
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
12
 
13
+ # Función para generar commit
14
  def generar_commit(diff):
15
  prompt = f"Escribí un mensaje de commit claro, corto y técnico basado en los siguientes cambios:\n{diff}"
16
  output = pipe(prompt, max_new_tokens=60, do_sample=True, temperature=0.7)
17
  return [output[0]["generated_text"]]
18
 
19
+ # Interfaz Gradio
20
  demo = gr.Interface(
21
  fn=generar_commit,
22
  inputs=gr.Textbox(lines=10, label="Código diff"),
requirements.txt CHANGED
@@ -1,4 +1,4 @@
1
  transformers
2
  torch
3
  gradio
4
- accelerate
 
1
  transformers
2
  torch
3
  gradio
4
+ accelerate