Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -2,11 +2,11 @@ import gradio as gr
|
|
2 |
from transformers import pipeline
|
3 |
|
4 |
# Modelos geradores otimizados
|
5 |
-
model_a = pipeline("text-generation", model="
|
6 |
-
model_b = pipeline("text-generation", model="
|
7 |
|
8 |
# Modelo juiz
|
9 |
-
arbiter = pipeline("text-classification", model="neuralmind/bert-base-portuguese-cased")
|
10 |
|
11 |
# Julgamento por batch
|
12 |
def judge_response(response_a, response_b):
|
@@ -28,8 +28,8 @@ iface = gr.Interface(
|
|
28 |
inputs=gr.Textbox(label="Digite sua pergunta:"),
|
29 |
outputs=[
|
30 |
gr.Textbox(label="Pergunta"),
|
31 |
-
gr.Textbox(label="Resposta do Modelo A (
|
32 |
-
gr.Textbox(label="Resposta do Modelo B (
|
33 |
gr.Textbox(label="Modelo Vencedor"),
|
34 |
gr.Textbox(label="Resposta Escolhida"),
|
35 |
],
|
|
|
2 |
from transformers import pipeline
|
3 |
|
4 |
# Modelos geradores otimizados
|
5 |
+
model_a = pipeline("text-generation", model="deepseek-ai/DeepSeek-R1-0528", trust_remote_code=True)
|
6 |
+
model_b = pipeline("text-generation", model="botbot-ai/CabraLlama3-8b")
|
7 |
|
8 |
# Modelo juiz
|
9 |
+
arbiter = pipeline("text-classification", model="neuralmind/bert-base-portuguese-cased", trust_remote_code=True)
|
10 |
|
11 |
# Julgamento por batch
|
12 |
def judge_response(response_a, response_b):
|
|
|
28 |
inputs=gr.Textbox(label="Digite sua pergunta:"),
|
29 |
outputs=[
|
30 |
gr.Textbox(label="Pergunta"),
|
31 |
+
gr.Textbox(label="Resposta do Modelo A (DeepSeek-R1-0528)"),
|
32 |
+
gr.Textbox(label="Resposta do Modelo B (CabraLlama3-8b)"),
|
33 |
gr.Textbox(label="Modelo Vencedor"),
|
34 |
gr.Textbox(label="Resposta Escolhida"),
|
35 |
],
|