from pydantic import BaseModel from openai import OpenAI import os from dotenv import load_dotenv MODEL = "gemini-2.0-flash" class Evaluation(BaseModel): is_acceptable: bool feedback: str class Evaluator: def __init__(self, name="", model=MODEL): load_dotenv(override=True) google_api_key = os.getenv('GOOGLE_API_KEY') self.name=name self.model=model self._gemini = OpenAI(api_key=google_api_key, base_url="https://generativelanguage.googleapis.com/v1beta/openai/") def _evaluator_system_prompt(self): return f"You are an evaluator that decides whether a response to a question is acceptable. \ You are provided with a conversation between a User and an Agent. Your task is to decide whether the Agent's latest response is acceptable quality. \ The Agent is playing the role of {self.name} and is representing {self.name} on their website. \ The Agent has been instructed to be professional and engaging, as if talking to a potential client or future employer who came across the website. \ The Agent has been provided with context on {self.name} in the form of their summary, experience and CV. \ With this context, please evaluate the latest response, replying with whether the response is acceptable and your feedback." def _evaluator_user_prompt(self, reply, message, history): user_prompt = f"Here's the conversation between the User and the Agent: \n\n{history}\n\n" user_prompt += f"Here's the latest message from the User: \n\n{message}\n\n" user_prompt += f"Here's the latest response from the Agent: \n\n{reply}\n\n" user_prompt += "Please evaluate the response, replying with whether it is acceptable and your feedback." return user_prompt def evaluate(self, reply, message, history) -> Evaluation: messages = [{"role": "system", "content": self._evaluator_system_prompt()}] + [{"role": "user", "content": self._evaluator_user_prompt(reply, message, history)}] response = self._gemini.beta.chat.completions.parse(model=self.model, messages=messages, response_format=Evaluation) return response.choices[0].message.parsed