|
import streamlit as st |
|
import os |
|
import google.generativeai as genai |
|
|
|
|
|
st.title("CHATBOT") |
|
|
|
os.environ['GOOGLE_API_KEY'] = "AIzaSyAh0J6mHZCEnYs0XDSB7QheB1QVZCjxO7k" |
|
genai.configure(api_key = os.environ['GOOGLE_API_KEY']) |
|
|
|
|
|
model = genai.GenerativeModel('gemini-pro') |
|
|
|
|
|
if "messages" not in st.session_state: |
|
st.session_state.messages = [ |
|
{ |
|
"role":"assistant", |
|
"content":"Ask me Anything" |
|
} |
|
] |
|
|
|
|
|
for message in st.session_state.messages: |
|
with st.chat_message(message["role"]): |
|
st.markdown(message["content"]) |
|
|
|
|
|
def llm_function(query): |
|
response = model.generate_content(query) |
|
|
|
|
|
with st.chat_message("assistant"): |
|
st.markdown(response.text) |
|
|
|
|
|
st.session_state.messages.append( |
|
{ |
|
"role":"user", |
|
"content": query |
|
} |
|
) |
|
|
|
|
|
st.session_state.messages.append( |
|
{ |
|
"role":"assistant", |
|
"content": response.text |
|
} |
|
) |
|
|
|
|
|
query = st.chat_input("What's up?") |
|
|
|
|
|
if query: |
|
|
|
with st.chat_message("user"): |
|
st.markdown(query) |
|
|
|
llm_function(query) |