def chat(message: str, history: list):
model = 'llama3.2'
# Update this system prompt to change the behavior of the model.
system_prompt = "You are a helpful assistant."
messages = [
{"role": "system", "content": system_prompt},
]
for user_message, assistant_message in history:
messages.append({"role": "user", "content": user_message})
messages.append({"role": "assistant", "content": assistant_message})
messages.append({"role": "user", "content": message})
print("History: ", history)
print("Messages: ", messages)
response = ollama.chat(model=model, messages=messages, stream=True)
all_content = ''
for chunk in response:
content = chunk.get("message", {}).get("content", "")
all_content += content
yield all_content
view = gr.ChatInterface(fn=chat)
view.launch()