nio-llm/nio_llm/client.py

125 lines
3.6 KiB
Python
Raw Normal View History

2023-05-20 19:19:14 +00:00
"""A Matrix client that uses Llama to respond to messages."""
import logging
2023-05-22 19:19:11 +00:00
import time
from pathlib import Path
2023-05-20 19:19:14 +00:00
from textwrap import dedent
from llama_cpp import Llama
from nio import AsyncClient, MatrixRoom, RoomMessageText
logger = logging.getLogger("nio-llm.client")
2023-05-20 19:19:14 +00:00
class LLMClient(AsyncClient):
"""A Matrix client that uses Llama to respond to messages."""
def __init__(
self,
2023-05-22 19:19:11 +00:00
username: str,
2023-05-20 19:19:14 +00:00
homeserver: str,
device_id: str,
2023-05-22 19:19:11 +00:00
preprompt: str,
ggml_path: Path,
room: str,
2023-05-20 19:19:14 +00:00
):
"""Create a new LLMClient instance."""
self.uid = f"@{username}:{homeserver.removeprefix('https://')}"
2023-05-22 19:19:11 +00:00
self.spawn_time = time.time() * 1000
self.username = username
self.preprompt = preprompt
self.room = room
# create the AsyncClient instance
super().__init__(
user=self.uid,
homeserver=homeserver,
device_id=device_id,
)
2023-05-20 19:19:14 +00:00
# create the Llama instance
self.llm = Llama(
model_path=str(ggml_path),
2023-05-20 19:19:14 +00:00
n_threads=12,
)
# add callbacks
self.add_event_callback(self.message_callback, RoomMessageText) # type: ignore
async def message_callback(self, room: MatrixRoom, event: RoomMessageText):
"""Process new messages as they come in."""
logger.debug(f"New RoomMessageText: {event.source}")
2023-05-22 19:19:11 +00:00
# ignore our own messages
if event.sender == self.user:
logger.debug("Ignoring our own message.")
return
# ignore messages pre-dating our spawn time
2023-05-22 19:19:11 +00:00
if event.server_timestamp < self.spawn_time:
logger.debug("Ignoring message pre-spawn.")
return
# ignore messages not in our monitored room
2023-05-22 19:19:11 +00:00
if room.room_id != self.room:
logger.debug("Ignoring message in different room.")
2023-05-20 19:19:14 +00:00
return
# ignore edited messages
if "m.new_content" in event.source["content"]:
logger.debug("Ignoring edited message.")
return
# ignore messages not mentioning us
if not (
"format" in event.source["content"]
and "formatted_body" in event.source["content"]
and event.source["content"]["format"] == "org.matrix.custom.html"
and f'<a href="https://matrix.to/#/{self.uid}">{self.username}</a>'
in event.source["content"]["formatted_body"]
):
2023-05-22 19:19:11 +00:00
logger.debug("Ignoring message not directed at us.")
return
# generate prompt from message
2023-05-22 19:19:11 +00:00
prompt = dedent(
f"""
{self.preprompt}
<{event.sender}>: {event.body}
<{self.username}>:
2023-05-22 19:19:11 +00:00
""",
).strip()
logger.debug(f"Prompt: {prompt}")
# enable typing indicator
await self.room_typing(
self.room,
typing_state=True,
timeout=100000000,
)
# generate response using llama.cpp
2023-05-22 19:19:11 +00:00
output = self.llm(
prompt,
max_tokens=100,
stop=[f"<{event.sender}>"],
2023-05-22 19:19:11 +00:00
echo=True,
)
# retreive the response
output = output["choices"][0]["text"] # type: ignore
output = output.removeprefix(prompt).strip()
# disable typing indicator
await self.room_typing(self.room, typing_state=False)
# send the response
await self.room_send(
room_id=self.room,
message_type="m.room.message",
content={
"msgtype": "m.text",
"body": output,
},
)