better logging
This commit is contained in:
parent
36ed6f0a95
commit
fe4cb78e8a
|
@ -32,8 +32,11 @@ command_prefix: '!c'
|
||||||
|
|
||||||
reply_in_thread: true
|
reply_in_thread: true
|
||||||
|
|
||||||
# Log the full response (prompt + response) at debug level.
|
logging:
|
||||||
log_full_response: false
|
log_level: 'info'
|
||||||
|
|
||||||
|
# Log the full response (prompt + response) at debug level.
|
||||||
|
log_full_response: false
|
||||||
|
|
||||||
logout_other_devices: false
|
logout_other_devices: false
|
||||||
|
|
||||||
|
|
49
main.py
49
main.py
|
@ -23,7 +23,6 @@ script_directory = os.path.abspath(os.path.dirname(__file__))
|
||||||
|
|
||||||
logging.basicConfig()
|
logging.basicConfig()
|
||||||
logger = logging.getLogger('MatrixGPT')
|
logger = logging.getLogger('MatrixGPT')
|
||||||
logger.setLevel(logging.INFO)
|
|
||||||
|
|
||||||
parser = argparse.ArgumentParser(description='MatrixGPT Bot')
|
parser = argparse.ArgumentParser(description='MatrixGPT Bot')
|
||||||
parser.add_argument('--config', default=Path(script_directory, 'config.yaml'), help='Path to config.yaml if it is not located next to this executable.')
|
parser.add_argument('--config', default=Path(script_directory, 'config.yaml'), help='Path to config.yaml if it is not located next to this executable.')
|
||||||
|
@ -52,6 +51,9 @@ check_config_value_exists(config_data, 'allowed_to_invite', allow_empty=True)
|
||||||
check_config_value_exists(config_data, 'command_prefix')
|
check_config_value_exists(config_data, 'command_prefix')
|
||||||
check_config_value_exists(config_data, 'data_storage')
|
check_config_value_exists(config_data, 'data_storage')
|
||||||
|
|
||||||
|
check_config_value_exists(config_data, 'logging')
|
||||||
|
check_config_value_exists(config_data['logging'], 'log_level')
|
||||||
|
|
||||||
check_config_value_exists(config_data, 'openai')
|
check_config_value_exists(config_data, 'openai')
|
||||||
check_config_value_exists(config_data['openai'], 'api_key')
|
check_config_value_exists(config_data['openai'], 'api_key')
|
||||||
check_config_value_exists(config_data['openai'], 'model')
|
check_config_value_exists(config_data['openai'], 'model')
|
||||||
|
@ -68,17 +70,23 @@ def retry(msg=None):
|
||||||
|
|
||||||
|
|
||||||
async def main():
|
async def main():
|
||||||
|
if config_data['logging']['log_level'] == 'info':
|
||||||
|
log_level = logging.INFO
|
||||||
|
elif config_data['logging']['log_level'] == 'debug':
|
||||||
|
log_level = logging.DEBUG
|
||||||
|
elif config_data['logging']['log_level'] == 'warning':
|
||||||
|
log_level = logging.WARNING
|
||||||
|
elif config_data['logging']['log_level'] == 'critical':
|
||||||
|
log_level = logging.CRITICAL
|
||||||
|
else:
|
||||||
|
log_level = logging.INFO
|
||||||
|
logger.setLevel(log_level)
|
||||||
|
|
||||||
# Logging in with a new device each time seems to fix encryption errors
|
# Logging in with a new device each time seems to fix encryption errors
|
||||||
device_id = config_data['bot_auth'].get('device_id', str(uuid4()))
|
device_id = config_data['bot_auth'].get('device_id', str(uuid4()))
|
||||||
|
|
||||||
matrix_helper = MatrixNioGPTHelper(
|
matrix_helper = MatrixNioGPTHelper(auth_file=Path(config_data['bot_auth']['store_path'], 'bot_auth.json'), user_id=config_data['bot_auth']['username'], passwd=config_data['bot_auth']['password'], homeserver=config_data['bot_auth']['homeserver'], store_path=config_data['bot_auth']['store_path'],
|
||||||
auth_file=Path(config_data['bot_auth']['store_path'], 'bot_auth.json'),
|
device_id=device_id, )
|
||||||
user_id=config_data['bot_auth']['username'],
|
|
||||||
passwd=config_data['bot_auth']['password'],
|
|
||||||
homeserver=config_data['bot_auth']['homeserver'],
|
|
||||||
store_path=config_data['bot_auth']['store_path'],
|
|
||||||
device_id=device_id,
|
|
||||||
)
|
|
||||||
client = matrix_helper.client
|
client = matrix_helper.client
|
||||||
|
|
||||||
openai.api_key = config_data['openai']['api_key']
|
openai.api_key = config_data['openai']['api_key']
|
||||||
|
@ -86,20 +94,9 @@ async def main():
|
||||||
storage = Storage(Path(config_data['data_storage'], 'matrixgpt.db'))
|
storage = Storage(Path(config_data['data_storage'], 'matrixgpt.db'))
|
||||||
|
|
||||||
# Set up event callbacks
|
# Set up event callbacks
|
||||||
callbacks = Callbacks(
|
callbacks = Callbacks(client, storage, openai_obj=openai, command_prefix=config_data['command_prefix'], openai_model=config_data['openai']['model'], reply_in_thread=config_data.get('reply_in_thread', False), allowed_to_invite=config_data['allowed_to_invite'],
|
||||||
client,
|
allowed_to_chat=config_data['allowed_to_chat'], log_full_response=config_data.get('log_full_response', False), system_prompt=config_data['openai'].get('system_prompt'), injected_system_prompt=config_data['openai'].get('injected_system_prompt', False),
|
||||||
storage,
|
hyper_temperature=config_data['openai'].get('temperature', 0))
|
||||||
openai_obj=openai,
|
|
||||||
command_prefix=config_data['command_prefix'],
|
|
||||||
openai_model=config_data['openai']['model'],
|
|
||||||
reply_in_thread=config_data.get('reply_in_thread', False),
|
|
||||||
allowed_to_invite=config_data['allowed_to_invite'],
|
|
||||||
allowed_to_chat=config_data['allowed_to_chat'],
|
|
||||||
log_full_response=config_data.get('log_full_response', False),
|
|
||||||
system_prompt=config_data['openai'].get('system_prompt'),
|
|
||||||
injected_system_prompt=config_data['openai'].get('injected_system_prompt', False),
|
|
||||||
hyper_temperature=config_data['openai'].get('temperature', 0)
|
|
||||||
)
|
|
||||||
client.add_event_callback(callbacks.message, RoomMessageText)
|
client.add_event_callback(callbacks.message, RoomMessageText)
|
||||||
client.add_event_callback(callbacks.invite_event_filtered_callback, InviteMemberEvent)
|
client.add_event_callback(callbacks.invite_event_filtered_callback, InviteMemberEvent)
|
||||||
client.add_event_callback(callbacks.decryption_failure, MegolmEvent)
|
client.add_event_callback(callbacks.decryption_failure, MegolmEvent)
|
||||||
|
@ -140,11 +137,7 @@ async def main():
|
||||||
device_list = [x.id for x in devices]
|
device_list = [x.id for x in devices]
|
||||||
if device_id in device_list:
|
if device_id in device_list:
|
||||||
device_list.remove(device_id)
|
device_list.remove(device_id)
|
||||||
x = await client.delete_devices(device_list, {
|
x = await client.delete_devices(device_list, {"type": "m.login.password", "user": config_data['bot_auth']['username'], "password": config_data['bot_auth']['password']})
|
||||||
"type": "m.login.password",
|
|
||||||
"user": config_data['bot_auth']['username'],
|
|
||||||
"password": config_data['bot_auth']['password']
|
|
||||||
})
|
|
||||||
logger.info(f'Logged out: {device_list}')
|
logger.info(f'Logged out: {device_list}')
|
||||||
|
|
||||||
await client.sync_forever(timeout=10000, full_state=True)
|
await client.sync_forever(timeout=10000, full_state=True)
|
||||||
|
|
|
@ -8,16 +8,7 @@ from nio import (AsyncClient, ErrorResponse, Event, MatrixRoom, MegolmEvent, Res
|
||||||
logger = logging.getLogger('MatrixGPT')
|
logger = logging.getLogger('MatrixGPT')
|
||||||
|
|
||||||
|
|
||||||
async def send_text_to_room(
|
async def send_text_to_room(client: AsyncClient, room_id: str, message: str, notice: bool = False, markdown_convert: bool = True, reply_to_event_id: Optional[str] = None, thread: bool = False, thread_root_id: str = None) -> Union[RoomSendResponse, ErrorResponse]:
|
||||||
client: AsyncClient,
|
|
||||||
room_id: str,
|
|
||||||
message: str,
|
|
||||||
notice: bool = False,
|
|
||||||
markdown_convert: bool = True,
|
|
||||||
reply_to_event_id: Optional[str] = None,
|
|
||||||
thread: bool = False,
|
|
||||||
thread_root_id: str = None
|
|
||||||
) -> Union[RoomSendResponse, ErrorResponse]:
|
|
||||||
"""Send text to a matrix room.
|
"""Send text to a matrix room.
|
||||||
|
|
||||||
Args:
|
Args:
|
||||||
|
@ -40,31 +31,16 @@ async def send_text_to_room(
|
||||||
# Determine whether to ping room members or not
|
# Determine whether to ping room members or not
|
||||||
msgtype = "m.notice" if notice else "m.text"
|
msgtype = "m.notice" if notice else "m.text"
|
||||||
|
|
||||||
content = {
|
content = {"msgtype": msgtype, "format": "org.matrix.custom.html", "body": message, }
|
||||||
"msgtype": msgtype,
|
|
||||||
"format": "org.matrix.custom.html",
|
|
||||||
"body": message,
|
|
||||||
}
|
|
||||||
|
|
||||||
if markdown_convert:
|
if markdown_convert:
|
||||||
content["formatted_body"] = markdown(message)
|
content["formatted_body"] = markdown(message)
|
||||||
|
|
||||||
if reply_to_event_id:
|
if reply_to_event_id:
|
||||||
if thread:
|
if thread:
|
||||||
content["m.relates_to"] = {
|
content["m.relates_to"] = {'event_id': thread_root_id, 'is_falling_back': True, "m.in_reply_to": {"event_id": reply_to_event_id}, 'rel_type': "m.thread"}
|
||||||
'event_id': thread_root_id,
|
|
||||||
'is_falling_back': True,
|
|
||||||
"m.in_reply_to": {
|
|
||||||
"event_id": reply_to_event_id
|
|
||||||
},
|
|
||||||
'rel_type': "m.thread"
|
|
||||||
}
|
|
||||||
else:
|
else:
|
||||||
content["m.relates_to"] = {
|
content["m.relates_to"] = {"m.in_reply_to": {"event_id": reply_to_event_id}}
|
||||||
"m.in_reply_to": {
|
|
||||||
"event_id": reply_to_event_id
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
return await client.room_send(room_id, "m.room.message", content, ignore_unverified_devices=True)
|
return await client.room_send(room_id, "m.room.message", content, ignore_unverified_devices=True)
|
||||||
|
@ -108,13 +84,7 @@ async def react_to_event(client: AsyncClient, room_id: str, event_id: str, react
|
||||||
Raises:
|
Raises:
|
||||||
SendRetryError: If the reaction was unable to be sent.
|
SendRetryError: If the reaction was unable to be sent.
|
||||||
"""
|
"""
|
||||||
content = {
|
content = {"m.relates_to": {"rel_type": "m.annotation", "event_id": event_id, "key": reaction_text}}
|
||||||
"m.relates_to": {
|
|
||||||
"rel_type": "m.annotation",
|
|
||||||
"event_id": event_id,
|
|
||||||
"key": reaction_text
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return await client.room_send(room_id, "m.reaction", content, ignore_unverified_devices=True, )
|
return await client.room_send(room_id, "m.reaction", content, ignore_unverified_devices=True, )
|
||||||
|
|
||||||
|
|
||||||
|
@ -160,19 +130,7 @@ async def get_thread_content(client: AsyncClient, room: MatrixRoom, base_event:
|
||||||
return messages
|
return messages
|
||||||
|
|
||||||
|
|
||||||
async def process_chat(
|
async def process_chat(client, room, event, command, store, openai_obj: ModuleType, openai_model: str, thread_root_id: str = None, system_prompt: str = None, log_full_response: bool = False, injected_system_prompt: str = False):
|
||||||
client,
|
|
||||||
room,
|
|
||||||
event,
|
|
||||||
command,
|
|
||||||
store,
|
|
||||||
openai_obj: ModuleType,
|
|
||||||
openai_model: str,
|
|
||||||
thread_root_id: str = None,
|
|
||||||
system_prompt: str = None,
|
|
||||||
log_full_response: bool = False,
|
|
||||||
injected_system_prompt: str = False
|
|
||||||
):
|
|
||||||
if not store.check_seen_event(event.event_id):
|
if not store.check_seen_event(event.event_id):
|
||||||
await client.room_typing(room.room_id, typing_state=True, timeout=3000)
|
await client.room_typing(room.room_id, typing_state=True, timeout=3000)
|
||||||
# if self.reply_in_thread:
|
# if self.reply_in_thread:
|
||||||
|
@ -196,6 +154,7 @@ async def process_chat(
|
||||||
if index != -9999:
|
if index != -9999:
|
||||||
messages.insert(index, {"role": "system", "content": injected_system_prompt})
|
messages.insert(index, {"role": "system", "content": injected_system_prompt})
|
||||||
|
|
||||||
|
logger.debug(f'Generating reply to event {event.event_id}')
|
||||||
response = openai_obj.ChatCompletion.create(model=openai_model, messages=messages, temperature=0, timeout=10)
|
response = openai_obj.ChatCompletion.create(model=openai_model, messages=messages, temperature=0, timeout=10)
|
||||||
text_response = response["choices"][0]["message"]["content"].strip().strip('\n')
|
text_response = response["choices"][0]["message"]["content"].strip().strip('\n')
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue