# MatrixGPT
_Chatbots for Matrix._
This bot supports OpenAI, Anthropic, and locally hosted models that use an OpenAI-compatible endpoint. It can run
multiple different models using different triggers, such as `!c4` for GPT4 and `!ca` for Anthropic, all through the same bot.
**Supported Services**
- OpenAI (+ vision)
- Anthropic (+ vision)
- Anthropic
- Bing Copilot
- ~~Dalle-3~~ (TODO)
## Install
1. Install requirements:
```bash
sudo apt install libolm-dev gcc python3-dev
pip install -r requirements.txt
```
2. Copy `config.sample.yaml` to `config.yaml` and fill it out with the bot's Matrix authentication and your OpenAI
and/or Anthropic API keys.
3. Start the bot with `python3 main.py`
[Pantalaimon](https://github.com/matrix-org/pantalaimon) is **required** for the bot to be able to talk in encrypted
rooms.
I included a sample Systemd service (`matrixgpt.service`).
## Use
First, invite your bot to a room. Then you can start a chat by prefixing your message with your trigger (for
example, `!c hello!`). The bot will create a thread when it replies. You don't need to use the trigger in the thread.
Use `!matrixgpt` to view the bot's help. The bot also responds to `!bots`.
- Don't try to use two bots in the same thread.
- You can DM the bot for a private chat.
- The bot will move its read marker whenever a message is sent in the room.
The bot can give helpful reactions:
- 🚫 means permission denied (not allowed to chat with the bot).
- 🕒 means the API timed out.
- ❌ means the bot encountered an exception.
- ❌ 🔐 means there was a decryption failure.
## TODO
- [ ] Dalle bot
- [ ] Add our own context mechanism to Copilot?
- [ ] Improve error messages sent with reactions to narrow down where the issue occurred.
- [ ] Allow replying to an image post which will give a vision model an image + text on the first message.
- [ ] Fix the typing indicator being removed when two responses are generating.
- [ ] ~~Add vision to Copilot~~ (not doing, API to unstable).