8d5c888cf9 | ||
---|---|---|
.gitignore | ||
Dockerfile | ||
LICENSE | ||
Modelfile | ||
README.md | ||
compose.yaml | ||
discollama.py | ||
poetry.lock | ||
pyproject.toml |
README.md
discollama
discollama
is a Discord bot powered by a local large language model backed by Ollama.
Dependencies
- Docker and Docker Compose
Run discollama.py
DISCORD_TOKEN=xxxxx docker compose up
Note: You must setup a Discord Bot and set environment variable
DISCORD_TOKEN
beforediscollama.py
can access Discord.
discollama.py
requires an Ollama server. Follow the steps in jmorganca/ollama repository to setup Ollama.
By default, it uses 127.0.0.1:11434
which can be overwritten with OLLAMA_HOST
.
Note: Deploying this on Linux requires updating network configurations and
OLLAMA_HOST
.
Customize discollama.py
The default LLM is mike/discollama
. A custom personality can be added by changing the SYSTEM
instruction in the Modelfile and running ollama create
:
ollama create mymodel -f Modelfile
This can be changed in compose.yaml
:
environment:
- OLLAMA_MODEL=mymodel
See jmorganca/ollama for more details.
Activating the Bot
Discord users can interact with the bot by mentioning it in a message to start a new conversation or in a reply to a previous response to continue an ongoing conversation.