mirror of
https://github.com/kevinthedang/discord-ollama.git
synced 2025-12-12 11:56:06 -05:00
Compare commits
27 Commits
v0.6.0
...
029e018dae
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
029e018dae | ||
|
|
e07e8fbf89 | ||
|
|
6d0a537540 | ||
|
|
0ddd59aea1 | ||
|
|
a5faca87aa | ||
|
|
4c96b3863a | ||
|
|
40783818b9 | ||
|
|
ed0d8600df | ||
|
|
03939ef562 | ||
|
|
456f70b9e1 | ||
|
|
5b542aca1a | ||
|
|
2a39e20fee | ||
|
|
2ea77c92f0 | ||
|
|
6c7e48d369 | ||
|
|
fe1f7ce5ec | ||
|
|
6ac45afb13 | ||
|
|
d570a50d46 | ||
|
|
1c8449d578 | ||
|
|
33152b33f3 | ||
|
|
1ccd1a012e | ||
|
|
68a5e097fe | ||
|
|
624ff2e5c8 | ||
|
|
9f61f6bc6c | ||
|
|
5d02800c3f | ||
|
|
5061dab335 | ||
|
|
947ff89958 | ||
|
|
6a9ee2d6d0 |
11
.env.sample
11
.env.sample
@@ -1,11 +1,8 @@
|
||||
# Discord token for the bot
|
||||
CLIENT_TOKEN = BOT_TOKEN
|
||||
|
||||
# model for the bot to query from (i.e. llama2 [llama2:13b], mistral, codellama, etc... )
|
||||
MODEL = MODEL_NAME
|
||||
|
||||
# discord bot user id for mentions
|
||||
CLIENT_UID = BOT_USER_ID
|
||||
# Default model for new users
|
||||
MODEL = DEFAULT_MODEL
|
||||
|
||||
# ip/port address of docker container, I use 172.18.0.3 for docker, 127.0.0.1 for local
|
||||
OLLAMA_IP = IP_ADDRESS
|
||||
@@ -16,3 +13,7 @@ DISCORD_IP = IP_ADDRESS
|
||||
|
||||
# subnet address, ex. 172.18.0.0 as we use /16.
|
||||
SUBNET_ADDRESS = ADDRESS
|
||||
|
||||
# redis port and ip, default redis port is 6379
|
||||
REDIS_IP = IP_ADDRESS
|
||||
REDIS_PORT = PORT
|
||||
26
.github/CONTRIBUTING.md
vendored
26
.github/CONTRIBUTING.md
vendored
@@ -1,31 +1,39 @@
|
||||
<!--
|
||||
Author: Kevin Dang
|
||||
Date: 1-30-2024
|
||||
Author: Kevin Dang
|
||||
Date: 1-30-2024
|
||||
Changes:
|
||||
10-01-2024 - Jonathan Smoley
|
||||
-->
|
||||
|
||||
## Naming Conventions
|
||||
* Branches
|
||||
* prefix your branch name with the type of contribution:
|
||||
* features: `'feature/**'`
|
||||
* releases: `'releases/**'`
|
||||
* bugs: `'bug/**'`
|
||||
* docs: `'docs/**'`
|
||||
|
||||
## Run the Bot
|
||||
* Refer to all sections below before running the bot.
|
||||
* You should now have `Ollama`, `NodeJS`, ran `npm install`.
|
||||
* You will also need a discord bot to run. Refer to the [developer portal](https://discord.com/developers/) to learn how to set one up and invite it to your server. If that does not help then look up a YouTube video like this [one](https://www.youtube.com/watch?v=KZ3tIGHU314&ab_channel=UnderCtrl).
|
||||
* Now run `npm run start` to run the client and ollama at the same time (this must be one in wsl or a Linux distro)
|
||||
|
||||
* Now run `npm run client` to run the client (this must be done in wsl or a Linux distro)
|
||||
|
||||
## Set up (Development-side)
|
||||
* Pull the repository using `https://github.com/kevinthedang/discord-ollama.git`.
|
||||
* Refer to `Ollama Setup` in the readme to set up Ollama.
|
||||
* Refer to `Environment Setup` in the readme to set up Ollama.
|
||||
* This must be set up in a Linux environment or wsl2.
|
||||
* Install NodeJS `v18.18.2`
|
||||
* You can check out `Resources` and `To Run` in the readme for a bit of help.
|
||||
* You can check out `Resources` in the readme for a bit of help.
|
||||
* You can also reference [NodeJS Setup](#nodejs-setup)
|
||||
* When you have the project pulled from github, open up a terminal and run `npm i` or `npm install` to get all of the packages for the project.
|
||||
* In some kind of terminal (`git bash` is good) to run the client. You can run Ollama but opening up wsl2 and typing `ollama serve`.
|
||||
* Refer to `Ollama Setup` if there are any issues.
|
||||
|
||||
## Environment
|
||||
* You will need two environment files:
|
||||
* You will need an environment file:
|
||||
* `.env`: for running the bot
|
||||
* Please refer to `.env.sample` for all environment variables to include
|
||||
* `.env.dev.local`: also runs the bot, but with development variables
|
||||
* Currently there are no differences between the two, but when needed, you may add environment variables as needed.
|
||||
|
||||
## NodeJS Setup
|
||||
* Install [nvm](https://github.com/nvm-sh/nvm?tab=readme-ov-file#installing-and-updating) using `curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.39.7/install.sh | bash`
|
||||
|
||||
22
.github/workflows/build.yml
vendored
22
.github/workflows/build.yml
vendored
@@ -13,10 +13,10 @@ jobs:
|
||||
- name: Checkout Repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Set up Node Environment lts/hydrogen
|
||||
- name: Set up Node Environment lts/jod
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: lts/hydrogen
|
||||
node-version: lts/jod
|
||||
cache: "npm"
|
||||
|
||||
- name: Install Project Dependencies
|
||||
@@ -31,10 +31,11 @@ jobs:
|
||||
run: |
|
||||
touch .env
|
||||
echo CLIENT_TOKEN = ${{ secrets.BOT_TOKEN }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo CLIENT_UID = ${{ secrets.CLIENT_UID }} >> .env
|
||||
echo OLLAMA_IP = ${{ secrets.OLLAMA_IP }} >> .env
|
||||
echo OLLAMA_PORT = ${{ secrets.OLLAMA_PORT }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo REDIS_IP = ${{ secrets.REDIS_IP }} >> .env
|
||||
echo REDIS_PORT = ${{ secrets.REDIS_PORT }} >> .env
|
||||
|
||||
# set -e ensures if nohup fails, this section fails
|
||||
- name: Startup Discord Bot Client
|
||||
@@ -49,20 +50,21 @@ jobs:
|
||||
- name: Checkout Repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Set up Node Environment lts/hydrogen
|
||||
- name: Set up Node Environment lts/jod
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: lts/hydrogen
|
||||
node-version: lts/jod
|
||||
cache: "npm"
|
||||
|
||||
- name: Create Environment Variables
|
||||
run: |
|
||||
touch .env
|
||||
echo CLIENT_TOKEN = ${{ secrets.BOT_TOKEN }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo CLIENT_UID = ${{ secrets.CLIENT_UID }} >> .env
|
||||
echo OLLAMA_IP = ${{ secrets.OLLAMA_IP }} >> .env
|
||||
echo OLLAMA_PORT = ${{ secrets.OLLAMA_PORT }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo REDIS_IP = ${{ secrets.REDIS_IP }} >> .env
|
||||
echo REDIS_PORT = ${{ secrets.REDIS_PORT }} >> .env
|
||||
|
||||
- name: Setup Docker Network and Images
|
||||
run: |
|
||||
@@ -70,8 +72,8 @@ jobs:
|
||||
|
||||
- name: Check Images Exist
|
||||
run: |
|
||||
(docker images | grep -q 'kevinthedang/discord-ollama' && docker images | grep -qE 'ollama/ollama') || exit 1
|
||||
(docker images | grep -q 'kevinthedang/discord-ollama' && docker images | grep -qE 'ollama/ollama' | docker images | grep -qE 'redis') || exit 1
|
||||
|
||||
- name: Check Containers Exist
|
||||
run: |
|
||||
(docker ps | grep -q 'ollama' && docker ps | grep -q 'discord') || exit 1
|
||||
(docker ps | grep -q 'ollama' && docker ps | grep -q 'discord' && docker ps | grep -q 'redis') || exit 1
|
||||
|
||||
52
.github/workflows/coverage.yml
vendored
Normal file
52
.github/workflows/coverage.yml
vendored
Normal file
@@ -0,0 +1,52 @@
|
||||
name: Coverage
|
||||
run-name: Code Coverage
|
||||
on:
|
||||
push:
|
||||
branches:
|
||||
- master
|
||||
|
||||
jobs:
|
||||
Discord-Node-Coverage:
|
||||
runs-on: ubuntu-latest
|
||||
environment: coverage
|
||||
timeout-minutes: 2
|
||||
steps:
|
||||
- name: Checkout Repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Set up Node Environment lts/jod
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: lts/jod
|
||||
cache: "npm"
|
||||
|
||||
- name: Install Project Dependencies
|
||||
run: |
|
||||
npm install
|
||||
|
||||
- name: Create Environment Variables
|
||||
run: |
|
||||
touch .env
|
||||
echo CLIENT_TOKEN = ${{ secrets.BOT_TOKEN }} >> .env
|
||||
echo OLLAMA_IP = ${{ secrets.OLLAMA_IP }} >> .env
|
||||
echo OLLAMA_PORT = ${{ secrets.OLLAMA_PORT }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo REDIS_IP = ${{ secrets.REDIS_IP }} >> .env
|
||||
echo REDIS_PORT = ${{ secrets.REDIS_PORT }} >> .env
|
||||
|
||||
- name: Collect Code Coverage
|
||||
run: |
|
||||
LINE_PCT=$(npm run coverage | tail -2 | head -1 | awk '{print $3}')
|
||||
echo "COVERAGE=$LINE_PCT" >> $GITHUB_ENV
|
||||
|
||||
- name: Upload Code Coverage
|
||||
uses: schneegans/dynamic-badges-action@v1.7.0
|
||||
with:
|
||||
auth: ${{ secrets.GIST_SECRET }}
|
||||
gistID: ${{ vars.GIST_ID }}
|
||||
filename: coverage.json
|
||||
label: Coverage
|
||||
message: ${{ env.COVERAGE }}
|
||||
valColorRange: ${{ env.COVERAGE }}
|
||||
maxColorRange: 100
|
||||
minColorRange: 0
|
||||
133
.github/workflows/deploy.yml
vendored
Normal file
133
.github/workflows/deploy.yml
vendored
Normal file
@@ -0,0 +1,133 @@
|
||||
name: Deploy
|
||||
run-name: Deploy Application Latest
|
||||
on:
|
||||
push:
|
||||
tags:
|
||||
- 'v*'
|
||||
|
||||
jobs:
|
||||
Deploy-Application:
|
||||
runs-on: self-hosted
|
||||
environment: deploy
|
||||
timeout-minutes: 5
|
||||
steps:
|
||||
- name: Checkout Repo
|
||||
uses: actions/checkout@v4
|
||||
|
||||
# Generate Secret File for Compose case
|
||||
- name: Create Environment Variables
|
||||
run: |
|
||||
touch .env
|
||||
echo CLIENT_TOKEN = ${{ secrets.CLIENT }} >> .env
|
||||
echo OLLAMA_IP = ${{ secrets.OLLAMA_IP }} >> .env
|
||||
echo OLLAMA_PORT = ${{ secrets.OLLAMA_PORT }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo DISCORD_IP = ${{ secrets.DISCORD_IP }} >> .env
|
||||
echo SUBNET_ADDRESS = ${{ secrets.SUBNET_ADDRESS }} >> .env
|
||||
echo REDIS_IP = ${{ secrets.REDIS_IP }} >> .env
|
||||
echo REDIS_PORT = ${{ secrets.REDIS_PORT }} >> .env
|
||||
|
||||
- name: Check if directory exists and delete it
|
||||
run: |
|
||||
if [ -d "${{ secrets.PATH }}" ]; then
|
||||
echo "Directory exists, deleting old version..."
|
||||
rm -rf ${{ secrets.PATH }}
|
||||
else
|
||||
echo "Directory does not exist."
|
||||
fi
|
||||
|
||||
- name: Clone Repo onto Server
|
||||
run: |
|
||||
git clone https://github.com/kevinthedang/discord-ollama.git ${{ secrets.PATH }}
|
||||
cd ${{ secrets.PATH }}
|
||||
|
||||
- name: Install nvm and Node.js lts/jod
|
||||
run: |
|
||||
curl -o- https://raw.githubusercontent.com/nvm-sh/nvm/v0.40.1/install.sh | bash
|
||||
export NVM_DIR="$HOME/.nvm"
|
||||
[ -s "$NVM_DIR/nvm.sh" ] && \. "$NVM_DIR/nvm.sh"
|
||||
echo "NVM installed successfully."
|
||||
nvm install lts/jod
|
||||
nvm alias default lts/jod
|
||||
node -v
|
||||
npm -v
|
||||
|
||||
- name: Build Application
|
||||
run: |
|
||||
export NVM_DIR="$HOME/.nvm"
|
||||
[ -s "$NVM_DIR/nvm.sh" ] && \. "$NVM_DIR/nvm.sh"
|
||||
npm install
|
||||
|
||||
IMAGE="kevinthedang/discord-ollama"
|
||||
REDIS="redis"
|
||||
OLLAMA="ollama/ollama"
|
||||
|
||||
if docker images | grep -q $IMAGE; then
|
||||
IMAGE_ID=$(docker images -q $IMAGE)
|
||||
CONTAINER_IDS=$(docker ps -q --filter "ancestor=$IMAGE_ID")
|
||||
|
||||
if [ ! -z "$CONTAINER_IDS" ]; then
|
||||
# Stop and remove the running containers
|
||||
docker stop $CONTAINER_IDS
|
||||
echo "Stopped and removed the containers using the image $IMAGE"
|
||||
fi
|
||||
docker rmi $IMAGE_ID
|
||||
echo "Old $IMAGE Image Removed"
|
||||
fi
|
||||
|
||||
if docker images | grep -q $REDIS; then
|
||||
IMAGE_ID=$(docker images -q $REDIS)
|
||||
CONTAINER_IDS=$(docker ps -q --filter "ancestor=$IMAGE_ID")
|
||||
|
||||
if [ ! -z "$CONTAINER_IDS" ]; then
|
||||
# Stop and remove the running containers
|
||||
docker stop $CONTAINER_IDS
|
||||
echo "Stopped and removed the containers using the image $REDIS"
|
||||
fi
|
||||
docker rmi $IMAGE_ID
|
||||
echo "Old $REDIS Image Removed"
|
||||
fi
|
||||
|
||||
if docker images | grep -q $OLLAMA; then
|
||||
IMAGE_ID=$(docker images -q $OLLAMA)
|
||||
CONTAINER_IDS=$(docker ps -q --filter "ancestor=$IMAGE_ID")
|
||||
|
||||
if [ ! -z "$CONTAINER_IDS" ]; then
|
||||
# Stop and remove the running containers
|
||||
docker stop $CONTAINER_IDS
|
||||
echo "Stopped and removed the containers using the image $OLLAMA"
|
||||
fi
|
||||
docker rmi $IMAGE_ID
|
||||
echo "Old $OLLAMA Image Removed"
|
||||
fi
|
||||
|
||||
docker network prune -f
|
||||
docker system prune -a -f
|
||||
|
||||
npm run docker:build-latest
|
||||
|
||||
- name: Start Application
|
||||
run: |
|
||||
docker network create --subnet=${{ secrets.SUBNET_ADDRESS }}/16 ollama-net || true
|
||||
docker run --rm -d \
|
||||
-v ollama:/root/.ollama \
|
||||
-p ${{ secrets.OLLAMA_PORT }}:${{ secrets.OLLAMA_PORT }} \
|
||||
--name ollama \
|
||||
--network ollama-net \
|
||||
--ip ${{ secrets.OLLAMA_IP }} \
|
||||
ollama/ollama:latest
|
||||
|
||||
docker run --rm -d \
|
||||
-v redis:/root/.redis \
|
||||
-p ${{ secrets.REDIS_PORT }}:${{ secrets.REDIS_PORT }} \
|
||||
--name redis \
|
||||
--network ollama-net \
|
||||
--ip ${{ secrets.REDIS_IP }} \
|
||||
redis:latest
|
||||
|
||||
docker run --rm -d \
|
||||
-v discord:/src/app \
|
||||
--name discord \
|
||||
--network ollama-net \
|
||||
--ip ${{ secrets.DISCORD_IP }} \
|
||||
kevinthedang/discord-ollama
|
||||
50
.github/workflows/release.yml
vendored
50
.github/workflows/release.yml
vendored
@@ -1,50 +0,0 @@
|
||||
name: Deploy
|
||||
run-name: Release Docker Image
|
||||
on:
|
||||
push:
|
||||
tags:
|
||||
- 'v*'
|
||||
|
||||
jobs:
|
||||
Release-Docker-Image:
|
||||
runs-on: ubuntu-latest
|
||||
environment: release
|
||||
timeout-minutes: 3
|
||||
steps:
|
||||
- name: Checkout Repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Set up Node Environment lts/hydrogen
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: lts/hydrogen
|
||||
cache: "npm"
|
||||
|
||||
- name: Create Environment Variables
|
||||
run: |
|
||||
touch .env
|
||||
echo CLIENT_TOKEN = NOT_REAL_TOKEN >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo CLIENT_UID = ${{ secrets.CLIENT_UID }} >> .env
|
||||
echo OLLAMA_IP = ${{ secrets.OLLAMA_IP }} >> .env
|
||||
echo OLLAMA_PORT = ${{ secrets.OLLAMA_PORT }} >> .env
|
||||
|
||||
- name: Get Version from package.json
|
||||
run: echo "VERSION=$(jq -r '.version' package.json)" >> $GITHUB_ENV
|
||||
|
||||
- name: Build Image
|
||||
run: |
|
||||
npm run docker:build
|
||||
|
||||
- name: Build Image as Latest
|
||||
run: |
|
||||
npm run docker:build-latest
|
||||
|
||||
- name: Log into Docker
|
||||
run: |
|
||||
docker login --username ${{ vars.DOCKER_USER }} --password ${{ secrets.DOCKER_PASS }}
|
||||
|
||||
- name: Release Docker Image
|
||||
run: |
|
||||
docker push ${{ vars.DOCKER_USER }}/discord-ollama:${{ env.VERSION }}
|
||||
docker push ${{ vars.DOCKER_USER }}/discord-ollama:latest
|
||||
11
.github/workflows/test.yml
vendored
11
.github/workflows/test.yml
vendored
@@ -25,10 +25,10 @@ jobs:
|
||||
- name: Checkout Repository
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- name: Set up Node Environment lts/hydrogen
|
||||
- name: Set up Node Environment lts/jod
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: lts/hydrogen
|
||||
node-version: lts/jod
|
||||
cache: "npm"
|
||||
|
||||
- name: Install Project Dependencies
|
||||
@@ -39,11 +39,12 @@ jobs:
|
||||
run: |
|
||||
touch .env
|
||||
echo CLIENT_TOKEN = ${{ secrets.BOT_TOKEN }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo CLIENT_UID = ${{ secrets.CLIENT_UID }} >> .env
|
||||
echo OLLAMA_IP = ${{ secrets.OLLAMA_IP }} >> .env
|
||||
echo OLLAMA_PORT = ${{ secrets.OLLAMA_PORT }} >> .env
|
||||
echo MODEL = ${{ secrets.MODEL }} >> .env
|
||||
echo REDIS_IP = ${{ secrets.REDIS_IP }} >> .env
|
||||
echo REDIS_PORT = ${{ secrets.REDIS_PORT }} >> .env
|
||||
|
||||
- name: Test Application
|
||||
run: |
|
||||
npm run test:run
|
||||
npm run tests
|
||||
|
||||
24
CODEOWNERS
Normal file
24
CODEOWNERS
Normal file
@@ -0,0 +1,24 @@
|
||||
# The further along the ownership is, the more precedence it has.
|
||||
# This to make sure the right people look at certain changes.
|
||||
# Last Edited: 11/23/2024
|
||||
# Author: Kevin Dang
|
||||
|
||||
# These owners will be the default owners
|
||||
# for everything in the repo. However it's
|
||||
# only for the rest of the files not declared by the
|
||||
# following ownerships below.
|
||||
* @kevinthedang @JT2M0L3Y
|
||||
|
||||
# Technical/Business Code Ownership
|
||||
/src/ @kevinthedang @JT2M0L3Y
|
||||
/tests/ @JT2M0L3Y
|
||||
/.github/ @kevinthedang
|
||||
|
||||
# Docker Ownership
|
||||
Dockerfile @kevinthedang
|
||||
docker-compose.yml @kevinthedang
|
||||
|
||||
# Documentation Ownership
|
||||
/docs/ @kevinthedang
|
||||
/imgs/ @kevinthedang
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
# use node LTS image for version 18
|
||||
FROM node:hydrogen-alpine
|
||||
# use node LTS image for version 22
|
||||
FROM node:jod-alpine
|
||||
|
||||
# set working directory inside container
|
||||
WORKDIR /app
|
||||
|
||||
26
README.md
26
README.md
@@ -5,8 +5,9 @@
|
||||
<p><a href="#"></a><a href="https://creativecommons.org/licenses/by/4.0/"><img alt="License" src="https://img.shields.io/badge/License-CC_BY_4.0-darkgreen.svg" /></a>
|
||||
<a href="#"></a><a href="https://github.com/kevinthedang/discord-ollama/releases/latest"><img alt="Release" src="https://img.shields.io/github/v/release/kevinthedang/discord-ollama?logo=github" /></a>
|
||||
<a href="#"></a><a href="https://github.com/kevinthedang/discord-ollama/actions/workflows/build.yml"><img alt="Build Status" src="https://github.com/kevinthedang/discord-ollama/actions/workflows/build.yml/badge.svg" /></a>
|
||||
<a href="#"></a><a href="https://github.com/kevinthedang/discord-ollama/actions/workflows/release.yml"><img alt="Release Status" src="https://github.com/kevinthedang/discord-ollama/actions/workflows/release.yml/badge.svg" /></a>
|
||||
<a href="#"></a><a href="https://github.com/kevinthedang/discord-ollama/actions/workflows/deploy.yml"><img alt="Deploy Status" src="https://github.com/kevinthedang/discord-ollama/actions/workflows/deploy.yml/badge.svg" /></a>
|
||||
<a href="#"></a><a href="https://github.com/kevinthedang/discord-ollama/actions/workflows/test.yml"><img alt="Testing Status" src="https://github.com/kevinthedang/discord-ollama/actions/workflows/test.yml/badge.svg" /></a>
|
||||
<a href="#"></a><a href="https://github.com/kevinthedang/discord-ollama/actions/workflows/coverage.yml"><img alt="Code Coverage" src="https://img.shields.io/endpoint?url=https://gist.githubusercontent.com/kevinthedang/bc7b5dcfa16561ab02bb3df67a99b22d/raw/coverage.json"></a>
|
||||
</div>
|
||||
|
||||
## About/Goals
|
||||
@@ -24,11 +25,17 @@ The project aims to:
|
||||
* [x] User vs. Server Preferences
|
||||
* [ ] Redis Caching
|
||||
* [x] Administrator Role Compatible
|
||||
* [x] Multi-User Chat Generation (Multiple users chatting at the same time) - This was built into from Ollama `v0.2.1+`
|
||||
* [ ] Automatic and Manual model pulling through the Discord client
|
||||
* [ ] Allow others to create their own models personalized for their own servers!
|
||||
* [ ] Documentation on creating your own LLM
|
||||
* [ ] Documentation on web scrapping and cleaning
|
||||
* [x] Multi-User Chat Generation (Multiple users chatting at the same time) - This was built in from Ollama `v0.2.1+`
|
||||
* [x] Automatic and Manual model pulling through the Discord client
|
||||
|
||||
Further, Ollama provides the functionality to utilize custom models or provide context for the top-layer of any model available through the Ollama model library.
|
||||
* [Customize a model](https://github.com/ollama/ollama#customize-a-model)
|
||||
* [Modelfile Docs](https://github.com/ollama/ollama/blob/main/docs/modelfile.md)
|
||||
|
||||
## Documentation
|
||||
These are guides to the features and capabilities of this app.
|
||||
* [User Slash Commands](./docs/commands-guide.md)
|
||||
* [Client Events](./docs/events-guide.md)
|
||||
|
||||
## Environment Setup
|
||||
* Clone this repo using `git clone https://github.com/kevinthedang/discord-ollama.git` or just use [GitHub Desktop](https://desktop.github.com/) to clone the repo.
|
||||
@@ -40,12 +47,15 @@ The project aims to:
|
||||
* [Docker Setup for Servers and Local Machines](./docs/setup-docker.md)
|
||||
* Nvidia is recommended for now, but support for other GPUs should be development.
|
||||
* Local use is not recommended.
|
||||
|
||||
## Resources
|
||||
* [NodeJS](https://nodejs.org/en)
|
||||
* This project runs on `lts\hydrogen`.
|
||||
* This project supports any NodeJS version above `16.x.x` to only allow ESModules.
|
||||
* This project runs on `lts\jod` and above.
|
||||
* This project requires the use of npm version `10.9.0` or above.
|
||||
* [Ollama](https://ollama.com/)
|
||||
* [Ollama Docker Image](https://hub.docker.com/r/ollama/ollama)
|
||||
* [Redis](https://redis.io/)
|
||||
* [Redis Docker Image](https://hub.docker.com/_/redis)
|
||||
* [Discord.js Docs](https://discord.js.org/docs/packages/discord.js/main)
|
||||
* [Setting up Docker (Ubuntu 20.04)](https://www.digitalocean.com/community/tutorials/how-to-install-and-use-docker-on-ubuntu-20-04)
|
||||
* [Setting up Nvidia Container Toolkit](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html)
|
||||
|
||||
@@ -7,13 +7,14 @@ services:
|
||||
build: ./ # find docker file in designated path
|
||||
container_name: discord
|
||||
restart: always # rebuild container always
|
||||
image: kevinthedang/discord-ollama:0.6.0
|
||||
image: kevinthedang/discord-ollama:0.8.5
|
||||
environment:
|
||||
CLIENT_TOKEN: ${CLIENT_TOKEN}
|
||||
MODEL: ${MODEL}
|
||||
CLIENT_UID: ${CLIENT_UID}
|
||||
OLLAMA_IP: ${OLLAMA_IP}
|
||||
OLLAMA_PORT: ${OLLAMA_PORT}
|
||||
MODEL: ${MODEL}
|
||||
REDIS_IP: ${REDIS_IP}
|
||||
REDIS_PORT: ${REDIS_PORT}
|
||||
networks:
|
||||
ollama-net:
|
||||
ipv4_address: ${DISCORD_IP}
|
||||
@@ -28,7 +29,6 @@ services:
|
||||
networks:
|
||||
ollama-net:
|
||||
ipv4_address: ${OLLAMA_IP}
|
||||
|
||||
runtime: nvidia # use Nvidia Container Toolkit for GPU support
|
||||
devices:
|
||||
- /dev/nvidia0
|
||||
@@ -37,7 +37,18 @@ services:
|
||||
ports:
|
||||
- ${OLLAMA_PORT}:${OLLAMA_PORT}
|
||||
|
||||
# Put Redis Container here?
|
||||
# setup redis container
|
||||
redis:
|
||||
image: redis:latest
|
||||
container_name: redis
|
||||
restart: always
|
||||
networks:
|
||||
ollama-net:
|
||||
ipv4_address: ${REDIS_IP}
|
||||
volumes:
|
||||
- redis:/root/.redis
|
||||
ports:
|
||||
- ${REDIS_PORT}:${REDIS_PORT}
|
||||
|
||||
# create a network that supports giving addresses withing a specific subnet
|
||||
networks:
|
||||
@@ -51,3 +62,4 @@ networks:
|
||||
volumes:
|
||||
ollama:
|
||||
discord:
|
||||
redis:
|
||||
|
||||
104
docs/commands-guide.md
Normal file
104
docs/commands-guide.md
Normal file
@@ -0,0 +1,104 @@
|
||||
## Commands Guide
|
||||
This is a guide to all of the slash commands for the app.
|
||||
|
||||
* Action Commands are commands that do not affect a user's `preference file`.
|
||||
* Guild Commands can also be considered action commands.
|
||||
|
||||
> [!NOTE]
|
||||
> Administrator commands are only usable by actual administrators on the Discord server.
|
||||
|
||||
### Guild Commands (Administrator)
|
||||
1. Disable (or Toggle Chat)
|
||||
This command will `enable` or `disable` whether or not the app will respond to users.
|
||||
|
||||
```
|
||||
/toggle-chat enabled true
|
||||
```
|
||||
|
||||
2. Shutoff
|
||||
This command will shutoff the app so no users can converse with it.
|
||||
The app must be manually restarted upon being shutoff.
|
||||
|
||||
Below shuts off the app by putting `true` in the `are-your-sure` field.
|
||||
|
||||
```
|
||||
/shutoff are-you-sure true
|
||||
```
|
||||
|
||||
### Action Commands
|
||||
1. Clear Channel (Message) History
|
||||
This command will clear the history of the current channel for the user that calls it.
|
||||
Running the command in any channel will clear the message history.
|
||||
|
||||
```
|
||||
/clear-user-channel-history
|
||||
```
|
||||
|
||||
2. Pull Model
|
||||
This command will pull a model that exists on the [Ollama Model Library](https://ollama.com/library). If it does not exist there, it will throw a hissy fit.
|
||||
|
||||
Below trys to pull the `codellama` model.
|
||||
|
||||
```
|
||||
/pull-model model-to-pull codellama
|
||||
```
|
||||
|
||||
3. Thread Create
|
||||
This command creates a public thread to talk with the app instead of using a `GuildText` channel.
|
||||
|
||||
```
|
||||
/thread
|
||||
```
|
||||
|
||||
4. (Private) Thread Create
|
||||
This command creates a private thread to talk with the bot privately.
|
||||
Invite others to the channel and they will be able to talk to the app as well.
|
||||
|
||||
```
|
||||
/private-thread
|
||||
```
|
||||
|
||||
### User Preference Commands
|
||||
1. Capacity
|
||||
This command changes how much context it will keep in conversations with the app.
|
||||
This is applied for all of existing chats when interacting with the app.
|
||||
|
||||
Below sets the message history capacity to at most 5 messages at once.
|
||||
|
||||
```
|
||||
/modify-capacity context-capacity 5
|
||||
```
|
||||
|
||||
2. Message Stream
|
||||
This command will toggle whether or not the app will "stream" a response.
|
||||
(think of how ChatGPT and other interfaces do this)
|
||||
|
||||
Below sets the `stream` to true to make the app respond in increments.
|
||||
|
||||
```
|
||||
/message-stream stream true
|
||||
```
|
||||
**This is very slow on Discord because "spamming" changes in a channel within a period of 5 seconds is not allowed.**
|
||||
|
||||
3. Message Style
|
||||
This command allows a user to select whether to embed the app's response.
|
||||
|
||||
```
|
||||
/message-style embed true
|
||||
```
|
||||
|
||||
This allows the app to respond as a user would normally respond.
|
||||
|
||||
```
|
||||
/message-style embed false
|
||||
```
|
||||
|
||||
4. Switch Model
|
||||
This command will switch the user-preferred model so long as it exists in within the local ollama service or from the [Ollama Model Library](https://ollama.com/library).
|
||||
If it cannot be found locally, it will attempt to find it in the model library.
|
||||
|
||||
Below we are trying to switch to a specific model size.
|
||||
|
||||
```
|
||||
/switch-model model-to-use llama3.2:1.3b
|
||||
```
|
||||
27
docs/events-guide.md
Normal file
27
docs/events-guide.md
Normal file
@@ -0,0 +1,27 @@
|
||||
## Events Guide
|
||||
This is a guide to all of the client events for the app.
|
||||
|
||||
> [!NOTE]
|
||||
> * Each of these is logged to the console for a developer to track.
|
||||
> * Possible interactions include commands, buttons, menus, etc.
|
||||
|
||||
1. ClientReady
|
||||
This event signifies that the Discord app is online.
|
||||
Here the app's activity is set and its commands are registered.
|
||||
|
||||
2. InteractionCreate
|
||||
This event signifies that a user interacted from Discord in some way.
|
||||
Here commands are selected from a knowledge bank and executed if found.
|
||||
|
||||
3. MessageCreate
|
||||
This event signifies that a message was sent.
|
||||
Here user questions and comments for the LLM are processed.
|
||||
1. check message is from a user and mentions the app
|
||||
2. check for interaction preferences
|
||||
3. add the message to a queue
|
||||
4. check the response for success
|
||||
5. send a response back to the user.
|
||||
|
||||
4. ThreadDelete
|
||||
This event signifies that a Discord Thread was deleted.
|
||||
Here any preferences set for interaction within the thread are cleared away.
|
||||
@@ -18,10 +18,6 @@
|
||||
|
||||

|
||||
|
||||
* You will also need your App's **Client ID**, navigate to **OAuth2** and copy your id.
|
||||
|
||||

|
||||
|
||||
* That should be all of the environment variables needed from Discord, now we need this app on your server.
|
||||
* Navigate to **Installation** and Copy the provided **Install Link** to allow your App to your server.
|
||||
* You should set the **Guild Install** permissions as you like, for this purpose we will allow admin priviledges for now. Ensure the **bot** scope is added to do this.
|
||||
|
||||
@@ -2,7 +2,9 @@
|
||||
* Follow this guide to setup [Docker](https://www.digitalocean.com/community/tutorials/how-to-install-and-use-docker-on-ubuntu-20-04)
|
||||
* If on Windows, download [Docker Desktop](https://docs.docker.com/desktop/install/windows-install/) to get the docker engine.
|
||||
* Please also install [Docker Compose](https://docs.docker.com/compose/install/linux/) for easy running. If not, there are [scripts](#manual-run-with-docker) to set everything up.
|
||||
* **IMPORTANT NOTE**: Currently, it seems like wsl does not like Nvidia Container Toolkit. It will work initially then reset it for some odd reason. For now, it is advised to use an actually Linux machine to run using Docker. If you do not care about utilizing your GPU or don't even have a Nvidia GPU then disregard this.
|
||||
|
||||
> [!IMPORTANT]
|
||||
> Currently, it seems like wsl does not like Nvidia Container Toolkit. It will work initially then reset it for some odd reason. For now, it is advised to use an actually Linux machine to run using Docker. If you do not care about utilizing your GPU or don't even have a Nvidia GPU then disregard this.
|
||||
|
||||
## Nvidia Container Toolkit Setup
|
||||
### Installation with Apt
|
||||
@@ -41,15 +43,15 @@ sudo systemctl restart docker
|
||||
* [GitHub repository](https://github.com/NVIDIA/nvidia-container-toolkit?tab=readme-ov-file) for Nvidia Container Toolkit
|
||||
|
||||
## To Run (with Docker and Docker Compose)
|
||||
* With the inclusion of subnets in the `docker-compose.yml`, you will need to set the `SUBNET_ADDRESS`, `OLLAMA_IP`, `OLLAMA_PORT`, and `DISCORD_IP`. Here are some default values if you don't care:
|
||||
* With the inclusion of subnets in the `docker-compose.yml`, you will need to set the `SUBNET_ADDRESS`, `OLLAMA_IP`, `OLLAMA_PORT`, `REDIS_IP`, `REDIS_PORT`, and `DISCORD_IP`. Here are some default values if you don't care:
|
||||
* `SUBNET_ADDRESS = 172.18.0.0`
|
||||
* `OLLAMA_IP = 172.18.0.2`
|
||||
* `OLLAMA_PORT = 11434`
|
||||
* `DISCORD_IP = 172.18.0.3`
|
||||
* `SUBNET_ADDRESS = 172.18.0.0`
|
||||
* `REDIS_IP = 172.18.0.4`
|
||||
* `REDIS_PORT = 6379`
|
||||
* Don't understand any of this? watch a Networking video to understand subnetting.
|
||||
* You also need all environment variables shown in [`.env.sample`](../.env.sample)
|
||||
* You will need a model in the container for this to work properly, on Docker Desktop go to the `Containers` tab, select the `ollama` container, and select `Exec` to run as root on your container. Now, run `ollama pull [model name]` to get your model.
|
||||
* For Linux Servers, you need another shell to pull the model, or if you run `docker compose build && docker compose up -d`, then it will run in the background to keep your shell. Run `docker exec -it ollama bash` to get into the container and run the samme pull command above.
|
||||
* Otherwise, there is no need to install any npm packages for this, you just need to run `npm run start` to pull the containers and spin them up.
|
||||
* For cleaning up on Linux (or Windows), run the following commands:
|
||||
* `docker compose stop`
|
||||
|
||||
@@ -11,14 +11,17 @@
|
||||
* You can now interact with the model you just ran (it might take a second to startup).
|
||||
* Response time varies with processing power!
|
||||
|
||||
> [!NOTE]
|
||||
> You can now pull models directly from the Discord client using `/pull-model <model-name>` or `/switch-model <model-name>`. They must exist from your local model library or from the [Ollama Model Library](https://ollama.com/library)
|
||||
|
||||
## To Run Locally (without Docker)
|
||||
* Run `npm install` to install the npm packages.
|
||||
* Ensure that your [.env](../.env.sample) file's `OLLAMA_IP` is `127.0.0.1` to work properly.
|
||||
* You only need your `CLIENT_TOKEN`, `MODEL`, `CLIENT_UID`, `OLLAMA_IP`, `OLLAMA_PORT`.
|
||||
* You only need your `CLIENT_TOKEN`, `OLLAMA_IP`, `OLLAMA_PORT`.
|
||||
* The ollama ip and port should just use it's defaults by nature. If not, utilize `OLLAMA_IP = 127.0.0.1` and `OLLAMA_PORT = 11434`.
|
||||
* Now, you can run the bot by running `npm run client` which will build and run the decompiled typescript and run the setup for ollama.
|
||||
* **IMPORTANT**: This must be ran in the wsl/Linux instance to work properly! Using Command Prompt/Powershell/Git Bash/etc. will not work on Windows (at least in my experience).
|
||||
* Refer to the [resources](../README.md#resources) on what node version to use.
|
||||
* If you are using wsl, open up a separate terminal/shell to startup the ollama service. Again, if you are running an older ollama, you must run `ollama serve` in that shell.
|
||||
* If you are on an actual Linux machine/VM there is no need for another terminal (unless you have an older ollama version).
|
||||
* If you do not have a model, you will need to run `ollama pull [model name]` in a separate terminal to get it.
|
||||
* If you do not have a model, you **can optionally** run `ollama pull [model name]` in wsl prior to application start. You are not required as it can be pulled from the Discord client.
|
||||
2022
package-lock.json
generated
2022
package-lock.json
generated
File diff suppressed because it is too large
Load Diff
33
package.json
33
package.json
@@ -1,45 +1,48 @@
|
||||
{
|
||||
"name": "discord-ollama",
|
||||
"version": "0.6.0",
|
||||
"version": "0.8.5",
|
||||
"description": "Ollama Integration into discord",
|
||||
"main": "build/index.js",
|
||||
"exports": "./build/index.js",
|
||||
"scripts": {
|
||||
"test:run": "vitest run",
|
||||
"test:coverage": "vitest run --coverage",
|
||||
"tests": "vitest run",
|
||||
"coverage": "vitest run --coverage",
|
||||
"watch": "tsx watch src",
|
||||
"build": "tsc",
|
||||
"prod": "node .",
|
||||
"client": "npm run build && npm run prod",
|
||||
"clean": "docker compose down && docker rmi $(docker images | grep kevinthedang | tr -s ' ' | cut -d ' ' -f 3) && docker rmi $(docker images --filter \"dangling=true\" -q --no-trunc)",
|
||||
"start": "docker compose build --no-cache && docker compose up -d",
|
||||
"docker:clean": "docker rm -f discord && docker rm -f ollama && docker network prune -f && docker rmi $(docker images | grep kevinthedang | tr -s ' ' | cut -d ' ' -f 3) && docker rmi $(docker images --filter \"dangling=true\" -q --no-trunc)",
|
||||
"docker:clean": "docker rm -f discord && docker rm -f ollama && docker rm -f redis && docker network prune -f && docker rmi $(docker images | grep kevinthedang | tr -s ' ' | cut -d ' ' -f 3) && docker rmi $(docker images --filter \"dangling=true\" -q --no-trunc)",
|
||||
"docker:network": "docker network create --subnet=172.18.0.0/16 ollama-net",
|
||||
"docker:build": "docker build --no-cache -t kevinthedang/discord-ollama:$(node -p \"require('./package.json').version\") .",
|
||||
"docker:build-latest": "docker build --no-cache -t kevinthedang/discord-ollama:latest .",
|
||||
"docker:client": "docker run -d -v discord:/src/app --name discord --network ollama-net --ip 172.18.0.3 kevinthedang/discord-ollama:$(node -p \"require('./package.json').version\")",
|
||||
"docker:redis": "docker run -d -v redis:/root/.redis -p 6379:6379 --name redis --network ollama-net --ip 172.18.0.4 redis:latest",
|
||||
"docker:ollama": "docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama --network ollama-net --ip 172.18.0.2 ollama/ollama:latest",
|
||||
"docker:ollama-cpu": "docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama --network ollama-net --ip 172.18.0.2 ollama/ollama:latest",
|
||||
"docker:start": "docker network prune -f && npm run docker:network && npm run docker:build && npm run docker:client && npm run docker:ollama",
|
||||
"docker:start-cpu": "docker network prune -f && npm run docker:network && npm run docker:build && npm run docker:client && npm run docker:ollama-cpu"
|
||||
"docker:start": "docker network prune -f && npm run docker:network && npm run docker:build && npm run docker:redis && npm run docker:client && npm run docker:ollama",
|
||||
"docker:start-cpu": "docker network prune -f && npm run docker:network && npm run docker:build && npm run docker:redis && npm run docker:client && npm run docker:ollama-cpu"
|
||||
},
|
||||
"author": "Kevin Dang",
|
||||
"license": "ISC",
|
||||
"dependencies": {
|
||||
"discord.js": "^14.15.3",
|
||||
"dotenv": "^16.4.5",
|
||||
"ollama": "^0.5.8"
|
||||
"discord.js": "^14.19.3",
|
||||
"dotenv": "^16.5.0",
|
||||
"ollama": "^0.5.15",
|
||||
"redis": "^4.7.0"
|
||||
},
|
||||
"devDependencies": {
|
||||
"@types/node": "^22.5.1",
|
||||
"@vitest/coverage-v8": "^2.0.5",
|
||||
"@types/node": "^22.13.14",
|
||||
"@vitest/coverage-v8": "^3.0.9",
|
||||
"ts-node": "^10.9.2",
|
||||
"tsx": "^4.19.0",
|
||||
"typescript": "^5.5.4",
|
||||
"vitest": "^2.0.5"
|
||||
"tsx": "^4.19.3",
|
||||
"typescript": "^5.8.2",
|
||||
"vitest": "^3.0.4"
|
||||
},
|
||||
"type": "module",
|
||||
"engines": {
|
||||
"node": ">=16.0.0"
|
||||
"npm": ">=10.9.0",
|
||||
"node": ">=22.12.0"
|
||||
}
|
||||
}
|
||||
@@ -1,13 +1,11 @@
|
||||
import { Client, GatewayIntentBits } from 'discord.js'
|
||||
import { UserMessage, registerEvents } from './utils/events.js'
|
||||
import Events from './events/index.js'
|
||||
import { Ollama } from 'ollama'
|
||||
import { createClient } from 'redis'
|
||||
import { Queue } from './queues/queue.js'
|
||||
|
||||
// Import keys/tokens
|
||||
import { UserMessage, registerEvents } from './utils/index.js'
|
||||
import Events from './events/index.js'
|
||||
import Keys from './keys.js'
|
||||
|
||||
|
||||
// initialize the client with the following permissions when logging in
|
||||
const client = new Client({
|
||||
intents: [
|
||||
@@ -16,31 +14,42 @@ const client = new Client({
|
||||
GatewayIntentBits.GuildMessages,
|
||||
GatewayIntentBits.MessageContent
|
||||
]
|
||||
});
|
||||
})
|
||||
|
||||
// initialize connection to redis
|
||||
const redis = createClient({
|
||||
url: `redis://${Keys.redisHost}:${Keys.redisPort}`,
|
||||
})
|
||||
|
||||
// initialize connection to ollama container
|
||||
const ollama = new Ollama({
|
||||
export const ollama = new Ollama({
|
||||
host: `http://${Keys.ipAddress}:${Keys.portAddress}`,
|
||||
})
|
||||
|
||||
// Create Queue managed by Events
|
||||
const messageHistory: Queue<UserMessage> = new Queue<UserMessage>
|
||||
|
||||
/**
|
||||
* register events for bot to listen to in discord
|
||||
* @param messageHistory message history for the llm
|
||||
* @param Events events to register
|
||||
* @param client the bot reference
|
||||
* @param Keys tokens from .env files
|
||||
*/
|
||||
registerEvents(client, Events, messageHistory, Keys, ollama)
|
||||
// register all events
|
||||
registerEvents(client, Events, messageHistory, ollama, Keys.defaultModel)
|
||||
|
||||
// Try to connect to redis
|
||||
await redis.connect()
|
||||
.then(response => {
|
||||
console.log('[Redis] Successfully Connected')
|
||||
})
|
||||
.catch(error => {
|
||||
console.error('[Redis] Connection Error. See error below:\n', error)
|
||||
console.warn('[Redis] Failed to connect to Redis Database, using local system')
|
||||
// TODO: create boolean flag that will probably be used in messageCreate.ts if redis database is down
|
||||
// When implementing this boolean flag, move connection to database BEFORE the registerEvents method
|
||||
})
|
||||
|
||||
// Try to log in the client
|
||||
await client.login(Keys.clientToken)
|
||||
.catch((error) => {
|
||||
console.error('[Login Error]', error)
|
||||
process.exit(1)
|
||||
})
|
||||
.catch((error) => {
|
||||
console.error('[Login Error]', error)
|
||||
process.exit(1)
|
||||
})
|
||||
|
||||
// queue up bots name
|
||||
messageHistory.enqueue({
|
||||
|
||||
@@ -1,16 +1,15 @@
|
||||
import { ChannelType, Client, CommandInteraction, ApplicationCommandOptionType } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { openConfig } from '../utils/index.js'
|
||||
import { Client, CommandInteraction, ApplicationCommandOptionType, MessageFlags } from 'discord.js'
|
||||
import { openConfig, SlashCommand, UserCommand } from '../utils/index.js'
|
||||
|
||||
export const Capacity: SlashCommand = {
|
||||
name: 'modify-capacity',
|
||||
description: 'number of messages bot will hold for context.',
|
||||
description: 'maximum amount messages bot will hold for context.',
|
||||
|
||||
// set available user options to pass to the command
|
||||
options: [
|
||||
{
|
||||
name: 'context-capacity',
|
||||
description: 'a number to set capacity',
|
||||
description: 'number of allowed messages to remember',
|
||||
type: ApplicationCommandOptionType.Number,
|
||||
required: true
|
||||
}
|
||||
@@ -20,14 +19,16 @@ export const Capacity: SlashCommand = {
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// fetch channel and message
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || channel.type !== (ChannelType.PrivateThread && ChannelType.PublicThread && ChannelType.GuildText)) return
|
||||
if (!channel || !UserCommand.includes(channel.type)) return
|
||||
|
||||
// set state of bot chat features
|
||||
openConfig(`${interaction.user.username}-config.json`, interaction.commandName, interaction.options.get('context-capacity')?.value)
|
||||
openConfig(`${interaction.user.username}-config.json`, interaction.commandName,
|
||||
interaction.options.get('context-capacity')?.value
|
||||
)
|
||||
|
||||
interaction.reply({
|
||||
content: `Message History Capacity has been set to \`${interaction.options.get('context-capacity')?.value}\``,
|
||||
ephemeral: true
|
||||
content: `Max message history is now set to \`${interaction.options.get('context-capacity')?.value}\``,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -1,34 +1,35 @@
|
||||
import { ChannelType, Client, CommandInteraction, TextChannel } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { clearChannelInfo } from '../utils/index.js'
|
||||
import { Channel, Client, CommandInteraction, MessageFlags, TextChannel } from 'discord.js'
|
||||
import { clearChannelInfo, SlashCommand, UserCommand } from '../utils/index.js'
|
||||
|
||||
export const ClearUserChannelHistory: SlashCommand = {
|
||||
name: 'clear-user-channel-history',
|
||||
description: 'clears history for user running this command in current channel',
|
||||
description: 'clears history for user in the current channel',
|
||||
|
||||
// Clear channel history for intended user
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// fetch current channel
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
const channel: Channel | null = await client.channels.fetch(interaction.channelId)
|
||||
|
||||
// if not an existing channel or a GuildText, fail command
|
||||
if (!channel || channel.type !== ChannelType.GuildText) return
|
||||
if (!channel || !UserCommand.includes(channel.type)) return
|
||||
|
||||
// clear channel info for user
|
||||
const successfulWipe = await clearChannelInfo(interaction.channelId,
|
||||
const successfulWipe = await clearChannelInfo(
|
||||
interaction.channelId,
|
||||
interaction.channel as TextChannel,
|
||||
interaction.user.username)
|
||||
interaction.user.username
|
||||
)
|
||||
|
||||
// check result of clearing history
|
||||
if (successfulWipe)
|
||||
interaction.reply({
|
||||
content: `Channel history in **${channel.name}** cleared for **${interaction.user.username}**.`,
|
||||
ephemeral: true
|
||||
content: `History cleared in **this channel** cleared for **${interaction.user.username}**.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
else
|
||||
interaction.reply({
|
||||
content: `Channel history could not be found for **${interaction.user.username}** in **${channel.name}**.\n\nPlease chat with **${client.user?.username}** to start a chat history.`,
|
||||
ephemeral: true
|
||||
content: `History was not be found for **${interaction.user.username}** in **this channel**.\n\nPlease chat with **${client.user?.username}** to start a chat history.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
}
|
||||
}
|
||||
72
src/commands/deleteModel.ts
Normal file
72
src/commands/deleteModel.ts
Normal file
@@ -0,0 +1,72 @@
|
||||
import { ApplicationCommandOptionType, Client, CommandInteraction, MessageFlags } from 'discord.js'
|
||||
import { UserCommand, SlashCommand } from '../utils/index.js'
|
||||
import { ollama } from '../client.js'
|
||||
import { ModelResponse } from 'ollama'
|
||||
|
||||
export const DeleteModel: SlashCommand = {
|
||||
name: 'delete-model',
|
||||
description: 'deletes a model from the local list of models. Administrator Only.',
|
||||
|
||||
// set available user options to pass to the command
|
||||
options: [
|
||||
{
|
||||
name: 'model-name',
|
||||
description: 'the name of the model to delete',
|
||||
type: ApplicationCommandOptionType.String,
|
||||
required: true
|
||||
}
|
||||
],
|
||||
|
||||
// Delete Model locally stored
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// defer reply to avoid timeout
|
||||
await interaction.deferReply()
|
||||
const modelInput: string = interaction.options.get('model-name')!!.value as string
|
||||
|
||||
// fetch channel and message
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || !UserCommand.includes(channel.type)) return
|
||||
|
||||
// Admin Command
|
||||
if (!interaction.memberPermissions?.has('Administrator')) {
|
||||
interaction.reply({
|
||||
content: `${interaction.commandName} is an admin command.\n\nPlease contact a server admin to pull the model you want.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
// check if model exists
|
||||
const modelExists = await ollama.list()
|
||||
.then(response => response.models.some((model: ModelResponse) => model.name.startsWith(modelInput)))
|
||||
.catch(error => {
|
||||
console.error(`[Command: delete-model] Failed to connect with Ollama service. Error: ${error.message}`)
|
||||
})
|
||||
|
||||
// Validate for any issue or if service is running
|
||||
if (!modelExists) {
|
||||
interaction.editReply({
|
||||
content: `The Ollama service is not running. Please turn on/download the [service](https://ollama.com/).`
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
|
||||
try {
|
||||
// call ollama to delete model
|
||||
if (modelExists) {
|
||||
await ollama.delete({ model: modelInput })
|
||||
interaction.editReply({
|
||||
content: `**${modelInput}** was removed from the the library.`
|
||||
})
|
||||
} else
|
||||
throw new Error()
|
||||
} catch (error) {
|
||||
// could not delete the model
|
||||
interaction.reply({
|
||||
content: `Could not delete the **${modelInput}** model. It probably doesn't exist or you spelled it incorrectly.\n\nPlease try again if this is a mistake.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1,10 +1,9 @@
|
||||
import { ChannelType, Client, CommandInteraction, ApplicationCommandOptionType } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { openConfig } from '../utils/index.js'
|
||||
import { Client, CommandInteraction, ApplicationCommandOptionType, MessageFlags } from 'discord.js'
|
||||
import { AdminCommand, openConfig, SlashCommand } from '../utils/index.js'
|
||||
|
||||
export const Disable: SlashCommand = {
|
||||
name: 'toggle-chat',
|
||||
description: 'toggle all chat features, Adminstrator Only.',
|
||||
description: 'toggle all chat features. Adminstrator Only.',
|
||||
|
||||
// set available user options to pass to the command
|
||||
options: [
|
||||
@@ -20,23 +19,25 @@ export const Disable: SlashCommand = {
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// fetch channel and message
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || channel.type !== ChannelType.GuildText) return
|
||||
if (!channel || !AdminCommand.includes(channel.type)) return
|
||||
|
||||
// check if runner is an admin
|
||||
if (!interaction.memberPermissions?.has('Administrator')) {
|
||||
interaction.reply({
|
||||
content: `${interaction.commandName} is an Administrator Command.\n\nYou, ${interaction.member?.user.username}, are not an Administrator in this server.\nPlease contact an admin to use this command.`,
|
||||
ephemeral: true
|
||||
content: `${interaction.commandName} is an admin command.\n\nPlease contact an admin to use this command for you.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
// set state of bot chat features
|
||||
openConfig(`${interaction.guildId}-config.json`, interaction.commandName, interaction.options.get('enabled')?.value)
|
||||
openConfig(`${interaction.guildId}-config.json`, interaction.commandName,
|
||||
interaction.options.get('enabled')?.value
|
||||
)
|
||||
|
||||
interaction.reply({
|
||||
content: `Chat features has been \`${interaction.options.get('enabled')?.value ? "enabled" : "disabled" }\``,
|
||||
ephemeral: true
|
||||
content: `${client.user?.username} is now **${interaction.options.get('enabled')?.value ? "enabled" : "disabled"}**.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -1,20 +1,24 @@
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { ThreadCreate } from './threadCreate.js'
|
||||
import { MessageStyle } from './messageStyle.js'
|
||||
import { MessageStream } from './messageStream.js'
|
||||
import { Disable } from './disable.js'
|
||||
import { Shutoff } from './shutoff.js'
|
||||
import { Capacity } from './capacity.js'
|
||||
import { PrivateThreadCreate } from './threadPrivateCreate.js'
|
||||
import { ClearUserChannelHistory } from './cleanUserChannelHistory.js'
|
||||
import { PullModel } from './pullModel.js'
|
||||
import { SwitchModel } from './switchModel.js'
|
||||
import { DeleteModel } from './deleteModel.js'
|
||||
|
||||
export default [
|
||||
ThreadCreate,
|
||||
PrivateThreadCreate,
|
||||
MessageStyle,
|
||||
MessageStream,
|
||||
Disable,
|
||||
Shutoff,
|
||||
Capacity,
|
||||
ClearUserChannelHistory
|
||||
ClearUserChannelHistory,
|
||||
PullModel,
|
||||
SwitchModel,
|
||||
DeleteModel
|
||||
] as SlashCommand[]
|
||||
@@ -1,16 +1,15 @@
|
||||
import { ApplicationCommandOptionType, ChannelType, Client, CommandInteraction } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { openConfig } from '../utils/index.js'
|
||||
import { ApplicationCommandOptionType, Client, CommandInteraction, MessageFlags } from 'discord.js'
|
||||
import { openConfig, SlashCommand, UserCommand } from '../utils/index.js'
|
||||
|
||||
export const MessageStream: SlashCommand = {
|
||||
name: 'message-stream',
|
||||
description: 'change preference on message streaming from ollama. WARNING: can be very slow.',
|
||||
description: 'change preference on message streaming from ollama. WARNING: can be very slow due to Discord limits.',
|
||||
|
||||
// user option(s) for setting stream
|
||||
options: [
|
||||
{
|
||||
name: 'stream',
|
||||
description: 'enable or disable stream preference',
|
||||
description: 'enable or disable message streaming',
|
||||
type: ApplicationCommandOptionType.Boolean,
|
||||
required: true
|
||||
}
|
||||
@@ -20,14 +19,16 @@ export const MessageStream: SlashCommand = {
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// verify channel
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || channel.type !== (ChannelType.PrivateThread && ChannelType.PublicThread && ChannelType.GuildText)) return
|
||||
if (!channel || !UserCommand.includes(channel.type)) return
|
||||
|
||||
// save value to json and write to it
|
||||
openConfig(`${interaction.user.username}-config.json`, interaction.commandName, interaction.options.get('stream')?.value)
|
||||
openConfig(`${interaction.user.username}-config.json`, interaction.commandName,
|
||||
interaction.options.get('stream')?.value
|
||||
)
|
||||
|
||||
interaction.reply({
|
||||
content: `Message streaming preferences set to: \`${interaction.options.get('stream')?.value}\``,
|
||||
ephemeral: true
|
||||
content: `Message streaming is now set to: \`${interaction.options.get('stream')?.value}\``,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -1,33 +0,0 @@
|
||||
import { ChannelType, Client, CommandInteraction, ApplicationCommandOptionType } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { openConfig } from '../utils/index.js'
|
||||
|
||||
export const MessageStyle: SlashCommand = {
|
||||
name: 'message-style',
|
||||
description: 'sets the message style to embed or normal',
|
||||
|
||||
// set available user options to pass to the command
|
||||
options: [
|
||||
{
|
||||
name: 'embed',
|
||||
description: 'toggle embedded or normal message',
|
||||
type: ApplicationCommandOptionType.Boolean,
|
||||
required: true
|
||||
}
|
||||
],
|
||||
|
||||
// Query for message information and set the style
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// fetch channel and message
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || channel.type !== (ChannelType.PrivateThread && ChannelType.PublicThread && ChannelType.GuildText)) return
|
||||
|
||||
// set the message style
|
||||
openConfig(`${interaction.user.username}-config.json`, interaction.commandName, interaction.options.get('embed')?.value)
|
||||
|
||||
interaction.reply({
|
||||
content: `Message style preferences for embed set to: \`${interaction.options.get('embed')?.value}\``,
|
||||
ephemeral: true
|
||||
})
|
||||
}
|
||||
}
|
||||
81
src/commands/pullModel.ts
Normal file
81
src/commands/pullModel.ts
Normal file
@@ -0,0 +1,81 @@
|
||||
import { ApplicationCommandOptionType, Client, CommandInteraction, MessageFlags } from "discord.js"
|
||||
import { ollama } from "../client.js"
|
||||
import { ModelResponse } from "ollama"
|
||||
import { UserCommand, SlashCommand } from "../utils/index.js"
|
||||
|
||||
export const PullModel: SlashCommand = {
|
||||
name: 'pull-model',
|
||||
description: 'pulls a model from the ollama model library. Administrator Only.',
|
||||
|
||||
// set available user options to pass to the command
|
||||
options: [
|
||||
{
|
||||
name: 'model-to-pull',
|
||||
description: 'the name of the model to pull',
|
||||
type: ApplicationCommandOptionType.String,
|
||||
required: true
|
||||
}
|
||||
],
|
||||
|
||||
// Pull for model from Ollama library
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// defer reply to avoid timeout
|
||||
await interaction.deferReply()
|
||||
const modelInput: string = interaction.options.get('model-to-pull')!!.value as string
|
||||
|
||||
// fetch channel and message
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || !UserCommand.includes(channel.type)) return
|
||||
|
||||
// Admin Command
|
||||
if (!interaction.memberPermissions?.has('Administrator')) {
|
||||
interaction.reply({
|
||||
content: `${interaction.commandName} is an admin command.\n\nPlease contact a server admin to pull the model you want.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
// check if model was already pulled, if the ollama service isn't running throw error
|
||||
const modelExists = await ollama.list()
|
||||
.then(response => response.models.some((model: ModelResponse) => model.name.startsWith(modelInput)))
|
||||
.catch(error => {
|
||||
console.error(`[Command: pull-model] Failed to connect with Ollama service. Error: ${error.message}`)
|
||||
})
|
||||
|
||||
// Validate for any issue or if service is running
|
||||
if (!modelExists) {
|
||||
interaction.editReply({
|
||||
content: `The Ollama service is not running. Please turn on/download the [service](https://ollama.com/).`
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
|
||||
try {
|
||||
// call ollama to pull desired model
|
||||
if (!modelExists) {
|
||||
interaction.editReply({
|
||||
content: `**${modelInput}** could not be found. Please wait patiently as I try to retrieve it...`
|
||||
})
|
||||
await ollama.pull({ model: modelInput })
|
||||
}
|
||||
} catch (error) {
|
||||
// could not resolve pull or model unfound
|
||||
interaction.editReply({
|
||||
content: `Could not retrieve the **${modelInput}** model. You can find models at [Ollama Model Library](https://ollama.com/library).\n\nPlease check the model library and try again.`
|
||||
})
|
||||
return
|
||||
}
|
||||
|
||||
// successful interaction
|
||||
if (modelExists)
|
||||
interaction.editReply({
|
||||
content: `**${modelInput}** is already available.`
|
||||
})
|
||||
else
|
||||
interaction.editReply({
|
||||
content: `Successfully added **${modelInput}**.`
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -1,50 +1,36 @@
|
||||
import { ChannelType, Client, CommandInteraction, ApplicationCommandOptionType } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { Client, CommandInteraction, MessageFlags } from 'discord.js'
|
||||
import { AdminCommand, SlashCommand } from '../utils/index.js'
|
||||
|
||||
export const Shutoff: SlashCommand = {
|
||||
name: 'shutoff',
|
||||
description: 'shutdown the bot. You will need to manually bring it online again. Administrator Only.',
|
||||
|
||||
// set available user options to pass to the command
|
||||
options: [
|
||||
{
|
||||
name: 'are-you-sure',
|
||||
description: 'true = yes, false = I\'m scared',
|
||||
type: ApplicationCommandOptionType.Boolean,
|
||||
required: true
|
||||
}
|
||||
],
|
||||
|
||||
// Query for message information and set the style
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// fetch channel and message
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || channel.type !== ChannelType.GuildText) return
|
||||
if (!channel || !AdminCommand.includes(channel.type)) return
|
||||
|
||||
// log this, this will probably be improtant for logging who did this
|
||||
console.log(`User -> ${interaction.user.tag} attempting to shutdown ${client.user!!.tag}`)
|
||||
console.log(`[Command: shutoff] User ${interaction.user.tag} attempting to shutdown ${client.user!!.tag}`)
|
||||
|
||||
// check if admin or false on shutdown
|
||||
if (!interaction.memberPermissions?.has('Administrator')) {
|
||||
interaction.reply({
|
||||
content: `**Shutdown Aborted:**\n\n${interaction.user.tag}, You do not have permission to shutoff **${client.user?.tag}**.`,
|
||||
ephemeral: true
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
return // stop from shutting down
|
||||
} else if (!interaction.options.get('are-you-sure')?.value) {
|
||||
interaction.reply({
|
||||
content: `**Shutdown Aborted:**\n\n${interaction.user.tag}, You didn't want to shutoff **${client.user?.tag}**.`,
|
||||
ephemeral: true
|
||||
})
|
||||
return // chickened out
|
||||
}
|
||||
|
||||
// Shutoff cleared, do it
|
||||
interaction.reply({
|
||||
content: `${client.user?.tag} is ${interaction.options.get('are-you-sure')?.value ? "shutting down now." : "not shutting down." }`,
|
||||
ephemeral: true
|
||||
content: `${client.user?.tag} is shutting down.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
|
||||
console.log(`[Command: shutoff] ${client.user?.tag} is shutting down.`)
|
||||
|
||||
// clean up client instance and stop
|
||||
client.destroy()
|
||||
}
|
||||
|
||||
73
src/commands/switchModel.ts
Normal file
73
src/commands/switchModel.ts
Normal file
@@ -0,0 +1,73 @@
|
||||
import { ApplicationCommandOptionType, Client, CommandInteraction } from "discord.js"
|
||||
import { ollama } from "../client.js"
|
||||
import { ModelResponse } from "ollama"
|
||||
import { openConfig, UserCommand, SlashCommand } from "../utils/index.js"
|
||||
|
||||
export const SwitchModel: SlashCommand = {
|
||||
name: 'switch-model',
|
||||
description: 'switches current model to use.',
|
||||
|
||||
// set available user options to pass to the command
|
||||
options: [
|
||||
{
|
||||
name: 'model-to-use',
|
||||
description: 'the name of the model to use',
|
||||
type: ApplicationCommandOptionType.String,
|
||||
required: true
|
||||
}
|
||||
],
|
||||
|
||||
// Switch user preferred model if available in local library
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
await interaction.deferReply()
|
||||
|
||||
const modelInput: string = interaction.options.get('model-to-use')!!.value as string
|
||||
|
||||
// fetch channel and message
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || !UserCommand.includes(channel.type)) return
|
||||
|
||||
try {
|
||||
// Phase 1: Switch to the model
|
||||
let switchSuccess = false
|
||||
await ollama.list()
|
||||
.then(response => {
|
||||
for (const model in response.models) {
|
||||
const currentModel: ModelResponse = response.models[model]
|
||||
if (currentModel.name.startsWith(modelInput)) {
|
||||
openConfig(`${interaction.user.username}-config.json`, interaction.commandName, modelInput)
|
||||
|
||||
// successful switch
|
||||
interaction.editReply({
|
||||
content: `Successfully switched to **${modelInput}** as the preferred model for ${interaction.user.username}.`
|
||||
})
|
||||
switchSuccess = true
|
||||
}
|
||||
}
|
||||
})
|
||||
.catch(error => {
|
||||
console.error(`[Command: switch-model] Failed to connect with Ollama service. Error: ${error.message}`)
|
||||
})
|
||||
// todo: problem can be here if async messes up
|
||||
if (switchSuccess) {
|
||||
// set model now that it exists
|
||||
openConfig(`${interaction.user.username}-config.json`, interaction.commandName, modelInput)
|
||||
return
|
||||
}
|
||||
|
||||
// Phase 2: Notify user of failure to find model.
|
||||
interaction.editReply({
|
||||
content: `Could not find **${modelInput}** in local model library.\n\nPlease contact an server admin for access to this model.`
|
||||
})
|
||||
} catch (error: any) {
|
||||
// could not resolve user model switch
|
||||
if (error.message.includes("fetch failed") as string)
|
||||
error.message = "The Ollama service is not running. Please turn on/download the [service](https://ollama.com/)."
|
||||
|
||||
interaction.editReply({
|
||||
content: `Unable to switch user preferred model to **${modelInput}**.\n\n${error.message}`
|
||||
})
|
||||
return
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -1,6 +1,5 @@
|
||||
import { ChannelType, Client, CommandInteraction, TextChannel, ThreadChannel } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { openChannelInfo } from '../utils/index.js'
|
||||
import { ChannelType, Client, CommandInteraction, MessageFlags, TextChannel, ThreadChannel } from 'discord.js'
|
||||
import { AdminCommand, openChannelInfo, SlashCommand } from '../utils/index.js'
|
||||
|
||||
export const ThreadCreate: SlashCommand = {
|
||||
name: 'thread',
|
||||
@@ -10,7 +9,7 @@ export const ThreadCreate: SlashCommand = {
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// fetch the channel
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || channel.type !== ChannelType.GuildText) return
|
||||
if (!channel || !AdminCommand.includes(channel.type)) return
|
||||
|
||||
const thread = await (channel as TextChannel).threads.create({
|
||||
name: `${client.user?.username}-support-${Date.now()}`,
|
||||
@@ -19,17 +18,15 @@ export const ThreadCreate: SlashCommand = {
|
||||
})
|
||||
|
||||
// Send a message in the thread
|
||||
thread.send(`Hello ${interaction.user} and others! \n\nIt's nice to meet you. Please talk to me by typing **@${client.user?.username}** with your prompt.`)
|
||||
thread.send(`Hello ${interaction.user} and others! \n\nIt's nice to meet you. Please talk to me by typing **@${client.user?.username}** with your message.`)
|
||||
|
||||
// handle storing this chat channel
|
||||
openChannelInfo(thread.id,
|
||||
thread as ThreadChannel,
|
||||
interaction.user.tag)
|
||||
openChannelInfo(thread.id, thread as ThreadChannel, interaction.user.tag)
|
||||
|
||||
// user only reply
|
||||
return interaction.reply({
|
||||
content: `I can help you in thread **${thread.id}** below.`,
|
||||
ephemeral: true
|
||||
content: `I can help you in <#${thread.id}> below.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -1,6 +1,5 @@
|
||||
import { ChannelType, Client, CommandInteraction, TextChannel, ThreadChannel } from 'discord.js'
|
||||
import { SlashCommand } from '../utils/commands.js'
|
||||
import { openChannelInfo } from '../utils/index.js'
|
||||
import { ChannelType, Client, CommandInteraction, MessageFlags, TextChannel, ThreadChannel } from 'discord.js'
|
||||
import { AdminCommand, openChannelInfo, SlashCommand } from '../utils/index.js'
|
||||
|
||||
export const PrivateThreadCreate: SlashCommand = {
|
||||
name: 'private-thread',
|
||||
@@ -10,7 +9,7 @@ export const PrivateThreadCreate: SlashCommand = {
|
||||
run: async (client: Client, interaction: CommandInteraction) => {
|
||||
// fetch the channel
|
||||
const channel = await client.channels.fetch(interaction.channelId)
|
||||
if (!channel || channel.type !== ChannelType.GuildText) return
|
||||
if (!channel || !AdminCommand.includes(channel.type)) return
|
||||
|
||||
const thread = await (channel as TextChannel).threads.create({
|
||||
name: `${client.user?.username}-private-support-${Date.now()}`,
|
||||
@@ -23,15 +22,12 @@ export const PrivateThreadCreate: SlashCommand = {
|
||||
|
||||
// handle storing this chat channel
|
||||
// store: thread.id, thread.name
|
||||
openChannelInfo(thread.id,
|
||||
thread as ThreadChannel,
|
||||
interaction.user.tag
|
||||
)
|
||||
openChannelInfo(thread.id, thread as ThreadChannel, interaction.user.tag)
|
||||
|
||||
// user only reply
|
||||
return interaction.reply({
|
||||
content: `I can help you in thread **${thread.id}**. Please refer to the private channel below this one.`,
|
||||
ephemeral: true
|
||||
content: `I can help you in <#${thread.id}>.`,
|
||||
flags: MessageFlags.Ephemeral
|
||||
})
|
||||
}
|
||||
}
|
||||
@@ -1,7 +1,9 @@
|
||||
import { embedMessage, event, Events, normalMessage, UserMessage } from '../utils/index.js'
|
||||
import { getChannelInfo, getServerConfig, getUserConfig, openChannelInfo, openConfig, UserConfig, getAttachmentData } from '../utils/index.js'
|
||||
import { clean } from '../utils/mentionClean.js'
|
||||
import { TextChannel } from 'discord.js'
|
||||
import { event, Events, normalMessage, UserMessage, clean } from '../utils/index.js'
|
||||
import {
|
||||
getChannelInfo, getServerConfig, getUserConfig, openChannelInfo,
|
||||
openConfig, UserConfig, getAttachmentData, getTextFileAttachmentData
|
||||
} from '../utils/index.js'
|
||||
|
||||
/**
|
||||
* Max Message length for free users is 2000 characters (bot or not).
|
||||
@@ -9,65 +11,100 @@ import { TextChannel } from 'discord.js'
|
||||
*
|
||||
* @param message the message received from the channel
|
||||
*/
|
||||
export default event(Events.MessageCreate, async ({ log, msgHist, tokens, ollama }, message) => {
|
||||
log(`Message \"${clean(message.content)}\" from ${message.author.tag} in channel/thread ${message.channelId}.`)
|
||||
export default event(Events.MessageCreate, async ({ log, msgHist, ollama, client, defaultModel }, message) => {
|
||||
const clientId = client.user!!.id
|
||||
let cleanedMessage = clean(message.content, clientId)
|
||||
log(`Message \"${cleanedMessage}\" from ${message.author.tag} in channel/thread ${message.channelId}.`)
|
||||
|
||||
// Do not respond if bot talks in the chat
|
||||
if (message.author.username === message.client.user.username) return
|
||||
|
||||
// Only respond if message mentions the bot
|
||||
if (!message.mentions.has(tokens.clientUid)) return
|
||||
if (!message.mentions.has(clientId)) return
|
||||
|
||||
// default stream to false
|
||||
let shouldStream = false
|
||||
|
||||
// Params for Preferences Fetching
|
||||
const maxRetries = 3
|
||||
const delay = 1000 // in millisecons
|
||||
|
||||
try {
|
||||
// Retrieve Server/Guild Preferences
|
||||
await new Promise((resolve, reject) => {
|
||||
getServerConfig(`${message.guildId}-config.json`, (config) => {
|
||||
// check if config.json exists
|
||||
if (config === undefined) {
|
||||
// Allowing chat options to be available
|
||||
openConfig(`${message.guildId}-config.json`, 'toggle-chat', true)
|
||||
reject(new Error('No Server Preferences is set up.\n\nCreating default server preferences file...\nPlease try chatting again.'))
|
||||
return
|
||||
}
|
||||
let attempt = 0
|
||||
while (attempt < maxRetries) {
|
||||
try {
|
||||
await new Promise((resolve, reject) => {
|
||||
getServerConfig(`${message.guildId}-config.json`, (config) => {
|
||||
// check if config.json exists
|
||||
if (config === undefined) {
|
||||
// Allowing chat options to be available
|
||||
openConfig(`${message.guildId}-config.json`, 'toggle-chat', true)
|
||||
reject(new Error('Failed to locate or create Server Preferences\n\nPlease try chatting again...'))
|
||||
}
|
||||
|
||||
// check if chat is disabled
|
||||
if (!config.options['toggle-chat']) {
|
||||
reject(new Error('Admin(s) have disabled chat features.\n\n Please contact your server\'s admin(s).'))
|
||||
return
|
||||
}
|
||||
// check if chat is disabled
|
||||
else if (!config.options['toggle-chat'])
|
||||
reject(new Error('Admin(s) have disabled chat features.\n\n Please contact your server\'s admin(s).'))
|
||||
else
|
||||
resolve(config)
|
||||
})
|
||||
})
|
||||
break // successful
|
||||
} catch (error) {
|
||||
++attempt
|
||||
if (attempt < maxRetries) {
|
||||
log(`Attempt ${attempt} failed for Server Preferences. Retrying in ${delay}ms...`)
|
||||
await new Promise(ret => setTimeout(ret, delay))
|
||||
} else
|
||||
throw new Error(`Could not retrieve Server Preferences, please try chatting again...`)
|
||||
}
|
||||
}
|
||||
|
||||
resolve(config)
|
||||
})
|
||||
})
|
||||
// Reset attempts for User preferences
|
||||
attempt = 0
|
||||
let userConfig: UserConfig | undefined
|
||||
|
||||
// Retrieve User Preferences
|
||||
const userConfig: UserConfig = await new Promise((resolve, reject) => {
|
||||
getUserConfig(`${message.author.username}-config.json`, (config) => {
|
||||
if (config === undefined) {
|
||||
openConfig(`${message.author.username}-config.json`, 'message-style', false)
|
||||
reject(new Error('No User Preferences is set up.\n\nCreating preferences file with \`message-style\` set as \`false\` for regular messages.\nPlease try chatting again.'))
|
||||
return
|
||||
}
|
||||
while (attempt < maxRetries) {
|
||||
try {
|
||||
// Retrieve User Preferences
|
||||
userConfig = await new Promise((resolve, reject) => {
|
||||
getUserConfig(`${message.author.username}-config.json`, (config) => {
|
||||
if (config === undefined) {
|
||||
openConfig(`${message.author.username}-config.json`, 'switch-model', defaultModel)
|
||||
reject(new Error(`No User Preferences is set up.\n\nCreating new preferences file for ${message.author.username}\nPlease try chatting again.`))
|
||||
return
|
||||
}
|
||||
|
||||
// check if there is a set capacity in config
|
||||
if (typeof config.options['modify-capacity'] !== 'number')
|
||||
log(`Capacity is undefined, using default capacity of ${msgHist.capacity}.`)
|
||||
else if (config.options['modify-capacity'] === msgHist.capacity)
|
||||
log(`Capacity matches config as ${msgHist.capacity}, no changes made.`)
|
||||
else {
|
||||
log(`New Capacity found. Setting Context Capacity to ${config.options['modify-capacity']}.`)
|
||||
msgHist.capacity = config.options['modify-capacity']
|
||||
}
|
||||
// check if there is a set capacity in config
|
||||
else if (typeof config.options['modify-capacity'] !== 'number')
|
||||
log(`Capacity is undefined, using default capacity of ${msgHist.capacity}.`)
|
||||
else if (config.options['modify-capacity'] === msgHist.capacity)
|
||||
log(`Capacity matches config as ${msgHist.capacity}, no changes made.`)
|
||||
else {
|
||||
log(`New Capacity found. Setting Context Capacity to ${config.options['modify-capacity']}.`)
|
||||
msgHist.capacity = config.options['modify-capacity']
|
||||
}
|
||||
|
||||
// set stream state
|
||||
shouldStream = config.options['message-stream'] as boolean || false
|
||||
// set stream state
|
||||
shouldStream = config.options['message-stream'] as boolean || false
|
||||
|
||||
resolve(config)
|
||||
})
|
||||
})
|
||||
if (typeof config.options['switch-model'] !== 'string')
|
||||
reject(new Error(`No Model was set. Please set a model by running \`/switch-model <model of choice>\`.\n\nIf you do not have any models. Run \`/pull-model <model name>\`.`))
|
||||
|
||||
resolve(config)
|
||||
})
|
||||
})
|
||||
break // successful
|
||||
} catch (error) {
|
||||
++attempt
|
||||
if (attempt < maxRetries) {
|
||||
log(`Attempt ${attempt} failed for User Preferences. Retrying in ${delay}ms...`)
|
||||
await new Promise(ret => setTimeout(ret, delay))
|
||||
} else
|
||||
throw new Error(`Could not retrieve User Preferences, please try chatting again...`)
|
||||
}
|
||||
}
|
||||
|
||||
// need new check for "open/active" threads/channels here!
|
||||
let chatMessages: UserMessage[] = await new Promise((resolve) => {
|
||||
@@ -99,11 +136,19 @@ export default event(Events.MessageCreate, async ({ log, msgHist, tokens, ollama
|
||||
})
|
||||
}
|
||||
|
||||
// response string for ollama to put its response
|
||||
let response: string
|
||||
if (!userConfig)
|
||||
throw new Error(`Failed to initialize User Preference for **${message.author.username}**.\n\nIt's likely you do not have a model set. Please use the \`switch-model\` command to do that.`)
|
||||
|
||||
// get message attachment if exists
|
||||
const messageAttachment: string[] = await getAttachmentData(message.attachments.first())
|
||||
const attachment = message.attachments.first()
|
||||
let messageAttachment: string[] = []
|
||||
|
||||
if (attachment && attachment.name?.endsWith(".txt"))
|
||||
cleanedMessage += await getTextFileAttachmentData(attachment)
|
||||
else if (attachment)
|
||||
messageAttachment = await getAttachmentData(attachment)
|
||||
|
||||
const model: string = userConfig.options['switch-model']
|
||||
|
||||
// set up new queue
|
||||
msgHist.setQueue(chatMessages)
|
||||
@@ -114,15 +159,12 @@ export default event(Events.MessageCreate, async ({ log, msgHist, tokens, ollama
|
||||
// push user response before ollama query
|
||||
msgHist.enqueue({
|
||||
role: 'user',
|
||||
content: clean(message.content),
|
||||
content: cleanedMessage,
|
||||
images: messageAttachment || []
|
||||
})
|
||||
|
||||
// undefined or false, use normal, otherwise use embed
|
||||
if (userConfig.options['message-style'])
|
||||
response = await embedMessage(message, ollama, tokens, msgHist, shouldStream)
|
||||
else
|
||||
response = await normalMessage(message, ollama, tokens, msgHist, shouldStream)
|
||||
// response string for ollama to put its response
|
||||
const response: string = await normalMessage(message, ollama, model, msgHist, shouldStream)
|
||||
|
||||
// If something bad happened, remove user query and stop
|
||||
if (response == undefined) { msgHist.pop(); return }
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
import { event, Events, registerCommands } from '../utils/index.js'
|
||||
import { ActivityType } from 'discord.js'
|
||||
import { event, Events, registerCommands } from '../utils/index.js'
|
||||
import commands from '../commands/index.js'
|
||||
|
||||
// Log when the bot successfully logs in and export it
|
||||
|
||||
@@ -21,7 +21,7 @@ export default event(Events.ThreadDelete, async ({ log }, thread: ThreadChannel)
|
||||
// filter files by thread id being deleted
|
||||
const filesToDiscard = files.filter(
|
||||
file => file.startsWith(`${thread.id}-`) &&
|
||||
file.endsWith('.json'))
|
||||
file.endsWith('.json'))
|
||||
|
||||
// remove files by unlinking
|
||||
filesToDiscard.forEach(file => {
|
||||
|
||||
11
src/keys.ts
11
src/keys.ts
@@ -1,11 +1,12 @@
|
||||
import { getEnvVar } from './utils/env.js'
|
||||
import { getEnvVar } from './utils/index.js'
|
||||
|
||||
export const Keys = {
|
||||
clientToken: getEnvVar('CLIENT_TOKEN'),
|
||||
model: getEnvVar('MODEL'),
|
||||
clientUid: getEnvVar('CLIENT_UID'),
|
||||
ipAddress: getEnvVar('OLLAMA_IP'),
|
||||
portAddress: getEnvVar('OLLAMA_PORT'),
|
||||
ipAddress: getEnvVar('OLLAMA_IP', '127.0.0.1'), // default ollama ip if none
|
||||
portAddress: getEnvVar('OLLAMA_PORT', '11434'), // default ollama port if none
|
||||
defaultModel: getEnvVar('MODEL', 'llama3.2'),
|
||||
redisHost: getEnvVar('REDIS_IP', '172.18.0.4'), // default redis host if none
|
||||
redisPort: parseInt(getEnvVar('REDIS_PORT', '6379')) // default redis port if none
|
||||
} as const // readonly keys
|
||||
|
||||
export default Keys
|
||||
@@ -17,7 +17,7 @@ export class Queue<T> implements IQueue<T> {
|
||||
* Set up Queue
|
||||
* @param capacity max length of queue
|
||||
*/
|
||||
constructor(public capacity: number = 5) {}
|
||||
constructor(public capacity: number = 5) { }
|
||||
|
||||
/**
|
||||
* Put item in front of queue
|
||||
|
||||
@@ -1,9 +1,10 @@
|
||||
import { UserMessage } from './events.js'
|
||||
import { ChannelType } from 'discord.js'
|
||||
import { UserMessage } from './index.js'
|
||||
|
||||
export interface UserConfiguration {
|
||||
'message-stream'?: boolean,
|
||||
'message-style'?: boolean,
|
||||
'modify-capacity': number
|
||||
'modify-capacity': number,
|
||||
'switch-model': string
|
||||
}
|
||||
|
||||
export interface ServerConfiguration {
|
||||
@@ -41,6 +42,21 @@ export interface Channel {
|
||||
messages: UserMessage[]
|
||||
}
|
||||
|
||||
/**
|
||||
* The following 2 types is allow for better readability in commands
|
||||
* Admin Command -> Don't run in Threads
|
||||
* User Command -> Used anywhere
|
||||
*/
|
||||
export const AdminCommand = [
|
||||
ChannelType.GuildText
|
||||
]
|
||||
|
||||
export const UserCommand = [
|
||||
ChannelType.GuildText,
|
||||
ChannelType.PublicThread,
|
||||
ChannelType.PrivateThread
|
||||
]
|
||||
|
||||
/**
|
||||
* Check if the configuration we are editing/taking from is a Server Config
|
||||
* @param key name of command we ran
|
||||
|
||||
@@ -1,21 +1,34 @@
|
||||
import { resolve } from 'path'
|
||||
import { config } from 'dotenv'
|
||||
|
||||
// Find config - ONLY WORKS WITH NODEMON
|
||||
const envFile = process.env.NODE_ENV === 'development' ? '.env.dev.local' : '.env'
|
||||
|
||||
// resolve config file
|
||||
const envFilePath = resolve(process.cwd(), envFile)
|
||||
const envFilePath = resolve(process.cwd(), '.env')
|
||||
const ipValidate: RegExp = /^((25[0-5]|(2[0-4]|1\d|[1-9]|)\d)\.?\b){4}$/
|
||||
|
||||
// set current environment variable file
|
||||
config({ path: envFilePath })
|
||||
|
||||
// Getter for environment variables
|
||||
/**
|
||||
* Method to validate if environment variables found in file utils/env.ts
|
||||
*
|
||||
* @param name Name of the environment variable in .env
|
||||
* @param fallback fallback value to set if environment variable is not set (used manually in src/keys.ts)
|
||||
* @returns environment variable value
|
||||
*/
|
||||
export function getEnvVar(name: string, fallback?: string): string {
|
||||
const value = process.env[name] ?? fallback
|
||||
if (value == undefined)
|
||||
if (!value)
|
||||
throw new Error(`Environment variable ${name} is not set.`)
|
||||
|
||||
// validate User-Generated Discord Application Tokens
|
||||
if (name === "CLIENT_TOKEN" && value.length > 72)
|
||||
throw new Error(`The "CLIENT_TOKEN" provided is not of at least length 72.
|
||||
This is probably an invalid token unless Discord updated their token policy. Please provide a valid token.`)
|
||||
|
||||
// validate IPv4 address found in environment variables
|
||||
if ((name.endsWith("_IP") || name.endsWith("_ADDRESS")) && !ipValidate.test(value))
|
||||
throw new Error(`Environment variable ${name} does not follow IPv4 formatting.`)
|
||||
|
||||
// return env variable
|
||||
return value
|
||||
}
|
||||
@@ -1,4 +1,4 @@
|
||||
import type { ClientEvents, Awaitable, Client, User } from 'discord.js'
|
||||
import type { ClientEvents, Awaitable, Client } from 'discord.js'
|
||||
import { Ollama } from 'ollama'
|
||||
import { Queue } from '../queues/queue.js'
|
||||
|
||||
@@ -8,17 +8,6 @@ export { Events } from 'discord.js'
|
||||
export type LogMethod = (...args: unknown[]) => void
|
||||
export type EventKeys = keyof ClientEvents // only wants keys of ClientEvents object
|
||||
|
||||
/**
|
||||
* Tokens to run the bot as intended
|
||||
* @param channel the channel where the bot will respond to queries
|
||||
* @param model chosen model for the ollama to utilize
|
||||
* @param clientUid the discord id for the bot
|
||||
*/
|
||||
export type Tokens = {
|
||||
model: string,
|
||||
clientUid: string
|
||||
}
|
||||
|
||||
/**
|
||||
* Parameters to run the chat query
|
||||
* @param model the model to run
|
||||
@@ -35,6 +24,7 @@ export type ChatParams = {
|
||||
* Format for the messages to be stored when communicating when the bot
|
||||
* @param role either assistant, user, or system
|
||||
* @param content string of the message the user or assistant provided
|
||||
* @param images array of images that the user or assistant provided
|
||||
*/
|
||||
export type UserMessage = {
|
||||
role: string,
|
||||
@@ -44,12 +34,18 @@ export type UserMessage = {
|
||||
|
||||
// Event properties
|
||||
export interface EventProps {
|
||||
client: Client
|
||||
log: LogMethod
|
||||
msgHist: Queue<UserMessage>
|
||||
tokens: Tokens,
|
||||
ollama: Ollama
|
||||
client: Client,
|
||||
log: LogMethod,
|
||||
msgHist: Queue<UserMessage>,
|
||||
ollama: Ollama,
|
||||
defaultModel: String
|
||||
}
|
||||
|
||||
/**
|
||||
* Format for the callback function tied to an event
|
||||
* @param props the properties of the event
|
||||
* @param args the arguments of the event
|
||||
*/
|
||||
export type EventCallback<T extends EventKeys> = (
|
||||
props: EventProps,
|
||||
...args: ClientEvents[T]
|
||||
@@ -61,6 +57,12 @@ export interface Event<T extends EventKeys = EventKeys> {
|
||||
callback: EventCallback<T>
|
||||
}
|
||||
|
||||
/**
|
||||
* Method to create an event object
|
||||
* @param key type of event
|
||||
* @param callback function to run when event is triggered
|
||||
* @returns event object
|
||||
*/
|
||||
export function event<T extends EventKeys>(key: T, callback: EventCallback<T>): Event<T> {
|
||||
return { key, callback }
|
||||
}
|
||||
@@ -70,15 +72,14 @@ export function event<T extends EventKeys>(key: T, callback: EventCallback<T>):
|
||||
* @param client initialized bot client
|
||||
* @param events all the exported events from the index.ts in the events dir
|
||||
* @param msgHist The message history of the bot
|
||||
* @param tokens the passed in environment tokens for the service
|
||||
* @param ollama the initialized ollama instance
|
||||
*/
|
||||
export function registerEvents(
|
||||
client: Client,
|
||||
events: Event[],
|
||||
msgHist: Queue<UserMessage>,
|
||||
tokens: Tokens,
|
||||
ollama: Ollama
|
||||
ollama: Ollama,
|
||||
defaultModel: String
|
||||
): void {
|
||||
for (const { key, callback } of events) {
|
||||
client.on(key, (...args) => {
|
||||
@@ -87,7 +88,7 @@ export function registerEvents(
|
||||
|
||||
// Handle Errors, call callback, log errors as needed
|
||||
try {
|
||||
callback({ client, log, msgHist, tokens, ollama }, ...args)
|
||||
callback({ client, log, msgHist, ollama, defaultModel }, ...args)
|
||||
} catch (error) {
|
||||
log('[Uncaught Error]', error)
|
||||
}
|
||||
|
||||
@@ -55,3 +55,12 @@ export async function getAttachmentData(attachment: Attachment | undefined): Pro
|
||||
const base64String = arrayBufferToBase64(buffer)
|
||||
return [base64String]
|
||||
}
|
||||
|
||||
/**
|
||||
* Method to retrieve the string data from the text file
|
||||
*
|
||||
* @param attachment the text file to convert to a string
|
||||
*/
|
||||
export async function getTextFileAttachmentData(attachment: Attachment): Promise<string> {
|
||||
return await (await fetch(attachment.url)).text()
|
||||
}
|
||||
@@ -53,7 +53,6 @@ export async function clearChannelInfo(filename: string, channel: TextChannel, u
|
||||
}
|
||||
})
|
||||
})
|
||||
console.log(cleanedHistory)
|
||||
return cleanedHistory
|
||||
}
|
||||
|
||||
@@ -81,7 +80,14 @@ export async function openChannelInfo(filename: string, channel: TextChannel | T
|
||||
}
|
||||
})
|
||||
} else { // file doesn't exist, create it
|
||||
const object: Configuration = JSON.parse(`{ \"id\": \"${channel?.id}\", \"name\": \"${channel?.name}\", \"user\": \"${user}\", \"messages\": []}`)
|
||||
const object: Configuration = JSON.parse(
|
||||
`{
|
||||
\"id\": \"${channel?.id}\",
|
||||
\"name\": \"${channel?.name}\",
|
||||
\"user\": \"${user}\",
|
||||
\"messages\": []
|
||||
}`
|
||||
)
|
||||
|
||||
const directory = path.dirname(fullFileName)
|
||||
if (!fs.existsSync(directory))
|
||||
|
||||
@@ -5,7 +5,7 @@ import { AbortableAsyncIterator } from "ollama/src/utils.js"
|
||||
/**
|
||||
* Method to query the Ollama client for async generation
|
||||
* @param params
|
||||
* @returns Asyn
|
||||
* @returns AsyncIterator<ChatResponse> generated by the Ollama client
|
||||
*/
|
||||
export async function streamResponse(params: ChatParams): Promise<AbortableAsyncIterator<ChatResponse>> {
|
||||
return await params.ollama.chat({
|
||||
|
||||
@@ -1,10 +1,10 @@
|
||||
// Centralized import index
|
||||
export * from './env.js'
|
||||
export * from './events.js'
|
||||
export * from './messageEmbed.js'
|
||||
export * from './messageNormal.js'
|
||||
export * from './commands.js'
|
||||
export * from './configInterfaces.js'
|
||||
export * from './mentionClean.js'
|
||||
|
||||
// handler imports
|
||||
export * from './handlers/chatHistoryHandler.js'
|
||||
|
||||
@@ -1,5 +1,3 @@
|
||||
import Keys from "../keys.js"
|
||||
|
||||
/**
|
||||
* Clean up the bot user_id so it only has the prompt
|
||||
*
|
||||
@@ -8,9 +6,9 @@ import Keys from "../keys.js"
|
||||
* - replace function works well for this
|
||||
*
|
||||
* @param message
|
||||
* @returns
|
||||
* @returns message without client id
|
||||
*/
|
||||
export function clean(message: string): string {
|
||||
const cleanedMessage: string = message.replace(`<@${Keys.clientUid}>`, '').trim()
|
||||
export function clean(message: string, clientId: string): string {
|
||||
const cleanedMessage: string = message.replace(`<@${clientId}>`, '').trim()
|
||||
return cleanedMessage
|
||||
}
|
||||
@@ -1,130 +0,0 @@
|
||||
import { EmbedBuilder, Message } from 'discord.js'
|
||||
import { ChatResponse, Ollama } from 'ollama'
|
||||
import { ChatParams, UserMessage, streamResponse, blockResponse } from './index.js'
|
||||
import { Queue } from '../queues/queue.js'
|
||||
import { AbortableAsyncIterator } from 'ollama/src/utils.js'
|
||||
|
||||
/**
|
||||
* Method to send replies as normal text on discord like any other user
|
||||
* @param message message sent by the user
|
||||
* @param tokens tokens to run query
|
||||
* @param msgHist message history between user and model
|
||||
*/
|
||||
export async function embedMessage(
|
||||
message: Message,
|
||||
ollama: Ollama,
|
||||
tokens: {
|
||||
model: string
|
||||
},
|
||||
msgHist: Queue<UserMessage>,
|
||||
stream: boolean
|
||||
): Promise<string> {
|
||||
// bot response
|
||||
let response: ChatResponse | AbortableAsyncIterator<ChatResponse>
|
||||
let result: string = ''
|
||||
|
||||
// initial message to client
|
||||
const botMessage = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription('Generating Response . . .')
|
||||
.setColor('#00FF00')
|
||||
|
||||
// send the message
|
||||
const sentMessage = await message.channel.send({ embeds: [botMessage] })
|
||||
|
||||
// create params
|
||||
const params: ChatParams = {
|
||||
model: tokens.model,
|
||||
ollama: ollama,
|
||||
msgHist: msgHist.getItems()
|
||||
}
|
||||
|
||||
try {
|
||||
// check if embed needs to stream
|
||||
if (stream) {
|
||||
response = await streamResponse(params)
|
||||
|
||||
for await (const portion of response) {
|
||||
result += portion.message.content
|
||||
|
||||
// exceeds handled length
|
||||
if (result.length > 5000) {
|
||||
const errorEmbed = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription(`Response length ${result.length} has exceeded Discord maximum.\n\nLong Stream messages not supported.`)
|
||||
.setColor('#00FF00')
|
||||
|
||||
// send error
|
||||
message.channel.send({ embeds: [errorEmbed] })
|
||||
break // cancel loop and stop
|
||||
}
|
||||
|
||||
// new embed per token...
|
||||
const streamEmbed = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription(result || 'No Content Yet...')
|
||||
.setColor('#00FF00')
|
||||
|
||||
// edit the message
|
||||
sentMessage.edit({ embeds: [streamEmbed] })
|
||||
}
|
||||
} else {
|
||||
response = await blockResponse(params)
|
||||
result = response.message.content
|
||||
|
||||
// long message, split into different embeds sadly.
|
||||
if (result.length > 5000) {
|
||||
const firstEmbed = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription(result.slice(0, 5000) || 'No Content to Provide...')
|
||||
.setColor('#00FF00')
|
||||
|
||||
// replace first embed
|
||||
sentMessage.edit({ embeds: [firstEmbed] })
|
||||
|
||||
// take the rest out
|
||||
result = result.slice(5000)
|
||||
|
||||
// handle the rest
|
||||
while (result.length > 5000) {
|
||||
const whileEmbed = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription(result.slice(0, 5000) || 'No Content to Provide...')
|
||||
.setColor('#00FF00')
|
||||
|
||||
message.channel.send({ embeds: [whileEmbed] })
|
||||
result = result.slice(5000)
|
||||
}
|
||||
|
||||
const lastEmbed = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription(result || 'No Content to Provide...')
|
||||
.setColor('#00FF00')
|
||||
|
||||
// rest of the response
|
||||
message.channel.send({ embeds: [lastEmbed] })
|
||||
} else {
|
||||
// only need to create 1 embed, handles 6000 characters
|
||||
const newEmbed = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription(result || 'No Content to Provide...')
|
||||
.setColor('#00FF00')
|
||||
|
||||
// edit the message
|
||||
sentMessage.edit({ embeds: [newEmbed] })
|
||||
}
|
||||
}
|
||||
} catch(error: any) {
|
||||
console.log(`[Util: messageEmbed] Error creating message: ${error.message}`)
|
||||
const errorEmbed = new EmbedBuilder()
|
||||
.setTitle(`Responding to ${message.author.tag}`)
|
||||
.setDescription(`**Response generation failed.**\n\nReason: ${error.message}`)
|
||||
.setColor('#00FF00')
|
||||
|
||||
// send back error
|
||||
sentMessage.edit({ embeds: [errorEmbed] })
|
||||
}
|
||||
|
||||
// Hope there is a response! undefined otherwie
|
||||
return result
|
||||
}
|
||||
@@ -1,4 +1,4 @@
|
||||
import { Message } from 'discord.js'
|
||||
import { Message, SendableChannels } from 'discord.js'
|
||||
import { ChatResponse, Ollama } from 'ollama'
|
||||
import { ChatParams, UserMessage, streamResponse, blockResponse } from './index.js'
|
||||
import { Queue } from '../queues/queue.js'
|
||||
@@ -7,26 +7,25 @@ import { AbortableAsyncIterator } from 'ollama/src/utils.js'
|
||||
/**
|
||||
* Method to send replies as normal text on discord like any other user
|
||||
* @param message message sent by the user
|
||||
* @param tokens tokens to run query
|
||||
* @param model name of model to run query
|
||||
* @param msgHist message history between user and model
|
||||
*/
|
||||
export async function normalMessage(
|
||||
message: Message,
|
||||
ollama: Ollama,
|
||||
tokens: {
|
||||
model: string
|
||||
},
|
||||
model: string,
|
||||
msgHist: Queue<UserMessage>,
|
||||
stream: boolean
|
||||
): Promise<string> {
|
||||
// bot's respnse
|
||||
let response: ChatResponse | AbortableAsyncIterator<ChatResponse>
|
||||
let result: string = ''
|
||||
const channel = message.channel as SendableChannels
|
||||
|
||||
await message.channel.send('Generating Response . . .').then(async sentMessage => {
|
||||
await channel.send('Generating Response . . .').then(async sentMessage => {
|
||||
try {
|
||||
const params: ChatParams = {
|
||||
model: tokens.model,
|
||||
model: model,
|
||||
ollama: ollama,
|
||||
msgHist: msgHist.getItems()
|
||||
}
|
||||
@@ -41,7 +40,8 @@ export async function normalMessage(
|
||||
result = portion.message.content
|
||||
|
||||
// new message block, wait for it to send and assign new block to respond.
|
||||
await message.channel.send("Creating new stream block...").then(sentMessage => { messageBlock = sentMessage })
|
||||
await channel.send("Creating new stream block...")
|
||||
.then(sentMessage => { messageBlock = sentMessage })
|
||||
} else {
|
||||
result += portion.message.content
|
||||
|
||||
@@ -63,17 +63,21 @@ export async function normalMessage(
|
||||
|
||||
// handle for rest of message that is >2000
|
||||
while (result.length > 2000) {
|
||||
message.channel.send(result.slice(0, 2000))
|
||||
channel.send(result.slice(0, 2000))
|
||||
result = result.slice(2000)
|
||||
}
|
||||
|
||||
// last part of message
|
||||
message.channel.send(result)
|
||||
channel.send(result)
|
||||
} else // edit the 'generic' response to new message since <2000
|
||||
sentMessage.edit(result)
|
||||
}
|
||||
} catch(error: any) {
|
||||
} catch (error: any) {
|
||||
console.log(`[Util: messageNormal] Error creating message: ${error.message}`)
|
||||
if (error.message.includes('fetch failed'))
|
||||
error.message = 'Missing ollama service on machine'
|
||||
else if (error.message.includes('try pulling it first'))
|
||||
error.message = `You do not have the ${model} downloaded. Ask an admin to pull it using the \`pull-model\` command.`
|
||||
sentMessage.edit(`**Response generation failed.**\n\nReason: ${error.message}`)
|
||||
}
|
||||
})
|
||||
|
||||
@@ -1,8 +1,20 @@
|
||||
// describe marks a test suite
|
||||
// expect takes a value from an expression
|
||||
// it marks a test case
|
||||
import { describe, expect, it } from 'vitest'
|
||||
import commands from '../src/commands'
|
||||
import { describe, expect, it, vi } from 'vitest'
|
||||
import commands from '../src/commands/index.js'
|
||||
|
||||
/**
|
||||
* Mocking redis found in client.ts because of the commands
|
||||
*/
|
||||
vi.mock('../src/client.js', () => ({
|
||||
redis: {
|
||||
createClient: vi.fn(),
|
||||
connect: vi.fn(),
|
||||
get: vi.fn(),
|
||||
set: vi.fn()
|
||||
}
|
||||
}))
|
||||
|
||||
/**
|
||||
* Commands test suite, tests the commands object
|
||||
@@ -12,7 +24,7 @@ import commands from '../src/commands'
|
||||
* @param name name of the test suite
|
||||
* @param fn function holding tests to run
|
||||
*/
|
||||
describe('#commands', () => {
|
||||
describe('Commands Existence', () => {
|
||||
// test definition of commands object
|
||||
it('references defined object', () => {
|
||||
// toBe compares the value to the expected value
|
||||
@@ -22,6 +34,52 @@ describe('#commands', () => {
|
||||
// test specific commands in the object
|
||||
it('references specific commands', () => {
|
||||
const commandsString = commands.map(e => e.name).join(', ')
|
||||
expect(commandsString).toBe('thread, private-thread, message-style, message-stream, toggle-chat, shutoff, modify-capacity, clear-user-channel-history')
|
||||
const expectedCommands = ['thread', 'private-thread', 'message-stream', 'toggle-chat', 'shutoff', 'modify-capacity', 'clear-user-channel-history', 'pull-model', 'switch-model', 'delete-model']
|
||||
expect(commandsString).toBe(expectedCommands.join(', '))
|
||||
})
|
||||
})
|
||||
|
||||
/**
|
||||
* User Commands Test suite for testing out commands
|
||||
* that would be run by users when using the application.
|
||||
*/
|
||||
describe('User Command Tests', () => {
|
||||
// test capacity command
|
||||
it('run modify-capacity command', () => {
|
||||
|
||||
})
|
||||
|
||||
it('run clear-user-channel-history command', () => {
|
||||
|
||||
})
|
||||
|
||||
it('run message-stream command', () => {
|
||||
|
||||
})
|
||||
|
||||
it('run message-style command', () => {
|
||||
|
||||
})
|
||||
|
||||
it('run thread command', () => {
|
||||
|
||||
})
|
||||
|
||||
it('run private-thread command', () => {
|
||||
|
||||
})
|
||||
})
|
||||
|
||||
/**
|
||||
* Admin Commands Test suite for running administrative
|
||||
* commands with the application.
|
||||
*/
|
||||
describe('Admin Command Tests', () => {
|
||||
it('run shutoff command', () => {
|
||||
|
||||
})
|
||||
|
||||
it('run toggle-chat command', () => {
|
||||
|
||||
})
|
||||
})
|
||||
@@ -1,15 +1,23 @@
|
||||
import { describe, expect, it } from 'vitest'
|
||||
import events from '../src/events'
|
||||
import { describe, expect, it, vi } from 'vitest'
|
||||
import events from '../src/events/index.js'
|
||||
|
||||
/**
|
||||
* Mocking ollama found in client.ts because pullModel.ts
|
||||
* relies on the existence on ollama. To prevent the mock,
|
||||
* we will have to pass through ollama to the commands somehow.
|
||||
*/
|
||||
vi.mock('../src/client.js', () => ({
|
||||
ollama: {
|
||||
pull: vi.fn() // Mock the pull method found with ollama
|
||||
}
|
||||
}))
|
||||
|
||||
/**
|
||||
* Events test suite, tests the events object
|
||||
* Each event is to be tested elsewhere, this file
|
||||
* is to ensure that the events object is defined.
|
||||
*
|
||||
* @param name name of the test suite
|
||||
* @param fn function holding tests to run
|
||||
*/
|
||||
describe('#events', () => {
|
||||
describe('Events Existence', () => {
|
||||
// test definition of events object
|
||||
it('references defined object', () => {
|
||||
expect(typeof events).toBe('object')
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
import { describe, expect, it } from 'vitest'
|
||||
import { getEnvVar } from '../src/utils'
|
||||
import { getEnvVar } from '../src/utils/index.js'
|
||||
|
||||
/**
|
||||
* getEnvVar test suite, tests the getEnvVar function
|
||||
@@ -7,7 +7,7 @@ import { getEnvVar } from '../src/utils'
|
||||
* @param name name of the test suite
|
||||
* @param fn function holding tests to run
|
||||
*/
|
||||
describe('#getEnvVar', () => {
|
||||
describe('Environment Setup', () => {
|
||||
// dummy set of keys
|
||||
const keys = {
|
||||
clientToken: 'CLIENT_TOKEN',
|
||||
|
||||
@@ -1,6 +1,8 @@
|
||||
import { describe, expect, it } from 'vitest'
|
||||
import { clean } from '../src/utils/mentionClean'
|
||||
import { getEnvVar } from '../src/utils'
|
||||
import { clean } from '../src/utils/index.js'
|
||||
|
||||
// Sample UID for testing
|
||||
const sampleId = '123456789'
|
||||
|
||||
/**
|
||||
* MentionClean test suite, tests the clean function
|
||||
@@ -8,10 +10,10 @@ import { getEnvVar } from '../src/utils'
|
||||
* @param name name of the test suite
|
||||
* @param fn function holding tests to run
|
||||
*/
|
||||
describe('#clean', () => {
|
||||
describe('Mentions Cleaned', () => {
|
||||
// test for id removal from message
|
||||
it('removes the mention from a message', () => {
|
||||
const message = `<@${getEnvVar('CLIENT_UID')}> Hello, World!`
|
||||
expect(clean(message)).toBe('Hello, World!')
|
||||
const message = `<@${sampleId}> Hello, World!`
|
||||
expect(clean(message, sampleId)).toBe('Hello, World!')
|
||||
})
|
||||
})
|
||||
@@ -1,5 +1,5 @@
|
||||
import { describe, expect, it } from 'vitest'
|
||||
import { Queue } from '../src/queues/queue'
|
||||
import { Queue } from '../src/queues/queue.js'
|
||||
|
||||
/**
|
||||
* Queue test suite, tests the Queue class
|
||||
@@ -7,8 +7,8 @@ import { Queue } from '../src/queues/queue'
|
||||
* @param name name of the test suite
|
||||
* @param fn function holding tests to run
|
||||
*/
|
||||
describe('#queue', () => {
|
||||
let queue= new Queue<string>()
|
||||
describe('Queue Structure', () => {
|
||||
let queue = new Queue<string>()
|
||||
|
||||
// test for queue creation
|
||||
it('creates a new queue', () => {
|
||||
|
||||
@@ -4,9 +4,10 @@ import { defineConfig, configDefaults } from 'vitest/config'
|
||||
export default defineConfig({
|
||||
test: {
|
||||
globals: true, // <-- reduces test file imports
|
||||
reporters: ['verbose'], // <-- verbose output
|
||||
coverage: {
|
||||
exclude: [...configDefaults.exclude, 'build/*', 'tests/*'], // <-- exclude JS build
|
||||
reporter: ['text', 'html'] // <-- reports in text, html
|
||||
reporter: ['text-summary'] // <-- report in text-summary
|
||||
}
|
||||
}
|
||||
})
|
||||
Reference in New Issue
Block a user