Continuation of Open-Chat as an effort to kick-start further open-source development without any open-source related funding.
Open-Chat will be built openly, while the more advanced bots and model connectors will be developed closed source until mature enough for release.
The plan for msgmate.io V2 beta is to host a fully usable version of Open-Chat and provide a wide range of model connectors and bots for experimentation.
Usage of these features will incur costs, which will be covered by token purchases. Token tracking and usage-related code will also remain closed source, but Open-Chat will always function seamlessly without them.
We are eager to get the infrastructure out there and used. A YouTube introduction series is planned to present the interface and its deployment.
If you have collaboration ideas, want a demo of Msgmate, or seek early user access, please email me at tim@msgmate.io
.
You can use open-chat for local LLM completions today by following these steps:
First clone the repository & initalize the submodules ( hal/msgmate
is not open-source yet but fully optional! )
git clone https://github.com/msgmate-io/open-chat
git submodule update --init --recursive
Setup .env
with your backends of choice e.g.:
GROQ_API_KEY="<API key for https://groq.com/>"
DEEPINFRA_API_KEY="<API key for https://deepinfra.com>"
OPENAI_API_KEY="<API key for https://api.openai.com/>"
Then build and run all services:
docker compose -f docker-compose.pro.yaml build
docker compose -f docker-compose.pro.yaml up
And you're ready to go at localhost
.
Future todos are to:
- provide pre-build docker images
- provide pre-build clients
- provide setup instruction to local bot connectors to e.g.: https://localai.io
Development generaly hapens with the default compose so just docker compose up
will do the trick.
git submodule update --init --recursive
docker compose up --build
Services:
backend
: Django backend with chat and user management featuresfrontend
: Vike.dev frontend with chat and user management featuresredis
: Redis for caching and chatingress
: Nginx ingress for routing, routes all traffic throughlocalhost:80
in developmenthal
: Default LLM model completion botredis-hal
: Redis for caching and chat for thehal
bot
For clearer seperation or to run the bot somewhere else, you migh want to run the bot and chat seperatly:
Chat Backend & Interface only
docker compose up backend frontend redis ingress
Hal Bot only
docker compose build hal redis-hal
docker compose up hal redis-hal
All services and modules should also work outside of docker, e.g.: for frontend only developement:
cd frontend
npm install
npm run build
npm run dev
This project is built and maintained by @tbscode and was initiated to offer a truly decentralized Chat GPT alternative, including user management, deployment, bots, voice chat, and more.
As I work full-time as CTO of the non-profit Little World, I will use the limited time I have to extend and manage this project. However, to truly advance this stack to where it needs to be for widespread adoption, it requires many more development hours and must remain open-source and applicable for small businesses or individuals to build and experiment with.
Efforts to secure development funding are ongoing but have so far been unsuccessful. I hope to find and encourage people to help with this project and am willing to invest in its open-source development.