A web interface for chatting with Alpaca through llama.cpp. Fully dockerized, with an easy to use API.
Serge - LLaMA made easy is a chat interface created with llama.cpp for running GGUF models. It is a self-hosted chat interface that does not require API keys. It uses a SvelteKit frontend and utilizes Redis for storing chat history and parameters. The API is built using FastAPI and LangChain, wrapping calls to llama.cpp using the python bindings.
To install Serge, you can use various methods depending on your setup:
Serge is a self-hosted chat interface that allows users to run GGUF models. It does not require API keys and utilizes a SvelteKit frontend. The chat history and parameters are stored in Redis, and the API is built using FastAPI and LangChain. Serge provides different installation methods such as Docker and Kubernetes. It is free and open-source software licensed under the MIT License and Apache-2.0.