catai
Chat UI and Local API for the Llama models
inference-server
Libraries and server to build AI applications. Adapters to various native bindings allowing local inference. Integrate it with your application, or use as a microservice.
ai-server
An OpenAI and Claude API compatible server using node-llama-cpp for local LLM models