inference-server

Libraries and server to build AI applications. Adapters to various native bindings allowing local inference. Integrate it with your application, or use as a microservice.

ai-server

An OpenAI and Claude API compatible server using node-llama-cpp for local LLM models