
node-llama-cpp
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
@basetenlabs/performance-client
This library provides a high-performance Node.js client for Baseten.co endpoints including embeddings, reranking, and classification. It was built for massive concurrent POST requests to any URL, also outside of baseten.co. The PerformanceClient is built
n8n-nodes-query-retriever-rerank
Advanced n8n community node for intelligent document retrieval with multi-step reasoning, reranking, and comprehensive debugging
n8n-nodes-contextualai
n8n community node for Contextual AI - enterprise-grade RAG agents, document parsing, querying, reranking, and evaluation
llama-cpp-capacitor
A native Capacitor plugin that embeds llama.cpp directly into mobile apps, enabling offline AI inference with chat-first API design. Supports both simple text generation and advanced chat conversations with system prompts, multimodal processing, TTS, LoRA