llama-cpp-capacitor

A native Capacitor plugin that embeds llama.cpp directly into mobile apps, enabling offline AI inference with chat-first API design. Supports both simple text generation and advanced chat conversations with system prompts, multimodal processing, TTS, LoRA

@huggingface/ollama-utils

Various utilities for maintaining Ollama compatibility with models on Hugging Face hub