{"package":"llama-cpp-capacitor","ecosystem":"npm","exists":true,"latest_version":"0.1.5","repository":"https://github.com/arusatech/llama-cpp","license":"MIT","description":"A native Capacitor plugin that embeds llama.cpp directly into mobile apps, enabling offline AI inference with chat-first API design. Complete iOS and Android support: text generation, chat, multimodal, TTS, LoRA, embeddings, and more.","downloads_weekly":5252,"deprecated":false,"health":{"score":67},"_cache":"db_only_bot","_partial":true,"_response_ms":0,"_powered_by":"depscope.dev — bot fast path (DB-only)","recommendation":{"action":"review"}}