{"package":"llm-guard","ecosystem":"npm","latest_version":"0.1.8","description":"A TypeScript library for validating and securing LLM prompts","license":"MIT","homepage":"https://therizwan.github.io/llm-guard/","repository":"https://github.com/therizwan/llm-guard","downloads_weekly":182,"health":{"score":46,"risk":"high","breakdown":{"maintenance":10,"popularity":3,"security":25,"maturity":6,"community":2},"deprecated":false,"max_score":100},"vulnerabilities":{"count":0,"critical":0,"high":0,"medium":0,"low":0,"details":[]},"versions":{"latest":"0.1.8","total_count":9,"recent":["0.1.0","0.1.1","0.1.2","0.1.3","0.1.4","0.1.5","0.1.6","0.1.7","0.1.8"]},"metadata":{"deprecated":false,"deprecated_message":null,"maintainers_count":1,"first_published":"2025-04-10T06:58:40.143Z","last_published":"2025-06-04T09:34:44.154Z","dependencies_count":0,"dependencies":[]},"bundle":{"size_kb":11.2,"gzip_kb":3.3,"dependency_count":0,"has_js_module":false,"has_side_effects":true,"scoped":false,"source":"bundlephobia"},"typescript":{"score":10,"has_types":true,"types_source":"bundled","types_package":null},"known_issues":{"bugs_count":0,"bugs_severity":{},"status_breakdown":{},"link":null,"scope":"none"},"recommendation":{"action":"safe_to_use","issues":[],"use_version":"0.1.8","version_hint":null,"summary":"llm-guard@0.1.8 is safe to use (health: 46/100)"},"requested_version":null,"_cache":"miss","_response_ms":911,"_powered_by":"depscope.dev — free package intelligence for AI agents","typosquat":{"is_suspected":false},"maintainer_trust":{"available":false},"malicious":{"is_malicious":false},"scorecard":{"available":false},"quality":{"available":false}}