{"package":"ai-dynamo-vllm","ecosystem":"pypi","exists":true,"latest_version":"0.8.4.post4","repository":"https://github.com/vllm-project/vllm","license":"Apache-2.0","description":"A high-throughput and memory-efficient inference and serving engine for LLMs","downloads_weekly":157,"deprecated":false,"health":{"score":71},"_cache":"db_only_bot","_partial":true,"_response_ms":0,"_powered_by":"depscope.dev — bot fast path (DB-only)","recommendation":{"action":"safe"}}