{"package":"flashinfer_python","ecosystem":"pypi","latest_version":"0.6.8.post1","description":"FlashInfer: Kernel Library for LLM Serving","license":"Apache-2.0","homepage":"https://pypi.org/project/flashinfer-python/","repository":"https://github.com/flashinfer-ai/flashinfer","downloads_weekly":899199,"health":{"score":81,"risk":"low","breakdown":{"maintenance":25,"popularity":14,"security":25,"maturity":15,"community":2},"deprecated":false,"max_score":100},"vulnerabilities":{"count":0,"critical":0,"high":0,"medium":0,"low":0,"details":[]},"versions":{"latest":"0.6.8.post1","total_count":63,"recent":["0.5.0","0.5.1","0.5.2","0.5.3","0.6.0rc1","0.6.0rc2","0.6.0","0.6.1","0.6.2","0.6.3","0.6.4","0.6.5","0.6.6","0.6.7","0.6.7.post1","0.6.7.post2","0.6.7.post3","0.6.8rc1","0.6.8","0.6.8.post1"]},"metadata":{"deprecated":false,"deprecated_message":null,"maintainers_count":1,"first_published":null,"last_published":"2026-04-18T18:28:10.285037Z","dependencies_count":16,"dependencies":["apache-tvm-ffi!=0.1.8,!=0.1.8.post0,<0.2,>=0.1.6","click","cuda-tile","einops","ninja","numpy","nvidia-cudnn-frontend>=1.13.0","nvidia-cutlass-dsl>=4.4.2","nvidia-ml-py","packaging>=24.2","requests","tabulate","torch","tqdm","nvidia-cutlass-dsl>=4.4.2; extra == \"cu12\"","nvidia-cutlass-dsl[cu13]>=4.4.2; extra == \"cu13\""]},"bundle":null,"typescript":null,"known_issues":{"bugs_count":0,"bugs_severity":{},"status_breakdown":{},"link":null,"scope":"none"},"recommendation":{"action":"safe_to_use","issues":[],"use_version":"0.6.8.post1","version_hint":null,"summary":"flashinfer_python@0.6.8.post1 is safe to use (health: 81/100)"},"requested_version":null,"_cache":"miss","_response_ms":407,"_powered_by":"depscope.dev — free package intelligence for AI agents","typosquat":{"is_suspected":false},"maintainer_trust":{"available":false},"malicious":{"is_malicious":false},"scorecard":{"available":false},"quality":{"available":false}}