{"package":"flash_attn","ecosystem":"pypi","latest_version":"2.8.3","description":"Flash Attention: Fast and Memory-Efficient Exact Attention","license":"BSD-3-Clause","homepage":"https://github.com/Dao-AILab/flash-attention","repository":"https://github.com/Dao-AILab/flash-attention","downloads_weekly":522601,"health":{"score":74,"risk":"moderate","breakdown":{"maintenance":10,"popularity":14,"security":25,"maturity":15,"community":10},"deprecated":false,"max_score":100},"vulnerabilities":{"count":0,"critical":0,"high":0,"medium":0,"low":0,"details":[]},"versions":{"latest":"2.8.3","total_count":75,"recent":["2.5.3","2.5.4","2.5.5","2.5.6","2.5.7","2.5.8","2.5.9.post1","2.6.0.post1","2.6.1","2.6.2","2.6.3","2.7.0.post2","2.7.1.post4","2.7.2.post1","2.7.3","2.7.4.post1","2.8.0.post2","2.8.1","2.8.2","2.8.3"]},"metadata":{"deprecated":false,"deprecated_message":null,"maintainers_count":1,"first_published":null,"last_published":"2025-08-15T08:28:12.911581Z","dependencies_count":2,"dependencies":["torch","einops"]},"bundle":null,"typescript":null,"known_issues":{"bugs_count":0,"bugs_severity":{},"status_breakdown":{},"link":null,"scope":"none"},"recommendation":{"action":"safe_to_use","issues":[],"use_version":"2.8.3","version_hint":null,"summary":"flash_attn@2.8.3 is safe to use (health: 74/100)"},"requested_version":null,"_cache":"miss","_response_ms":615,"_powered_by":"depscope.dev — free package intelligence for AI agents","typosquat":{"is_suspected":false},"maintainer_trust":{"available":false},"malicious":{"is_malicious":false},"scorecard":{"available":false},"quality":{"available":false}}