Summary
A significant performance bug has been discovered in cuBLAS, causing a 60% performance loss for batched FP32 matrix multiplications on RTX GPUs, including the new RTX 5090. The library dispatches inefficient kernels, utilizing only about 40% of available compute, even with the latest CUDA and cuBLAS versions. This issue affects a fundamental operation critical for many AI/ML workloads.
Continue Reading
Explore related coverage about community news and adjacent AI developments: [r/ML] [D] MYTHOS-INVERSION STRUCTURAL AUDIT, [r/LocalLLaMA] karpathy / autoresearch, [HN] Show HN: Ship of Theseus License, [r/ML] [R] Agentic AI and Occupational Displacement: A Multi-Regional Task Exposure Analysis (236 occupations, 5 US metros).
Related Articles
Comments
Sign in to leave a comment.
Loading comments...