S-STE: Continuous Pruning Function for Efficient 2: 4 Sparse Pre-training
–Neural Information Processing Systems
Training deep neural networks (DNNs) is costly. Fortunately, Nvidia Ampere and Hopper GPUs can accelerate matrix multiplications twice as fast as a dense equivalent by implementing 2:4 sparsity.
Neural Information Processing Systems
May-29-2025, 04:57:31 GMT
- Country:
- North America > United States
- Pennsylvania (0.14)
- Texas (0.14)
- North America > United States
- Genre:
- Research Report > Experimental Study (0.93)
- Industry:
- Information Technology (0.48)
- Technology: