SparseML

About

SparseML enables network sparsity with a few lines of code, through open-source libraries, to boost neural network inference performance.

SparseML allows you to create inference-optimized sparse models using state-of-the-art pruning and quantization algorithms. Models trained with SparseML can then be exported to ONNX and deployed with DeepSparse for GPU-class performance on CPU hardware.


Media

Products media viewer

  • Intro to SparseML

    Learn about SparseML, open-source libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models.


Similar products

Neural Magic (Acquired by Red Hat) products