Capability
Model Compression Through Pruning And Structured Sparsity Support
4 artifacts provide this capability.
Want a personalized recommendation?
Find the best match →Top Matches
Lightweight ML inference for mobile and edge devices.
Unique: Runtime support for pruned and sparsified models that skip zero-valued weights and use sparse tensor formats, enabling compression beyond quantization for models trained with sparsity constraints.
vs others: Complementary to quantization for additional compression; however, requires training-time support and sparse tensor format standardization which are not fully documented.