Capability
Quantization And Model Compression For Efficient Deployment
20 artifacts provide this capability.
Want a personalized recommendation?
Find the best match →Top Matches
via “model quantization and compression for edge deployment”
fill-mask model by undefined. 6,06,75,227 downloads.
Unique: Post-training quantization via ONNX Runtime or PyTorch quantization APIs requires no retraining while achieving 4x model size reduction; supports multiple quantization schemes (symmetric, asymmetric, per-channel) for fine-grained accuracy-efficiency control
vs others: Simpler than quantization-aware training (no retraining required) and more portable than framework-specific quantization due to ONNX support