Capability
Inference Latency Profiling And Analysis
15 artifacts provide this capability.
Want a personalized recommendation?
Find the best match →Top Matches
via “performance profiling and latency measurement”
Cross-platform ONNX inference for mobile devices.
Unique: Implements per-operator profiling that is execution-provider-aware — profiling data shows which operators ran on CPU vs accelerator, enabling developers to understand why certain operators didn't accelerate as expected. This is more detailed than TensorFlow Lite's profiling, which is less granular.
vs others: More detailed profiling than PyTorch Mobile because it includes per-operator timing and memory usage; more accessible than native profiling tools (Instruments on iOS, Android Profiler) because profiling is built into the runtime and doesn't require external tools.