@neuralmagic
Neural Magic
3 years
We carried a 4-core laptop around Boston, comparing runs of sparsified #YOLOv5 object detection model running on the #DeepSparse Engine and #ONNXRuntime . End result: Pruning + INT8 quantization = 10x faster and 12x smaller model. Replicate our results:
3
8
41

Replies

@Paul78615431
Drunk Donkey
3 years
@neuralmagic What were your thoughts on AI Day and Dojo?
1
0
0
@nkanauzu
Nkanauzu
3 years
0
0
0