Speed-up ONNX models with TensorRT

ONNX models ONNX is a great initiative to standardize the structure and storage of deep neural networks. Almost all frameworks export to ONNX in one way or another. Implementations may then vary, though it’s expected to converge eventually. ONNX models are of great interest since: They can easily be exported around with their weights They can be optimized and converted to a variety of CPUs and GPUs Both interoperability and hardware access are key to replace or enhance modern software stacks with deep neural networks.