FP8 versus INT8 for efficient deep learning inference

Open in new window