FP8 versus INT8 for efficient deep learning inference