Tesla does INT8 inference. Way more efficient than FP16。。。
Tesla does INT8 inference. Way more efficient than FP16, but took us a lot of effort to overcome quantization errors.
— Elon Musk (@elonmusk) February 28, 2023
戳这里 Claim your page
来源: 文学城-蓝调