快要看不懂了。。。 NVIDIA’s Deep Learning Accelerator (DLA) With this context in mind, then, it makes sense for NVIDIA to build a fixed function accelerator that acts as an efficient inference engine as part of a larger solution. NVIDIA announced that its next generation DrivePX platform for autonomous vehicles, the Xavier SOC, would consist of ARM CPU cores, Volta GPU cores, and a fixed function Deep Learning Accelerator (DLA) for inference. This approach, the company says, will result in higher performance at lower power, while maintaining the flexibility for customization that its automotive OEMs demand. NVIDIA has always focused on solving very hard, computationally complex problems. So it has no interest in designing, as Huang puts it, a deep learning chip for smart lawn mowers, or a deep learning chip for refrigerators, or a deep learning chip for streetlamps. (All might be different.) By deciding to open source the DLA, NVIDIA is enabling its rich deep learning ecosystem to extend to include low cost, high volume and low power ASICs and SOCs, allowing other companies and researchers to build their own chips using this accelerator. And of course it all runs the same CUDA software used by NVIDIA GPUs. The company, in effect, is saying, “OK, if you want to build a TPU for your little widget, it’s probably best to build it on our technology, since nobody knows more about accelerating AI than NVIDIA.” Meanwhile NVIDIA can focus on building high margin, high value platforms needed in the datacenter and at the edge. 如果inference部分统一了,狗狗的TPU嗝屁了。那智能电冰箱,智能割草机的training 怎么弄?总不能放到cloud上去做training吧。。。
【在 p**f 的大作中提到】 : 快要看不懂了。。。 : NVIDIA’s Deep Learning Accelerator (DLA) : With this context in mind, then, it makes sense for NVIDIA to build a fixed : function accelerator that acts as an efficient inference engine as part of a : larger solution. NVIDIA announced that its next generation DrivePX platform : for autonomous vehicles, the Xavier SOC, would consist of ARM CPU cores, : Volta GPU cores, and a fixed function Deep Learning Accelerator (DLA) for : inference. This approach, the company says, will result in higher : performance at lower power, while maintaining the flexibility for : customization that its automotive OEMs demand.
C*5
21 楼
当然是用GPU训练好了再deploy到DLA啊,而且deploy之前要用TensorRT优化一下。
fixed a platform
【在 p**f 的大作中提到】 : 快要看不懂了。。。 : NVIDIA’s Deep Learning Accelerator (DLA) : With this context in mind, then, it makes sense for NVIDIA to build a fixed : function accelerator that acts as an efficient inference engine as part of a : larger solution. NVIDIA announced that its next generation DrivePX platform : for autonomous vehicles, the Xavier SOC, would consist of ARM CPU cores, : Volta GPU cores, and a fixed function Deep Learning Accelerator (DLA) for : inference. This approach, the company says, will result in higher : performance at lower power, while maintaining the flexibility for : customization that its automotive OEMs demand.