Speeding Up Deep Learning Inference on Edge Devices

When it comes to AI based applications, there is a need to counter latency constraints and strategize to speed up the inference. There are few techniques that can be leveraged namely Weight Pruning, Quantization, and Weight sharing among others that can help in speeding up an inference on edge.

Reading Time: 5 minutes
Read the article  

ABOUT THE AUTHOR

Anand Borad

Anand Borad works as senior marketing executive and takes care of digital and content marketing efforts for Medical devices, Connected Retail & Healthcare and New product development initiatives. He enjoys learning newer technologies and adopting it into everyday marketing practices.