Apple Explains Its On-device Deep Neural Network for Face Detection

Apple Explains Its On-device Deep Neural Network for Face Detection

Posted by · 6299 views · Translate

Apple has published a new entry on its machine learning blog that explains how it uses an on-device deep neural network for face detection.

We faced several challenges. The deep-learning models need to be shipped as part of the operating system, taking up valuable NAND storage space. They also need to be loaded into RAM and require significant computational time on the GPU and/or CPU. Unlike cloud-based services, whose resources can be dedicated solely to a vision problem, on-device computation must take place while sharing these system resources with other running applications. Finally, the computation must be efficient enough to process a large Photos library in a reasonably short amount of time, but without significant power usage or thermal increase.

In the blog entry, Apple engineers discuss:
● how we fully leverage our GPU and CPU (using BNNS and Metal)
● memory optimizations for network inference, and image loading and caching
● how we implemented the network in a way that did not interfere with the multitude of other simultaneous tasks expected of iPhone.

Hit the link below for the full post!

Read More


Apple Explains Its On-device Deep Neural Network for Face Detection

Recent