Hi,
I have an Iphone 8 model no
mq6h2gh/a
This phone has an A11 bionic chip. I created a toy caffe model with just two convolutional layers.
I calculated the computing needs for this net is ~150 million MAC.
The Neural Engine is supposed to have 600 billion MAC capacity (or half of it if one MAC is 2 operations).
However my toy network runs for 40-50 msecond! This implies 3-4 GMAC/sec computing capacity, way below the advertised 600GMAC/sec.
I checked and the useCPUOnly feature of MLModel is set to false.
How do I know that CoreML uses the built in neural engine?
For reference the same model did 18 ms forward pass on a 960M Nvidia GPU.
Regards
Gabor