You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Jan 3, 2023. It is now read-only.
The performance of the VGG16 network imported to ngraph via TF frontend appears to be very very slow. To get the perspective, I have 3 implementations of VGG16:
Neon-Ng : Neon frontend in ngraph ( VGG16 using layers of neon frontend)
TF : TensorFlow code (VGG16 using tf.nn.conv2d, tf.nn.bias_add, tf.nn.relu/max_pool etc)
TF-Ng : I import the checkpoint files of the implementation 2 here .
And for batch size of 64 on a skylake machine I get the following performances:
Implementation 1: 1561 GFlops/s
Implementation 2: 1309 GFlops/s ( Using tensorflow 1.4.0-dev on intel python)
Implementation 3: 51.82 GFlops/s.
I am trying to investigate why the TF frontend in ngraph is slow. Any inputs would be very helpful.