You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
As we not have a theoretical computational cost, if we have a theoretical specification of a device, we can estimate theoretical (upper-bound) of inference speed on the device.
It'd be nice if chainer_computational_cost can have a feature to receive a hardware profile and calculate layer-wise and total theoretical speed.
Problems
Interface design
Embed inference time inside layer report?
Pros: implementation is the simplest
Cons: Need to specify only one hardware profile to Hook instance.
Layer-specific accelerator
NVIDIA Volta has TensorCore, in V100 case convolution operation is 125TFLOPs, although other general arithmetics are 15TFLOPs.
If we use only one value as the hardware's theoretical performance, the estimated inference time is totally useless.
The text was updated successfully, but these errors were encountered:
belltailjp
changed the title
Theoretical speed estimation
[idea] Theoretical speed estimation
Oct 17, 2018
As we not have a theoretical computational cost, if we have a theoretical specification of a device, we can estimate theoretical (upper-bound) of inference speed on the device.
It'd be nice if chainer_computational_cost can have a feature to receive a hardware profile and calculate layer-wise and total theoretical speed.
Problems
The text was updated successfully, but these errors were encountered: