-
Notifications
You must be signed in to change notification settings - Fork 3
Open
Description
As per the architecture definition provided here, it is shown that the batch normalization is used in inverted-residual blocks as below.
- Bottleneck layer which is normal convolutional layer is equipped with batch norm
- Depthwise layer is also equipped with batch norm
- Pointwise layer is also equipped with batch norm.
Even in their paper they mentioned that they use batch norm after every layer.
But when we download a pretrained model from tensorflow and visualize it in Netron, it is as per below.
- Bottleneck layer doesnt use batch norm instead use bias.
- Depthwise is equipped with batch norm.
- Pointwise layer doesnt use batch norm instead use bias.
So this make a huge difference in number of parameters and final accuracy.
Metadata
Metadata
Assignees
Labels
No labels