- Notifications
You must be signed in to change notification settings - Fork265
-
Hi, 1st) QuantReLU(act_quant=Uint8ActPerTensorFloat,bit_width=act_bit_width,per_channel_broadcastable_shape=(1,oup,1,1),scaling_stats_permute_dims=(1,0,2,3),scaling_per_output_channel=True) I tried many absorb/reorder transformations to convert the remaining Mul node, with no success, as activations QuantNodes are not scalar/1D anymore, but tensors. Is activation per channel supported by FINN? If so, which is the sequence of transformations to streamline the model and absorb those Mul nodes, as it happens when activation is per tensor? 2nd) Somehow, I tried the per channel activation to solve it, but it did not work. Is there any way to limit the decimals of those values? Using BatchNorm2dToQuantScaleBias in Brevitas? Limiting in Pytorch? It worked for me configuring QuantReLU as relu6, with min and max val and fixed scale, but causing undesirable accuracy drop. |
BetaWas this translation helpful?Give feedback.
All reactions
👀 1