Module: tfmot.quantization.keras.experimental.default_n_bit.default_n_bit_transforms

Module containing N-bit default transforms.

Classes

class ConcatTransform: Transform for Concatenate. Quantize only after concatenation.

class ConcatTransform3Inputs: Transform for 3 inputs Concatenate.

class ConcatTransform4Inputs: Transform for 4 inputs Concatenate.

class ConcatTransform5Inputs: Transform for 5 inputs Concatenate.

class ConcatTransform6Inputs: Transform for 6 inputs Concatenate.

class Conv2DBatchNormActivationQuantize: Ensure FQ does not get placed between Conv, BatchNorm and ReLU.

class Conv2DBatchNormQuantize: Ensure FQ does not get placed between Conv and BatchNorm.

class Conv2DBatchNormReLUQuantize: Ensure FQ does not get placed between Conv, BatchNorm and ReLU.

class Conv2DReshapeBatchNormActivationQuantize: Ensure FQ does not get placed between Conv, BatchNorm and ReLU.

class Conv2DReshapeBatchNormQuantize: Ensure FQ does not get placed between Conv, Reshape and BatchNorm.

class Conv2DReshapeBatchNormReLUQuantize: Ensure FQ does not get placed between Conv, BatchNorm and ReLU.

class InputLayerQuantize: Quantizes InputLayer, by adding QuantizeLayer after it.

class LayerReLUQuantize: Ensure FQ does not get placed between Add and ReLU.

class LayerReluActivationQuantize: Ensure FQ does not get placed between Add and ReLU.

class SeparableConv1DQuantize: Add QAT support for Keras SeparableConv1D layer.

class SeparableConvQuantize: Break SeparableConv into a DepthwiseConv and Conv layer.