Module: tfm.optimization.optimizer_factory

Optimizer factory class.

Classes

class OptimizerFactory: Optimizer factory class.

Functions

register_optimizer_cls(...): Register customize optimizer cls.

LEGACY_OPTIMIZERS_CLS

{
 'adafactor': 'Unimplemented',
 'adagrad': <class 'keras.src.optimizers.legacy.adagrad.Adagrad'>,
 'adam': <class 'keras.src.optimizers.legacy.adam.Adam'>,
 'adam_experimental': <class 'keras.src.optimizers.adam.Adam'>,
 'adamw': <class 'official.modeling.optimization.legacy_adamw.AdamWeightDecay'>,
 'adamw_experimental': <class 'keras.src.optimizers.adamw.AdamW'>,
 'lamb': <class 'official.modeling.optimization.lamb.LAMB'>,
 'lars': <class 'official.modeling.optimization.lars.LARS'>,
 'rmsprop': <class 'keras.src.optimizers.legacy.rmsprop.RMSprop'>,
 'sgd': <class 'keras.src.optimizers.legacy.gradient_descent.SGD'>,
 'sgd_experimental': <class 'keras.src.optimizers.sgd.SGD'>,
 'slide': 'Unimplemented'
}

LR_CLS

{
 'cosine': <class 'official.modeling.optimization.lr_schedule.CosineDecayWithOffset'>,
 'exponential': <class 'official.modeling.optimization.lr_schedule.ExponentialDecayWithOffset'>,
 'polynomial': <class 'official.modeling.optimization.lr_schedule.PolynomialDecayWithOffset'>,
 'power': <class 'official.modeling.optimization.lr_schedule.DirectPowerDecay'>,
 'power_linear': <class 'official.modeling.optimization.lr_schedule.PowerAndLinearDecay'>,
 'power_with_offset': <class 'official.modeling.optimization.lr_schedule.PowerDecayWithOffset'>,
 'step_cosine_with_offset': <class 'official.modeling.optimization.lr_schedule.StepCosineDecayWithOffset'>,
 'stepwise': <class 'official.modeling.optimization.lr_schedule.PiecewiseConstantDecayWithOffset'>
}

NEW_OPTIMIZERS_CLS

{
 'adafactor': 'Unimplemented',
 'adagrad': <class 'keras.src.optimizers.adagrad.Adagrad'>,
 'adam': <class 'keras.src.optimizers.adam.Adam'>,
 'adam_experimental': <class 'keras.src.optimizers.adam.Adam'>,
 'adamw': <class 'official.modeling.optimization.legacy_adamw.AdamWeightDecay'>,
 'adamw_experimental': <class 'keras.src.optimizers.adamw.AdamW'>,
 'lamb': <class 'official.modeling.optimization.lamb.LAMB'>,
 'lars': <class 'official.modeling.optimization.lars.LARS'>,
 'rmsprop': <class 'keras.src.optimizers.rmsprop.RMSprop'>,
 'sgd': <class 'keras.src.optimizers.sgd.SGD'>,
 'sgd_experimental': <class 'keras.src.optimizers.sgd.SGD'>,
 'slide': 'Unimplemented'
}

SHARED_OPTIMIZERS

{
 'adafactor': 'Unimplemented',
 'adam_experimental': <class 'keras.src.optimizers.adam.Adam'>,
 'adamw': <class 'official.modeling.optimization.legacy_adamw.AdamWeightDecay'>,
 'adamw_experimental': <class 'keras.src.optimizers.adamw.AdamW'>,
 'lamb': <class 'official.modeling.optimization.lamb.LAMB'>,
 'lars': <class 'official.modeling.optimization.lars.LARS'>,
 'sgd_experimental': <class 'keras.src.optimizers.sgd.SGD'>,
 'slide': 'Unimplemented'
}

WARMUP_CLS

{
 'linear': <class 'official.modeling.optimization.lr_schedule.LinearWarmup'>,
 'polynomial': <class 'official.modeling.optimization.lr_schedule.PolynomialWarmUp'>
}