Get started with TensorFlow model optimization

Stay organized with collections Save and categorize content based on your preferences.

1. Choose the best model for the task

Depending on the task, you will need to make a tradeoff between model complexity and size. If your task requires high accuracy, then you may need a large and complex model. For tasks that require less precision, it is better to use a smaller model because they not only use less disk space and memory, but they are also generally faster and more energy efficient.

2. Pre-optimized models

See if any existing TensorFlow Lite pre-optimized models provide the efficiency required by your application.

3. Post-training tooling

If you cannot use a pre-trained model for your application, try using TensorFlow Lite post-training quantization tools during TensorFlow Lite conversion, which can optimize your already-trained TensorFlow model.

See the post-training quantization tutorial to learn more.

Next steps: Training-time tooling

If the above simple solutions don't satisfy your needs, you may need to involve training-time optimization techniques. Optimize further with our training-time tools and dig deeper.