This document introduces Estimators--a high-level TensorFlow API that greatly simplifies machine learning programming. Estimators encapsulate the following actions:
- export for serving
You may either use the pre-made Estimators we provide or write your
own custom Estimators. All Estimators--whether pre-made or custom--are
classes based on the
Advantages of Estimators
Estimators provide the following benefits:
- You can run Estimators-based models on a local host or on a distributed multi-server environment without changing your model. Furthermore, you can run Estimators-based models on CPUs, GPUs, or TPUs without recoding your model.
- Estimators simplify sharing implementations between model developers.
- You can develop a state of the art model with high-level intuitive code, In short, it is generally much easier to create models with Estimators than with the low-level TensorFlow APIs.
- Estimators are themselves built on tf.layers, which simplifies customization.
- Estimators build the graph for you. In other words, you don't have to build the graph.
- Estimators provide a safe distributed training loop that controls how and
- build the graph
- initialize variables
- start queues
- handle exceptions
- create checkpoint files and recover from failures
- save summaries for TensorBoard
When writing an application with Estimators, you must separate the data input pipeline from the model. This separation simplifies experiments with different data sets.
Pre-made Estimators enable you to work at a much higher conceptual level
than the base TensorFlow APIs. You no longer have to worry about creating
the computational graph or sessions since Estimators handle all
the "plumbing" for you. That is, pre-made Estimators create and manage
Session objects for you. Furthermore, pre-made Estimators
let you experiment with different model architectures by making only minimal
DNNClassifier, for example, is a pre-made Estimator class that
trains classification models through dense, feed-forward neural networks.
Structure of a pre-made Estimators program
A TensorFlow program relying on a pre-made Estimator typically consists of the following four steps:
Write one or more dataset importing functions. For example, you might create one function to import the training set and another function to import the test set. Each dataset importing function must return two objects:
- a dictionary in which the keys are feature column names and the values are Tensors (or SparseTensors) containing the corresponding feature data
- a Tensor containing one or more labels
For example, the following code illustrates the basic skeleton for an input function:
def input_fn(dataset): ... # manipulate dataset, extracting feature names and the label return feature_dict, label
See Using the
DatasetAPI for TensorFlow Input Pipelines for full details.)
Define the feature columns. Each
tf.feature_columnidentifies a feature name, its type, and any input pre-processing. For example, the following snippet creates three feature columns that hold integer or floating-point data. The first two feature columns simply identify the feature's name and type. The third feature column also specifies a lambda the program will invoke to scale the raw data:
# Define three numeric feature columns. population = tf.feature_column.numeric_column('population') crime_rate = tf.feature_column.numeric_column('crime_rate') median_education = tf.feature_column.numeric_column('median_education', normalizer_fn='lambda x: x - global_education_mean')
Instantiate the relevant pre-made Estimator. For example, here's a sample instantiation of a pre-made Estimator named
# Instantiate an estimator, passing the feature columns. estimator = tf.estimator.Estimator.LinearClassifier( feature_columns=[population, crime_rate, median_education], )
Call a training, evaluation, or inference method. For example, all Estimators provide a
trainmethod, which trains a model.
# my_training_set is the function created in Step 1 estimator.train(input_fn=my_training_set, steps=2000)
Benefits of pre-made Estimators
Pre-made Estimators encode best practices, providing the following benefits:
- Best practices for determining where different parts of the computational graph should run, implementing strategies on a single machine or on a cluster.
- Best practices for event (summary) writing and universally useful summaries.
If you don't use pre-made Estimators, you must implement the preceding features yourself.
The heart of every Estimator--whether pre-made or custom--is its model function, which is a method that builds graphs for training, evaluation, and prediction. When you are using a pre-made Estimator, someone else has already implemented the model function. When relying on a custom Estimator, you must write the model function yourself. A companion document explains how to write the model function.
We recommend the following workflow:
- Assuming a suitable pre-made Estimator exists, use it to build your first model and use its results to establish a baseline.
- Build and test your overall pipeline, including the integrity and reliability of your data with this pre-made Estimator.
- If suitable alternative pre-made Estimators are available, run experiments to determine which pre-made Estimator produces the best results.
- Possibly, further improve your model by building your own custom Estimator.