RSVP for your your local TensorFlow Everywhere event today!

tf.config.experimental.get_memory_info

Get memory info for the chosen device, as a dict.

This function returns a dict containing information about the device's memory usage. For example:

if tf.config.list_physical_devices('GPU'):
  # Returns a dict in the form {'current': <current mem usage>,
  #                             'peak': <peak mem usage>}
  tf.config.experimental.get_memory_info('GPU:0')

Currently returns the following keys: 'current': The current memory used by the device, in bytes. 'peak': The peak memory used by the device across the run of the program, in bytes.

More keys may be added in the future, including device-specific keys.

Currently raises an exception for the CPU.

For GPUs, TensorFlow will allocate all the memory by default, unless changed with tf.config.experimental.set_memory_growth. The dict specifies only the current and peak memory that TensorFlow is actually using, not the memory that TensorFlow has allocated on the GPU.

device Device string to get the memory information for, e.g. "GPU:0". See https://www.tensorflow.org/api_docs/python/tf/device for specifying device strings.

A dict with keys 'current' and 'peak', specifying the current and peak memory usage respectively.

ValueError Non-existent or CPU device specified.