This callback writes a log for TensorBoard, which allows you to visualize dynamic graphs of your training and test metrics, as well as activation histograms for the different layers in your model.
callback_tensorboard(log_dir = NULL, histogram_freq = 0, batch_size = 32, write_graph = TRUE, write_grads = FALSE, write_images = FALSE, embeddings_freq = 0, embeddings_layer_names = NULL, embeddings_metadata = NULL, embeddings_data = NULL, update_freq = "epoch")
The path of the directory where to save the log files to be
parsed by Tensorboard. The default is
frequency (in epochs) at which to compute activation histograms for the layers of the model. If set to 0, histograms won't be computed.
size of batch of inputs to feed to the network for histograms computation.
whether to visualize the graph in Tensorboard. The log
file can become quite large when write_graph is set to
whether to visualize gradient histograms in TensorBoard.
whether to write model weights to visualize as image in Tensorboard.
frequency (in epochs) at which selected embedding layers will be saved.
a list of names of layers to keep eye on. If
a named list which maps layer name to a file name in which metadata for this embedding layer is saved. See the details about the metadata file format. In case if the same metadata file is used for all embedding layers, string can be passed.
Data to be embedded at layers specified in
TensorBoard is a visualization tool provided with TensorFlow.
You can find more information about TensorBoard here.
When using a backend other than TensorFlow, TensorBoard will still work (if you have TensorFlow installed), but the only feature available will be the display of the losses and metrics plots.