It is always useful to investigate what causes an experiment to succeed. The ClearML Web UI provides a deep experiment comparison, allowing to locate, visualize, and analyze differences including:
- Artifacts - Input model, output model, and model design.
- Execution - Installed packages and source code.
- Configuration - Configuration objects used by the experiment.
- Info - General experiment details.
- Plots are combined to have multiple lines from different experiments (for example multiple RoC curves laid on top of each other).
- Debug samples
- Debug samples by each iteration
- Examine samples with a viewer (for images and video), and a player (for audio) (see reporting media).
The ClearML experiment comparison provides comparison features making it easy to compare experiments.
To select experiments to compare:
- Go to an experiments table, which includes the experiments to be compared.
- Select the experiments to compare, select the checkboxes individually or select the top checkbox for all experiments. After selecting the second checkbox, a bottom bar appears.
- In the bottom bar, click COMPARE. The comparison page appears, showing a column for each experiment and differences with a highlighted background color. The experiment on the left is the base experiment. Other experiments compare to the base experiment.
The DETAILS tab includes deep comparisons of the following:
- Artifacts - Input/output models and their configuration, and other artifacts, if any.
- Execution Details - Source code, uncommitted changes, and installed python packages. See details here.
- Configuration - Configuration objects used by the experiment, sorted by sections. See details here.
- Info - General experiment details. This includes information describing the stored experiment. See details here.
To Locate the Source Differences:
Click the DETAILS tab > Expand highlighted sections, or, in the header, click (Previous diff) or (Next diff).
For example, in the image below, expanding ARTIFACTS > Output Model > Model shows that the model names are different.
Compare hyperparameters as values, or compare by metric (hyperparameter parallel coordinate comparison).
The Values mode is a side-by-side comparison that shows hyperparameter value differences highlighted line-by-line.
To view a side by side values comparison:
- Click the HYPER PARAMETERS tab.
- In the dropdown menu (on the upper left, next to + Add Experiments), choose Values.
- To show only differences, move the Hide Identical Fields slider to on.
- Locate differences by either:
- Clicking (Previous diff) or (Next diff).
- Scrolling to see highlighted hyperparameters.
For example, expanding General shows that the
epochs differ between the experiments.
In the Parallel Coordinates mode, compare a metric to any combination of hyperparameters using a parallel coordinates plot.
To compare by metric:
- Click the HYPER PARAMETERS tab.
- In the dropdown menu (on the upper left, next to + Add Experiments), choose Parallel Coordinates.
- In Performance Metric, expand a metric or monitored resource, and then click a variant.
- Select the metric values to use. Choose one of the following:
- LAST - The final value, or the most recent value, for in-progress experiments.
- MIN - Minimal value.
- MAX - Maximal value.
- In Parameters, select the hyperparameter checkboxes to compare.
- To view one experiment on the plot, hover over the experiment name in the legend.
For example, plot the metric/variant
validation: epoch_accuracy against the hyperparameters
Hover over one of the experiment names in the legend, and the plot shows only that data.
Visualize the comparison of scalars, which includes metrics and monitored resources in the SCALARS tab.
To compare specific values:
- Click the SCALARS tab.
- In the dropdown menu (upper right of the left sidebar), choose either:
- Last Values (the final or most recent value)
- Min Values (the minimal values)
- Max Values (the maximal values)
- Sort by variant.
Compare scalar series in plots and analyze differences using ClearML Web UI plot tools.
To compare scalar series:
- Click the SCALARS tab.
- In the dropdown menu (upper right of the left sidebar), choose Graph.
To improve scalar series analysis:
- In Group by, select one of these option:
- Metric - all variants for a metric on the same plot.
- Metric+Variant - every variant appears on its own plot.
- Horizontal axis options:
- Relative time since the experiment began,
- Wall (clock time).
- Smooth a curve - move the Smoothing slider or type in a smoothing number from 0 to 0.999.
- Use plot controls, which appear when you hover over the top of a plot.
- Hide / show scalar plots - Click Hide all and .
- Filter scalars by full or partial scalar name.
This image shows scalars grouped by metric.
This image shows scalars grouped by metric and variant.
Visualize the comparison of any data that ClearML automatically captures or that is explicitly reported in experiments, in the PLOTS tab.
To compare plots:
Click the PLOTS tab.
To improve your comparison, use either of the following:
- To locate scalars, click HIDE ALL, and then (show) to choose which scalars to see. Scalars can also be filtered by full or partial scalar name, using the search bar.
- Use any of the plot controls that appear when hovering over the top of a plot, including:
- Downloading the image
- Downloading the data as JSON
- Switching between logarithmic / linear scale.
Compare debug samples at any iteration to verify that an experiment is running as expected. The most recent iteration appears first. Use the viewer / player to inspect images, audio, video samples and do any of the following:
- Move to the same sample in a different iteration (move the iteration slider).
- Show the next or previous iteration's sample.
- Download the file .
- View the sample's iteration number, width, height, and coordinates.
To compare debug samples:
Click the DEBUG SAMPLES tab. The most recent iteration appears at the top.
Locate debug samples by doing the following:
- Filter by metric. In the Metric list, choose a metric.
- Show other iterations. Click (Older images), (New images), or (Newest images).
To open a debug sample (image, audio, or video) in the viewer or player, click the thumbnail.
To move to the same sample in another iteration, click (previous), (next), or move the slider.
To view a debug sample in the viewer / player:
Click the debug sample thumbnail.
Do any of the following:
- Move to the same sample in another iteration - Move the slider, or click < (previous) or > (next).
- Download the file - Click .
- For images, locate a position on the sample - Hover over the sample and the X, Y coordinates appear in the legend below the sample.
To assist in experiment analysis, the comparison page supports:
- Changing compared experiments selection
- Finding the next or previous difference.
- Hiding identical fields
- Searching all text
- Choosing a different base experiment
- Dynamic ordering of the compared experiments
- Sharing experiments
- Auto refresh
- Click + Add Experiment in the top left corner of any of the comparison pages. This will open up a window with an experiment table with the currently compared experiments at the top.
- Find the experiments to add by sorting and filtering the experiments with the appropriate column header controls. Alternatively, use the search bar to find experiments by name.
- Select experiments to include in the comparison (and / or clear the selection of any experiment you wish to remove).
- Click APPLY.
- Find the previous difference , or the next difference .
Move the Hide Identical Fields slider to "on" mode to see only fields that are different.
Search all text in the comparison.
Show differences in other experiments in reference to a new base experiment. To set a new base experiment, do one of the following:
- Click on on the top right of the experiment that will be the new base.
- Click on the new base experiment and drag it all the way to the left
To reorder the experiments being compared, press on the top right of the experiment that needs to be moved, and drag the experiment to its new position.
Remove an experiment from the comparison, by pressing on the top right of the experiment that needs to be removed.
To share a comparison table, copy the full URL from the address bar and send it to a teammate to collaborate. They will get the exact same page (including selected tabs etc.).