Configure Metrics (STL)
Metrics are used to monitor the training and validation process and to evaluate the model and algorithm during testing. If you are not familiar with continual learning metrics, feel free to learn more from my article: A Summary of Continual Learning Metrics.
Under the PyTorch Lightning framework, callbacks add additional actions at different points in the experiment, including before, during, or after training, validation, or testing. The metrics in CLArena are implemented as metric callbacks, which can:
- Calculate metrics and save their data to files.
- Visualize metrics as plots from the saved data.
- Log additional metrics during training process. (Note the majority of training metrics are handled by Lightning Loggers. See Configure Lightning Loggers (CL Main) section)
The details of these actions are configured by the metric callbacks. Each group of metrics is organized as one metric callback. For example, CLAccuracy
and CLLoss
correspond to accuracy and loss metrics for continual learning. We can apply multiple metrics at the same time.
Metrics are a sub-config under the experiment index config (CL Main). To configure custom metrics, create a YAML file in the metrics/
folder. At the moment, we only support a uniform metrics setting across all tasks. Below is an example of the metrics config.
Example
configs
βββ __init__.py
βββ entrance.yaml
βββ experiment
β βββ example_clmain_train.yaml
β βββ ...
βββ metrics
β βββ cl_default.yaml
...
configs/experiment/example_clmain_train.yaml
defaults:
...
- /metrics: cl_default.yaml
...
The metrics config is a list of metric callback objects:
configs/metrics/cl_default.yaml
- _target_: clarena.metrics.CLAccuracy
save_dir: ${output_dir}/results/
test_acc_csv_name: acc.csv
test_acc_matrix_plot_name: acc_matrix.png
test_ave_acc_plot_name: ave_acc.png
- _target_: clarena.metrics.CLLoss
save_dir: ${output_dir}/results/
test_loss_cls_csv_name: loss_cls.csv
test_loss_cls_matrix_plot_name: loss_cls_matrix.png
test_ave_loss_cls_plot_name: ave_loss_cls.png
Supported Metrics & Required Config Fields
In CLArena, we have implemented many metric callbacks as Python classes in the clarena.metrics
module that you can use for your experiments.
To choose a metric callback, assign the _target_
field to the corresponding class name, such as clarena.metrics.CLAccuracy
for CLAccuracy
. Each metric callback has its own hyperparameters and configurations, which means it has its own required fields. The required fields are the same as the arguments of the class specified by _target_
. The arguments for each metric callback class can be found in the API documentation.
Below is the full list of supported metric callbacks. These callbacks can only be applied to CL Main experiments. Note that the names in the βMetric Callbackβ column are the exact class names that you should assign to _target_
.
General
These metrics can be generally used unless noted otherwise.
Metric Callback | Description | Required Config Fields |
---|---|---|
CLAccuracy | Provides all actions that are related to CL accuracy metric, which include:
The callback is able to produce the following outputs: |
Same as CLAccuracy class arguments |
CLLoss | Provides all actions that are related to CL loss metrics, which include:
The callback is able to produce the following outputs:
|
Same as CLLoss class arguments |
Each CL algorithm may have their own metrics and variables to log. We have implemented specialized metrics for different CL algorithms.
HAT
These metrics should be used with CL algorithm HAT and its extensions AdaHAT, FGAdaHAT. Please refer to Configure CL Algorithm (CL Main) section.
Metric Callback | Description | Required Config Fields |
---|---|---|
HATMasks | Provides all actions that are related to masks of HAT (Hard Attention to the Task) algorithm and its extensions, which include:
The callback is able to produce the following outputs:
|
Same as HATMasks class arguments |
HATAdjustmentRate | Provides all actions that are related to adjustment rate of HAT (Hard Attention to the Task) algorithm and its extensions, which include:
The callback is able to produce the following outputs:
|
Same as HATAdjustmentRate class arguments |
HATNetworkCapacity | Provides all actions that are related to network capacity of HAT (Hard Attention to the Task) algorithm and its extensions, which include:
|
Same as HATNetworkCapacity class arguments |
Metrics are used to monitor training and validation process, and evaluate the model and algorithm during testing process.
Under the framework of PyTorch Lightning, callbacks are used to add additional actions and functionalities integrated in different timing of the experiment, which includes before, during, or after training, validating, or testing process. The metrics in our packages are implemented as metric callbacks, which can do:
- Calculate metrics and save their data to files.
- Visualize metrics as plots from the saved data.
The details of the actions can be configured by the metric callbacks. Each group of metrics is organized as one metric callback, for example, CULDistributionDistance
and CULAccuracyDifference
correspond to DD and AD metrics of continual unlearning. We can apply multiple metrics at the same time.
Continual unlearning is an experiment on top of continual learning with unlearning capabilities; therefore, it shares the same metrics with CL to measure regular CL performance, please refer to Configure Metrics (CL Main) section. The metrics to measure unlearning performance must be used in CUL full evaluation experiment, please refer to CUL Full Evaluation.
Metrics is a sub-config under the experiment index config (CUL Main), as well as experiment index config (CUL full evaluation). To configure custom metrics, you need to create a YAML file in metrics/
folder. At the moment, we only support uniform metrics across all tasks. Below shows examples of the metrics config.
Example
configs
βββ __init__.py
βββ entrance.yaml
βββ experiment
β βββ example_culmain_train.yaml
β βββ ...
βββ metrics
β βββ cl_default.yaml
...
configs/experiment/example_culmain_train.yaml
defaults:
...
- /metrics: cl_default.yaml
...
The metrics config is a list of metric callback objects:
configs/metrics/cl_default.yaml
- _target_: clarena.metrics.CLAccuracy
save_dir: ${output_dir}/results/
test_acc_csv_name: acc.csv
test_acc_matrix_plot_name: acc_matrix.png
test_ave_acc_plot_name: ave_acc.png
- _target_: clarena.metrics.CLLoss
save_dir: ${output_dir}/results/
test_loss_cls_csv_name: loss_cls.csv
test_loss_cls_matrix_plot_name: loss_cls_matrix.png
test_ave_loss_cls_plot_name: ave_loss_cls.png
Supported Metrics & Required Config Fields
In CLArena, we implemented many metric callbacks in clarena.metrics
module that you can use for CUL main and full evaluation experiment.
The _target_
field of each callback must be assigned to the corresponding class name, such as clarena.metrics.CLAccuracy
for CLAccuracy
. Each metric callback has its own required fields, which are the same as the arguments of the class specified by _target_
. The arguments of each metric callback class can be found in API documentation.
Below is the full list of supported metric callbacks. Note that the βMetric Callbackβ is exactly the class name that the _target_
field is assigned.
CL Metrics
These metric callbacks can be applied to CUL main experiment, because they can be applied to CL main experiment.
Please refer to Supported Metrics & Required Config Fields in Configure Metrics (CL Main) for more details.
Unlearning Metrics
These metric callbacks can be applied to CUL full evaluation experiment only.
Metric Callback | Description | Required Config Fields |
---|---|---|
CULDistributionDistance | Provides all actions that are related to CUL distribution distance (DD) metric, which include:
The callback is able to produce the following outputs:
|
Same as CULDistributionDistance class arguments |
CULAccuracyDifference | Provides all actions that are related to CUL accuracy difference (AD) metric, which include:
The callback is able to produce the following outputs:
|
Same as CULAccuracyDifference class arguments |
Under the framework of PyTorch Lightning, we use callbacks to add additional actions and functionalities integrated in different timing of the experiment. This includes before, during, or after training, validating, or testing process.
Each type of callback is designed for one set of actions, such as save sample images, generating logging information, etc. we can apply multiple callbacks to enables different sets of actions at the same time.
This section is only about configuring callbacks other than metric callbacks. To configure metric callbacks, please refer to Configure Metrics (CL Main) section.
Callbacks is a sub-config under the experiment index config (CUL Main) and experiment index config (CUL full evaluation). To configure custom callbacks, you need to create a YAML file in callbacks/
folder. At the moment, we only support uniform callbacks setting across all tasks. Below shows examples of the callbacks config.
Example
configs
βββ __init__.py
βββ entrance.yaml
βββ experiment
β βββ example_clmain_train.yaml
β βββ ...
βββ callback
β βββ cul_default.yaml
...
configs/experiment/example_culmain_train.yaml
defaults:
...
- /callbacks: cul_default.yaml
...
The callbacks config is a list of callback objects:
configs/callbacks/cul_default.yaml
- _target_: clarena.callbacks.CULPylogger
- _target_: clarena.callbacks.CLRichProgressBar
- _target_: clarena.callbacks.SaveFirstBatchImages
save_dir: ${output_dir}/samples/
- _target_: clarena.callbacks.SaveModels
save_dir: ${output_dir}/saved_models/
Supported Callbacks & Required Config Fields
All Lightning built-in callbacks are supported. In CLArena, we also implemented many callbacks in clarena.callbacks
module that you can use for your CUL Main experiment.
The _target_
field of each callback must be assigned to the corresponding class name, such as lightning.pytorch.callbacks.EarlyStopping
for EarlyStopping
. Each callback has its own required fields, which are the same as the arguments of the class specified by _target_
. The arguments of each callback class can be found in PyTorch Lightning documentation and CLArena API documentation.
PyTorch Lightning Documentation (Built-In Callbacks) API Reference (Callbacks) Source Code (Callbacks)
Below is the full list of supported callbacks that can be applied to CUL Main experiment. Note that the βCallbackβ is exactly the class name that the _target_
field is assigned.
Callback | Description | Required Config Fields |
---|---|---|
SaveModels | Saves the model at the end of each training task. Please refer to Save and Evaluate Model (CLMain) section. | Same as SaveModels class arguments |
CULPylogger | Provides additional logging messages for during continual unlearning progress. For example, at the start and end of each training and testing task, log the task ID and other relevant information. | Same as CULPylogger class arguments |
SaveFirstBatchImages | Saves images and labels of the first batch of training data into files. Applies to all tasks. | Same as SaveFirstBatchImages class arguments |
CLRichProgressBar | Customised RichProgressBar for continual learning. |
Same as CLRichProgressBar class arguments |
The callbacks that can be applied to CUL full evaluation experiment include CULPylogger
and CLRichProgressBar
.
Metrics are used to monitor training and validation process, and evaluate the model and algorithm during testing process.
Under the framework of PyTorch Lightning, callbacks are used to add additional actions and functionalities integrated in different timing of the experiment, which includes before, during, or after training, validating, or testing process. The metrics in our packages are implemented as metric callbacks, which can do:
- Calculate metrics and save their data to files.
- Visualize metrics as plots from the saved data.
- Log additional metrics during training process. (Note the majority of training metrics are handled by Lightning Loggers. See Configure Lightning Loggers section)
The details of the actions can be configured by the metric callbacks. Each group of metrics is organized as one metric callback, for example, MTLAccuracy
and MTLLoss
correspond to accuracy and loss metrics of multi-task learning. We can apply multiple metrics at the same time.
Metrics is a sub-config under the experiment index config (MTL). To configure custom metrics, you need to create a YAML file in metrics/
folder. At the moment, we only support uniform metrics across all tasks. Below shows examples of the metrics config.
Example
configs
βββ __init__.py
βββ entrance.yaml
βββ experiment
β βββ example_mtl_train.yaml
β βββ ...
βββ metrics
β βββ mtl_default.yaml
...
configs/experiment/example_mtl_train.yaml
defaults:
...
- /metrics: mtl_default.yaml
...
The metrics config is a list of metric callback objects:
configs/metrics/mtl_default.yaml
- _target_: clarena.metrics.MTLAccuracy
save_dir: ${output_dir}/results/
test_acc_csv_name: acc.csv
test_ave_acc_plot_name: ave_acc.png
- _target_: clarena.metrics.MTLLoss
save_dir: ${output_dir}/results/
test_loss_cls_csv_name: loss_cls.csv
test_ave_loss_cls_plot_name: ave_loss_cls.png
Supported Metrics & Required Config Fields
In CLArena, we implemented many metric callbacks in clarena.metrics
module that you can use for MTL experiment.
The _target_
field of each callback must be assigned to the corresponding class name, such as clarena.metrics.MTLAccuracy
for MTLAccuracy
. Each metric callback has its own required fields, which are the same as the arguments of the class specified by _target_
. The arguments of each metric callback class can be found in API documentation.
Below is the full list of supported metric callbacks. These callbacks can only be applied to MTL experiment. Note that the βMetric Callbackβ is exactly the class name that the _target_
field is assigned.
General
These metrics can be generally used unless noted otherwise.
Callback | Description | Required Config Fields |
---|---|---|
MTLAccuracy | Provides all actions that are related to MTL accuracy metric, which include:
The callback is able to produce the following outputs:
|
Same as MTLAccuracy class arguments |
MTLLoss | Provides all actions that are related to MTL loss metrics, which include:
The callback is able to produce the following outputs:
|
Same as MTLLoss class arguments |
Metrics are used to monitor training and validation process, and evaluate the model and algorithm during testing process.
Under the framework of PyTorch Lightning, callbacks are used to add additional actions and functionalities integrated in different timing of the experiment, which includes before, during, or after training, validating, or testing process. The metrics in our packages are implemented as metric callbacks, which can do:
- Calculate metrics and save their data to files.
- Visualize metrics as plots from the saved data.
- Log additional metrics during training process. (Note the majority of training metrics are handled by Lightning Loggers. See Configure Lightning Loggers section)
The details of the actions can be configured by the metric callbacks. Each group of metrics is organized as one metric callback, for example, STLAccuracy
and STLLoss
correspond to accuracy and loss metrics of single-task learning. We can apply multiple metrics at the same time.
Metrics is a sub-config under the experiment index config (STL). To configure custom metrics, you need to create a YAML file in metrics/
folder. At the moment, we only support uniform metrics across all tasks. Below shows examples of the metrics config.
Example
configs
βββ __init__.py
βββ entrance.yaml
βββ experiment
β βββ example_stl_train.yaml
β βββ ...
βββ metrics
β βββ stl_default.yaml
...
configs/experiment/example_stl_train.yaml
defaults:
...
- /metrics: stl_default.yaml
...
The metrics config is a list of metric callback objects:
configs/metrics/mtl_default.yaml
- _target_: clarena.metrics.STLAccuracy
save_dir: ${output_dir}/results/
test_acc_csv_name: acc.csv
- _target_: clarena.metrics.STLLoss
save_dir: ${output_dir}/results/
test_loss_cls_csv_name: loss_cls.csv
Supported Metrics & Required Config Fields
In CLArena, we implemented many metric callbacks in clarena.metrics
module that you can use for STL experiment.
The _target_
field of each callback must be assigned to the corresponding class name, such as clarena.metrics.STLAccuracy
for STLAccuracy
. Each metric callback has its own required fields, which are the same as the arguments of the class specified by _target_
. The arguments of each metric callback class can be found in API documentation.
Below is the full list of supported metric callbacks. These callbacks can only be applied to STL experiment. Note that the βMetric Callbackβ is exactly the class name that the _target_
field is assigned.
General
These metrics can be generally used unless noted otherwise.
Callback | Description | Required Config Fields |
---|---|---|
STLAccuracy | Provides all actions that are related to STL accuracy metric, which include:
The callback is able to produce the following outputs:
|
Same as STLAccuracy class arguments |
STLLoss | Provides all actions that are related to STL loss metrics, which include:
The callback is able to produce the following outputs:
|
Same as STLLoss class arguments |