MetricTask¶
MetricTask
is a base class for generating lsst.verify.Measurement
given input data.
Each MetricTask
class accepts specific type(s) of datasets and produces measurements for a specific metric or family of metrics.
MetricTask
is a PipelineTask
and can be executed as part of pipelines.
Python API summary¶
from lsst.verify.tasks.metricTask import MetricTask
-
class
MetricTask
(**kwargs) A base class for tasks that compute one metric from input datasets
...
- attributeconfig
Access configuration fields and retargetable subtasks.
See also
See the MetricTask
API reference for complete details.
Butler datasets¶
Output datasets¶
measurement
The value of the metric. The dataset type should not be configured directly, but should be set changing the
package
andmetric
template variables to the metric’s namespace (package, by convention) and in-package name, respectively.MetricTask
subclasses that only support one metric should set these variables automatically.
Retargetable subtasks¶
No subtasks.
Configuration fields¶
connections¶
- Data type
lsst.pipe.base.config.MetricConfigConnections
- Field type
Configurations describing the connections of the PipelineTask to datatypes
saveLogOutput¶
Flag to enable/disable saving of log output for a task, enabled by default.
saveMetadata¶
Flag to enable/disable metadata saving for a task, enabled by default.
In Depth¶
Subclassing¶
MetricTask
is primarily customized using the run
method.
The task config should use lsst.pipe.base.PipelineTaskConnections
to identify input datasets; MetricConfig
handles the output dataset.
Error Handling¶
In general, a MetricTask
may run in three cases:
the task can compute the metric without incident.
the task does not have the data required to compute the metric. This can happen with metadata- or table-based metrics if the user runs generic metric configurations on arbitrary pipelines, or if they make changes to the pipeline configuration that enable or disable processing steps. Middleware automatically handles the case where an entire dataset is missing.
the task has the data it needs, but cannot compute the metric. This could be because the data are corrupted, because the selected algorithm fails, or because the metric is ill-defined given the data.
A MetricTask
must distinguish between these cases so that calling frameworks can handle them appropriately.
A task for a metric that does not apply to a particular pipeline run (case 2) must either raise NoWorkFound
or return None
; it must not return a dummy value or raise a different exception.
A task that cannot give a valid result (case 3) must raise MetricComputationError
.
In grey areas, developers should choose a MetricTask
’s behavior based on whether the root cause is closer to case 2 or case 3.
For example, TimingMetricTask
accepts top-level task metadata as input, but returns None
if it can’t find metadata for the subtask it is supposed to time.
The subtask metadata are most likely missing because the subtask was never run, making the situation equivalent to case 2.
On the other hand, metadata with nonsense values falls squarely under case 3.