ConsolidateResourceUsageConnections¶
- class lsst.analysis.tools.tasks.ConsolidateResourceUsageConnections(*, config: PipelineTaskConfig | None = None)¶
- Bases: - PipelineTaskConnections- Connection definitions for - ConsolidateResourceUsageTask.- Attributes Summary - Mapping holding all connection attributes. - Set of dimension names that define the unit of work for this task. - Set with the names of all - InitInputconnection attributes.- Set with the names of all - InitOutputconnection attributes.- Set with the names of all - connectionTypes.Inputconnection attributes.- Connection for output dataset. - Set with the names of all - Outputconnection attributes.- Set with the names of all - PrerequisiteInputconnection attributes.- Methods Summary - adjustQuantum(inputs, outputs, label, data_id)- Override to make adjustments to - lsst.daf.butler.DatasetRefobjects in the- lsst.daf.butler.Quantumduring the graph generation stage of the activator.- buildDatasetRefs(quantum)- Build - QuantizedConnectioncorresponding to input- Quantum.- Return the names of regular input and output connections whose data IDs should be used to compute the spatial bounds of this task's quanta. - Return the names of regular input and output connections whose data IDs should be used to compute the temporal bounds of this task's quanta. - Attributes Documentation - allConnections: Mapping[str, BaseConnection] = {'output_table': Output(name='ResourceUsageSummary', storageClass='DataFrame', doc='Consolidated table of resource usage statistics. One row per task label', multiple=False, deprecated=None, _deprecation_context='', dimensions=(), isCalibration=False)}¶
- Mapping holding all connection attributes. - This is a read-only view that is automatically updated when connection attributes are added, removed, or replaced in - __init__. It is also updated after- __init__completes to reflect changes in- inputs,- prerequisiteInputs,- outputs,- initInputs, and- initOutputs.
 - defaultTemplates = {}¶
 - deprecatedTemplates = {}¶
 - dimensions: set[str] = {}¶
- Set of dimension names that define the unit of work for this task. - Required and implied dependencies will automatically be expanded later and need not be provided. - This may be replaced or modified in - __init__to change the dimensions of the task. After- __init__it will be a- frozensetand may not be replaced.
 - initInputs: set[str] = frozenset({})¶
- Set with the names of all - InitInputconnection attributes.- See - inputsfor additional information.
 - initOutputs: set[str] = frozenset({})¶
- Set with the names of all - InitOutputconnection attributes.- See - inputsfor additional information.
 - inputs: set[str] = frozenset({})¶
- Set with the names of all - connectionTypes.Inputconnection attributes.- This is updated automatically as class attributes are added, removed, or replaced in - __init__. Removing entries from this set will cause those connections to be removed after- __init__completes, but this is supported only for backwards compatibility; new code should instead just delete the collection attributed directly. After- __init__this will be a- frozensetand may not be replaced.
 - output_table¶
- Connection for output dataset. 
 - outputs: set[str] = frozenset({'output_table'})¶
- Set with the names of all - Outputconnection attributes.- See - inputsfor additional information.
 - prerequisiteInputs: set[str] = frozenset({})¶
- Set with the names of all - PrerequisiteInputconnection attributes.- See - inputsfor additional information.
 - Methods Documentation - adjustQuantum(inputs: dict[str, tuple[lsst.pipe.base.connectionTypes.BaseInput, collections.abc.Collection[lsst.daf.butler._dataset_ref.DatasetRef]]], outputs: dict[str, tuple[lsst.pipe.base.connectionTypes.Output, collections.abc.Collection[lsst.daf.butler._dataset_ref.DatasetRef]]], label: str, data_id: DataCoordinate) tuple[collections.abc.Mapping[str, tuple[lsst.pipe.base.connectionTypes.BaseInput, collections.abc.Collection[lsst.daf.butler._dataset_ref.DatasetRef]]], collections.abc.Mapping[str, tuple[lsst.pipe.base.connectionTypes.Output, collections.abc.Collection[lsst.daf.butler._dataset_ref.DatasetRef]]]]¶
- Override to make adjustments to - lsst.daf.butler.DatasetRefobjects in the- lsst.daf.butler.Quantumduring the graph generation stage of the activator.- Parameters:
- inputsdict
- Dictionary whose keys are an input (regular or prerequisite) connection name and whose values are a tuple of the connection instance and a collection of associated - DatasetRefobjects. The exact type of the nested collections is unspecified; it can be assumed to be multi-pass iterable and support- lenand- in, but it should not be mutated in place. In contrast, the outer dictionaries are guaranteed to be temporary copies that are true- dictinstances, and hence may be modified and even returned; this is especially useful for delegating to- super(see notes below).
- outputsMapping
- Mapping of output datasets, with the same structure as - inputs.
- labelstr
- Label for this task in the pipeline (should be used in all diagnostic messages). 
- data_idlsst.daf.butler.DataCoordinate
- Data ID for this quantum in the pipeline (should be used in all diagnostic messages). 
 
- inputs
- Returns:
- adjusted_inputsMapping
- Mapping of the same form as - inputswith updated containers of input- DatasetRefobjects. Connections that are not changed should not be returned at all. Datasets may only be removed, not added. Nested collections may be of any multi-pass iterable type, and the order of iteration will set the order of iteration within- PipelineTask.runQuantum.
- adjusted_outputsMapping
- Mapping of updated output datasets, with the same structure and interpretation as - adjusted_inputs.
 
- adjusted_inputs
- Raises:
- ScalarError
- Raised if any - Inputor- PrerequisiteInputconnection has- multipleset to- False, but multiple datasets.
- NoWorkFound
- Raised to indicate that this quantum should not be run; not enough datasets were found for a regular - Inputconnection, and the quantum should be pruned or skipped.
- FileNotFoundError
- Raised to cause QuantumGraph generation to fail (with the message included in this exception); not enough datasets were found for a - PrerequisiteInputconnection.
 
 - Notes - The base class implementation performs important checks. It always returns an empty mapping (i.e. makes no adjustments). It should always called be via - superby custom implementations, ideally at the end of the custom implementation with already-adjusted mappings when any datasets are actually dropped, e.g.:- def adjustQuantum(self, inputs, outputs, label, data_id): # Filter out some dataset refs for one connection. connection, old_refs = inputs["my_input"] new_refs = [ref for ref in old_refs if ...] adjusted_inputs = {"my_input", (connection, new_refs)} # Update the original inputs so we can pass them to super. inputs.update(adjusted_inputs) # Can ignore outputs from super because they are guaranteed # to be empty. super().adjustQuantum(inputs, outputs, label_data_id) # Return only the connections we modified. return adjusted_inputs, {} - Removing outputs here is guaranteed to affect what is actually passed to - PipelineTask.runQuantum, but its effect on the larger graph may be deferred to execution, depending on the context in which- adjustQuantumis being run: if one quantum removes an output that is needed by a second quantum as input, the second quantum may not be adjusted (and hence pruned or skipped) until that output is actually found to be missing at execution time.- Tasks that desire zip-iteration consistency between any combinations of connections that have the same data ID should generally implement - adjustQuantumto achieve this, even if they could also run that logic during execution; this allows the system to see outputs that will not be produced because the corresponding input is missing as early as possible.
 - buildDatasetRefs(quantum: Quantum) tuple[lsst.pipe.base.connections.InputQuantizedConnection, lsst.pipe.base.connections.OutputQuantizedConnection]¶
- Build - QuantizedConnectioncorresponding to input- Quantum.- Parameters:
- quantumlsst.daf.butler.Quantum
- Quantum object which defines the inputs and outputs for a given unit of processing. 
 
- quantum
- Returns:
- retValtupleof (InputQuantizedConnection,
- OutputQuantizedConnection) Namespaces mapping attribute names (identifiers of connections) to butler references defined in the input- lsst.daf.butler.Quantum.
 
- retVal
 
 - getSpatialBoundsConnections() Iterable[str]¶
- Return the names of regular input and output connections whose data IDs should be used to compute the spatial bounds of this task’s quanta. - The spatial bound for a quantum is defined as the union of the regions of all data IDs of all connections returned here, along with the region of the quantum data ID (if the task has spatial dimensions). - Returns:
- connection_namescollections.abc.Iterable[str]
- Names of collections with spatial dimensions. These are the task-internal connection names, not butler dataset type names. 
 
- connection_names
 - Notes - The spatial bound is used to search for prerequisite inputs that have skypix dimensions. The default implementation returns an empty iterable, which is usually sufficient for tasks with spatial dimensions, but if a task’s inputs or outputs are associated with spatial regions that extend beyond the quantum data ID’s region, this method may need to be overridden to expand the set of prerequisite inputs found. - Tasks that do not have spatial dimensions that have skypix prerequisite inputs should always override this method, as the default spatial bounds otherwise cover the full sky. 
 - getTemporalBoundsConnections() Iterable[str]¶
- Return the names of regular input and output connections whose data IDs should be used to compute the temporal bounds of this task’s quanta. - The temporal bound for a quantum is defined as the union of the timespans of all data IDs of all connections returned here, along with the timespan of the quantum data ID (if the task has temporal dimensions). - Returns:
- connection_namescollections.abc.Iterable[str]
- Names of collections with temporal dimensions. These are the task-internal connection names, not butler dataset type names. 
 
- connection_names
 - Notes - The temporal bound is used to search for prerequisite inputs that are calibration datasets. The default implementation returns an empty iterable, which is usually sufficient for tasks with temporal dimensions, but if a task’s inputs or outputs are associated with timespans that extend beyond the quantum data ID’s timespan, this method may need to be overridden to expand the set of prerequisite inputs found. - Tasks that do not have temporal dimensions that do not implement this method will use an infinite timespan for any calibration lookups.