Butler¶
-
class
lsst.daf.butler.Butler(config=None, butler=None, collection=None, run=None, searchPaths=None)¶ Bases:
objectMain entry point for the data access system.
Parameters: - config :
ButlerConfig,Configorstr, optional. Configuration. Anything acceptable to the
ButlerConfigconstructor. If a directory path is given the configuration will be read from abutler.yamlfile in that location. IfNoneis given default values will be used.- butler :
Butler, optional. If provided, construct a new Butler that uses the same registry and datastore as the given one, but with the given collection and run. Incompatible with the
configandsearchPathsarguments.- collection :
str, optional Collection to use for all input lookups, overriding config[“collection”] if provided.
- run :
str,Run, optional Collection associated with the
Runto use for outputs, overriding config[“run”]. If aRunassociated with the given Collection does not exist, it will be created. If “collection” is None, this collection will be used for input lookups as well; if not, it must have the same value as “run”.- searchPaths :
listofstr, optional Directory paths to search when calculating the full Butler configuration. Not used if the supplied config is already a
ButlerConfig.
Raises: - ValueError
Raised if neither “collection” nor “run” are provided by argument or config, or if both are provided and are inconsistent.
Attributes: - config :
str,ButlerConfigorConfig, optional (filename to) configuration. If this is not a
ButlerConfig, defaults will be read. If astr, may be the path to a directory containing a “butler.yaml” file.- datastore :
Datastore Datastore to use for storage.
- registry :
Registry Registry to use for lookups.
Attributes Summary
GENERATIONThis is a Generation 3 Butler. Methods Summary
datasetExists(datasetRefOrType[, dataId])Return True if the Dataset is actually present in the Datastore. export(*, directory, filename, format, transfer)Export datasets from the repository represented by this Butler.get(datasetRefOrType[, dataId, parameters])Retrieve a stored dataset. getDeferred(datasetRefOrType, …)Create a DeferredDatasetHandlewhich can later retrieve a datasetgetDirect(ref[, parameters])Retrieve a stored dataset. getUri(datasetRefOrType[, dataId, predict])Return the URI to the Dataset. import_(*, directory, filename, format, transfer)Import datasets exported from a different butler repository. ingest(*datasets, transfer)Store and register one or more datasets that already exist on disk. makeRepo(root[, config, standalone, …])Create an empty data repository by adding a butler.yaml config to a repository root directory. put(obj, datasetRefOrType[, dataId, producer])Store and register a dataset. remove(datasetRefOrType[, dataId, delete, …])Remove a dataset from the collection and possibly the repository. transaction()Context manager supporting Butlertransactions.validateConfiguration([logFailures, …])Validate butler configuration. Attributes Documentation
-
GENERATION= 3¶ This is a Generation 3 Butler.
This attribute may be removed in the future, once the Generation 2 Butler interface has been fully retired; it should only be used in transitional code.
Methods Documentation
-
datasetExists(datasetRefOrType, dataId=None, **kwds)¶ Return True if the Dataset is actually present in the Datastore.
Parameters: - datasetRefOrType :
DatasetRef,DatasetType, orstr When
DatasetRefthedataIdshould beNone. Otherwise theDatasetTypeor name thereof.- dataId :
dictorDataCoordinate A
dictofDimensionlink name, value pairs that label theDatasetRefwithin a Collection. WhenNone, aDatasetRefshould be provided as the first argument.- kwds
Additional keyword arguments used to augment or construct a
DataCoordinate. SeeDataCoordinate.standardizeparameters.
Raises: - LookupError
Raised if the Dataset is not even present in the Registry.
- datasetRefOrType :
-
export(*, directory: Optional[str] = None, filename: Optional[str] = None, format: Optional[str] = None, transfer: Optional[str] = None) → AbstractContextManager[lsst.daf.butler.core.repoTransfers.RepoExport]¶ Export datasets from the repository represented by this
Butler.This method is a context manager that returns a helper object (
RepoExport) that is used to indicate what information from the repository should be exported.Parameters: - directory :
str, optional Directory dataset files should be written to if
transferis notNone.- filename :
str, optional Name for the file that will include database information associated with the exported datasets. If this is not an absolute path and
directoryis notNone, it will be written todirectoryinstead of the current working directory. Defaults to “export.{format}”.- format :
str, optional File format for the database information file. If
None, the extension offilenamewill be used.- transfer :
str, optional Transfer mode passed to
Datastore.export.
Raises: - TypeError
Raised if the set of arguments passed is inconsistent.
- directory :
-
get(datasetRefOrType, dataId=None, parameters=None, **kwds)¶ Retrieve a stored dataset.
Parameters: - datasetRefOrType :
DatasetRef,DatasetType, orstr When
DatasetRefthedataIdshould beNone. Otherwise theDatasetTypeor name thereof.- dataId :
dictorDataCoordinate A
dictofDimensionlink name, value pairs that label theDatasetRefwithin a Collection. WhenNone, aDatasetRefshould be provided as the first argument.- parameters :
dict Additional StorageClass-defined options to control reading, typically used to efficiently read only a subset of the dataset.
- kwds
Additional keyword arguments used to augment or construct a
DataCoordinate. SeeDataCoordinate.standardizeparameters.
Returns: - obj :
object The dataset.
- datasetRefOrType :
-
getDeferred(datasetRefOrType: Union[lsst.daf.butler.core.datasets.DatasetRef, lsst.daf.butler.core.datasets.DatasetType, str], dataId: Union[lsst.daf.butler.core.dimensions.coordinate.DataCoordinate, Mapping[str, Any], None] = None, parameters: Optional[dict] = None, **kwds) → lsst.daf.butler.core.deferredDatasetHandle.DeferredDatasetHandle¶ Create a
DeferredDatasetHandlewhich can later retrieve a datasetParameters: - datasetRefOrType :
DatasetRef,DatasetType, orstr When
DatasetRefthedataIdshould beNone. Otherwise theDatasetTypeor name thereof.- dataId :
dictorDataCoordinate, optional A
dictofDimensionlink name, value pairs that label theDatasetRefwithin a Collection. WhenNone, aDatasetRefshould be provided as the first argument.- parameters :
dict Additional StorageClass-defined options to control reading, typically used to efficiently read only a subset of the dataset.
- kwds
Additional keyword arguments used to augment or construct a
DataId. SeeDataIdparameters.
Returns: - obj :
DeferredDatasetHandle A handle which can be used to retrieve a dataset at a later time
- datasetRefOrType :
-
getDirect(ref, parameters=None)¶ Retrieve a stored dataset.
Unlike
Butler.get, this method allows datasets outside the Butler’s collection to be read as long as theDatasetRefthat identifies them can be obtained separately.Parameters: - ref :
DatasetRef Reference to an already stored dataset.
- parameters :
dict Additional StorageClass-defined options to control reading, typically used to efficiently read only a subset of the dataset.
Returns: - obj :
object The dataset.
- ref :
-
getUri(datasetRefOrType, dataId=None, predict=False, **kwds)¶ Return the URI to the Dataset.
Parameters: - datasetRefOrType :
DatasetRef,DatasetType, orstr When
DatasetRefthedataIdshould beNone. Otherwise theDatasetTypeor name thereof.- dataId :
dictorDataCoordinate A
dictofDimensionlink name, value pairs that label theDatasetRefwithin a Collection. WhenNone, aDatasetRefshould be provided as the first argument.- predict :
bool If
True, allow URIs to be returned of datasets that have not been written.- kwds
Additional keyword arguments used to augment or construct a
DataCoordinate. SeeDataCoordinate.standardizeparameters.
Returns: - uri :
str URI string pointing to the Dataset within the datastore. If the Dataset does not exist in the datastore, and if
predictisTrue, the URI will be a prediction and will include a URI fragment “#predicted”. If the datastore does not have entities that relate well to the concept of a URI the returned URI string will be descriptive. The returned URI is not guaranteed to be obtainable.
Raises: - FileNotFoundError
A URI has been requested for a dataset that does not exist and guessing is not allowed.
- datasetRefOrType :
-
import_(*, directory: Optional[str] = None, filename: Optional[str] = None, format: Optional[str] = None, transfer: Optional[str] = None)¶ Import datasets exported from a different butler repository.
Parameters: - directory :
str, optional Directory containing dataset files. If
None, all file paths must be absolute.- filename :
str, optional Name for the file that containing database information associated with the exported datasets. If this is not an absolute path, does not exist in the current working directory, and
directoryis notNone, it is assumed to be indirectory. Defaults to “export.{format}”.- format :
str, optional File format for the database information file. If
None, the extension offilenamewill be used.- transfer :
str, optional Transfer mode passed to
Datastore.export.
Raises: - TypeError
Raised if the set of arguments passed is inconsistent.
- directory :
-
ingest(*datasets, transfer: Optional[str] = None)¶ Store and register one or more datasets that already exist on disk.
Parameters: - datasets :
FileDataset Each positional argument is a struct containing information about a file to be ingested, including its path (either absolute or relative to the datastore root, if applicable), a
DatasetRef, and optionally a formatter class or its fully-qualified string name. If a formatter is not provided, the formatter that would be used forputis assumed. On return, allFileDataset.refattributes will have theirDatasetRef.idattribute populated and allFileDataset.formatterattributes will be set to the formatter class used.FileDataset.pathattributes may be modified to put paths in whatever the datastore considers a standardized form.- transfer :
str, optional If not
None, must be one of ‘move’, ‘copy’, ‘hardlink’, or ‘symlink’, indicating how to transfer the file.
Raises: - TypeError
Raised if the butler was not constructed with a Run, and is hence read-only.
- NotImplementedError
Raised if the
Datastoredoes not support the given transfer mode.- DatasetTypeNotSupportedError
Raised if one or more files to be ingested have a dataset type that is not supported by the
Datastore..- FileNotFoundError
Raised if one of the given files does not exist.
- FileExistsError
Raised if transfer is not
Nonebut the (internal) location the file would be moved to is already occupied.
- datasets :
-
static
makeRepo(root, config=None, standalone=False, createRegistry=True, searchPaths=None, forceConfigRoot=True, outfile=None)¶ Create an empty data repository by adding a butler.yaml config to a repository root directory.
Parameters: - root :
str Filesystem path to the root of the new repository. Will be created if it does not exist.
- config :
Configorstr, optional Configuration to write to the repository, after setting any root-dependent Registry or Datastore config options. Can not be a
ButlerConfigor aConfigSubset. IfNone, default configuration will be used. Root-dependent config options specified in this config are overwritten ifforceConfigRootisTrue.- standalone :
bool If True, write all expanded defaults, not just customized or repository-specific settings. This (mostly) decouples the repository from the default configuration, insulating it from changes to the defaults (which may be good or bad, depending on the nature of the changes). Future additions to the defaults will still be picked up when initializing
Butlersto repos created withstandalone=True.- createRegistry :
bool, optional If
Truecreate a new Registry.- searchPaths :
listofstr, optional Directory paths to search when calculating the full butler configuration.
- forceConfigRoot :
bool, optional If
False, any values present in the suppliedconfigthat would normally be reset are not overridden and will appear directly in the output config. This allows non-standard overrides of the root directory for a datastore or registry to be given. If this parameter isTruethe values forrootwill be forced into the resulting config if appropriate.- outfile :
str, optional If not-
None, the output configuration will be written to this location rather than into the repository itself.
Returns: Raises: - ValueError
Raised if a ButlerConfig or ConfigSubset is passed instead of a regular Config (as these subclasses would make it impossible to support
standalone=False).- os.error
Raised if the directory does not exist, exists but is not a directory, or cannot be created.
Notes
Note that when
standalone=False(the default), the configuration search path (seeConfigSubset.defaultSearchPaths) that was used to construct the repository should also be used to construct any Butlers to avoid configuration inconsistencies.- root :
-
put(obj, datasetRefOrType, dataId=None, producer=None, **kwds)¶ Store and register a dataset.
Parameters: - obj :
object The dataset.
- datasetRefOrType :
DatasetRef,DatasetType, orstr When
DatasetRefis provided,dataIdshould beNone. Otherwise theDatasetTypeor name thereof.- dataId :
dictorDataCoordinate A
dictofDimensionlink name, value pairs that label theDatasetRefwithin a Collection. WhenNone, aDatasetRefshould be provided as the second argument.- producer :
Quantum, optional The producer.
- kwds
Additional keyword arguments used to augment or construct a
DataCoordinate. SeeDataCoordinate.standardizeparameters.
Returns: - ref :
DatasetRef A reference to the stored dataset, updated with the correct id if given.
Raises: - TypeError
Raised if the butler was not constructed with a Run, and is hence read-only.
- obj :
-
remove(datasetRefOrType, dataId=None, *, delete=True, remember=True, **kwds)¶ Remove a dataset from the collection and possibly the repository.
The identified dataset is always at least removed from the Butler’s collection. By default it is also deleted from the Datastore (e.g. files are actually deleted), but the dataset is “remembered” by retaining its row in the dataset and provenance tables in the registry.
If the dataset is a composite, all components will also be removed.
Parameters: - datasetRefOrType :
DatasetRef,DatasetType, orstr When
DatasetRefthedataIdshould beNone. Otherwise theDatasetTypeor name thereof.- dataId :
dictorDataId A
dictofDimensionlink name, value pairs that label theDatasetRefwithin a Collection. WhenNone, aDatasetRefshould be provided as the first argument.- delete :
bool If
True(default) actually delete the dataset from the Datastore (i.e. actually remove files).- remember :
bool If
True(default), retain dataset and provenance records in theRegistryfor this dataset.- kwds
Additional keyword arguments used to augment or construct a
DataId. SeeDataIdparameters.
Raises: - datasetRefOrType :
-
validateConfiguration(logFailures=False, datasetTypeNames=None, ignore=None)¶ Validate butler configuration.
Checks that each
DatasetTypecan be stored in theDatastore.Parameters: - logFailures :
bool, optional If
True, output a log message for every validation error detected.- datasetTypeNames : iterable of
str, optional The
DatasetTypenames that should be checked. This allows only a subset to be selected.- ignore : iterable of
str, optional Names of DatasetTypes to skip over. This can be used to skip known problems. If a named
DatasetTypecorresponds to a composite, all component of thatDatasetTypewill also be ignored.
Raises: - ButlerValidationError
Raised if there is some inconsistency with how this Butler is configured.
- logFailures :
- config :