PosixDatastore

class lsst.daf.butler.datastores.posixDatastore.PosixDatastore(config, registry, butlerRoot=None)

Bases: lsst.daf.butler.datastores.fileLikeDatastore.FileLikeDatastore

Basic POSIX filesystem backed Datastore.

Parameters:
config : DatastoreConfig or str

Configuration. A string should refer to the name of the config file.

registry : Registry

Registry to use for storing internal information about the datasets.

butlerRoot : str, optional

New datastore root to use to override the configuration value.

Raises:
ValueError

If root location does not exist and create is False in the configuration.

Notes

PosixDatastore supports all transfer modes for file-based ingest: "move", "copy", "symlink", "hardlink", "relsymlink" and None (no transfer).

Attributes Summary

containerKey
defaultConfigFile Path to configuration defaults.
isEphemeral
names Names associated with this datastore returned as a list.

Methods Summary

addStoredItemInfo(refs, infos) Record internal storage information associated with one or more datasets.
computeChecksum(filename[, algorithm, …]) Compute the checksum of the supplied file.
emptyTrash([ignore_errors]) Remove all datasets from the trash.
exists(ref) Check if the dataset exists in the datastore.
export(refs, *, directory, transfer) Export datasets for transfer to another data repository.
fromConfig(config, registry, butlerRoot) Create datastore from type specified in config file.
get(ref[, parameters]) Load an InMemoryDataset from the store.
getLookupKeys() Return all the lookup keys relevant to this datastore.
getStoredItemInfo(ref) Retrieve information associated with file stored in this Datastore.
getUri(ref[, predict]) URI to the Dataset.
ingest(*datasets, transfer) Ingest one or more files into the datastore.
makeTableSpec()
put(inMemoryDataset, ref) Write a InMemoryDataset with a given DatasetRef to the store.
remove(ref) Indicate to the Datastore that a dataset can be removed.
removeStoredItemInfo(ref) Remove information about the file associated with this dataset.
setConfigRoot(root, config, full[, overwrite]) Set any filesystem-dependent config options for this Datastore to be appropriate for a new empty repository with the given root.
transaction() Context manager supporting Datastore transactions.
transfer(inputDatastore, ref) Retrieve a dataset from an input Datastore, and store the result in this Datastore.
trash(ref[, ignore_errors]) Indicate to the datastore that a dataset can be removed.
validateConfiguration(entities[, logFailures]) Validate some of the configuration for this datastore.
validateKey(lookupKey, entity) Validate a specific look up key with supplied entity.

Attributes Documentation

containerKey = None
defaultConfigFile = 'datastores/posixDatastore.yaml'

Path to configuration defaults. Relative to $DAF_BUTLER_DIR/config or absolute path. Can be None if no defaults specified.

isEphemeral = False
names

Names associated with this datastore returned as a list.

Some datastores can have child datastores.

Methods Documentation

addStoredItemInfo(refs, infos)

Record internal storage information associated with one or more datasets.

Parameters:
refs : sequence of DatasetRef

The datasets that have been stored.

infos : sequence of StoredDatastoreItemInfo

Metadata associated with the stored datasets.

static computeChecksum(filename, algorithm='blake2b', block_size=8192)

Compute the checksum of the supplied file.

Parameters:
filename : str

Name of file to calculate checksum from.

algorithm : str, optional

Name of algorithm to use. Must be one of the algorithms supported by :py:class`hashlib`.

block_size : int

Number of bytes to read from file at one time.

Returns:
hexdigest : str

Hex digest of the file.

emptyTrash(ignore_errors=True)

Remove all datasets from the trash.

Parameters:
ignore_errors : bool

If True return without error even if something went wrong. Problems could occur if another process is simultaneously trying to delete.

exists(ref)

Check if the dataset exists in the datastore.

Parameters:
ref : DatasetRef

Reference to the required dataset.

Returns:
exists : bool

True if the entity exists in the Datastore.

export(refs: Iterable[DatasetRef], *, directory: Optional[str] = None, transfer: Optional[str] = None) → Iterable[FileDataset]

Export datasets for transfer to another data repository.

Parameters:
refs : iterable of DatasetRef

Dataset references to be exported.

directory : str, optional

Path to a directory that should contain files corresponding to output datasets. Ignored if transfer is None.

transfer : str, optional

Mode that should be used to move datasets out of the repository. Valid options are the same as those of the transfer argument to ingest, and datastores may similarly signal that a transfer mode is not supported by raising NotImplementedError.

Returns:
dataset : iterable of DatasetTransfer

Structs containing information about the exported datasets, in the same order as refs.

Raises:
NotImplementedError

Raised if the given transfer mode is not supported.

static fromConfig(config: Config, registry: Registry, butlerRoot: Optional[str] = None) → 'Datastore'

Create datastore from type specified in config file.

Parameters:
config : Config

Configuration instance.

registry : Registry

Registry to be used by the Datastore for internal data.

butlerRoot : str, optional

Butler root directory.

get(ref, parameters=None)

Load an InMemoryDataset from the store.

Parameters:
ref : DatasetRef

Reference to the required Dataset.

parameters : dict

StorageClass-specific parameters that specify, for example, a slice of the dataset to be loaded.

Returns:
inMemoryDataset : object

Requested dataset or slice thereof as an InMemoryDataset.

Raises:
FileNotFoundError

Requested dataset can not be retrieved.

TypeError

Return value from formatter has unexpected type.

ValueError

Formatter failed to process the dataset.

getLookupKeys()

Return all the lookup keys relevant to this datastore.

Returns:
keys : set of LookupKey

The keys stored internally for looking up information based on DatasetType name or StorageClass.

getStoredItemInfo(ref)

Retrieve information associated with file stored in this Datastore.

Parameters:
ref : DatasetRef

The dataset that is to be queried.

Returns:
info : StoredFilenfo

Stored information about this file and its formatter.

Raises:
KeyError

Dataset with that id can not be found.

getUri(ref, predict=False)

URI to the Dataset.

Parameters:
ref : DatasetRef

Reference to the required Dataset.

predict : bool

If True, allow URIs to be returned of datasets that have not been written.

Returns:
uri : str

URI string pointing to the dataset within the datastore. If the dataset does not exist in the datastore, and if predict is True, the URI will be a prediction and will include a URI fragment “#predicted”. If the datastore does not have entities that relate well to the concept of a URI the returned URI string will be descriptive. The returned URI is not guaranteed to be obtainable.

Raises:
FileNotFoundError

A URI has been requested for a dataset that does not exist and guessing is not allowed.

Notes

When a predicted URI is requested an attempt will be made to form a reasonable URI based on file templates and the expected formatter.

ingest(*datasets, transfer: Optional[str] = None)

Ingest one or more files into the datastore.

Parameters:
datasets : FileDataset

Each positional argument is a struct containing information about a file to be ingested, including its path (either absolute or relative to the datastore root, if applicable), a complete DatasetRef (with dataset_id not None), and optionally a formatter class or its fully-qualified string name. If a formatter is not provided, the one the datastore would use for put on that dataset is assumed.

transfer : str, optional

How (and whether) the dataset should be added to the datastore. If None (default), the file must already be in a location appropriate for the datastore (e.g. within its root directory), and will not be modified. Other choices include “move”, “copy”, “link”, “symlink”, “relsymlink”, and “hardlink”. “link” is a special transfer mode that will first try to make a hardlink and if that fails a symlink will be used instead. “relsymlink” creates a relative symlink rather than use an absolute path. Most datastores do not support all transfer modes. “auto” is a special option that will let the data store choose the most natural option for itself.

Raises:
NotImplementedError

Raised if the datastore does not support the given transfer mode (including the case where ingest is not supported at all).

DatasetTypeNotSupportedError

Raised if one or more files to be ingested have a dataset type that is not supported by the datastore.

FileNotFoundError

Raised if one of the given files does not exist.

FileExistsError

Raised if transfer is not None but the (internal) location the file would be moved to is already occupied.

Notes

Subclasses should implement _prepIngest and _finishIngest instead of implementing ingest directly. Datastores that hold and delegate to child datastores may want to call those methods as well.

Subclasses are encouraged to document their supported transfer modes in their class documentation.

classmethod makeTableSpec()
put(inMemoryDataset, ref)

Write a InMemoryDataset with a given DatasetRef to the store.

Parameters:
inMemoryDataset : object

The dataset to store.

ref : DatasetRef

Reference to the associated Dataset.

Raises:
TypeError

Supplied object and storage class are inconsistent.

DatasetTypeNotSupportedError

The associated DatasetType is not handled by this datastore.

Notes

If the datastore is configured to reject certain dataset types it is possible that the put will fail and raise a DatasetTypeNotSupportedError. The main use case for this is to allow ChainedDatastore to put to multiple datastores without requiring that every datastore accepts the dataset.

remove(ref)

Indicate to the Datastore that a dataset can be removed.

Warning

This method deletes the artifact associated with this dataset and can not be reversed.

Parameters:
ref : DatasetRef

Reference to the required Dataset.

Raises:
FileNotFoundError

Attempt to remove a dataset that does not exist.

Notes

This method is used for immediate removal of a dataset and is generally reserved for internal testing of datastore APIs. It is implemented by calling trash() and then immediately calling emptyTrash(). This call is meant to be immediate so errors encountered during removal are not ignored.

removeStoredItemInfo(ref)

Remove information about the file associated with this dataset.

Parameters:
ref : DatasetRef

The dataset that has been removed.

classmethod setConfigRoot(root, config, full, overwrite=True)

Set any filesystem-dependent config options for this Datastore to be appropriate for a new empty repository with the given root.

Parameters:
root : str

URI to the root of the data repository.

config : Config

A Config to update. Only the subset understood by this component will be updated. Will not expand defaults.

full : Config

A complete config with all defaults expanded that can be converted to a DatastoreConfig. Read-only and will not be modified by this method. Repository-specific options that should not be obtained from defaults when Butler instances are constructed should be copied from full to config.

overwrite : bool, optional

If False, do not modify a value in config if the value already exists. Default is always to overwrite with the provided root.

Notes

If a keyword is explicitly defined in the supplied config it will not be overridden by this method if overwrite is False. This allows explicit values set in external configs to be retained.

transaction()

Context manager supporting Datastore transactions.

Transactions can be nested, and are to be used in combination with Registry.transaction.

transfer(inputDatastore, ref)

Retrieve a dataset from an input Datastore, and store the result in this Datastore.

Parameters:
inputDatastore : Datastore

The external Datastore from which to retreive the Dataset.

ref : DatasetRef

Reference to the required dataset in the input data store.

trash(ref, ignore_errors=True)

Indicate to the datastore that a dataset can be removed.

Parameters:
ref : DatasetRef

Reference to the required Dataset.

ignore_errors : bool

If True return without error even if something went wrong. Problems could occur if another process is simultaneously trying to delete.

Raises:
FileNotFoundError

Attempt to remove a dataset that does not exist.

validateConfiguration(entities, logFailures=False)

Validate some of the configuration for this datastore.

Parameters:
entities : iterable of DatasetRef, DatasetType, or StorageClass

Entities to test against this configuration. Can be differing types.

logFailures : bool, optional

If True, output a log message for every validation error detected.

Raises:
DatastoreValidationError

Raised if there is a validation problem with a configuration. All the problems are reported in a single exception.

Notes

This method checks that all the supplied entities have valid file templates and also have formatters defined.

validateKey(lookupKey, entity)

Validate a specific look up key with supplied entity.

Parameters:
lookupKey : LookupKey

Key to use to retrieve information from the datastore configuration.

entity : DatasetRef, DatasetType, or StorageClass

Entity to compare with configuration retrieved using the specified lookup key.

Raises:
DatastoreValidationError

Raised if there is a problem with the combination of entity and lookup key.

Notes

Bypasses the normal selection priorities by allowing a key that would normally not be selected to be validated.