Butler v26.0.0 2023-09-22¶
Now supports Python 3.11.
New Features¶
Added the ability to remove multiple dataset types at once, including expansion of wildcards, with
Registry.removeDatasetType
andbutler remove-dataset-type
. (DM-34568)Added the
ArrowNumpyDict
storage class to Parquet formatter. (DM-37279)Added support for columns with array values (1D and multi-dimensional) in Parquet tables accessed via arrow/astropy/numpy. Pandas does not support array-valued columns. (DM-37425)
Integrated an experimental Butler server into distribution.
lsst.daf.butler.server
will likely not be in this location permanently. The interface is also evolving and should be considered extremely unstable. Some testing of the remote registry code has been included. (DM-37609)Added support for writing/reading masked columns in astropy tables. This also adds support for masked columns in pandas dataframes, with limited support for conversion between the two. (DM-37757)
Dimension records are now available via attribute access on
DataCoordinate
instances, allowing syntax likedata_id.exposure.day_obs
. (DM-38054)Added default row groups (targeting a size of <~ 1GB) for Parquet files. (DM-38063)
Butler.get()
andButler.put()
can now be used with resolvedDatasetRef
. (DM-38210)Butler.transfer_from()
can now be used in conjunction with aChainedDatastore
. Additionally, datastore constraints are now respected. (DM-38240)Modified
Butler.import_()
(and by extension thebutler import
command-line) to accept URIs for the directory and export file.Modified
butler ingest-files
to accept a remote URI for the table file. (DM-38492)
Added support for multi-index dataframes with
DataFrameDelegate
andInMemoryDatastore
. (DM-38642)Added new APIs to support the deprecation of
LimitedButler.datastore
:LimitedButler.get_datastore_roots
can be used to retrieve any root URIs associated with attached datastores. If a datastore does not support the concept it will returnNone
for its root URI.LimitedButler.get_datastore_names
can be used to retrieve the names of the internal datastores.LimitedButler.get_many_uris
allows for the bulk retrieval of URIs from a list of refs.Also made
getURI
andgetURIs
available forLimitedButler
. (DM-39915)
Modified to fully support Pydantic version 2.x and version 1.x. (DM-40002; DM-40303)
API Changes¶
Added new APIs for checking dataset existence.
stored
checks whether the datastore artifact(s) exists for a singleDatasetRef
.stored_many
is a bulk version ofstored
that can be used for manyDatasetRef
.exists
checks whether registry and datastore know about a singleDatasetRef
and can optionally check for artifact existence. The results are returned in anFlag
object (specificallyDatasetExistence
) that evaluates toTrue
if the dataset is available for retrieval.
Additionally
DatasetRef
now has a new method for checking whether twoDatasetRef
only differ by compatible storage classes. (DM-32940)lsst.daf.Butler.transfer_from
method now acceptsLimitedButler
as a source Butler. In cases when a full butler is needed as a source it will try to cast it to aButler
. (DM-33497)Creating an unresolved dataset reference now issues an
UnresolvedRefWarning
and is deprecated (and subsequently removed).A resolved
DatasetRef
can now be created by specifying the run without the ID – the constructor will now automatically issue an ID. Previously this was an error. To support ID generation a new optional parameterid_generation_mode
can now be given to the constructor to allow the ID to be constructed in different ways. (DM-37703)
DatasetRef
constructor now requiresrun
argument in all cases and always constructs a resolved reference.Methods
DatasetRef.resolved()
,DatasetRef.unresolved()
, andDatasetRef.getCheckedId()
were removed. (DM-37704)
Added
StorageClassDelegate.copy()
method. By default this method callscopy.deepcopy()
but subclasses can override as needed. (DM-38694)Database.fromUri
andDatabase.makeEngine
methods now acceptsqlalchemy.engine.URL
instances in addition to strings. (DM-39484)Added new parameter
without_datastore
to theButler
andButlerConfig
constructors to allow a butler to be created that can not access a datastore. This can be helpful if you want to query registry without requiring the overhead of the datastore. (DM-40120)
Bug Fixes¶
Fixed race condition in datastore cache involving the possibility of multiple processes trying to retrieve the same file simultaneously and one of those processes deleting the file on exit of the context manager. (DM-37092)
Made
Registry.findDataset
respect the storage class of aDatasetType
that is passed to it. This also makes directPipelineTask
execution respect storage class conversions in the same way that execution butler already did. (DM-37450)Can now properly retrieve astropy full table metadata with
butler.get
. (DM-37530)Fixed an order-of-operations bug in the query system (and as a result,
QuantumGraph
generation) that manifested as a “Custom operation find_first not supported by engine iteration” message. (DM-37625)Butler.put
is fixed to raise a correct exception for duplicate put attempts forDatasetRef
with the same dataset ID. (DM-37704)Fixed parsing of order by terms to treat direct references to dimension primary key columns as references to the dimensions. (DM-37855)
Fixed bugs involving CALIBRATION-collection skipping and long dataset type names that were introduced on DM-31725. (DM-37868)
Now check for big-endian arrays when serializing to Parquet. This allows astropy FITS tables to be easily serialized. (DM-37913)
Fixed bugs in spatial query constraints introduced in DM-31725. (DM-37930)
Fixed additional bugs in spatial query constraints introduced in DM-31725. (DM-37938)
Fixed occasional crashes in
Butler
refresh()
method due to a race condition in dataset types refresh. (DM-38305)Fixed query manipulation logic to more aggressively move operations from Python postprocessing to SQL.
This fixes a bug in
QuantumGraph
generation that occurs when a dataset type that is actually present in an input collection has exactly the same dimensions as the graph as a whole, manifesting as a mismatch betweendaf_relation
engines. (DM-38402)Add check for
ListType
when pandas converts a list object into Parquet. (DM-38845)Few registry methods treated empty collection list in the same way as
None
, meaning that Registry-default run collection was used. This has been fixed now to mean that queries always return empty result set, with explicit “doomed by” messages. (DM-38915)Fixed a bug in
butler query-data-ids
that caused a cryptic “the query has deferred operations…” error message when a spatial join is involved. (DM-38943)Fixed more issues with storage class conversion. (DM-38952)
Fixed a SQL generation bug for queries that involve the common
skypix
dimension and at least two other spatial dimensions. (DM-38954)Fixed bugs in storage class conversion in
FileDatastore
, as used byQuantumBackedButler
. (DM-39198)Fixed the bug in initializing PostgreSQL registry which resulted in “password authentication failed” error. The bug appeared during the SQLAlchemy 2.0 transition which changed default rendering of URL to string. (DM-39484)
Fixed a rare bug in follow-up dataset queries involving relation commutators.
This occurred when building QuantumGraphs where a “warp” dataset type was an overall input to the pipeline and present in more than one input RUN collection. (DM-40184)
Ensureed
Datastore
record exports (as used in quantum-backed butler) are deduplicated when necessary. (DM-40381)
Performance Enhancement¶
When passing lazy query-results objects directly to various registry methods (
associate
,disassociate
,removeDatasets
, andcertify
), query and process one dataset type at a time instead of querying for all of them and grouping by type in Python. (DM-39939)
Other Changes and Additions¶
Rewrote the registry query system, using the new
daf_relation
package.This change should be mostly invisible to users, but there are some subtle behavior changes:
Registry.findDatasets
now respects the given storage class when passed a fullDatasetType
instance, instead of replacing it with storage class registered with that dataset type. This causes storage class overrides inPipelineTask
input connections to be respected in more contexts as well; in at least some cases these were previously being incorrectly ignored.Registry.findDatasets
now utilizes cached summaries of which dataset types and governor dimension values are present in each collection. This should result in fewer and simpler database calls, but it does make the result vulnerable to stale caches (which, likeRegistry
methods more generally, must be addressed manually via calls toRegistry.refresh
.The diagnostics provided by the
explain_no_results
methods on query result object (used prominently in the reporting on empty quantum graph builds) have been significantly improved, though they now usedaf_relation
terminology that may be unfamiliar to users.Registry
is now more consistent about raisingDataIdValueError
when given invalid governor dimension values, while not raising (but providingexplain_no_results
diagnostics) for all other invalid dimension values, as per RFC-878.Registry
methods that take awhere
argument are now typed to expect astr
that is notNone
, with the default no-op value now an empty string (before either an emptystr
orNone
could be passed, and meant the same thing). This should only affect downstream type checking, as the runtime code still just checks for whether the argument evaluates asFalse
in a boolean context. (DM-31725)
Added dimensions config entries that declare that the
visit
dimension “populates” various dimension elements that define many-to-many relationships.In the future, this will be used to ensure the correct records are included in exports of dimension records. (DM-34589)
Added converter config to allow
lsst.ip.isr.IntermediateTransmissionCurve
and subclasses to be used forlsst.afw.image.TransmissionCurve
. (DM-36597)Butler.getURIs
no longer checks the file system to see if the file exists before returning a URI if the datastore thinks it knows about the file. This does mean that if someone has removed the file from the file system without deleting it from datastore that a URI could be retrieved for something that does not exist. (DM-37173)Enhanced the JSON and YAML formatters so that they can both handle dataclasses and Pydantic models (previously JSON supported Pydantic and YAML supported dataclasses).
Rationalized the storage class conversion handling to always convert from a
dict
to the original type even if the caller is requesting adict
. Without this change it was possible to have some confusion where a Pydantic model’s serialization did not match thedict
-like view it was emulating. (DM-37214)
Added an
obsCoreTableManager
property toRegistry
for access to the ObsCore table manager. This will be set toNone
when repository lacks an ObsCore table. It should only be used by a limited number of clients, e.g.lsst.obs.base.DefineVisitsTask
, which need to update the table. (DM-38205)Modified
Butler.ingest()
such that it can now ingest resolvedDatasetRef
. If unresolved refs are given (which was the previous requirement for ingest and is no longer possible) they are resolved internally but a warning is issued.Added
repr()
support forRegistryDefaults
class. (DM-38779)
The behavior of
FileDatastore.transfer_from()
has been clarified regarding what to do when an absolute URI (from a direct ingest) is found in the source butler. Iftransfer="auto"
(the default) the absolute URI will be stored in the target butler. If any other transfer mode is used the absolute URI will be copied/linked into the target butler. (DM-38870)Made minor modifications to the StorageClass system to support mock storage classes (in
pipe_base
) for testing. (DM-38952)Replaced the use of
lsst.utils.ellipsis
mypy workaround with the native typetype.EllipsisType
available since Python 3.10. (DM-39410)Moved Butler repository aliasing resolution into
ButlerConfig
so that it is available everywhere without having to do the resolving each time. (DM-39563)Added ability for some butler primitives to be cached and re-used on deserialization through a special interface. (DM-39582)
Replaced usage of
Butler.registry.dimensions
withButler.dimensions
.Modernized type annotations.
Fixed some documentation problems.
Made some Minor modernizations to use set notation and f-strings. (DM-39605)
Changed all Butler code and tests to use conforming DataIDs. Removed the fake
DataCoordinate
classes from the datastore tests. Improved type annotations in some test files. (DM-39665)Added various optimizations to
QuantumGraph
loading. (DM-40121)Fixed docs on referring to timespans in queries, and made related error messages more helpful. (DM-38084)
Clarified that
butler prune-datasets --purge
always removes dataset entries and clarified when the run argument is used. (DM-39086)
An API Removal or Deprecation¶
Deprecated methods for constructing or using
DimensionPacker
instances.The
DimensionPacker
interface is not being removed, but all concrete implementations will now be downstream ofdaf_butler
and will not satisfy the assumptions of the current interfaces for constructing them. (DM-31924)Butler.datasetExists
has been deprecated and will be removed in a future release. It has been replaced byButler.stored()
(specifically to check if the datastore has the artifact) andButler.exists()
which will check registry and datastore and optionally check whether the artifact exists. (DM-32940)Removed the
Spectraction
storage class. This was a temporary storage class added for convenience during development, which was a roll-up-and-pickle of all the potentially relevant parts of the extraction. All the necessary information is now stored inside theSpectractorSpectrum
storage class. (DM-33932)Removed deprecated
ButlerURI
(uselsst.resources.ResourcePath
instead).Removed deprecated
kwargs
parameter fromDeferredDatasetHandle
.Removed the deprecated
butler prune-collection
command.Removed the deprecated
checkManagerDigests
from butler registry. (DM-37534)
Deprecated
Butler.getDirect()
andButler.putDirect()
. We have modified theget()
andput()
variants to recognize the presence of a resolvedDatasetRef
and use it directly. Forget()
we no longer unpack theDatasetRef
and re-run the query, but return exactly the dataset being requested.Removed
Butler.pruneCollections
. This method was replaced byButler.removeRuns
andRegistry.removeCollections
a long time ago and the command-line interface was removed previously. (DM-38210)
Code that calculates schema digests was removed, registry will no longer store digests in the database. Previously we saved schema digests, but we did not verify them since w_2022_22 in v24.0. (DM-38235)
Support for integer dataset IDs in registry has now been removed. All dataset IDs must now be
uuid.UUID
. (DM-38280)Removed support for non-UUID dataset IDs in
Butler.transfer_from()
. Theid_gen_map
parameter has been removed and thelocal_refs
parameter has been removed fromDatastore.transfer_from()
. (DM-38409)Deprecated
reconstituteDimensions
argument fromQuantum.from_simple
. (DM-39582)The semi-public
Butler.datastore
property has now been deprecated. TheLimitedButler
API has been expanded such that there is no longer any need for anyone to access the datastore class directly. (DM-39915)lsst.daf.butler.registry.DbAuth
class has been moved to thelsst-utils
package and can be imported from thelsst.utils.db_auth
module. (DM-40462)
Butler v25.0.0 2023-02-27¶
This is the last release that can access data repositories using integer dataset IDs. Please either recreate these repositories or convert them to use UUIDs using the butler migrate tooling.
New Features¶
Added
StorageClass.is_type
method to compare a type with that of the storage class itelf.Added keys, values, items, and iterator for
StorageClassFactory
. (DM-29835)
Updated parquet backend to use Arrow Tables natively, and add converters to and from pandas DataFrames, Astropy Tables, and Numpy structured arrays. (DM-34874)
Butler.transfer_from()
can now copy dimension records as well as datasets. This significantly enhances the usability of this method when transferring between disconnected Butlers. Thebutler transfer-datasets
command will transfer dimension records by default but this can be disabled with the--no-transfer-dimensions
option (which can be more efficient if you know that the destination Butler contains all the records). (DM-34887)butler query-data-ids
will now determine default dimensions to use if a dataset type and collection is specified. The logical AND of all supplied dataset types will be used. Additionally, if no results are returned a reason will now be given in many cases. (DM-35391)Added
DataFrameDelegate
to allow DataFrames to be used withlsst.pipe.base.InMemoryDatasetHandle
. (DM-35803)Add
StorageClass.findStorageClass
method to find a storage class from a python type. (DM-35815)The optional dependencies of
lsst-resources
can be requested as optional dependencies oflsst-daf-butler
and will be passed down to the underlying package. This allows callers oflsst.daf.butler
to specify the type of resources they want to be able to access without being aware of the role oflsst.resources
as an implementation detail. (DM-35886)Requires Python 3.10 or greater for better type annotation support. (DM-36174)
Bind values in Registry queries can now specify list/tuple of numbers for identifiers appearing on the right-hand side of
IN
expression. (DM-36325)It is now possible to override the python type returned by
butler.get()
(if the types are compatible with each other) by using the newreadStorageClass
parameter. Deferred dataset handles can also be overridden.For example, to return an
astropy.table.Table
from something that usually returns anlsst.afw.table.Catalog
you would do:table = butler.getDirect(ref, readStorageClass="AstropyTable")
Any parameters given to the
get()
must still refer to the native storage class. (DM-4551)
API Changes¶
Deprecate support for accessing data repositories with integer dataset IDs, and disable creation of new data repositories with integer dataset IDs, as per RFC-854. (DM-35063)
DimensionUniverse
now has aisCompatibleWith()
method to check if two universes are compatible with each other. The initial test is very basic but can be improved later. (DM-35082)Deprecated support for components in
Registry.query*
methods, per RFC-879. (DM-36312)Multiple minor API changes to query methods from RFC-878 and
RFC-879 <https://jira.lsstcorp.org/browse/RFC-879>_
.This includes:
CollectionSearch
is deprecated in favor ofSequence[str]
and the newCollectionWildcard
class.queryDatasetTypes
andqueryCollections
now returnIterable
(representing an unspecified in-memory collection) andSequence
, respectively, rather than iterators.DataCoordinateQueryResults.findDatasets
now raisesMissingDatasetTypeError
when the given dataset type is not registered.Passing regular expressions and other patterns as dataset types to
queryDataIds
andqueryDimensionRecords
is deprecated.Passing unregistered dataset types
queryDataIds
andqueryDimensionRecords
is deprecated; in the future this will raiseMissingDatasetTypeError
instead of returning no query results.Query result class
explain_no_results
now returnsIterable
instead ofIterator
. (DM-36313)
A method has been added to
DatasetRef
andDatasetType
, namedoverrideStorageClass
, to allow a new object to be created that has a different storage class associated with it. (DM-4551)
Bug Fixes¶
Fixed a bug in the parquet reader where a single string column name would be interpreted as an iterable. (DM-35803)
Fixed bug in
elements
argument to various export methods that prevented it from doing anything. (DM-36111)A bug has been fixed in
DatastoreCacheManager
that triggered if two processes try to cache the same dataset simultaneously. (DM-36412)Fixed bug in pandas
dataframe
to arrow conversion that would crash with some pandas object data types. (DM-36775)Fixed bug in pandas
dataframe
to arrow conversion that would crash with partially nulled string columns. (DM-36795)
Other Changes and Additions¶
For command-line options that split on commas, it is now possible to specify parts of the string not to split by using
[]
to indicate comma-separated list content. (DM-35917)Moved the typing workaround for the built-in
Ellipsis
() singleton to
lsst.utils
. (DM-36108)Now define regions for data IDs with multiple spatial dimensions to the intersection of those dimensions’ regions. (DM-36111)
Added support for in-memory datastore to roll back a call to
datastore.trash()
. This required that thebridge.moveToTrash()
method now takes an additionaltransaction
parameter (that can beNone
). (DM-36172)Restructured internal Registry query system methods to share code better and prepare for more meaningful changes. (DM-36174)
Removed unnecessary table-locking in dimension record insertion.
Prior to this change, we used explicit full-table locks to guard against a race condition that wasn’t actually possible, which could lead to deadlocks in rare cases involving insertion of governor dimension records. (DM-36326)
Chained Datastore can now support “move” transfer mode for ingest. Files are copied to each child datastore unless only one child datastore is accepting the incoming files, in which case “move” is used. (DM-36410)
DatastoreCacheManager
can now use an environment variable,$DAF_BUTLER_CACHE_DIRECTORY_IF_UNSET
, to specify a cache directory to use if no explicit directory has been specified by configuration or by the$DAF_BUTLER_CACHE_DIRECTORY
environment variable. Additionally, aDatastoreCacheManager.set_fallback_cache_directory_if_unset()
class method has been added that will set this environment variable with a suitable value. This is useful for multiprocessing where each forked or spawned subprocess needs to share the same cache directory. (DM-36412)Added support for
ChainedDatastore.export()
. (DM-36517)Reworked transaction and connection management for compatibility with transaction-level connection pooling on the server.
Butler clients still hold long-lived connections, via delegation to SQLAlchemy’s connection pooling, which can handle disconnections transparently most of the time. But we now wrap all temporary table usage and cursor iteration in transactions. (DM-37249)
An API Removal or Deprecation¶
Removed deprecated filterLabel exposure component access. (DM-27811)
Butler v24.0.0 2022-08-26¶
New Features¶
Support LSST-style visit definitions where a single exposure is part of a set of related exposures all taken with the same acquisition command. Each exposure knows the “visit” it is part of.
Modify the
exposure
dimension record to includeseq_start
andseq_end
metadata.Modify
visit
record to include aseq_num
field.Remove
visit_system
dimension and addvisit_system_membership
record to allow a visit to be associated with multiple visit systems. (DM-30948)
butler export-calibs
now takes a--transfer
option to control how data are exported (usedirect
to do in-place export) and a--datasets
option to limit the dataset types to be exported. It also now takes a default collections parameter (all calibration collections). (DM-32061)Iterables returned from registry methods
queryDataIds
andqueryDimensionRecords
have two new methods -order_by
andlimit
. (DM-32403)Builds using
setuptools
now calculate versions from the Git repository, including the use of alpha releases for those associated with weekly tags. (DM-32408)Butler can now support lookup of repositories by label if the user environment is correctly configured. This is done using the new
get_repo_uri()
andget_known_repos()
APIs. (DM-32491)Add a butler command line command called
butler remove-collections
that can remove non-RUN collections. (DM-32687)Add a butler command line command called
butler remove-runs
that can remove RUN collections and contained datasets. (DM-32831)It is now possible to register type conversion functions with storage classes. This can allow a dataset type definition to change storage class in the registry whilst allowing datasets that have already been serialized using one python type to be returned using the new python type. The
storageClasses.yaml
definitions can now look like:TaskMetadata: pytype: lsst.pipe.base.TaskMetadata converters: lsst.daf.base.PropertySet: lsst.pipe.base.TaskMetadata.from_metadata
Declares that if a
TaskMetadata
is expected then aPropertySet
can be converted to the correct python type. (DM-32883)Dimension record imports now ignore conflicts (without checking for consistency) instead of failing. (DM-33148)
Storage class converters can now also be used on
put
. (DM-33155)If a
DatasetType
has been constructed that differs from the registry definition, but in a way that is compatible throughStorageClass
conversion, then using that in alsst.daf.butler.Butler.get()
call will return a python type that matches the user-specifiedStorageClass
instead of the internal python type. (DM-33303)The dataset ID can now be used in a file template for datastore (using
{id}
). (DM-33414)Add
Registry.getCollectionParentChains
to find theCHAINED
collections that another collection belongs to. (DM-33643)Added
has_simulated
to theexposure
record to indicate that some content of this exposure was simulated. (DM-33728)The command-line tooling has changed how it sets the default logger when using
--log-level
. Now only the default logger(s) (lsst
and the colon-separated values stored in the$DAF_BUTLER_ROOT_LOGGER
) will be affected by using--log-level
without a specific logger name. By default only this default logger will be set toINFO
log level and all other loggers will remain asWARNING
. Use--log-level '.=level'
to change the root logger (this will not change the default logger level and so an additional call to--log-level DEBUG
may be needed to turn on debugging for all loggers). (DM-33809)Added
azimuth
to theexposure
andvisit
records. (DM-33859)If repository aliases have been defined for the site they can now be used in place of the Butler repository URI in both the
Butler
constructor and command-line tools. (DM-33870)Added
visit_system
toinstrument
record and allowed it to be used as a tie breaker in dataset determination if a dataId is given usingseq_num
andday_obs
and it matches multiple visits.Modify export YAML format to include the dimension universe version and namespace.
Allow export files with older visit definitions to be read (this does not fill in the new metadata records).
DimensionUniverse
now supports thein
operator to check if a dimension is part of the universe. (DM-33942)
Added a definition for using healpix in skypix definitions.
Change dimension universe caching to support a namespace in addition to a version number. (DM-33946)
Added a formatter for
lsst.utils.packages.Packages
Python types inlsst.daf.butler.formatters.packages.PackagesFormatter
. (DM-34105)Added an optimization that speeds up
butler query-datasets
when using--show-uri
. (DM-35120)
API Changes¶
Many internal utilities from
lsst.daf.butler.core.utils
have been relocated to thelsst.utils
package. (DM-31722)The
ButlerURI
class has now been removed from this package. It now exists aslsst.resources.ResourcePath
. All code should be modified to use the new class name. (DM-31723)lsst.daf.butler.Registry.registerRun
andlsst.daf.butler.Registry.registerCollection
now return a Booelan indicating whether the collection was created or already existed. (DM-31976)A new optional parameter,
record_validation_info
has been added toingest
(and related datastore APIs) to allow the caller to declare that file attributes such as the file size or checksum should not be recorded. This can be useful if the file is being monitored by an external system or it is known that the file might be compressed in-place after ingestion. (DM-33086)Added a new
DatasetType.is_compatible_with
method. This method determines if two dataset types are compatible with each other, taking into account whether the storage classes allow type conversion. (DM-33278)The
run
parameter has been removed from Butler methodlsst.daf.butler.Butler.pruneDatasets
. It was never used in Butler implementation, client code should simply remove it. (DM-33488)Registry methods now raise exceptions belonging to a class hierarchy rooted at
lsst.daf.butler.registry.RegistryError
. See also Error handling with Registry methods for details. (DM-33600)Added
DatasetType.storageClass_name
property to allow the name of the storage class to be retrieved without requiring that the storage class exists. This is possible if people have used local storage class definitions or a testDatasetType
was created temporarily. (DM-34460)
Bug Fixes¶
butler export-calibs
can now copy files that require the use of a file template (for example if a direct URI was stored in datastore) with metadata records. File templates that use metadata records now complain if the record is not attached to theDatasetRef
. (DM-32061)Make it possible to run
queryDimensionRecords
while constraining on the existence of a dataset whose dimensions are not a subset of the record element’s dependencies (e.g.raw
andexposure
). (DM-32454)Butler constructor can now take a
os.PathLike
object when thebutler.yaml
is not included in the path. (DM-32467)In the butler presets file (used by the
--@
option), use option names that match the butler CLI command option names (without leading dashes). Fail if option names used in the presets file do not match options for the current butler command. (DM-32986)The butler CLI command
remove-runs
can now unlink RUN collections from parent CHAINED collections. (DM-33619)Improves
butler query-collections
:TABLE output formatting is easier to read.
Adds INVERSE modes for TABLE and TREE output, to view CHAINED parent(s) of collections (non-INVERSE lists children of CHAINED collections).
Sorts datasets before printing them. (DM-33902)
Fix garbled printing of raw-byte hashes in query-dimension-records. (DM-34007)
The automatic addition of
butler.yaml
to the Butler configuration URI now also happens when aResourcePath
instance is given. (DM-34172)Fix handling of “doomed” (known to return no results even before execution) follow-up queries for datasets. This frequently manifested as a
KeyError
with a message about dataset type registration duringQuantumGraph
generation. (DM-34202)Fix
queryDataIds
bug involving dataset constraints with no dimensions. (DM-34247)The
click.Path
API changed, change from ordered arguments to keyword arguments when calling it. (DM-34261)Fix
queryCollections
bug in which children of chained collections were being alphabetically sorted instead of ordered consistently with the order in which they would be searched. (DM-34328)Fixes the bug introduced in DM-33489 (appeared in w_2022_15) which causes not-NULL constraint violation for datastore component column. (DM-34375)
Fixes an issue where the command line tools were caching argument and option values but not separating option names from option values correctly in some cases. (DM-34812)
Other Changes and Additions¶
Add a
NOT NULL
constraint to dimension implied dependency columns.NULL
values in these columns already cause the query system to misbehave. (DM-21840)Update parquet writing to use default per-column compression. (DM-31963)
Tidy up
remove-runs
subcommand confirmation report by sorting dataset types and filtering out those with no datasets in the collections to be deleted. (DM-33584)The constraints on collection names have been relaxed. Previously collection names were limited to ASCII alphanumeric characters plus a limited selection of symbols (directory separator, @-sign). Now all unicode alphanumerics can be used along with emoji. (DM-33999)
File datastore now always writes a temporary file and renames it even for local file system datastores. This minimizes the risk of a corrupt file being written if the process writing the file is killed at the wrong time. (DM-35458)
An API Removal or Deprecation¶
The
butler prune-collections
command line command is now deprecated. Please consider usingremove-collections
orremove-runs
instead. Will be removed after v24. (DM-32499)All support for reading and writing
Filter
objects has been removed. The oldfilter
component for exposures has been removed, and replaced with a newfilter
component backed byFilterLabel
. It functions identically to thefilterLabel
component, which has been deprecated. (DM-27177)
Butler v23.0.0 2021-12-10¶
New Features¶
Add ability to cache datasets locally when using a remote file store. This can significantly improve performance when retrieving components from a dataset. (DM-13365)
Add a new
butler retrieve-artifacts
command to copy file artifacts from a Butler datastore. (DM-27241)Add
butler transfer-datasets
command-line tool and associatedButler.transfer_from()
API.This can be used to transfer datasets between different butlers, with the caveat that dimensions and dataset types must be pre-defined in the receiving butler repository. (DM-28650)
Add
amp
parameter to the Exposure StorageClass, allowing single-amplifier subimage reads. (DM-29370)Add new
butler collection-chain
subcommand for creating collection chains from the command line. (DM-30373)Add
butler ingest-files
subcommand to simplify ingest of any external file. (DM-30935)Add class representing a collection of log records (
ButlerLogRecords
).Allow this class to be stored and retrieved from a Butler datastore.
Add special log handler to allow JSON log records to be stored.
Add
--log-file
option to command lines to redirect log output to file.Add
--no-log-tty
to disable log output to terminal. (DM-30977)
Registry methods that previously could raise an exception when searching in calibrations collections now have an improved logic that skip those collections if they were not given explicitly but only appeared in chained collections. (DM-31337)
Add a confirmation step to
butler prune-collection
to help prevent accidental removal of collections. (DM-31366)Add
butler register-dataset-type
command to register a new dataset type. (DM-31367)Use cached summary information to simplify queries involving datasets and provide better diagnostics when those queries yield no results. (DM-31583)
Add a new
butler export-calibs
command to copy calibrations and write an export.yaml document from a Butler datastore. (DM-31596)Support rewriting of dataId containing dimension records such as
day_obs
andseq_num
inbutler.put()
. This matches the behavior ofbutler.get()
. (DM-31623)Add
--log-label
option tobutler
command to allow extra information to be injected into the log record. (DM-31884)The
Butler.transfer_from
method no longer registers new dataset types by default.Add the related option
--register-dataset-types
to thebutler transfer-datasets
subcommand. (DM-31976)
Support UUIDs as the primary keys in registry and allow for reproducible UUIDs.
This change will significantly simplify transferring of data between butler repositories. (DM-29196)
Allow registry methods such as
queryDatasets
to use a glob-style string when specifying collection or dataset type names. (DM-30200)Add support for updating and replacing dimension records. (DM-30866)
API Changes¶
A new method
Datastore.knows()
has been added to allow a user to ask the datastore whether it knows about a specific dataset but without requiring a check to see if the artifact itself exists. UseDatastore.exists()
to check that the datastore knows about a dataset and the artifact exists. (DM-30335)
Bug Fixes¶
Fix handling of ingest_date timestamps.
Previously there was an inconsistency between ingest_date database-native UTC handling and astropy Time used for time literals which resulted in 37 second difference. This updates makes consistent use of database-native time functions to resolve this issue. (DM-30124)
Fix butler repository creation when a seed config has specified a registry manager override.
Previously only that manager was recorded rather than the full set. We always require a full set to be recorded to prevent breakage of a butler when a default changes. (DM-30372)
Stop writing a temporary datastore cache directory every time a
Butler
object was instantiated. Now only create one when one is requested. (DM-30743)Fix
Butler.transfer_from()
such that it registers any missing dataset types and also skips any datasets that do not have associated datastore artifacts. (DM-30784)Add support for click 8.0. (DM-30855)
Replace UNION ALL with UNION for subqueries for simpler query plans. (DM-31429)
Fix parquet formatter error when reading tables with no indices.
Previously, this would cause butler.get to fail to read valid parquet tables. (DM-31700)
Fix problem in ButlerURI where transferring a file from one URI to another would overwrite the existing file even if they were the same actual file (for example because of soft links in the directory hierarchy). (DM-31826)
Performance Enhancement¶
Make collection and dataset pruning significantly more efficient. (DM-30140)
Add indexes to make certain spatial join queries much more efficient. (DM-31548)
Made 20x speed improvement for
Butler.transfer_from
. The main slow down is asking the datastore whether a file artifact exists. This is now parallelized and the result is cached for later. (DM-31785)Minor efficiency improvements when accessing
lsst.daf.butler.Config
hierarchies. (DM-32305)FileDatastore: Improve removing of datasets from the trash by at least a factor of 10. (DM-29849)
Other Changes and Additions¶
Enable serialization of
DatasetRef
and related classes to JSON format. (DM-28678)ButlerURI
http
schemes can now handle non-WebDAV endpoints. Warnings are only issued if WebDAV functionality is requested. (DM-29708)Switch logging such that all logging messages are now forwarded to Python
logging
fromlsst.log
. Previously all Pythonlogging
messages were being forwarded tolsst.log
. (DM-31120)Add formatter and storageClass information for FocalPlaneBackground. (DM-22534)
Add formatter and storageClass information for IsrCalib. (DM-29531)
Change release note creation to use [Towncrier](https://towncrier.readthedocs.io/en/actual-freaking-docs/index.html). (DM-30291)
Add a Butler configuration for an execution butler that has pre-defined registry entries but no datastore records.
The
Butler.put()
will return the pre-existing dataset ref but will still fail if a datastore record is found. (DM-30335)If an unrecognized dimension is used as a look up key in a configuration file (using the
+
syntax) a warning is used suggesting a possible typo rather than a confusingKeyError
. This is no longer a fatal error and the key will be treated as a name. (DM-30685)Add
split
transfer mode that can be used when some files are inside the datastore and some files are outside the datastore. This is equivalent to usingNone
anddirect
mode dynamically. (DM-31251)
Butler v22.0 2021-04-01¶
New Features¶
A Butler instance can now be configured with dataId defaults such as an instrument or skymap. [DM-27153]
Add
butler prune-datasets
command. [DM-26689]Add
butler query-dimension-records
command [DM-27344]Add
--unlink
option tobutler prune-collection
command. [DM-28857]Add progress reporting option for long-lived commands. [DM-28964]
Add
butler associate
command to add existing datasets to a tagged collection. [DM-26688]Add officially-supported JSON serialization for core Butler classes. [DM-28314]
Allow
butler.get()
to support dimension record values such as exposure observing day or detector name in the dataID. [DM-27152]Add “direct” ingest mode to allow a file to be ingested retaining the full path to the original file. [DM-27478]
Bug Fixes¶
Fix temporal queries and clarify
Timespan
behavior. [DM-27985]
Other Changes and Additions¶
Make
ButlerURI
class immutable. [DM-29073]Add
ButlerURI.findFileResources
method to walk the directory tree and return matching files. [DM-29011]Improve infrastructure for handling test repositories. [DM-23862]
Butler Datastores¶
New Features¶
Implement basic file caching for use with remote datastores. [DM-29383]
Require that a DataId always be available to a
Formatter
. This allows formatters to do a consistency check such as comparing the physical filter in a dataId with that read from a file. [DM-28583]Add special mode to datastore to instruct it to ignore registry on
get
. This is useful for Execution Butlers where registry knows in advance about all datasets but datastore does not. [DM-28648]Add
forget
method to instruct datastore to remove all knowledge of a dataset without deleting the file artifact. [DM-29106]
Butler Registry¶
New Features¶
Avoid long-lived connections to database. [DM-26302]
Add option to flatten when setting a collection chain. [DM-29203]