Exporting results#
You can save the catalogs that result from running your workflow to disk, in parquet format, using the write_catalog call.
You must provide a base_catalog_path, which is the output path for your catalog directory, and (optionally) a name for your catalog, catalog_name. The catalog_name is the catalog’s internal name and therefore may differ from the catalog’s base directory name. If the directory already exists and you want to overwrite its content, set the overwrite flag to True. Do not forget to provide the necessary credentials, as storage_options to the UPath construction, when trying to
export the catalog to protected remote storage.
For example, to save a catalog that contains the results of crossmatching Gaia with ZTF to "./my_catalogs/gaia_x_ztf", one could run:
gaia_x_ztf_catalog.write_catalog(base_catalog_path="./my_catalogs/gaia_x_ztf",
catalog_name="gaia_x_ztf")
The HATS catalogs on disk follow a well-defined directory structure:
gaia_x_ztf/
├── dataset
│ ├── _common_metadata
│ ├── _metadata
│ ├── data_thumbnail.parquet
│ ├── Norder={}/
│ │ └── Dir={}/
│ │ ├── Npix={}.parquet
| | └── ...
│ └── Norder={}/
│ └── Dir={}/
│ ├── Npix={}.parquet
| └── ...
├── hats.properties
├── partition_info.csv
└── skymap.fits
You can read more about the HATS formation on the HATS ReadTheDocs.
About#
Authors: Neven Caplar
Last updated on: April 4, 2025
If you use lsdb for published research, please cite following instructions.