This project is Copyright (c) STScI and licensed under the terms of the BSD 3-Clause license. This package is based upon the Astropy package template which is licensed under the BSD 3-clause license. See the licenses folder for more information.
We love contributions! caldp is open source, built on open source, and we'd love to have you hang out in our community.
Imposter syndrome disclaimer: We want your help. No, really.
There may be a little voice inside your head that is telling you that you're not ready to be an open source contributor; that your skills aren't nearly good enough to contribute. What could you possibly offer a project like this one?
We assure you - the little voice in your head is wrong. If you can write code at all, you can contribute code to open source. Contributing to open source projects is a fantastic way to advance one's coding skills. Writing perfect code isn't the measure of a good developer (that would disqualify all of us!); it's trying to create something, making mistakes, and learning from those mistakes. That's how we all improve, and we are happy to help others learn.
Being an open source contributor doesn't just mean writing code, either. You can help out by writing documentation, tests, or even giving feedback about the project (and yes - that includes giving feedback about the contribution process). Some of these contributions may be the most valuable to the project as a whole, because you're coming to the project with fresh eyes, so you can see the errors and assumptions that seasoned contributors have glossed over.
Note: This disclaimer was originally written by Adrienne Lowe for a PyCon talk, and was adapted by caldp based on its use in the README file for the MetPy project.
CALDP is used to integrate fundamental HST calibration programs (e.g. calacs.e) with input data, output data, and calibration reference files (CRDS). Ultimately,
CALDP does end-to-end calibration of HST data in a manner similar to the archive pipeline, including the generation of preview images.
CALDP is primarily a Python package with some installable scripts, but also includes infrastructure for building a Docker container with everything needed to fully calibrate raw HST data to produce pipeline-like products.
CALDP has two basic ways it can be run:
- CALDP can be run native in a conda environment.
- CALDP can be run inside a Docker container.
A variation of running CALDP inside the Docker container is:
3. Run arbitrary numbers of CALDP containers on AWS compute clusters, pulling inputs from Astroquery, and writing outputs to AWS S3 storage. This can vastly accelerate large processing runs.
The core logic of CALDP is implemented in the caldp Python package in the process and create_preview modules. CALDP also includes convenience scripts to make it simpler to configure and call these modules. Since it is primarily Python, nothing precludes running CALDP outside a container provided you install prerequisites.
WARNING: By default this install method will completely replace any installation you already have at $HOME/miniconda3 unlless you supply additional parameters.
The following commands will install:
- Miniconda
- The stable version of HSTCAL
- Fitscut
- Whichever version of CALDP you clone and/or checkout
Parameters specified below in [ ] are optional, but must be specified in order, i.e. to change the CONDA_DIR you must specify all four parameters explicitly.
git clone https://github.com/spacetelescope/caldp.git
cd caldp
scripts/caldp-install-all [HSTCAL] [PY_VER] [CONDA_ENV] [CONDA_DIR]
Parameter | Default | Description |
---|---|---|
HSTCAL | stable | Version of base calibration packages, nominally stable or latest. |
PY_VER | 3.6.10 | Python version for CALDP conda environment. |
CONDA_ENV | caldp_stable | Conda environment which will be created |
CONDA_DIR | ${HOME}/miniconda3 | Location of Miniconda Installation. |
This section breaks down the Everything installation into different functional steps so that you can omit steps or customize as needed, e.g. if you already have a miniconda3 installation and just want to add to it.
git clone https://github.com/spacetelescope/caldp.git
cd caldp
scripts/caldp-install-conda [CONDA_DIR]
source ~/.bashrc
scripts/caldp-install-cal [HSTCAL] [PY_VER] [CONDA_ENV] [CONDA_DIR]
source $CONDA_DIR/etc/profile.d/conda.sh
conda activate [CONDA_ENV]
scripts/caldp-install-fitscut ${CONDA_DIR}/envs/${CONDA_ENV}
pip install .[dev,test]
While doing CALDP development you can of course just iterate changing, re-installing, and testing CALDP itself.
The abstract command for running CALDP natively is:
caldp-process <ipppssoot> [<input_path>] [<output_path>] [<config>]
Parameter | Default Value | Description |
---|---|---|
ipppssoot | N/A | HST dataset identifier, you must always specify this |
input_path | file:. | can be file:<relative_path> or astroquery: or (probably coming s3://input-bucket/subdirs...) |
output_path | file:. | can be file:<relative_path> or s3://output-bucket/subdirs... |
config | caldp-config-onsite | can be caldp-config-offsite, caldp-config-onsite, caldp-config-aws, <custom> |
Running natively, file paths for CALDP work normally with the exception that they're specified using a URI-like notation which begins with file:. Absolute paths work here.
Below are some parameter examples for running CALDP natively with different input and output modes. caldp-process is configured to run using local files by default.
# All file access defaults to current working directory. Inputs must pre-exist.
# Inputs: Finds raw files matching j8cb010b0 in current working directory
# Outputs: Puts output product trees under current working directory as data and messages subdirectories.
# CRDS configuration: VPN configuration, no CRDS server required, /grp/crds/cache must be visible.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-process j8cb010b0
# ----------------------------------------------------------------------------------------
# File access in subdirectories, inputs must pre-exist.
# Inputs: Finds raw files matching j8cb010b0 in subdirectory j8cb010b0_inputs.
# Outputs: Copies output product tree under subdirectory j8cb010b0_outputs.
# CRDS configuration: VPN configuration, no CRDS server required, /grp/crds/cache must be visible.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-process j8cb010b0 file:j8cb010b0_inputs file:j8cb010b0_outputs
# ----------------------------------------------------------------------------------------
# Download inputs from astroquery as neeed
# Inputs: Downloads raw files matching j8cb010b0 from astroquery to current working directory / CALDP_HOME.
# Outputs: Copies output product tree under subdirectory j8cb010b0_outputs.
# CRDS configuration: VPN configuration, no CRDS server required, /grp/crds/cache must be visible.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-process j8cb010b0 astroquery: file:j8cb010b0_outputs
# ----------------------------------------------------------------------------------------
# Download inputs from astroquery, upload outputs to S3, current AWS Batch configuration minus Docker.
# Inputs: Downloads raw files matching j8cb010b0 from astroquery to current working directory / CALDP_HOME.
# Outputs: Copies output product tree to AWS S3 storage bucket, AWS credentials and permission required.
# CRDS configuration: VPN configuration, no CRDS server required, /grp/crds/cache must be visible.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-process j8cb010b0 astroquery: s3://calcloud-hst-pipeline-outputs
# ----------------------------------------------------------------------------------------
# Download inputs from s3, upload outputs to S3 (AWS credentials and permission required)
# Inputs: Downloads compressed (tar.gz) file matching j8cb010b0 from s3 and extracts to folder in the current working directory / CALDP_HOME/j8cb010b0/.
# Outputs: Copies output product tree to AWS S3 storage bucket.
# CRDS configuration: VPN configuration, no CRDS server required, /grp/crds/cache must be visible.
# Scratch files: Extra processing artifacts appear in CALDP_HOME/j8cb010b0/. Export CALDP_HOME to move them somewhere else.
caldp-process j8cb010b0 s3://calcloud-hst-pipeline-inputs s3://calcloud-hst-pipeline-outputs
While CALDP is a natively installable Python package, its roots are as a Docker container used to perform HST calibrations on AWS Batch. CALDP has subsequently been enhanced to run using inputs and outputs from a local file system rather than cloud resources like Astroquery and AWS S3 storage. The primary difference from running natively is that some portion of your native file system must be mounted inside the container to pass files in and out as naturally as possible. By default, your current working directory becomes $HOME (/home/developer)
If you want to run CALDP as a container then the equivalent of installing it is either building or pulling the container (i.e. from an AWS elastic container registry, ECR). This section will cover building your own CALDP image. To complete this section for personal use, all you need is a local installation of Docker and the supplied scripts should run it for you even more easily than normal. This section doesn't cover using Docker in general, or hosting your own images on Docker Hub or AWS Elastic Container Registry (ECR) where you can make them available to others.
- Clone this repo to a local directory and CD to it.
1. Edit scripts/caldp-image-config to set your Docker repo and default tag. Unless you're ready to push an image, you can use any name for your respository. Leave the default tag set to "latest" until you're familiar with the scripts and ready to modify or improve them.
git clone https://github.com:/spacetelescope/caldp.git
cd caldp
Configure and build:
# Edit scripts/caldp-image-config to set the Docker image config variables for # your currrent build. These will include the repo and image tag your want to # build and/or push. vim scripts/caldp-image-config # and customize as needed, particularly ECR config.
# Install CALDP natively to get convenience scripts and your configuration from (1). pip install .
# This script executes docker build to create the image with your configuration caldp-image-build
At this stage you can proceed to running your image if you wish.
3. (optional) When you're ready to share your image with others and have done the corresponding Docker Hub or ECR setup, you can log in from your shell and then:
caldp-ecr-login <hst_repro_role> # Log into the ECR repo prior to pushing
caldp-image-push
This will push your image to the repo and tag your configured above.
As part of the Docker build the conda SSL certs are replaced with certs supplied by STScI (tls-ca-bundle.pem) using the fix-certs script. These certs are required to build and run in the AWS CALCLOUD environment.
The following command configures CALDP to run from a container locally. It has the advantage that the entire HST calibration environment is included within the container so there are no other preliminary setup steps other than setting up Docker. The same container can be run locally or on pipeline cluster systems like AWS Batch.
caldp-docker-run-pipeline <ipppssoot> [<input_path>] [<output_path>] [<caldp_process_config>]
This should look very similar to the caldp-process command shown in the Native CALDP section above because it is. The primary differences are that absolute native paths do not work.
NOTE: The config file specified to caldp-docker-run-pipeline is used to configure processing, not to select the image. caldp-docker-run-pipeline automatically uses caldp-image-config to select the image to run.
Below are some parameter examples for running CALDP inside Docker with different input and output modes. caldp-process is still configured to run using local files by default.
# All file access defaults to current working directory. Inputs must pre-exist.
# Inputs: Finds raw files matching j8cb010b0 in current working directory
# Outputs: Puts output product trees under current working directory as data and messages subdirectories.
# CRDS configuration: Remote configuration, server https://hst-crds.stsci.edu must be up, files downloaded to crds_cache.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-docker-run-pipeline j8cb010b0
# ----------------------------------------------------------------------------------------
# File access in subdirectories, inputs must pre-exist.
# Inputs: Finds raw files matching j8cb010b0 in subdirectory j8cb010b0_inputs.
# Outputs: Copies output product tree under subdirectory j8cb010b0_outputs.
# CRDS configuration: Remote configuration, server https://hst-crds.stsci.edu must be up, files downloaded to crds_cache.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-docker-run-pipeline j8cb010b0 file:j8cb010b0_inputs file:j8cb010b0_outputs
# ----------------------------------------------------------------------------------------
# Download inputs from astroquery as neeed
# Inputs: Downloads raw files matching j8cb010b0 from astroquery to current working directory / CALDP_HOME.
# Outputs: Copies output product tree under subdirectory j8cb010b0_outputs.
# CRDS configuration: Remote configuration, server https://hst-crds.stsci.edu must be up, files downloaded to crds_cache.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-docker-run-pipeline j8cb010b0 astroquery: file:j8cb010b0_outputs
# ----------------------------------------------------------------------------------------
# Download inputs from astroquery, upload outputs to S3, current AWS Batch configuration minus Docker.
# Inputs: Downloads raw files matching j8cb010b0 from astroquery to current working directory / CALDP_HOME.
# CRDS configuration: Remote configuration, server https://hst-crds.stsci.edu must be up, files downloaded to crds_cache.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-docker-run-pipeline j8cb010b0 astroquery: s3://calcloud-hst-pipeline-outputs/batch-22
# ----------------------------------------------------------------------------------------
# Download inputs from s3, upload outputs to S3, current AWS Batch configuration minus Docker.
# Inputs: Downloads raw files matching j8cb010b0 from s3 to current working directory / CALDP_HOME.
# CRDS configuration: Remote configuration, server https://hst-crds.stsci.edu must be up, files downloaded to crds_cache.
# Scratch files: Extra processing artifacts appear in the current working directory. Export CALDP_HOME to move them somewhere else.
caldp-docker-run-pipeline j8cb010b0 s3://calcloud-hst-pipeline-inputs s3://calcloud-hst-pipeline-outputs
After configuring Docker, caldp-docker-run-pipeline runs caldp-process inside the docker container with the parameters given on the command line. While file: paths are defined relative to your native file system, within the Docker container they will nominally be interpreted relative to /home/developer. Since the CALDP_HOME directory is mounted read/write inside Docker, files needed to process a dataset will be reflected back out of the Docker container to CALDP_HOME, defaulting to your current working directory.
NOTE: Running the final cloud-like configuration above does not produce results idenitical to AWS Batch processing because it is only processing a single dataset and skips batch tracking and organization actions normally performed by the batch trigger lambda which operates on a list of datasets.
Below is the calling sequence used to run CALDP on AWS Batch. This command is specified in the AWS Batch job definition and used to run all queued jobs. The calling sequence uses more customized input parameters in the outermost wrapper script specifying only the S3 output bucket and dataset name.
caldp-process-aws <s3_output_path> <ipppssoot>
Internally, caldp-process-aws runs caldp-process automatically configured to use:
- astroquery: to obtain raw data.
- the specified S3 output path which typically includes a batch "subdirectory".
- the specified dataset (ipppssoot) to define which data to fetch and process.
- a serverless CRDS configuration dependent only on S3 files.
Despite supporting a containerized use case, since AWS Batch (or equivalent) normally runs Docker, caldp-process-aws is effectively a native mode command when run by itself. There is no wrapper script equivalent to caldp-docker-run-pipeline to configure and run caldp-process-aws inside Docker automatically, but since it really requires no additional file mounts or ports, it is simple to run with Docker.
Running caldp-process-aws does require access to the CRDS and the output bucket on AWS S3 storage, i.e. appropriate credentials and permissions.
Sometimes you want to execute commands other than caldp-process in the container environment. You can run any command using caldp-docker-run-container which is itself normally wrapped by caldp-docker-run-pipeline.
Before running, the environment variable CALDP_DOCKER_RUN_PARS needs to be defined to add Docker command line switches which precede the CALDP image on the docker run command line. It should be defined as follows to e.g. enable the interactive debug:
export CALDP_DOCKER_RUN_PARS="-it"
Once CALDP_DOCKER_RUN_PARS is defined, you can start an interactive session inside the container like this:
caldp-docker-run-container /bin/bash
The same method can be used to add additional docker configuration parameters for any reason.
CALDP_DOCKER_RUN_PARS defaults to --rm to do automatic container cleanup during normal non-debug operation. It could also be used to e.g. make a port mapping for JupyterLab by adding: -p 8888:8888.
The CALDP_HOME environment variable defines which native directory caldp-docker-run-pipeline will mount inside the running Docker container at $HOME as read/write. If not exported, CALDP_HOME defaults to the directory you run caldp-docker-run-pipeline from. Since caldp-process runs at $HOME within the Docker container, any scratch files used during processing will appear externally within CALDP_HOME. Note that using caldp-docker-run-pipeline is not a requirement, it is just a script used to establish standard Docker configuration for local CALDP execution.
One technique for enabling AWS access inside the container is to put a .aws configuration directory in your CALDP_HOME directory.
Since caldp-docker-run-pipeline mounts CALDP_HOME inside the container at $HOME, AWS will see them where it expects to find them. AWS Batch nominally runs worker nodes which have the necessary permissions attached so no .aws directory is needed on AWS Batch.
CALDP and CALCLOUD output data in a form desgined to help track the state of individual datasets.
As such, the output directory is organized into two subdirectories:
- messages
- data
A key difference between CALDP and CALCLOUD is that the former is designed for processing single datasets, while the latter is designed for processing batches of datasets which are run individually by CALCLOUD. In this context, normally files downloaded from CALCLOUD's S3 storage to an onsite directory are placed in a "batch directory", and the CALDP equivalent of that batch directory is the output directory. The same messages and data appearing in the CALDP output directory would also appeaar in the sync'ed CALCLOUD batch directory.
The messages subdirectory is used to record the status of individual datasets as they progress through processing, data transfer, and archiving. Each dataset has a similarly named state file which moves between state directories as it starts or completes various states. The dataset file can be used to record metadata but its primary use is to enable simple indentification dataset state without the use of a database, queues, etc. Only a local file system is needed to track state using this scheme. A mirror of this same scheme is used on the cloud on S3 storage to help guide file downloads from AWS.
<output_path>/
messages/
datasets-processed/
<ipppssoots...> # CALDP, normally running on AWS batch, leaves messages here. they're empty.
dataset-synced/
<ipppssoots...> # CALCLOUD's downloader leaves messages here, normally containing abspaths of files to archive.
dataset-archived/
<ipppssoots...> # The archive can acknowledge archive completion here, file contents should be preserved.
The data subdirectory parallels but has a different structure than the messages subdirectory. For every ipppssoot message, there is a data directory and subdirectories which contain output files from processsing that ipppssoot. In the current implementation, the ipppssoot message file is empty, it is normally populated by CALCLOUD's downloader with the paths of files to archive when it is output to dataset-synced.
<output_path>/
data/
<instrument>/
<ipppssoots...>/ # one dir per ipppssoot
science data files for one ipppssoot...
logs/
log and metrics files for one ipppssoot...
previews/
preview images for one ipppssoot...
CALDP runs a sequence of steps and programs to fully process each dataset. Every program has its own methods of error handling and reporting failures. One limitation of AWS Batch is that the only CALDP status communicated directly back to Batch is the numerical program exit code. There is a universal convention that a program which exits with a non-zero return status has failed; conversely a status of zero indicates success. There is no convention about what non-zero exit code values should be, they vary program by program. It should be noted that Python and Batch have different methods of displaying the same one byte exit code, unsigned byte for Python, integer for Batch.
CALDP error code meanings can only be found in the program logs or in caldp/exit_codes.py. In contrast, AWS Batch reports text descriptions in addition to numerical exit codes, but only for failures at the Batch level, such as Docker failures.
A CALCLOUD Batch event handler is triggered upon CALDP job failure. The event handler interprets the combination of CALDP exit code, Batch exit code, and Batch exit reason to determine the error type and react appropriately. Reactions include automatically rescuing jobs with memory errors, retrying Docker failures, recording error-ipppssoot messages, etc.
Because there is uncertainty about how each subprogram chooses to define exit codes, and to give the batch event handler more information for decision making, CALDP often brackets blocks of code like this:
with sysexit.exit_on_exception(caldp_exit_code, "descriptive message"):
... python statements ...
such that an exception raised by the nested statements is caught and thrown to the exit_receiver() handler, typically at the highest program level:
with sysexit.exit_reciever():
main()
The exit_receiver() intercepts the chain of unwinding handlers, squelches the traceback between exit_on_exception() and exit_receiver(), then calls sys._exit(caldp_exit_code) to exit immediately. In this manner, caldp reports the error code caldp_exit_code rather than any code assigned by a subprogram.
Currently three different failure modes involving memory errors are mapped onto the same CALCLOUD job rescue handling: Python MemoryError, Unreported but logged subprogram Python MemoryError, Container memory error. This illustrates how characterization and handling are sometimes just... ugly.
Codes are assigned to specific functional blocks in the hope that as new failure modes are observed, handling can be added to CALCLOUD without changing CALDP. However, when necessary, exception bracketing should be revised, new error codes should be added, and the modified exit_codes.py module should be copied to CALCLOUD which may also need handling updates.
NOTE: AWS Batch also issues numerical exit codes so while there are no known cases of overlap, there is a potential for amiguity between Batch and CALDP, but not for CALDP subprograms.
As explained previously, each of the 3 CALDP use cases has a different CRDS configuration. This implementation is described here in case it is necessary to write additional configurations or add variables to these. At present, unlike caldp-image-config, these config scripts don't generally need customization, they are used as-is to support their use cases.
CALDP configuration scripts set environment variables which will be defined within the scope of caldp-process. These configuration scripts are installed alongside other CALDP scripts so they can be sourced directly without knowing where they are installed. The name of the configuration script is passed as a 4th generally defaulted parameter to caldp-process:
Top Level Script | Config Script | Description |
---|---|---|
caldp-process | caldp-config-onsite | Configures CRDS to operate from Central Store /grp/crds/cache. Should scale. |
caldp-docker-run-pipeline | caldp-config-offsite | Configures CRDS to download from CRDS server. This may not scale well. |
caldp-process-aws | caldp-config-aws | Configures CRDS to operate from S3 storage with no server dependency. Should scale. |
CALDP has a tox configuration which can be run to check for source code related issues including code quality (flake8), whitespace formatting (Black), and CALDP security scanning (bandit). Once you have fully installed your system, you can run all tox checks by:
$ tox
or individually:
$ tox -e flake8
$ tox -e bandit
$ tox -e black
Initial runs of tox are slow until most the environment is built and cached; afterwards CALDP package updates are relatively quick. The tox cache will be treated as part of the Docker build context and uploaded to Docker whenever it changes prior to the next build;
The CALDP repo is set up for GitHub Actions with the following workflows:
- build: native install and CALDP overall pytest with code coverage, no S3 testing
- docker: Docker build and test with one astroquery dataset, null outputs
- check: flake8, black, and bandit checks
Whenever you do a PR or merge to spacetelescope/caldp, GitHub will automatically run CI tests for CALDP.
It's common to do testing on a development machine prior to pushing. This can basically be accomplished by installing caldp, configuring your environment, and then running pytest similar to how it will be run by Travis.
# FIRST: Setup a conda environment for CALDP as discussed above in native installs.
# Don't use the "everything install" if you have an existing conda environment you
# don't want to wipe out. Make sure to activate it.
# THEN: configure your environment and run pytest as Travis would:
source caldp-config-offsite
pytest caldp --cov=caldp --cov-fail-under 80 --capture=tee-sys
NOTE: Not all CALDP code and capabilities are tested, particularly the wrapper scripts currently associated with running the Python package inside and outside Docker.
Because S3 inputs and outputs require AWS credentials to enable access, and specific object paths to use, testing of S3 modes is controlled by two environment variables which define where to locate S3 inputs and outputs:
export CALDP_S3_TEST_INPUTS=s3://caldp-hst-test/inputs/test-batch
export CALDP_S3_TEST_OUTPUTS=s3://caldp-hst-test/outputs/test-batch
If either or both of the above variables is defined, pytest will also execute tests which utilize the S3 input or output modes. You must also have AWS credentials for this. Currently S3 is not tested on Travis.