LATISS Combined Calibrations Generation Procedure¶
Overview¶
This page assumes the reader is familiar with the content explained in the Daytime Calibrations section:
Post-Condition¶
Individual calibration images will be taken and verified using the cp_verify framework.
If
generate_calibrations
isTrue
, a (daily) combined calibration image per detector will be certified in a butlerCALIBRATION
collection.
Procedure Steps¶
Once you are logged into LOVE, click on the ATQueue
panel, as circled on the left side of the figure below (for reference, MTQueue
to launch LSSTComCam
scripts is circled on the right):
Load the Script¶
After clicking on the ATQueue
panel, search for the script maintel/make_latiss_calibrations.py
under AVAILABLE SCRIPTS
on the left, as shown in the figure below:
Load the script by clicking on the button in front of the name of the script that has a triangle.
Enter configuration parameters¶
After loading the script, a window that contains two sections, SCHEMA
(top) and CONFIG
(bottom), will appear.
The former will show the available configuration options (and the default values of some of them) that should be entered in the latter. The configuration options are as follows:
script_mode
: Currently, the script can be run in three modes, in which it will produce only biases (BIAS
), biases and darks (BIAS_DARK
), or biases, darks, and flats (BIAS_DARK_FLAT
). Default:BIAS_DARK_FLAT
n_bias
: Number of biases to take. Default:20
.n_discard_bias
: Additional number of bias images to take and discard before starting the sequence. Default:1
.n_dark
: Number of darks to take. Default:22
.n_discard_dark
: Additional number of dark images to take and discard before starting the sequence. Default:1
.exp_times_dark
: The exposure time of each dark image (sec). If a single value, then the same exposure time is used for each exposure. This default set of exposure times allows for one discarded dark plus four darks to use at 5s (to match with the previous defaults), two darks at 15s (to allow the linearity of the dark signal to be monitored), and sixteen darks at 30s (to provide darks that directly match the default observation exposure time). Default:[5, 5, 5, 5, 5, 15, 15, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30, 30]
.n_flat
: Number of flats to take. Default:20
.n_discard_flat
: Additional number of flat images to take and discard before starting the sequence. Default:1
.exp_times_flat
: The exposure time of each flat image (sec). If a single value, then the same exposure time is used for each exposure. Ifdo_ptc
isTrue
, the exposure times should form an adecuate secuence of flat pairs, each pair with the same exposure time. Ifdo_gain_from_flat_pairs
isTrue
, at least two flats with the same exposure time should be taken. Default:5
.filter
: Filter name or ID; if omitted the filter is not changed. Default:null
.grating
: Grating name; if omitted the grating is not changed. Default:null
.detectors
: Detector IDs that will be pased to the pipeline tasks, given as an array of integers, e.g.,[0,1,2,3]
. The default value is an empty array, which will translate in using all the detectors (a single detector for LATISS). Default:[]
do_verify
: Should the combined calibrations be verified? (c.f.,cp_verify
). Default:True
.generate_calibrations
: Should the combined calibrations be generated from the images taken and used as references for image verification? (“internal verification”). IfFalse
and anddo_verify
isTrue
, pre-existing calibrations will be used as reference for verification (“external verification”), and they should be provided in the input collections for the verification pipetasks. Default:False
.config_options_bias
: Options to be passed to the command-line bias pipetask. They will overwrite the values incpBias.yaml
. Default:-c isr:doDefect=False
.config_options_dark
: Options to be passed to the command-line dark pipetask. They will overwrite the values incpDark.yaml
. Default:-c isr:doDefect=False
.config_options_flat
: Options to be passed to the command-line flat pipetask. They will overwrite the values incpFlat.yaml
. Default:-c isr:doDefect=False
.do_defects
: Should defects be built using darks and flats?.script_mode
must beBIAS_DARK_FLAT
. Default:False
config_options_defects
: Options to be passed to the command-line defects pipetask. They will overwrite the values infindDefects.yaml
. Default:-c isr:doDefect=False
.do_ptc
: Should a Photon Transfer Curve be constructed from the flats taken?script_mode
must beBIAS_DARK_FLAT
. Default:False
.config_options_ptc
: Options to be passed to the command-line PTC pipetask. They will overwrite the values incpPtc.yaml
. Default:-c isr:doCrosstalk=False
.do_gain_from_flat_pairs
: Should the gain be estimated from each pair of flats taken at the same exposure time? Runs thecpPtc.yaml#generateGainFromFlatPair
pipeline. Since this pipeline is a subset of the PTC pipeline, you can use use theconfig_options_ptc
parameter to pass options to theISR
(Instrument Signature Removal) andcpExtract
tasks which form this pipeline. IfTrue
, the script mode should beBIAS_DARK_FLAT
. Default:True
.input_collections_bias
: List of additional (theOCPS
already addsLATISS/raw/all
as a default) comma-separated input collections for the bias pipetask. The pipetask is called via theOCPS
after enabling it with theLATISS
configuration. Default:LATISS/calib
.input_collections_verify_bias
: Additional comma-separated input collections to pass to the verify (bias) pipetask. Default:LATISS/calib
.input_collections_dark
: Additional comma-separarted input collections to pass to the dark pipetask. Default:LATISS/calib
.input_collections_verify_dark
: Additional comma-separated input collections to pass to the verify (dark) pipetask. Default:LATISS/calib
.input_collections_flat
: Additional comma-separated input collections to pass to the flat pipetask. Default:LATISS/calib
.input_collections_verify_flat
: Additional comma-separated input collections to pass to the verify (flat) pipetask. Default:LATISS/calib
.input_collections_defects
: Additional comma-separated input collections to pass to the defects pipetask. Default:LATISS/calib
.input_collections_ptc
: Additional comma-separated input collections to pass to the Photon Transfer Curve pipetask. Default:LATISS/calib
.calib_collection
:CALIBRATION
collection where the calibrations will be certified into, for example,LATISS/calib/u/plazas/YYYYMMMDD.test
orLATISS/calib/daily
. Default:LATISS/calib/daily
.repo
: Butler repository. Default:/repo/LATISS
.n_processes
: Number of processes that the pipetasks will use. Default:8
.certify_calib_begin_date
: The beginning date for the validity range of the certified calibration, for example,2021-07-15
. Default:1950-01-01
.certify_calib_end_date
: The end date for the validity range of the certified calibration, for example,2021-07-16
. Default:2050-01-01
.oods_timeout
: Timeout value, in seconds, for the Observatory Operations Data Service (OODS
). Default:120
.
Configuration examples¶
Daily Default¶
Preferred daily script mode to be run: if no configuration parameters are passed to LOVE and the default parameters are used, the script will take 21 biases, 23 darks at 5 seconds (5 exposures), 15 seconds (2 exposures), and 30 seconds (16 exposures), and finally 21 flats of 5 seconds each one.
In each case, the first image will be discarded. New combined calibrations will not be generated, and verification of the images taken will be performed using the existing combined calibrations in the LATISS/calib
collection (i.e., the script will do external verification
).
In this case, no defects will be made.
Following DMTN-222, a gain estimate will be produced from each of the 10 flat pairs taken.
Users should adjust parameters when needed, for example, the exposure times or the number of exposures taken.
If the exposure times need to change, it can be done as follows:
exp_times_dark: 20
exp_times_flats: 30
Changing the exposure times and the number of exposures¶
If both the number of exposures and exposure times need to change, it can be done like this:
n_bias: 30
n_dark: 5
exp_times_dark: [5, 10, 15, 20, 25]
n_flat: 10
exp_times_flat: [5, 10, 15, 20, 25, 30, 35, 40, 45, 50]
Example of a configuration file for internal_verification
.¶
Note that the newly-generated combined calibrations
will be certified in the calib_collection
collection, so this parameter must be specified, and new validity ranges should be provided (spanning one day for daily calibrations).
The name of the collection needs to be changed if the script needs to be run again (or the validity range), as it is not possible to certify the same type of calibration in the same collection with the same validity range:
generate_calibrations: True
calibration_collection: LATISS/calib/daily/calib.2022NOV04.1
certify_calib_begin_date: "2022-11-04"
certify_calib_begin_date: "2022-11-05"
Including a Photon Transfer Curve (PTC)¶
In the following example, a new set of calibrations is generated, including a PTC (note that the exposure times need to be given by pairs and the total length must correspond to n_flat
) and defects.
If the individual images taken pass verification using as reference the newly generated combined bias, dark, and flat, the combined calibrations will be certified in the calib_collection
collection with the validity range given by certify_calib_begin_date
and certify_calib_end_date
.
There is the option to take flats with a particular filter and grating (the appropiate names/ID should be replaced in ${FILTER_NAME_OR_ID}
and ${GRATING_NAME_OR_ID}
below):
script_mode: BIAS_DARK_FLAT
n_flat: 14
exp_times_flat: [5, 5, 10, 10, 15, 15, 20, 20, 25, 25, 30, 30, 35, 35]
filter: ${FILTER_NAME_OR_ID}
grating: ${GRATING_NAME_OR_ID}
generate_calibrations=True
calib_collection: "LATISS/calib/daily/calibs.2022NOV04.1"
certify_calib_begin_date: "2022-11-04"
certify_calib_end_date: "2022-11-05"
do_defects: True
do_ptc: True
Another example including PTC and defects generation¶
Another example set of configuration parameters is as follows:
n_bias: 6
n_dark: 6
exp_times_dark: [5, 5, 5, 10, 15, 20]
n_flat: 14
exp_times_flat: [0.1, 0.1, 0.35, 0.35, 0.6, 0.6, 0.8, 0.8, 1.0, 1.0, 1.35, 1.35, 1.6, 1.6]
calib_collection: "LATISS/calib/u/plazas/daily.2021SEP13.test1"
do_verify: True
input_collections_verify_bias: "u/czw/DM-28920/calib.20210720,LATISS/calib"
input_collections_verify_dark: "u/czw/DM-28920/calib.20210720,LATISS/calib"
input_collections_verify_flat: "u/czw/DM-28920/calib.20210720,LATISS/calib"
certify_calib_begin_date: "2021-07-15"
certify_calib_end_date: "2021-07-17"
script_mode: BIAS_DARK_FLAT
do_defects: True
do_ptc: True
Notes:
The
detectors
parameters was omitted, therefore, by default, the single LATISS detector will be passed to the LSST Science Pipelines pipetasks.The
generate_calibrations
parameters was omitted, and therefore combined calibrations will not be generated from the individual images taken (biases, darks, and flats sincescript_mode
isBIAS_DARK_FLAT
), as its default value isFalse
. Pipetasks that require combined calibrations to run will search for them in their input collections. For example, sincedo_verify
isTrue
, the bias, dark, and flat verification tasks will look for combined reference calibrations in their input collections, given by theinput_collections_verify_bias
,input_collections_verify_dark
, andinput_collections_verify_flat
parameters. Since the collectionu/czw/DM-28920/calib.20210720
is located before the standard collectionLATISS/calib
in these parameters, the verification tasks will look there first. On the other hand, sincedo_ptc
isTrue
andinput_collections_ptc
is omitted, the PTC task will look for combined calibrations (e.g., bias, dark) in the standard calibration collectionLATISS/calib
, which is the default for this parameter.Sometimes running the PTC can take a long time. In order to obtain a quick estimation for the gain (and monitor, for example, its stability with time), the parameter
do_gain_from_flat_pairs
can be set toTrue
. In that case, only one pair of flats is required, so the parameterexp_times_flat
could be set to, e.g.,[1.2, 1.2]
. However, the task will estimate a gain for every flat pair that has been taken (LOVE
will report the values per exposure pair per detector per amplifier). For example, ifexp_times_flat
is[0.1, 0.1, 0.35, 0.35, 0.6, 0.6, 1, 1.5, 1.7, 2.1, 2.3]
, gains will be estimated from the first three flat pairs.See DMTN-222 for a discussion on calibration generation, verification, acceptance, and certfication, including suggested naming conventions for parameters such as
calib_collection
.
Launch the script¶
When the configuration options have been entered and the script is ready to be launched, click on the ADD
button in the lower right of the screen (refer to image above).
Accessing the calibrations¶
If generate_calibrations
is True
, the certified combined calibrations will be available via the collection specified by the calib_collection parameter.
They could be retrieved from a notebook for manipulation and visualization:
import lsst.daf.butler as dB
butler = dB.Butler("/repo/LATISS", collections=["LATISS/calib/daily.2021SEP13.test1"])
detector = 0
exposure = [bias1ID, bias2ID] # e.g., [2021071500001, 2021071500002]
# For detector "0":
bias = butler.get('bias', detector=detector, exposure=exposure[0], instrument='LATISS')
dark = butler.get('dark', detector=detector, exposure=exposure[0], instrument='LATISS')
flat = butler.get('flat', detector=detector, exposure=exposure[0], instrument='LATISS')
defects = butler.get('defects', detector=detector, exposure=exposure[0], instrument='LATISS')
ptc = butler.get('ptc', detector=detector, exposure=exposure[0], instrument='LATISS')
If do_gain_from_flat_pair
is True
, the estimated gains (as well as the measured empirical readout noise from the overscan during Instrument Signature Removal) can be found by requesting the cpPtcExtract
data structure.
In this case, the exposure ID should be one of the two flats used to estimate the gain:
cpCovs = butler.get('cpPtcExtract', detector=detector[0], exposure=flat1ID, instrument='LATISS')
gain_values = cpCov.gain
noise_values = cpCov.noise
The gain estimated in this way (from single pairs of flats) is an approximation that is likely to be more accurate at lower fluxes. This method has the advantage that it allows to obtain a quick estimate of the gain without having to take multiple flat pairs to construct a full PTC and to fit a model to it.
In addition, the statistics produced by the verification step can be analized by running the Jupyter notebooks in the examples
folder in cp_verify
.
As it is shown in these notebooks, useful statistics and information about the results of the cp_verify
tests can be retrieved from the butler via (using flat verification as an example):
runStats = butler.get('verifyFlatStats', instrument='LATISS')
runDetStats = butler.get('verifyFlatDetStats', instrument='LATISS', detector=0, exposure=flatExposureID)
The images processed by cp_verify
can also be retrieved for visual inspection:
import lsst.afw.display as afwDisplay
afwDisplay.setDefaultBackend("matplotlib")
imProc = butler.get('verifyFlatProc', detector=0, exposure=flatExposureID, instrument='LATISS')
calibArray = imProc.getImage().getArray()
# Get simple stats
q25, q50, q75 = np.percentile(calibArray.flatten(), [25, 50, 75])
sigma = 0.74 * (q75 - q25)
display = afwDisplay.Display(dims=(1000, 1000))
display.scale('asinh', 'zscale')
display.scale('linear', (q50 - 3.0 * sigma), (q50 + 3.0* sigma), "")
display.mtv(imProc)
DMTN-222 recomends taking 20 individual images for each calibration type. The following example failed verification because only 3 images of each type were taken:
script_mode: BIAS_DARK
n_bias: 3
n_dark: 3
exp_times_dark: 15
calib_collection: "LATISS/calib/u/plazas/daily.2022SEP29.2"
generate_calibrations: True
do_verify: True
certify_calib_begin_date: "1980-01-01"
certify_calib_end_date: "2051-12-31"
This failed verification, in particular, the NOISE
test, likely because of using only 3 images to build the combined images.
The script, at the end, printed in the LOVE output interface:
BIAS calibration failed verification and will not be certified.
Because we are generating combined calibrations, and because we need a certified bias to make a combined dark, the script stopped at this point, as we can’t continue.
In addition, the script issued a WARNING
in the LOVE
interface. The warning tells us the exposures that have tests that failed, the tests which failed, and the amps with tests that failed, per exposure and test type. In the end, we also get the generation and verification collection.
All this information is useful to follow up: look at the images, make plots and tables in a notebook (using copies of the notebooks that can be currently found in the examples
of cp_verify
), etc.
The observer or user will find this info in LOVE as the script runs.
The warning, in this case, is as follows:
Script WARNING: Exposures with verification tests that failed:
2022092900001 2022092900002 2022092900003
Number of tests that failed per test type:
Exposure ID: 2022092900001
CR_NOISE: 8
NOISE: 16
MEAN: 5
Exposure ID: 2022092900002
NOISE: 16
CR_NOISE: 1
Exposure ID: 2022092900003
NOISE: 16
CR_NOISE: 1
Test types that failed verification per exposure,
detector, and amplifier:
Exposure ID: 2022092900001
RXX_S00 C00 CR_NOISE
RXX_S00 C00 NOISE
RXX_S00 C01 CR_NOISE
RXX_S00 C01 MEAN
RXX_S00 C01 NOISE
RXX_S00 C02 NOISE
RXX_S00 C03 NOISE
RXX_S00 C04 CR_NOISE
RXX_S00 C04 MEAN
RXX_S00 C04 NOISE
RXX_S00 C05 CR_NOISE
RXX_S00 C05 MEAN
RXX_S00 C05 NOISE
RXX_S00 C06 NOISE
RXX_S00 C07 NOISE
RXX_S00 C10 CR_NOISE
RXX_S00 C10 NOISE
RXX_S00 C11 CR_NOISE
RXX_S00 C11 MEAN
RXX_S00 C11 NOISE
RXX_S00 C12 NOISE
RXX_S00 C13 NOISE
RXX_S00 C14 CR_NOISE
RXX_S00 C14 NOISE
RXX_S00 C15 CR_NOISE
RXX_S00 C15 MEAN
RXX_S00 C15 NOISE
RXX_S00 C16 NOISE
RXX_S00 C17 NOISE
Exposure ID: 2022092900002
RXX_S00 C00 NOISE
RXX_S00 C01 NOISE
RXX_S00 C02 NOISE
RXX_S00 C03 NOISE
RXX_S00 C04 NOISE
RXX_S00 C05 CR_NOISE
RXX_S00 C05 NOISE
RXX_S00 C06 NOISE
RXX_S00 C07 NOISE
RXX_S00 C10 NOISE
RXX_S00 C11 NOISE
RXX_S00 C12 NOISE
RXX_S00 C13 NOISE
RXX_S00 C14 NOISE
RXX_S00 C15 NOISE
RXX_S00 C16 NOISE
RXX_S00 C17 NOISE
Exposure ID: 2022092900003
RXX_S00 C00 NOISE
RXX_S00 C01 NOISE
RXX_S00 C02 NOISE
RXX_S00 C03 NOISE
RXX_S00 C04 NOISE
RXX_S00 C05 CR_NOISE
RXX_S00 C05 NOISE
RXX_S00 C06 NOISE
RXX_S00 C07 NOISE
RXX_S00 C10 NOISE
RXX_S00 C11 NOISE
RXX_S00 C12 NOISE
RXX_S00 C13 NOISE
RXX_S00 C14 NOISE
RXX_S00 C15 NOISE
RXX_S00 C16 NOISE
RXX_S00 C17 NOISE
Threshold values:
Acceptable maximum number of failures per detector per test type: 8
This value is controlled by the configuration parameter: 'number_verification_tests_threshold_<IMGTYPE>'
Acceptable maximum number of failed detectors: 1
Acceptable maximum number of failed tests per exposure: 8
Acceptable maximum number of failed exposures: 3
Final number of exposures that failed verification: 3
Verification failure criterium: if, for at least une type of test,
the majority of tests fail in the majority of detectors and the
the majority of exposures, verification will fail and the calibration
will not be certified. In terms of the threshold values, this amounts for the condition that
the final number of exposures that failed verification is greater than
or equal to the acceptable maximum number of failed exposures.
Generation collection: u/ocps/329cb5f2f9ab4e1ebacf1577435d2eed
Verification collection: u/ocps/5579591802f0470e88b7cdf6add8e1bf
Trying again with 20 images per calibration type, the majority of verification tests pass in this case. The warning message displayed on the LOVE output will be as follows:
Script WARNING: BIAS calibration passed the overall verification criteria and will be certified, but the are tests that did not pass:
BIAS calibration passed the overall verification criteria and will be certified, but the are tests that did not pass:
Script WARNING: Exposures with verification tests that failed:
2022092900004
Number of tests that failed per test type:
Exposure ID: 2022092900004
CR_NOISE: 6
MEAN: 4
Test types that failed verification per exposure,
detector, and amplifier:
Exposure ID: 2022092900004
RXX_S00 C01 CR_NOISE
RXX_S00 C01 MEAN
RXX_S00 C04 CR_NOISE
RXX_S00 C04 MEAN
RXX_S00 C05 CR_NOISE
RXX_S00 C11 CR_NOISE
RXX_S00 C11 MEAN
RXX_S00 C13 CR_NOISE
RXX_S00 C14 CR_NOISE
RXX_S00 C14 MEAN
Exposure ID: 2022092900005
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900006
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900007
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900008
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900009
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900010
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900011
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900012
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900013
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900014
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900015
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900016
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900017
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900018
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900019
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900020
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900021
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900022
No failures in 'verify_stats' for this exposure.
Exposure ID: 2022092900023
No failures in 'verify_stats' for this exposure.
Threshold values:
Acceptable maximum number of failures per detector per test type: 8
This value is controlled by the configuration parameter: 'number_verification_tests_threshold_<IMGTYPE>'
Acceptable maximum number of failed detectors: 1
Acceptable maximum number of failed tests per exposure: 8
Acceptable maximum number of failed exposures: 11
Final number of exposures that failed verification: 0
Verification failure criterium: if, for at least une type of test,
the majority of tests fail in the majority of detectors and the
the majority of exposures, verification will fail and the calibration
will not be certified. In terms of the threshold values, this amounts for the condition that
the final number of exposures that failed verification is greater than
or equal to the acceptable maximum number of failed exposures.
Generation collection: u/ocps/def62fc0cc6645d089edee4eb797e3f1
Verification collection: u/ocps/85f23b5af0ba44cf967f183c59b1073e
If we take the generation and verificaton collections at the end of the message above, we can make a copy of the bias verification notebook in examples
of cp_verify
and get a table with a summary of the results from running cp_verify
:
We can see an example of failed verification tests by using images taken on September 27, 2022.
In this case, “external verification” was used (i.e., generate_calibrations
was set to False
and the reference combined calibrations used were the ones in LATISS/calib
since no other input collections were specified).
As mentioned before, external calibration is currently the default, per DMTN-222.
In order to make the cp_verify
summary table, we use the following verification collections (note that we don’t have generation collections in this case, as “external verification” was used and there were not newly generated combined calibrations):
Verification collections:
flat: a8382fe83cd24376afe0e3c1a03892c0
dark: c60a2bf9b07a49c19b22fd641f558f65
bias: 71a65a29339140c5917e296b0b05f312
In this case, the script informed that many exposures had too many tests that failed, so the overall verification process failed.
It could be due to the fact that a new LATISS sequencer file was used.
Using the verification collections from above in copies of the cpVerify
notebooks, we can get summary tables that confirm that verification did not pass:
Troubleshooting¶
After checking the configuration options and the LOVE
error messages, the file /scratch/uws/${jobId}/outs/ocps.log
will contain additional technical information on which pipetask failed, if any.
{jobId}
is returned by the OCPS and can be retrieved from the LOVE
output messages.
This procedure was last modified on Nov 20, 2024.