JWST Noise Floor. I. Random Error Sources in JWST NIRCam Time Series

, , , , , , and

Published 2020 October 29 © 2020. The American Astronomical Society. All rights reserved.
, , Citation Everett Schlawin et al 2020 AJ 160 231 DOI 10.3847/1538-3881/abb811

Download Article PDF
DownloadArticle ePub

You need an eReader or compatible software to experience the benefits of the ePub3 file format.

1538-3881/160/5/231

Abstract

James Webb Space Telescope (JWST) transmission and emission spectra will provide invaluable glimpses of transiting exoplanet atmospheres, including possible biosignatures. This promising science from JWST, however, will require exquisite precision and understanding of systematic errors that can impact the time series of planets crossing in front of and behind their host stars. Here, we provide estimates of the random noise sources affecting JWST Near-Infrared Camera (NIRCam) time-series data on the integration-to-integration level. We find that 1/f noise can limit the precision of grism time series for two groups (230–1000 ppm depending on the extraction method and extraction parameters), but will average down like the square root of N frames/reads. The current NIRCam grism time-series mode is especially affected by 1/f noise because its GRISMR dispersion direction is parallel to the detector fast-read direction, but could be alleviated in the GRISMC direction. Care should be taken to include as many frames as possible per visit to reduce this 1/f noise source: thus, we recommend the smallest detector subarray sizes one can tolerate, four output channels, and readout modes that minimize the number of skipped frames (RAPID or BRIGHT2). We also describe a covariance-weighting scheme that can significantly lower the contributions from 1/f noise as compared to sum extraction. We evaluate the noise introduced by pre-amplifier offsets, random telegraph noise, and high dark current resistor capacitor (RC) pixels and find that these are correctable below 10 ppm once background subtraction and pixel masking are performed. We explore systematic error sources in a companion paper.

Export citation and abstract BibTeX RIS

1. Introduction

The James Webb Space Telescope (JWST) will provide powerful new measurements of exoplanet atmospheres with spectroscopy from 0.6 to 28 μm (Beichman et al. 2014; Barstow et al. 2015; Greene et al. 2016; Howe et al. 2017; Schlawin et al. 2018). The sensitivity of JWST combined with its unprecedented wavelength coverage will enable it to measure the abundance of carbon-bearing molecules (CO, CO2, and CH4) in exoplanet atmospheres as well as study cooler, smaller planets than previously characterized.

JWST has also been considered for observations of potentially habitable Earth-like planets that orbit their stars at a distance where water could be in liquid form on their surface. The TRAPPIST-1 system (Gillon et al. 2016, 2017) presents an exciting opportunity to study Earth-sized planets orbiting a 0.1 R M-type star. By putting together 10–30 transits, it will be possible to collect enough photons to detect CO2 (Barstow & Irwin 2016; Krissansen-Totton et al. 2018; Lustig-Yaeger et al. 2019). Optimistically, 30 transits may contain enough photons to detect O3 in the atmosphere of TRAPPIST-1d with JWST if clouds have a minimal impact on its atmosphere (Barstow & Irwin 2016). More detailed modeling of atmospheric evolution indicates that biogenic oxygen may be too difficult to detect with any JWST instrument but desiccated oceans may produce detectable abiogenic oxygen (Lustig-Yaeger et al. 2019). Alternatively, 10 transits of TRAPPIST-1e could be sufficient if the planet's atmosphere has CO2 and CH4 biosignatures similar to early Earth (Krissansen-Totton et al. 2018). A critical question in studying planets like TRAPPIST-1d/e with small spectroscopic signatures, however, is whether photon-limited performance is possible with JWST observations. For difficult biosignatures, noise levels less than 10 ppm (parts per million) are required and this value shrinks in the presence of clouds.

Experience with the Hubble Space Telescope (HST), Spitzer, and Kepler shows that many systematics can affect high-precision time series and prevent photon-limited performance unless corrected (e.g., Beichman et al. 2014). Many of these effects are detector-related including charge trapping in HST's detector (Berta et al. 2012; Zhou et al. 2017), intrapixel sensitivity on the short wavelength (SW) bands of Spitzer IRAC (Morales-Calderón et al. 2006), and sensitivity to pointing jitter with Kepler 4 (e.g., Vanderburg & Johnson 2014; Beichman et al. 2014).

The Near-Infrared Camera (NIRCam; Rieke et al. 2005) contains a slitless grism within the instrument's long wavelength (LW) channel (Greene et al. 2017). This slitless grism mode is similar to the WFC3 grism on HST that has been successfully employed on many transiting planets (e.g., Deming et al. 2013; Kreidberg et al. 2014; Sing et al. 2016; Wakeford et al. 2017). NIRCam grism observations will collect moderate resolution R ∼ 1100–1700 spectra over the wavelength range from 2.4 to 5.0 μm. A dichroic beamsplitter allows simultaneous operations of the SW weak lens imaging or (in future cycles) Dispersed Hartmann Sensor spectroscopy (DHS; Schlawin et al. 2017). After approval and implementation, the DHS mode will permit spectroscopy from 1.0 to 2.0 μm using a dispersive device that measures relative mirror phases. The DHS mode will also permit time-series spectroscopy on very bright targets (KS  ∼ 1 mag in the Vega system) inaccessible by other JWST instrument modes.

Given the finite lifetime of the James Webb mission, it is important to characterize the random and systematic errors before launch to maximize the science return of the observatory. We focus this paper on the random errors and a companion paper on the systematic errors that can affect time-series observations with the NIRCam instrument. Parallel efforts to characterize other instrument's noise sources are being carried out, such as with the detectors of the Mid-Infrared Instrument (MIRI; Matsuo et al. 2019).

Section 2 lists the known random effects that can degrade the precision of time-series measurements compared to the photon noise limit. Sections 3 through 5 detail these noise contributions. We conclude in Section 6 that 1/f noise is the largest random error to impact NIRCam grism time-series measurements. Paper II will address the systematic errors like pointing jitter, nonlinearities, variable aperture losses, and charge trapping that degrade the precision of time series measurements.

2. List of Known Detector Effects

An ideal HgCdTe photodiode array would simply count the photons incident on its surface (as described in Appendix A). The NIRCam, Near InfraRed Spectrograph (NIRSpec), and Near-Infrared Imager and Slitless Spectrograph (NIRISS) instruments on JWST use Teledyne HAWAII-2RG (H2RG) HgCdTe detectors (e.g., Gardner et al. 2006; Rauscher et al. 2012). For NIRCam, the eight SW detectors have a 2.5 μm cutoff while the two LW detectors have a 5.2 μm cutoff (STScI 2017). Each of the detectors is connected to a System Image Digitizing Enhancing Controlling And Retrieving Application Specific Integrated Circuit (SIDECAR ASIC) that controls how the pixels are clocked, reset, and read (Loose et al. 2006). Each detector has a default full-frame configuration with four amplifier outputs that are read out simultaneously on four vertical subsections of the detectors (depicted in Figure 1), or alternatively are read out with a single amplifier at about one-fourth the data rate.

Figure 1.

Figure 1. In FULL frame imaging and STRIPE mode, four amplifiers are used to simultaneously read out four vertical stripes of pixels. A  four-pixel wide boundary of reference pixels can be used to subtract pre-amplifier offsets and some 1/f noise. The fast-read directions are marked by red arrows, while the slow-read directions are marked by green arrows. Sources dispersed by the GRISMR element cross the amplifier boundaries and are parallel to the fast-read direction, making spectra susceptible to residual 1/f noise not removed through reference pixel correction. Note: the image is not to scale—the reference pixels have been exaggerated for visibility.

Standard image High-resolution image

Each active pixel's Negatively doped-type Positively doped-type (NP) semiconductor junction may be measured nondestructively multiple times (referred to as reads or samples) before the pixel is reset, ending an integration. The NP semiconductors are discussed in more detail in Appendix A. An analog-to-digital converter (ADC) turns the number of electrons collected in the detector pixel into a count of data numbers (DN) from 0 to 65,535 (i.e., a 16-bit number) with a conversion gain of ∼1.8 e/DN for the LW detectors and ∼2.0 e/DN for the SW detectors. One pass through all the pixels constitutes a frame. An integration consists of a collection of frames that ends when the pixels are reset. The frames may be coadded into groups of one or more frames to reduce data bandwidth, and the integrations are packaged together into a single exposure. In this paper, we discuss the RAPID read mode (with one frame per group), BRIGHT1 (with one frame per group and a skipped frame between groups), and BRIGHT2 (with two frames per group). STScI (2020a) contains more information on the readout patterns. In addition to active pixels, there is a boundary of 4 pixels that are electrically identical to the active pixels but are not sensitive to light (Loose et al. 2003; Rieke 2007). These can be used to correct for amplifier voltage drifts.

In the ideal case, the number of DNs measured by the detector will be proportional to incident flux from an astrophysical source such as a star–planet system. If there are no astrophysical variations, the only variability of a light curve will be the Poisson photon shot noise. There are many nonideal effects, however, that can affect the stability of a light curve measurement of a science target. Here, we review many of the known noise sources that could impact the stability of a time series beyond photon counting statistics. The detector effects are listed briefly and then discussed in more detail as related to the JWST NIRCam detectors below that.

The following sources of noise introduce random variations to time series (on top of the photon noise) on an integration-by-integration level.

  • 1.  
    Pre-amplifier reset offsets (a SIDECAR ASIC–related effect). When the electronic pre-amplifiers in NIRCam's SIDECAR ASICs are reset, there is a noise source during the reset process. This noise will add a random offset to the pedestal level to all pixels corresponding to one output amplifier. Pre-amplifier resets and voltage drifts can also appear in slope images (DN s−1). The noise introduced by pre-amplifier offsets cannot be decreased by averaging more pixels within an amplifier region of a frame, but reference pixel or background subtraction can efficiently remove the pre-amplifier offsets. For scientists familiar with CCDs, the pre-amplifier offsets look similar to a bias frame that changes in time. These are discussed further in Section 3.
  • 2.  
    Amplifier boundary discontinuities. The NIRCam detectors can be read out in FULL frame, STRIPE, and WINDOW modes. The FULL frame and STRIPE modes make use of four amplifiers to simultaneously record or reset pixels. The parallel operation of the amplifiers reduces the frame time and reset time by a factor of ∼4 but can introduce subtle voltage biases between the readout channels. If these voltage biases are not corrected, they can potentially introduce discontinuities in a spectrum which crosses amplifier boundaries. The amplifier boundary discontinuities are a consequence of the interplay between the detector video output lines with pre-amplifier reset offsets described above as well as the ADCs. Slight gain differences can exist between the four vertical stripes read out by the four amplifiers, but these are stable in time and will come out with flat fielding.
  • 3.  
    1/ f noise. The JWST HgCdTe detector readout system adds correlated read noise to the digital images. This correlated noise is caused by (1) the readout integrated circuits (ROICs), which have a p-type metal-oxide semiconductor field-effect transistor (PFET) source follower, as well as (2) direct current (DC) biases in the SIDECAR ASIC electronics (Rauscher et al. 2011). This electronic read noise has the property that most of the noise power is concentrated at low frequencies. 1/f noise causes spatial correlations primarily in the fast-read direction (along detector rows). The 1/f noise from the SIDECAR ASICs is highly correlated between amplifiers because it is caused by a common reference voltage. However, the 1/f noise from the PFETs can vary between amplifiers because each video output has its own PFET (Rauscher et al. 2011). The 1/f noise can be reduced by subtracting the values from background pixels or reference pixels that are read closely in time. The NIRSpec instrument on JWST has a mode called IRS2 specifically designed to reduce 1/f noise (Rauscher et al. 2011), but it is not available on NIRCam.
  • 4.  
    Even/odd offsets. Alternating columns have different bias levels. Additionally, there are offsets between columns that can change from frame to frame, especially on NIRCam's LW detector. These can be reduced with reference pixels or background subtraction.
  • 5.  
    kTC noise. Each reset at the beginning of an integration introduces thermal noise associated with the unknown amount of charge stored in a pixel at the reset voltage level. Fortunately, the bias offset produced by kTC noise will completely subtract out when fitting a slope image from multiple samples up the ramp. This noise would only become relevant when attempting to make use of the first group of a ramp, such as when only one group is commanded.
  • 6.  
    Elevated columns during an integration. Entire columns appear to spontaneously increase by a few to 10 s of DN. The behavior of these elevated columns can be seen to move to adjacent columns in subsequent integrations. The root cause for the column transitions and its movement is still being investigated.
  • 7.  
    Random telegraph noise (RTN). Some of the pixels will exhibit spontaneous jumps in signal even with no illumination (Bacon et al. 2005). Fortunately, RTN appears in only ∼1000 pixels out of 4 × 106 on the array. The RTN pixels in the Long Wavelength A-side (ALONG) detector used for grism time series include many jumps, of order ∼50 DN, up and down along a 20 minute exposure.

We discuss the above noise contributions in more detail in Sections 3 through 5.

In addition to these random effects, there are also systematic effects that do not average out when more integrations are collected. These can also be functions of other variables such as temperature, telescope pointing, or the number of filled charge traps in pixels' photodiodes. Paper II will discuss these systematic effects, including intrapixel sensitivity, detector temperature fluctuations, charge trapping, nonlinearities, variable aperture losses, and reciprocity failure. Here, we will focus on the random effects that are largely independent from one integration to another.

3. Pre-amplifier Offsets and Amplifier Boundaries

3.1. Pre-amplifier Offsets

The NIRCam detectors' pre-amplifiers in the SIDECAR ASIC have a reference voltage that will drift with time, which will affect the pedestal level of all pixels referenced to that voltage. This was introduced as error source 1 in Section 2. To address the reference voltage drift, the pre-amplifiers can be reset to the reference voltage periodically, which will introduce a random uncertainty into the pedestal level. The SIDECAR ASICs support resetting to the reference voltage either at frame boundaries (between frames) or at integration boundaries (between integrations). In early ASICs the pre-amplifier drift between resets was seen to be high enough to allow the pedestal to drift off the range of the ADC during longer integrations; hence, the NIRCam team elected to reset the pre-amplifiers on frame boundaries rather then integration boundaries. Subsequent delivery of flight ASICs showed much lower drift rates. To characterize the impact of reset per frame (RPF) versus reset per integration (RPI), a set of darks was collected during NIRCam cryogenic vacuum (CV) pre-delivery testing at Lockheed-Martin in both RPI and RPF mode. In these tests, the pedestal level of all pixels could be seen to drift over 19.3 minute dark integrations by 5–100 DN in RPI mode. The drift was usually linear over these 19 minute exposures, but in one case was observed to drift downward from 100 DN to −50 DN and back to 100 DN (Robberto 2014) while RPF introduces a random uncertainty of the order of 30 DN rms at each frame boundary in the sample up the ramp. Reference pixel correction effectively removes both the random RPF jumps and the systematic RPI drift; analysis (Robberto 2014) showed that either approach was roughly equivalent. Operationally, NIRCam elected to continue using RPF since there was no benefit to switching to RPI and significant legacy characterization data had been obtained in the RPF mode. Reference pixel correction or careful background subtraction can efficiently remove the signature of the pre-amplifier reset or drifts. The NIRCam grism subarrays and dispersed source location are placed at the edge of the NIRCam LW detector to ensure that the reference pixels are nearby and saved with active pixels.

Figure 2 shows an example time series of the reference pixels for a dark exposure. The pedestal/bias level of the reference pixels undergoes a sharp jump between each frame (10.7 s in duration). Fortunately, the pixels within an amplifier move together, so reference pixel subtraction or background subtraction will efficiently remove the offsets.

Figure 2.

Figure 2. Reference pixel time series for a two frame integration (left) and a long dark integration of 108 frames (right). Each of the plots show that the SIDECAR ASIC pre-amplifier resets between frames in an integration cause discontinuities in the reference signal. Amplifiers 1 and 4 track the time all through a frame with the side reference pixels, whereas amplifiers 2 and 3 in the middle only contain information on the bottom and top reference pixels at the beginning and ending of a frame. See Figure 1 for a layout of the detector. The long dark exposure example of 108 frames shows the behavior over long timescales.

Standard image High-resolution image

3.2. Amplifier Boundary Discontinuities

NIRCam's STRIPE mode and full-frame mode both employ four amplifiers to simultaneously read out or reset 4 pixels at once. This simultaneous use of four amplifiers reduces the frame time (to scan through all pixels in a 2048-pixel-wide subarray or full frame) by a factor of about 4. Thus, the STRIPE and full-frame modes increase the brightness of objects that can be detected before saturation as well as increase the number of reads for a detector pixel before saturation to average down readout noise. Users have the option to select one or four amplifiers with the number of output channels in the Astronomer's Proposal Tool (APT; STScI 2020b). On the flip side, the four amplifiers will have different behaviors such as separate gains and offsets in the bias level. The differential biases between the amplifiers will show up as step function changes in the slope of the image at the amplifier boundaries (512, 1024, and 1536), as visible in Figure 3 (top panel). If a spectrum of a source is extracted without any vertical background extraction or reference pixel correction, the amplifier offsets cause sharp discontinuities in the summed spectrum. The image shown here is from a grism exposure during a cryogenic vacuum testing 3 (CV3) test at NASA Goddard. The locations of amplifier boundaries and reference pixels are visible in Figure 1.

Figure 3.

Figure 3. An example slope image from a CV3 grism test, which consisted of 10 groups in RAPID read mode for an integration time of 107 s, and trimmed to a 2048 × 256 subarray. The median profile of the image at ∼3.5 μm is shown to the left of each image. First panel: no reference pixel correction is applied, leaving offsets at the amplifier boundaries. Second panel: averaging all the reference pixels within an amplifier and subtracting the result reduces the offsets between amplifiers but there is still a subtle boundary at X = 1536. Third panel: reference pixel correction using the four nonilluminated pixels in each column largely removes amplifier offsets, but introduces read noise to each column. Fourth panel: background subtraction with the mean value of the pixels below Y = 30 and above Y = 180 eliminates offsets between amplifiers without introducing any substantial new noise source.

Standard image High-resolution image

Fortunately, the amplifier offsets are efficiently removed by reference pixels and background subtraction. For example, subtracting the average reference pixel in each amplifier will reduce the offsets between amplifiers, shown in Figure 3 (second from the top panel). One could subtract the mean reference pixel in each column, shown in Figure 3 (third from the top panel), but given that there are only four reference pixels at the bottom of each column, this introduces extra noise in the subtraction. Instead, one can use the background region for each column (∼105 pixels used here) to improve the removal of gradients and the amplifier offsets. In Figure 3 (bottom panel), we determine the median of all pixels from Y = 0 to 30 and Y = 180 to 255 in each column and subtract this value from the whole column. This is repeated for all columns. This background-subtraction method can efficiently remove the pre-amplifier offsets to the level where amplifier discontinuities become smaller than the combined read and photon noise in the image. This background-subtraction method can be applied to the STRIPE subarrays that are 64, 128, 256, or the full 2048 pixel tall. We note that some amount of self-subtraction occurs from the extended wings of the point-spread function (PSF) for a 64 pixel tall subarray. A 20 pixel aperture contains ∼6% less flux than a 128 pixel aperture at 3.5 μm, but will not produce a significant time dependence to the flux. We discuss aperture losses more in Paper II.

4. 1/f Noise

The readout circuitry for NIRCam and other JWST near-infrared instruments introduces correlated noise from pixel to pixel. This is visible when examining the time series of pixels as they are read out within a frame in the order they are addressed, as shown in Figure 1. 5 The spatial coordinates are relabeled as time coordinates assuming a 10 μs pixel address time plus overheads between rows, so there is 1 pixel per 10 μs within a row with 120 μs gaps between rows. The power spectrum of this pixel-by-pixel time series reveals high power at low frequencies that drop approximately like a 1/f power law.

Figure 4 shows the average Lomb–Scargle periodogram from output amplifier 2 (covering X = 512 pixels to X = 1023 pixels) over 108 dark groups in the RAPID read mode. The dark frames were part of a single integration taken at NASA Johnson during the OTIS (Optical Telescope element + Integrated Science instrument module) cryogenic test. To create the periodogram, each group is subtracted by the average of all groups to remove the bias and all pixels that deviate by more than 80 DN from the average are masked to eliminate bad pixels.

Figure 4.

Figure 4. Left: the average Lomb–Scargle periodogram from the time series of amplifier 2 over 108 RAPID groups. The noise in the spectrum is close to a 1/f power law (orange line) from 10 to 1000 Hz. Above 1000 Hz, it is closer to white noise (green line) with spikes at specific frequencies. The top x-axis shows the inverse frequency as measured in the number of clock cycles, which corresponds to the number of pixels read within a row. Right: this 1/f noise introduced in the readout circuit causes prominent correlated noise along the fast-read direction (horizontal) on a dark frame. The amplifier offsets have been removed, but the 1/f noise from these adjacent amplifiers (separated at X = 512) shows different behavior.

Standard image High-resolution image

Figure 4 shows that the noise power drops nearly like 1/f from the lowest frequency in a frame to ∼1000 Hz, where the curve flattens out. This indicates that the noise is closer to white noise, where each pixel's noise would be statistically independent from the others, at high frequencies. The periodogram also shows many spikes at high frequency on top of the 1/f noise. One of these prominent spikes occurs at the line reset time of 524 clock cycles, which is the time to read out all the pixels in a row for the amplifier and line overheads before moving to the next row.

The 1/f noise is compared to a simulated time series with the same robust standard deviation but with Gaussian white noise where each pixel is independent and identically distributed as a Gaussian. It is clear from Figure 4 that the 1/f noise and the periodic signatures significantly increase the power over uncorrelated white noise (green line) at low frequencies. 1/f rises above the white noise level at 70 cycles/1400 Hz for amplifier 2 and the intersection ranges from 70 to 135 across the four amplifiers. This power spectrum shows that the read noise is significantly correlated from pixel to pixel and this will affect both high-precision spectroscopy and high-precision photometry.

The 1/f noise can be reduced by subtracting reference pixels and/or active pixels surrounding the aperture of interest. As the power spectrum shows, the 1/f noise begins to contribute significantly for frequencies below 1000 Hz (i.e., timescales longer than 100 cycles, where a cycle is the 10 μs time to address a pixel). One hundred cycles correspond to 100 pixels, except between rows and frames due to overheads. Ideally, source apertures would be less than 100 pixels wide (equivalent to 1 ms of time) so that background subtraction along rows will remove the 1/f noise but this is impossible for grism spectra dispersed along the rows. The NIRCam grism time series uses the GRISMR element in the pupil wheel to disperse along the rows. NIRCam has a GRISMC element that will disperse along the column direction. We discuss the GRISMC element further in Section 4.3.1.

It is also possible to modify the detector readout to reduce the 1/f noise by reading reference pixels more regularly. This is accomplished by the NIRSpec IRS2 mode, which reduces the read noise from 25 e to 10 e and also reduces the correlations between pixels (Rauscher et al. 2011). However, the way that the NIRCam focal plane electronics transmits telemetry packets will not work for IRS2.

4.1. Long Darks

It is instructive to look at dark frames to isolate the contributions of the readout electronics, including the SIDECAR ASIC, and how much they contribute to the error budget. The dark frames do not have issues with the photometric stability of illumination lamps, which is a persistent challenge with ground-based tests. They also do not contain photon noise in the background, which can make electronic noise sources harder to see. We study here the contributions from read noise, including the 1/f noise from the electronics. This experiment with dark frames does not test subtle changes in behavior of the electronic read noise under illumination, such as crosstalk where pixels that have no direct flux from a source in an image correlate with other regions of the detector with high flux rates.

A standard long dark exposure used for detector characterization is a single integration that is 108 groups of one frame each (RAPID mode). The detectors are read in the full-frame mode so that each frame is 10.73677 s in duration with the four output amplifiers in use, thus the total time for the exposure is 19.3 minutes. We use one of these long dark integrations to create simulated time series as if it were observing many consecutive integrations in the full-frame mode. The long darks are broken into 54 sequential pairs of images that are subtracted from each other (group 1 minus group 0, 3 minus 2, 5 minus 4, etc. in a zero-based counting scheme). This method of breaking up the sequence precludes the ability to study the effects of detector resets, which only occur at the beginning of an integration. We find that the detector reset does affect the dark current in an integration, but we do not expect detector resets to significantly impact the time series. This is discussed further in Appendix D. We also relabel the times to include the reset time that would normally occur after each integration, so the duration of the exposure from start to finish is 29.0 minutes. The relabeling of the times makes the simulated exposure similar to a real time-series observation but does not impact the analysis of the time series, which focuses on the scatter in the flux between read pairs.

Figure 5 shows an example dark read pair from frame 55 minus frame 54. In the raw data, there are four prominent vertical stripes that are 512 pixels wide and 2048 pixels tall due to offsets in the readout amplifiers. These amplifier offsets are caused by drifts in the reference voltage that are reset once per frame and vary from amplifier to amplifier. Additionally, there are obvious correlations along the fast-read (X) direction due to 1/f noise that appear as horizontal stripes in the image.

Figure 5.

Figure 5. Time-series data of a 70 pixel radius photometric aperture obtained during a long dark exposure for detector ALONG = A5. The long dark exposure of 108 frames is used to construct 54 pairs of reads that are subtracted, in analogy with a series of integrations. Top: example read pair subtraction with an overlay of the apertures for three different reduction steps from raw data using no reference pixel correction (left), reference pixel correction that uses the top, bottom, and side reference pixels (middle), and finally a method using a row-by-row median subtraction (right). Bottom: the time series for each circular aperture is calculated from the above images with photutils using the source circular aperture and a circular background annulus. The raw frame photometry (left) has a standard deviation of 23,500 DN, which is reduced to 10,800 DN with reference pixel correction (middle). Row-by-row subtraction lowers the contribution from 1/f noise to 6000 DN (right).

Standard image High-resolution image

The ramps-to-slope pipeline ncdhas or the STScI JWST pipeline will apply reference pixel correction, bad pixel masks, and nonlinearity correction before fitting the groups up the ramp to a line. In this work, we use ncdhas. The reference pixels are comprised of a 4 pixel boundary along the edges of the detector, which are not connected to detector photodiodes and can therefore be used to measure common-mode electronic noise without being affected by photon counting noise. As seen in Figure 5, reference pixels also share the amplifier offsets, so they can be used to reduce the large offsets between amplifiers.

The reference pixels are most useful for subtracting the offsets between amplifiers because there are eight rows of 512 pixels or a total of 4096 reference pixels to average for each amplifier at the bottom and top of the arrays, as shown in Figure 1. There are an additional four columns on the left and right side boundaries (8176 pixels) that can be used to further correct the left and right amplifier offsets. However, the 1/f noise that correlates along the fast-read (X) direction is more difficult to remove with reference pixels. Only 4 pixels on the left and 4 pixels on the right side may be used to subtract the 1/f noise in a given row and the four amplifiers can exhibit different 1/f noise. Therefore, significant residuals remain along the fast-read (X) direction in the reference-corrected read pair shown in Figure 5. These side reference pixels also cannot correct for read noise correlations that happen on shorter length scales than 512 pixels (5 ms in time).

4.2. Photometric Aperture

Exoplanet transit measurements are dependent on the photometric stability within an aperture or spectrum. We therefore perform aperture photometry on the dark frame as if it were a weak lens image with a 70 pixel radius circular source aperture and a 75 to 100 pixel background annulus illustrated in Figure 5. We place three different source apertures with the same radii on the detector for comparison of different regions. The background subtraction will largely remove the offsets affecting the whole 512 × 2048 vertical stripe read out by one amplifier. However, the correlated 1/f noise along the X direction will not be subtracted efficiently by an annulus. Instead, a row-by-row line fit to the background can be more effective.

For a read pair subtraction, the correlated double sample (CDS) read noise is 13.3 e or 7.2 DN for the A5 detector studied here. If we assume that each pixel's read noise is uncorrelated with its neighbors, the total noise within the source aperture is then $\sqrt{{N}_{\mathrm{px}}}\times \mathrm{RN},$ where Npx is the number of pixels and RN is the read noise of a single pixel. If we also assume that the background pixels are uncorrelated with each other and the source pixels, then the total noise in background-subtracted photometry is 1310 DN.

This theoretical read noise is smaller than the photon noise expected inside an aperture during photometric stability time-series applications. To calculate the photon noise as well as convert the noise into ppm, we need to assume a well filling fraction for the integrations. Here, we assume a peak pixel count of 30,000 DN or about 60% well depth. With this detector well filling fraction and a weak lens plus 8 waves (WLP8) PSF, the number of DN collected within a 70 pixel radius aperture is Stot = 3.9 × 107 DN during the integration. The Poisson shot noise for photons is ${\sigma }_{\mathrm{phot}}=\sqrt{{S}_{\mathrm{tot}}}/\sqrt{G}=4400$ DN, where G is the gain (G ≈ 1.8 for the LW detectors). In terms of fractional noise, the photon shot noise should be 113 ppm, and the read noise should be 33 ppm. Therefore, the read noise is expected to be about one-fourth the shot noise for a two-group RAPID read mode (i.e., CDS). When averaging together 111 full frames over 60 minutes (an order-of-magnitude transit duration for most planets), the photon noise would contribute 11 ppm and the read noise would contribute 3 ppm.

In our experimental time series, however, the standard deviation in the time series in these three aperture source locations is 22,000 to 25,000 DN (without any attempt to correct for amplifier offsets or 1/f noise using reference pixels or otherwise). If uncorrected, this extra read noise will contribute 560 to 630 ppm to the time series per read pair! This read noise would be ∼5× the ∼113 ppm photon noise. However, there are ways to significantly reduce the pre-amplifier offsets and 1/f noise.

These large correlations in read noise can be reduced with the reference pixels, which are designed to track electronic noise without the presence of background photon noise. We use the ncdhas pipeline to apply reference pixel correction, which dramatically reduces the amplifier offsets and also reduces 1/f noise correlations along the fast-read (X) direction. The standard deviation of the flux for the three source apertures considered drops down to levels from 10,000 DN to 11,500 DN (depending on the aperture), but still falls short of the expectation of 1310 DN if the read noise is uncorrelated.

The background annular aperture can affect the precision of a light curve, but there is a balance between photon noise, background noise, and 1/f noise. The aperture radius studied here (70 pixels) has an area (15,400 pixels2) that is larger than the area of background annulus (13,700 pixels). If background and photon noise were the only factor, the background annular aperture could be larger to increase the precision. However, a large background becomes decreasingly correlated with the source pixels and 1/f noise makes using a background annulus much larger than 100 pixels less helpful for foreground-limited exoplanet system observations.

4.2.1. Row-by-row Subtraction

It is important to subtract the values from either background pixels or reference pixels that are read out as close in time to the science target pixels as possible, where the correlations are strong. Therefore, the column-by-column background subtraction shown in Figure 3 along the slow-read direction is not very good at removing 1/f noise, even though it efficiently removes amplifier offsets. Instead, pixels of the same row (i.e., along the fast-read direction) are preferable to reduce 1/f noise.

The eight reference pixels in each row (as shown in Figure 1) are not sufficient to remove all 1/f noise in each row because the 8 pixels have their own read noise and they do not fully encapsulate the row's 1/f behavior. Therefore, we perform another correction step that uses more pixels. We find the median of each group up the ramp along the fast-read (X) direction outside the source aperture and subtract this from every row. We call this a row-by-row background subtraction. The algorithm to perform row-by-row background subtraction is as follows.

  • 1.  
    Choose one group of an exposure.
    • (a)  
      Remove the pre-amplifier offsets within each amplifier by subtracting the median reference pixel.
    • (b)  
      Select a row of the group image.
      • i.  
        Select all pixels outside the source aperture and outside other bright sources.
      • ii.  
        Calculate the median pixel count of the selected pixels.
      • iii.  
        Subtract this median pixel value from all pixels in the row including the source aperture.
    • (c)  
      Repeat Step 1b for all rows in the group.
  • 2.  
    Repeat Step 1 for all groups in an exposure because each group will have its own pre-amplifier offsets.

With row-by-row subtraction, the standard deviation of the time series drops to 6000 DN, but is still larger than the expectation of 1310 DN if all noise were uncorrelated. Row-by-row subtraction is shown in Figure 5, both for an example image as well as a time series. We note that Steps 1a–1c nearly commutes with the slope fit for an integration (within 1 ppm), so one can also apply Steps 1a–1c on a slope image rather than group by group as in Step 2.

4.2.2. Row-by-row Amplifier-by-amplifier Subtraction

A modification to the above strategy is to find the median of each amplifier individually, which we call row-by-row amplifier-by-amplifier subtraction. This takes care of any differences between amplifiers. The algorithm is the same as in Section 4.2.1, with a modification to Step 1b to select the pixels of a row that correspond to one amplifier (512 pixels per amplifier for FULL or STRIPE subarrays) and to repeat this for all four amplifiers.

Row-by-row amplifier-by-amplifier subtraction dramatically decreases the errors as compared to reference pixels alone. The standard deviations of the background-subtracted photometry range from 2540 to 3300 DN or 65 to 85 ppm across detectors, as listed in Table 1. Therefore, this simple approach can dramatically decrease the 1/f noise to well below the photon limit. The row-by-row amplifier-by-amplifier subtraction is possible for weak lens photometry but not GRISMR spectroscopy, which is dispersed parallel to detector rows (the fast-read direction). This is because the spectrum will span across the entirety of one or more amplifiers, as shown in Figure 1, so there are no source-free pixels in one or more amplifiers to calculate the median 1/f noise from. We note that row-by-row amplifier-by-amplifier subtraction can be difficult in dense stellar fields. If there are bright sources along the same rows as a source and also within the same amplifier, it can make finding clean background regions in a 512-pixel-wide amplifier challenging. In this case, one may have to revert to the row-by-row subtraction discussed above in Section 4.2.1.

Table 1. Summary of Noise Sources in Long Dark Test

Aperture TypeWLP8Grism 0.17 μm bin
Source Aperture Size70 px RadiusΔX = 165 px, ΔY = 14 px
Noise Description σ σ σ σ
 DNppmDNppm
Ideal Photon Noise44001131430388
Ideal Uncorrelated Read Noise13003335094
Measured Read Noise with Amplifier Offsets (No Correction)23,50060080,93021,900
Measured Read Noise (Reference Pixel Correction)10,80028099002680
Measured Read Noise (Row-by-row Subtraction)600015038601044
Measured Read Noise (Row-by-row Amplifier-by-amplifier Mean Subtraction)290075  
Measured Read Noise (Smoothed 161 pixel Kernel Interpolating Over Aperture)330085  
Measured Read Noise (Reference Channel Subtraction)9700250  
Measured Read Noise (20 Component PCA Subtraction)5400140  
Measured Read Noise (5 Component PCA on Each Amplifier)320081  
Measured Read Noise (Variance-weighted Summation)  3060830
Measured Read Noise (Covariance-weighted Summation)  850230
Measured Read Noise (GRISMC Row-by-row Subtraction)  683205

Note. We evaluate the theoretical and measured noise sources from 54 read pairs contained in a long dark integration. The ppm units and photon uncertainties are calculated by assuming a peak 60% well depth or an aperture sum of 3.9 × 107 DN during an integration for the weak lens PSF and 3.7 × 106 DN during and integration for the grism PSF. The noise is reported for the aperture/spectral extraction (in DN) or relative to the photon signal (in ppm). The integration is assumed to be a bright source with two groups of 1 integration each up the ramp (CDS mode). For the grism, statistics are shown for a representative 0.165 μm wide (165 pixel) wavelength bin centered on 3.34 μm with a spatial extent of 14 pixels in the Y-direction. The principal component analysis (PCA)-based methods and reference channel subtraction are discussed in Appendix B.

Download table as:  ASCIITypeset image

4.2.3. More Sophisticated 1/f Noise Reduction Techniques

We also consider more sophisticated ways to reduce read noise than the median row subtraction discussed in Sections 4.2.1 and 4.2.2. These include a smoothing kernel, using an amplifier as a reference channel and PCA. None of these techniques yielded better results than the row-by-row amplifier-by-amplifier subtraction in Section 4.2.2, but the ideas are discussed in Appendix B to be used as a reference for future work to reduce 1/f noise.

4.3. Grism Extraction

The current NIRCam grism time-series mode, which uses the GRISMR element in the pupil wheel, is particularly susceptible to 1/f noise because the dispersion direction is parallel to the detector's fast-read direction as shown in Figure 1. This is also visible in Figure 6 where the grism apertures (in red) are parallel to the 1/f noise correlations. The background subtraction along the spatial direction (vertically) will efficiently remove pre-amplifier offsets from frame to frame and amplifier to amplifier. However, there are only limited background regions along the horizontal direction available to subtract out 1/f noise. This will result in increased noise on the right three amplifiers for the F322W2 filter (2.4–4.0 μm over ∼1600 pixels) and the left two amplifiers for the F444W filter (3.9–5.0 μm over ∼1100 pixels), in the native raw detector pixel coordinate system used in this paper.

Figure 6.

Figure 6. Top row: the read pairs have significant pre-amplifier offsets and 1/f noise (left), which are reduced by reference pixels (middle) and column-by-column, row-by-row subtraction (right). The spectrum (outlined in red aperture box) is dispersed along the fast-read (horizontal) direction, so it is particularly susceptible to 1/f noise. We extract the full spectrum and create normalized time series for three example wavelength bins with aperture sizes with ΔX = 165 px (0.165 μm) and ΔY = 14 px (088). Bottom row: the standard deviation of the time series dramatically decreases from 21,900 ppm to 2680 ppm with reference pixels and 1044 ppm with column-by-column, row-by-row subtraction for the 3.34 μm bin. 1/f noise still dominates over the theoretical photon noise limit of 388 ppm for the 3.34 μm bin. Here, the simulated grism image for an F322W2 filter spans three amplifier channels (each 512 pixels wide) and part of the leftmost amplifier, so only reference pixels and horizontal pixels with X < 300 can be used to reduce the 1/f noise.

Standard image High-resolution image

We give some illustrative numbers for a 0.165 μm wide (R ∼ 20) bin of a grism spectrum centered on 3.34 μm as listed in Table 1. Without reference pixel subtraction, the grism apertures have standard deviations of 80,930 DN. Reference pixel subtraction drops this value to 9900 DN.

As in Section 4.2.1, we perform row-by-row subtraction to further reduce the noise to 3860 DN. We use the pixels from X = 5 to X = 301 for the row-by-row subtraction to ensure there is essentially no source flux being subtracted. Note this is only possible when there are no nearby sources at X = 5 to X = 301. For a well depth of 60%, there are 3.7 × 106 DN of source counts in the 0.17 μm bin, which amounts to a photon error of 1,430 DN assuming a gain of 1.8. Therefore, 1/f noise can potentially dominate the error budget for broadband grism time series. Expressed in ppm, the read noise after processing with row-by-row subtraction is 1044 ppm as compared to a photon noise of 390 ppm for a CDS subtraction (two groups in ramp).

Given the high 1/f read noise, it is important to maximize the number of samples up the ramp. The 1/f read noise averages down like $\sqrt{N}$ statistics after background subtraction (i.e., each read can be assumed independent and identically distributed). Therefore, for the example bin size of 0.17 μm, there should be at least seven reads to average per integration to ensure that the read noise is less than the photon noise. For RAPID and BRIGHT1 modes, there should therefore be seven groups and for BRIGHT2, there should be at least four groups (of two frames each). When deciding between BRIGHT1 and BRIGHT2, we recommend BRIGHT2 because it coadds reads together rather than skip frames. Using >7 groups for RAPID and >4 groups for BRIGHT2 would make the photon noise about equal to the read noise for this aperture size but ideally the read noise should be a small fraction of the photon noise.

4.3.1. Using the GRISMC Element

One way to mitigate 1/f noise is to use a different grism element (GRISMC) in the pupil wheel, which disperses the spectra along the column (vertical) direction. The NIRISS and NIRSpec instruments' spectroscopic modes are designed this way so that the dispersion direction is approximately perpendicular to the fast-read direction. This allows spatial background subtraction along the fast-read direction to mitigate 1/f noise. GRISMC is being discussed as a science enhancement for a future JWST cycle, most likely beyond cycle 1.

The trade-off with the GRISMC pupil element is that vertical subarrays will only be read with one amplifier (WINDOW mode), and thus will be about 4 to 4.5 times longer than four amplifiers (STRIPE mode), depending on line overheads. The overheads come into play the most for small subarrays. A 2048 × 64 STRIPE subarray has a frame time equal to 0.34061 s whereas a 64 × 2048 WINDOW mode has a frame time equal to 1.558 s. Therefore, the maximum brightness observable without saturation with a GRISMC pupil element and 64 × 2048 WINDOW mode would be K ≈ 5.7 compared to the GRISMR 2048 × 64 STRIPE mode with K ≈ 4.0. For this reason, the GRISMC is not currently available for time-series observations. If enabled as a new mode, GRISMC observations could have substantially reduced 1/f noise for the majority of known transiting planets, which are fainter than K ≈ 5.7.

Figure 7 demonstrates the advantage of the GRISMC dispersing element in NIRCam's LW pupil wheel. Before background subtraction (left panel), prominent 1/f noise can be seen as horizontal lines and bands approximately perpendicular to the dispersion direction. We show an example row-by-row linear background subtraction using the pixels from X = 1537 to X = 1760 and X = 1774 to X = 2044. This efficiently reduces the 1/f noise along the rows, but leaves a small residual nearby the source. The residuals are difficult to see by eye in the image, but are apparent in the covariance matrix in Appendix C.

Figure 7.

Figure 7. Left panel: the GRISMC element can be used to disperse the spectrum along the column direction so that it is approximately perpendicular to the fast-read direction. Middle panel: with the GRISMC disperser, row-by-row background subtraction (outside the source aperture) and within an amplifier efficiently reduces the 1/f noise. Right panel: the simulated time series with the GRISMC disperser dramatically lowers the detector noise in the time series (205 ppm) as compared to the GRISMR disperser (1044 ppm) because of the reduced 1/f noise. In this extraction, the background-subtraction aperture was optimized to reduce the 1/f noise.

Standard image High-resolution image

Background subtraction can alternatively be performed on short pixel distances (≲70 px), where the source flux is low (≳10 spatial pixels from the source center) with the GRISMC element. As seen in Figure 4, the 1/f noise begins to appear for more than ∼70 clock cycles corresponding to distances more than about 70 pixels along a row and frequencies below 1400 Hz. Subtracting out 1/f noise at 70 clock cycle timescales is not possible with the GRISMR element, where the spectrum extends more than 1100 pixels along the fast-read direction for the F444W filter. Furthermore, the background subtraction with the GRISMC element can be performed within one amplifier rather than extrapolating one amplifier's 1/f noise behavior from other amplifiers as was done with the GRISMR element. We found that a background from 7 to 30 pixels from the source efficiently removed 1/f noise down to 683 DN (205 ppm) in a 0.165 m-wide wavelength bin as shown in Figure 7 (right). The GRISMC therefore allows a standard background subtraction and summation source extraction methods, which efficiently reduce 1/f noise to below the photon noise level.

4.3.2. Small Spatial Apertures

Given the high level of correlation between pixels, another way to reduce 1/f noise is with a small aperture in the spatial (Y) direction. The spatial point spread is tightly concentrated with a FWHM of 1.3 to 2.6 pixels depending on wavelength, so smaller apertures are possible. For example, a 14 pixel (spatial Y-direction) tall aperture results in a measured read noise of 3860 DN (1044 ppm) for the same 0.17 μm wide bin centered at 3.34 μm whereas a 3 pixel tall aperture results in a measured read noise of 1070 DN (290 ppm), which is below the expected photon noise. However, this will sacrifice photons, so there must be a balance between reducing 1/f noise while capturing as many photons as possible. Small spatial apertures will also be more susceptible to pointing jitter or wave front variations that manifest as variable aperture losses. Therefore, a weighting scheme expanded from optimal extraction (Horne 1986) is preferred.

4.3.3. Optimal Covariance-weighted Flux

Here, we provide a modification to the conventional optimal extraction formula from Horne (1986), that uses the covariance matrix of correlated pixels. The goal is to find a scheme that estimates the total flux by summing the value in each pixel j. We start by calculating a profile function Pj in the spatial direction that describes the fraction of the total flux in pixel j. The profile is normalized so that

The summation for NIRCam would be along the Y (spatial) direction for a fixed wavelength. There is a small tilt of the grism spectrum of 0.11 deg (i.e., 1:480), so summing along the Y-direction will slightly mix together different wavelengths at the subpixel level. The exoplanet science applications we focus on here are concentrated on measuring broad molecular features, where the maximum possible resolution of the instrument is not as critical. Therefore, we allow for a small amount of wavelength mixing by studying summations along the Y-direction instead of a more complicated interpolated model. The technique below can also be modified to include summation along the spatial and spectral directions.

Any pixel can be used to estimate the total flux by dividing by the profile

where Sj is the background-subtracted flux. Therefore, the flux total is a weighted average of each pixel's estimate of the total flux fj . The goal becomes a weighting scheme to determine which weights wj minimize the variance in the total background estimate

The weights wj can take on any value while still giving an unbiased estimator of the flux (Horne 1986). The weights are not normalized like the profile, Pj .

If one assumes that all pixels are independent thereby having zero correlation, the optimal weights (ones that minimize the variance in fopt) are

Equation (1)

where Var() is the variance or the expectation of the squared deviation from the mean and Vj  = Var(Sj ) (Horne 1986).

Exoplanet observations, where high signal-to-noise is needed to separate the planet signal from the stellar signal, are generally in the foreground limit (where the variance approaches the number of photons), so Vj  = Nphot,j in this limit. The profile is also proportional to the number of photons from the source, so

where k is a proportionality constant

In this case, the weights become

Equation (2)

Therefore, using this traditional formula (Equation (1)) for bright sources will provide nearly the same extraction as straight summation,

Equation (3)

In other words, in the limit where the read noise and background noise goes to zero, conventional variance weighting approaches straight summation.

However, in the presence of noise correlations, the formulae become modified. The optimal weights for correlated data are the sums of the inverse covariance matrix along one dimension (Schmelling 1995),

Equation (4)

where Cij is the covariance (Cov()) of the variable Si /Pi with Sj /Pj ,

Equation (5)

The variance in the total flux is

Equation (6)

for both conventional optimal extraction and covariance-weighted extraction. In Section 4.3.4, we compare the results of conventional variance weights (Equation (1)) and profile weights (Equation (2)) that result in straight summation and covariance weights (Equation (4)).

4.3.4. Empirical Covariance Matrix

The ideal weighting of the spectrum from Equation (4) depends on the covariance matrix Cij . We use a long dark exposure from the JWST Optical Telescope Element and Integrated Science Instrument Module (OTIS) test performed at NASA's Johnson Space Center on 2017 August 23 to estimate the read noise in Cij . We process the data by performing CDS read pairs of long dark data (108 frames) with reference pixel corrections, column-by-column median subtraction, and row-by-row median subtraction. These steps will be performed on grism time series before extraction on an observation of a bright source with two groups per integration in the RAPID read mode (one frame per group). The resulting images from the column-by-column and row-by-row subtraction are similar to Figure 6 (right). We consider a wavelength bin that is 100 pixels in the dispersion direction (0.1 μm) and 14 pixels in the spatial direction.

Figure 8 shows the covariance matrix of this wavelength bin, taken from the median covariance matrix of 20 regions of the detector, where each region is 100 pixels in the dispersion direction (X) and 14 pixels in the spatial direction (Y). The diagonal elements ∼150 (e)2 are smaller than published NIRCam detector performance tables (σ2 ≈ 132 (e)2 =169 (e)2 for CDS mode). However, if we skip the row-by-row and column-by-column subtraction step, they are consistent with σ2 ≈ 132 (e)2 = 169 (e)2. It is clear that 1/f noise causes covariance most prominently along the spectral direction (the fast-read direction), with a covariance near 20 (e)2. Surprisingly, there is also a covariance between spatial pixels (the slow read direction) of ∼10 (e)2. The average covariance profile in the dispersion direction shows jaggedness to it in Figure 8 where the covariance rises and falls every other pixel with a peak-to-peak amplitude of ∼4 (e)2. This is the even/odd column offset discussed as noise contribution 4. Note that the read-noise covariance matrix described here is for the CDS mode, equivalent to when there are two groups per integration. The covariance matrix elements will decrease by a factor of approximately $\sqrt{2/{N}_{G}}$ where NG is the number of groups.

Figure 8.

Figure 8. Left: the read noise shows significant correlations between pixels—unlike the traditional assumption that each pixel's noise is statistically independent. The data were computed from a long dark exposure of 54 read pairs where reference pixel corrections, column-by-column subtraction and row-by-row subtraction, have been performed. The detector's diagonal elements are slightly below the published detector read noise (132 = 169 (e)2) due to the row-by-row and column-by-column subtraction step. Right: the spectral (X) and spatial (Y) directions from the matrix are averaged to build up a higher signal-to-noise covariance profile. Spectrally, the pixels have a covariance of ∼20 (e)2 expected from 1/f noise. Spatially, they have a somewhat smaller covariance of ∼10(e)2.

Standard image High-resolution image

The empirical covariance matrix from Figure 8 is used to estimate the optimal pixel weights using Equation (4). We reduce the noise and jaggedness of the empirical covariance matrix in Figure 8 by averaging the value along the diagonal, across all spectral pixels and across the spatial pixels with the same value of i − j. Here, the profile is the value from pynrc (Leisenring 2019), which uses webbpsf (Perrin et al. 2014). We assume that the photon and background noise is $\sqrt{{N}_{\mathrm{phot}}}$ and that they are completely uncorrelated from one pixel to the next, so the covariance matrix is the sum, in quadrature, of the empirical read-noise covariance matrix and a diagonal photon noise matrix. To calculate Nphoton we assume a source is bright enough that the maximum well depth in the spectrum image is 60% of the well depth.

The resulting optimal weighting function wj /Pj is plotted in Figure 9. This optimal weighting significantly increases the contributions from the inner pixels compared to variance weights. The optimal weighting scheme also negatively weights some pixels outside the core of the PSF to reduce correlated noise.

Figure 9.

Figure 9. Left: conventional variance weighting (black line) provides nearly the same weights as simple summation (purple line) for targets that are significantly brighter than the background flux, such as exoplanet host stars. However, covariance weighting from Equation (4) (red) strongly weights the core and negatively weights the wings to reduce the contribution from correlated 1/f noise. Here, we assume a maximum well depth in the image of 60% and a worst-case scenario of two groups of RAPID (one frame per group) in the integration. Right: covariance weighting will dramatically lower the scatter in the simulated spectrum plus dark frame time series compared to variance weighting.

Standard image High-resolution image

Conventional optimal variance weighting slightly lowers the contribution of read noise to the overall error budget. We test this conclusion with a series of grism images comprised of a constant A0V star spectrum added to the CDS read pairs, but do not add photon noise to focus on the read noise contribution. We do background subtraction along the dispersion and spatial directions, as would be done in a spectroscopic extraction pipeline. We perform sum extraction over a 14 spatial pixels and then combine together 165 spectral pixels into a 0.165 μm wavelength bin. The standard deviation of sum extraction (black weights in Figure 9) is 3860 DN, which is 2.7 times the read noise, as shown in Table 1. The standard deviation of the extracted flux drops to 3060 DN if conventional variance weighting (purple weights in Figure 9) is used.

Optimal covariance weighting significantly lowers the contribution of read noise to the overall error budget. We use Equation (4) (red weights in Figure 9) to reduce the read noise. For the covariance weights, we assume constant off-diagonal matrix elements of 15 (e)2 and do not include the spectral covariance. The resulting standard deviation of 850 DN is now 60% of the photon noise, but still above ideal read noise limit of 350 DN, due to remaining pixel correlations. Thus, we recommend optimal covariance weighting when extracting grism time series to dramatically decrease the 1/f noise contribution. As the number of groups of the ramp is increased, the read noise should drop as 1/$\sqrt{{N}_{G}}$ so the covariance weighting is most important when NG is small.

The methodology in calculating Pj and Cov(Si , Sj ) can significantly affect the precision on the time series. Typically Pj and the photon noise contribution to Cov(Si , Sj ) can be found by fitting smooth functions (polynomials) along the dispersion direction. If Pj is calculated from each integration individually, it will inherit the 1/f noise variations of each integration. Therefore, we hold Pj constant along the time series with a value saved from the middle image at all wavelengths. Alternatively, Pj could be calculated from a PSF library or theoretically from webbpsf (Perrin et al. 2014). As we discussed in Section 4.3.2, one can also use a very small spatial aperture to reduce the contributions from 1/f noise. However, the optimal covariance approach allows one to use a wider aperture that will be less sensitive to pointing jitter and wave front variations.

4.4. The Impact of Correlated Read Noise on Time Series

So far, we have discussed the magnitude of correlated read noise on an integration-by-integration basis, but here we explore how it will impact science targets. We calculate the expected noise in the transit depth or eclipse depth of a generic exoplanet. For this calculation, we assume a planet orbits a G-type 5800 K star with a 3 hour transit duration is observed using a 2048 × 256 STRIPE subarray and is read out in RAPID mode with one frame per group and four amplifiers (number of output channels = 4). We consider a range of K magnitudes from 6 to 11. We then calculate the maximum number of groups up the ramp that can be commanded while keeping the peak pixel count below 50% well depth to safely avoid severe nonlinearities. Higher percentages of well depth, such as 80% or 90%, are possible to fit more groups in and thus increase the signal-to-1/f-noise ratio. The trade-off is that detector nonlinearities may go uncorrected by the pipeline or there could be reciprocity failure. Here, we assume a conservative 50% value to also ensure that inaccuracies in the JWST NIRCam throughput or a source's absolute calibration does not risk saturating the brightest pixels. From there, we calculate the noise due to photon counting statistics (i.e., $\sqrt{{N}_{\mathrm{phot}}}$) and compare it to the read noise. As in Section 4.3, we focus on a 0.17 μm wide bin centered at 3.34 μm

Figure 10 shows the photon noise as a function of K Vega magnitude. The photon noise scales as $\sqrt{f}$, where f is the flux because it is governed by Poisson statistics. The read noise drops slowly with K magnitude because the time to reset between integrations takes a smaller and smaller fraction of the total exposure as the number of groups per integration grows. We show the read noise assuming sum extraction using Equation (3) (1044 ppm for CDS integration) and covariance-weighted extraction using Equation (4) (230 ppm for CDS integration). If a sum extraction is used, the read noise will dominate for K ≲ 8.7. However, covariance weights can ensure that read noise is smaller than photon shot noise for K > 6.0.

Figure 10.

Figure 10. Bright sources, where only a handful of groups per integration are possible without saturation, can be dominated by read noise if simple sum extraction is used. A covariance-weighted extraction can substantially decrease the contribution of correlated read noise to below the photon limit. The errors here are for a grism time-series wavelength bin that is 0.165 μm wide centered on 3.34 μm for a G-type host star and a 3 hour transit duration. We assume a 2048 × 256 subarray and four amplifiers (number of output channels = 4).

Standard image High-resolution image

Some of the most interesting targets will be nearby bright sources, where the photon noise is small and the feature size is small. For context, the feature size of atmospheric signatures on archetypical planets are expected to be ∼300 ppm for hot Jupiters and Neptunes but only 10 ppm for cool super-Earths like K2-3 b if they have high-metallicity atmospheres (Greene et al. 2016). Therefore, observations of transiting cool super-Earths will require careful treatment of the correlated read noise, such as the covariance-weighting scheme, to achieve the precision needed to probe their atmospheres.

We note that Figure 10 shows a somewhat counter-intuitive result: read noise dominates for bright sources and becomes less important for faint sources. The reason is that for exoplanet transit observations, the detector well depth is filled to approximately the same level (∼32,000 DN or 50% well depth per pixel) regardless if the system's magnitude is 6 or 11. For the bright source with a few groups up the ramp, the read noise will be a larger fraction of 32,000 DN than for a faint source with many groups up the ramp.

5. Random Telegraph Noise and RC pixels

Several pixels exhibit RC (exponential) behavior (Rowlands et al. 2012) and others exhibit unpredictable jumps in their DC levels. The jumps are called RTN (Bacon et al. 2005; Rauscher et al. 2014) and some examples are shown in Figure 11. RTN is also referred to as burst noise and popcorn noise (Bacon et al. 2005). If RTN occurs on a frequency similar to the frame rate, they can be more pronounced (Smith & Hale 2012). These are rare events (400–1000 pixels out of 4 × 106) so they are unlikely to land in a grism or in-focus aperture.

Figure 11.

Figure 11. Pixels sometimes exhibit unusual behaviors that are not related to counting photon flux. Examples are shown for RTN and RC pixels.

Standard image High-resolution image

For a weak lens photometric aperture (such as a 70 pixel radius aperture), there will be of the order of 2–4 pixels exhibiting RTN. If these are not identified, the RTN can contribute ∼200 counts in 4 × 107 or 5 ppm changes. However, the RTN pixels shown in Figure 11 have large enough jumps that they can be identified as noisy pixels. If these RTN are directly in the middle of the source, the photon levels approach ∼3000 DN per group, so the photon shot noise is ∼40 DN. In these pixels, the photon noise can make identifying RTN more difficult. Additionally, if there are only two groups per integration, the RTN cannot be identified from a single integration and they would have to be identified at the exposure level. RTN are not currently flagged in the pipeline as "do not use," although they are marked as having increased noise. There is ongoing work to improve the reference files to include flags for RTN pixels.

There are also RC pixels (as shown in Figure 11) that have exponential curves like a resister-capacitor circuit. These pixels consistently show RC behavior and are flagged as "do not use" for the pipeline so they may be skipped in the aperture summation. In this case, the spectrum will have sharp dips in flux relative to nearby bins. Exoplanet transmission and emission spectra are differential measurements that essentially compare the flux in eclipse with out of eclipse so the dips will not change the percentage difference. A spectroscopic pipeline can also add corrections to wavelength bins that contain "do no use" pixels by assuming a smooth spatial profile or a theoretical PSF. There are 19 RC pixels in a 1100 × 20 pixel F444W aperture and 32 RC pixels in a 1600 × 20 pixel F322W2 aperture.

6. Conclusions

JWST time-series modes will be a valuable tool to characterize the atmospheres of exoplanets. In order to reap the benefits of NIRCam's grism time-series mode, however, it is critical to understand the various random and systematic error sources that will impact the science of transiting planets. We studied the random errors that can impact NIRCam photometric and grism time-series modes in this paper that add to the photon counting error.

We list the known sources of random error that can impact time-series stability in Section 2. We conclude the following about these random sources of error.

  • 1.  
    1/f noise is a major contributor to noise, especially in the grism time series. It is important to subtract along the fast-read (horizontal) direction to reduce the correlated noise. Fortunately, 1/f noise averages like $1/\sqrt{{N}_{\mathrm{int}}}$ from integration to integration after background subtraction. We found that subtracting the median value from each row in each amplifier efficiently reduces the 1/f noise. This will be possible for imaging time series but not grism spectroscopy, where one must use the outer background regions in the spectral dispersion direction to reduce 1/f noise. For a small number of groups up the ramp (<7 groups), the 1/f noise will be larger than the photon noise using sum extraction along a 14 pixel tall spatial extraction region and a 165 pixel wide spectral (0.165 μm) wavelength bin. A covariance-weighting scheme described in Section 4.3.3 can significantly lower the read noise from ∼1000 ppm to ∼250 ppm per double-correlated sampling by positively weighting the central pixels. In addition, the GRISMC element can disperse the spectra perpendicular to the 1/f noise correlations so that they can be naturally subtracted out with spatial background subtraction. A GRISMC mode is being discussed as a future science enhancement.
  • 2.  
    Background subtraction efficiently removes most other effects. We find that background subtraction efficiently removes many random detector effects including:
    • (a)  
      pre-amplifier reset offsets,
    • (b)  
      amplifier discontinuities, and
    • (c)  
      even–odd offsets.
    These effects all become smaller than the 1/f noise with column-by-column (i.e., spatial) background subtraction.
  • 3.  
    RTN and RC pixels only affect a small fraction of pixels. We find that if RTNs are ignored by a pipeline, they could contribute about 5 ppm per integration, but the value will be less if the jumps are detected by a pipeline. The RC pixels affect about 0.1% of pixels but are marked as "do no use" in the pipeline so they should not affect time-series stability when ignored in aperture summation.

These random error sources are only one class of noise sources to consider and tend to average down with the number of integrations. In Paper II, we will discuss the systematic noise that can arise from other time-variable behavior such as pointing drifts and detector charge trapping.

Funding for the E.S. is provided by the JWST project at NASA Goddard Space Flight Center. The authors thank Rafia Bushra for the aperture radius exploration that led to several insights about 1/f noise. The authors also than Nikolay Nikolov and John Stansberry for helpful discussions about 1/f noise and Doug Kelly for planning many of the time-series tests we used to assess stability. We also thank the anonymous referee for helpful and prompt comments that improved this work.

Facilities: JWST(NIRCam) - James Webb Space Telescope, NASA Goddard Space Environment Simulator - , NASA Johnson Chamber A. -

Software: astropy package (Astropy Collaboration et al. 2013), pynrc (Leisenring 2019), webbpsf (Perrin et al. 2014), numpy (van der Walt et al. 2011), scipy (Virtanen et al. 2020), pysynphot (Lim et al. 2015), ncdhas, photutils (Bradley et al. 2016), matplotlib (Hunter 2007).

Appendix A: A Primer on Detector Operation

A.1. Photosensitive Operation

It is helpful to review the basic operation of a near-infrared detector (e.g., Rieke 2007) in order to understand the systematics affecting time-series data. Figure 12 shows a schematic of an NP semiconductor junction for a H2RG detector inspired by Smith et al. (2008). The crystalline HgCdTe structure has four valence electrons that form covalent bonds in a tetrahedral structure. By varying the relative amount of Hg, Cd, and Te (doping), it is possible to increase or decrease the number of electrons to create mobile charge carriers that move along the crystaline lattice. The N-type semiconductor is a negatively doped HgCdTe crystalline structure that has negative charge carriers (electrons). The P-type semiconductor is a positively doped HgCdTe crystalline structure that has positive charge carriers (vacant holes in the lattice structure than can move like electrons).

Figure 12.

Figure 12. A cartoon schematic of a H2RG photodiode detector inspired by Smith et al. (2008) and Tulloch (2018). The orange and blue/purple colors depict the N-type and P-type semiconductors, respectively. The undepleted zones of the semiconductor (darker colors with − and + signs) contain mobile charge carriers whereas the depletion region (lighter color without − and +) has no mobile carriers and all electrons are in the valence state. The depletion region is widest after the detector resets and shrinks as the detector is illuminated with photons that create electron–hole pairs. The electron–hole pairs migrate across the depletion region toward the undepleted zones due to an electric field in the depletion zone. The voltage can be measured nondestructively (for many reads) as the well fills.

Standard image High-resolution image

Here, we discuss how the N-type and P-type semiconductors operate as a photodiode. The junction of these two HgCdTe materials enables electrons to flow from the N-type material to the P-type material where it can complete a valence shell in the semiconductor material. This produces a negative charge on the P-type side. Conversely, the positively charged holes from the P-type material flow to the N-type material to complete the valence shell and this produces a net positive charge on the N-type side. The resulting voltage difference is called the contact potential.

In the region where electrons have completed the valence shells in the P-type material and holes in the N-type material, the valence shells are complete. This region spanning across the two semiconductor materials is called the depletion region (pictured in Figure 12), because it is depleted of mobile charge carriers. The positive net charge on the N-type material and negative charge in the P-type material causes a voltage (potential) difference across the detector. It should be noted that the depletion region is depleted in mobile charge carriers, but is not depleted in net charge because it has a net positive charge in the N-type depletion region and negative charge in the P-type depletion region.

The NP junction in Figure 12 is reverse biased with a positive potential on the N-type material and negative potential on the P-type. The positive potential on the N-type material attracts the negative charge carriers (electrons) whereas the negative potential on the P-type material attracts the positive charge carriers (holes). The reverse bias thus reduces the number of mobile charge carriers and increases the size of the depletion region. The increased depletion region size also increases the well depth or capacity of electrons that the detector may collect before saturation. The reverse bias is applied at each reset at the beginning of a detector's integration. Figure 12 shows the initial reverse voltage that is applied during a pixel reset. The physical size of the N-type HgCdTe semiconductor is much larger than the P-type HgCdTe because the density of carriers is lower in the N-type semiconductor.

Note that Figure 12 shows the mobile charge carriers only. Textbook NP junction schematics often show the net charge and thus symbolize the regions with charge carriers as the empty and the depletion regions with "+" symbols on the N-type semiconductor and "−" symbols on the P-type (e.g., Halliday et al. 2004). The schematic in Figure 12 instead shows the mobile charge carriers as "+" and "−" because it is easier to visualize the charge trapping effect and motion of mobile charges this way (e.g., Smith et al. 2008).

When an incoming photon strikes the N-type depletion region, it will excite an electron from the valence to conduction band which creates an electron–hole pair (Rieke 2007). The depletion region has a net electric field pointed toward the P-type side, which controls the direction electrons and holes will migrate. The electron moves to the extremity of the depletion zone in the N-type material whereas the positive hole generated crosses the NP junction and adds a mobile charge to the P-type side. The incoming photon has the effect of decreasing the size of the depletion region. A second photon will excite another electron–hole pair so that a hole moves to the P-type side. This continues and decreases the size of the depletion region. The diffusion of charge carriers across the junction changes the voltage across the NP junction, which can be measured with a sensitive amplifier circuit.

As the size of the depletion region decreases, the detector becomes decreasingly sensitive to new photons. Thus, the NIRCam HgCdTe detectors are never strictly linear. Eventually, with enough photons, the depletion region is too small to create electron–hole pairs with incoming photons. This is where the detector approaches saturation and no longer changes voltage with more light.

A.2. Readout Circuit

The accumulated charge from absorbed photons in the detector is measured with a multiplexor (MUX) that forms a readout integrated circuit (ROIC; Rieke 2007). This readout circuitry enables the signal to be measured nondestructively up the ramp in so-called multiaccum mode.

A sensitive amplifier converts the measured voltage potential (as compared with the bias level after reset) into a signal. The electronic circuit that digitizes the voltages is tuned to ensure the 16 bits (65,536 possible DNs) measure the voltage across the NP junction from the value after reset (and before most photons are detected) to the saturation voltage where the maximum number of photons are collected and no more electron–hole pairs can be produced in the depletion region. The minimum voltage is several thousands of counts above zero to ensure that fluctuations in the bias signal do not go below zero. The 65,536 DN level is tuned to be just above the HgCdTe maximum well capacity to measure the full well capacity of a pixel.

Appendix B: More Sophisticated 1/f Noise Reduction Techniques

Here we discuss some more sophisticated 1/f noise reduction techniques. These were applied to the same time series explored in Section 4.1 These methods were not as effective as the simpler row-by-row amplifier-by-amplifier method from Section 4.2.2, but are worth discussing in the search for better ways to reduce 1/f noise.

B.1. Smoothed Kernel Subtraction

We start by applying a smoothing kernel to each group (one read for RAPID mode) up the ramp of the raw data. The kernel is a uniform normalized kernel 1 pixel in height and 161 pixels along a row. This enables subtraction of 1/f noise that has length scales shorter than the line time (i.e., timescales shorter than 512 clock cycles or 5.12 ms), which the median subtraction cannot remove. Care must be taken not to apply the kernel through the source aperture, which would subtract and distort any signal from a star. The kernel should be wider than the source aperture diameter and interpolate through the source points.

We use the astropy.convolution.convolve() routine, which calculates a convolution of the kernel by the image. astropy.convolution.convolve() ignores all NaN values in the mask and renormalizes the kernel by the sum of the remaining kernel points. This efficiently interpolates over bad pixels and can also interpolate over the source aperture by creating a circle of NaN values around the source. This is similar to the row-by-row method where the pixels are selected outside the source aperture with no bright sources but it operates on timescales shorter than 5.12 ms. Our smoothing kernel is a flat boxcar normalized to 1.0. The resulting time series has a standard deviation of 2800 DN to 3700 ppm. This is slightly larger than the much simpler row-by-row amplifier-by-amplifier subtraction in Section 4.2.2.

B.2. Reference Channel Subtraction

We note that the four channels can sometimes exhibit the same 1/f correlation behavior, as shown in Figure 13 because they share a common reference voltage. Another option to remove 1/f noise is to subtract the three out of four output amplifiers that have the source on them by the one output amplifier without the target. We subtract the first amplifier's time series (pixels X = 1 through X = 512) from all amplifiers by shifting and flipping the 2048 × 256 image subsection to the corners of the four amplifiers at (X = 1, X = 513, X = 1025, and X = 1537). Of course, this means that this amplifier is perfectly subtracted and no noise or signal remains in this one-fourth of the image. The subtraction step should increase the read noise on an individual pixel (for the remaining three output amplifiers) by $\sim \sqrt{2}$ because the errors from the subtracted pixel should add in quadrature to the original pixels. This increase in read noise on a per pixel level might be acceptable if it decreases the correlated errors along the rows. On the other hand, if each amplifier exhibits its own individual correlated noise behavior, the reference amplifier will fail to subtract out all correlated noise and may introduce more noise.

Figure 13.

Figure 13. The time series for all four amplifiers share common-mode behaviors for some intervals but not always (left and right plots). This is visible in a short segment of the time series of all pixels for an example group up the ramp 53 in a dark ramp where the bias has already been subtracted. The pixels in each time series are smoothed with a 51-pixel-wide median filter to show the low-frequency noise. The left panel shows the a time near the beginning of a frame, where amplifiers 1 and 2 share some common-mode read noise. The right panel shows another part of the time series where amplifiers 1 and 2 have more disparate noise signatures.

Standard image High-resolution image

The net effect of this method is to increase the noise over a row-by-row median subtraction. The standard deviations in the time series range from 8500 DN to 10,800 DN for this reference amplifier method as opposed to the 5500 DN to 6600 DN for a row-by-row median. This means the correlations within an amplifier are not efficiently subtracted and may be propagated to the remaining amplifiers.

B.3. Principal Component Analysis Correction

Another approach to treating the 1/f noise is dimensionality reduction with PCA. Here, we assume that the counts for each pixel along a column is a random variable that can be correlated with pixels in other columns. Each row represents a different noise instantiation. First, we apply reference pixel correction to remove the pre-amplifier resets. Next, we calculate the mean value of each pixel in the 108 group stack and subtract all groups by this mean image. This removes all bias levels and focuses on the time variability of the pixels. Here, we assume that the dark current is negligible. Next, we mask all bad pixels by setting all pixels with an absolute value larger than 200 DN and replace them with zero DN to remove bad pixels that can drive the principal components. The first 10 principal components are plotted in Figure 14 for two example groups (54 and 55).

Figure 14.

Figure 14. The first 10 principal component eigenvectors identified from group 54 (left) and group 55 (right). Pre-amplifier reset values are seen in the first principal component, and change from one group (one read in RAPID) to another. These are residual values that are not fully subtracted by reference pixels. The principal components generally increase in frequency with principal component number. In other words, the highest frequency principal components explain the least amount of the variance as would be expected for 1/f noise.

Standard image High-resolution image

The first 10 principal components explain 14% of the total variance in each group up the ramp. The principal components from one group to another are different, but the ordering of these components generally follows a pattern of increasing frequency with higher-order principal components (smaller eigenvalues). This is expected for 1/f noise where the lowest frequencies have most of the power. Since the first principal components are different from one group to another, we calculated the principal components for each group independently. We then create a noise model image which is the matrix multiplication of the eigenvectors by the principal components for 10 components. This model is subtracted from each group to create a 1/f-corrected image.

The PCA subtraction is applied to each group up the ramp and then these groups are subtracted in pairs to create a time series. We use 10 principal components as a starting point for this method. The final time series has a measured standard deviation of 5300–5500 DN within the background-subtracted aperture. This is comparable to the row and column subtraction standard deviation of 5500–6500 DN. Adding 10 more principal components for a total of 20 eigenvectors only reduces the time-series noise to 5100–5400 for the three apertures.

In another approach, we treat each amplifier separately so that the principal components are calculated independently. This approach means that a noise signature in three amplifiers but not the fourth does not propagate to the fourth amplifier when subtracting the PCA-based model. This individual-amplifier model has four times the number of variables as the combined PCA approach, so we fit with only the first five principal components in each amplifier. The resulting eigenvectors are shown in Figure 15.

Figure 15.

Figure 15. The first five principal component eigenvectors identified from group 54 (left) and group 55 (right) where each amplifier has been analyzed independently. The amplifier boundaries are demarcated by solid vertical lines.

Standard image High-resolution image

If we treat each amplifier separately, the PCA model is more effective in reducing 1/f noise. We begin by keeping five principal components within each amplifier and subtracting these from each amplifier separately. The resulting standard deviation of the time series is 2500–3820 DN. This improvement is beyond what is possible on a real source because the principal components include pixels over the source aperture. To simulate a real photometry scenario, it is necessary to interpolate over the source aperture points as done in Appendix B.1 for the smoothing kernel. However, we stop the PCA at this point because it does not do as well as the simple row-by-row amplifier-by-amplifier mean subtraction performed in Section 4.2.2.

Appendix C: Covariance Matrix with Row-by-row Subtraction

Here, we examine the covariance matrix profile for GRISMC images to show the effect of row-by-row median subtraction within an amplifier. This uses the same image and simulated grism spectrum as in Section 4.3.1. We calculate the covariance matrix for the pixels in amplifier 4, which is labeled in Figure 1. For the covariance, we treat each row in the image as an independent noise instance and find the covariance between pixels within that row. We exclude all pixels that include the bright source region and then average all covariance matrix elements Cij that have the same value for i − j to create a differential covariance profile. As shown in Figure 16, the covariance level is 15–30 (e)2 due to 1/f noise and rises for nearby pixels (as ΔX is reduced).

Figure 16.

Figure 16. 1/f noise causes read noise covariance between pixels, but can be reduced in GRISMC spectroscopy considerably with row-by-row subtraction. Before any row-by-row subtraction (blue line), all pixels within an amplifier are positive correlated. These can be reduced with row-by-row linear subtraction using all pixels within a row that are outside the source and within an amplifier. The resulting images have smaller overall covariance (orange line), but still have correlations for closely spaced pixels ($| {\rm{\Delta }}X| \lt 50\,\mathrm{px}$). Therefore, the optimal background subtraction will use a tighter background-subtraction region ($| {\rm{\Delta }}X| \lt 50\,\mathrm{px}$) to reduce 1/f noise.

Standard image High-resolution image

Row-by-row linear subtraction for all pixels outside the source substantially lowers the covariance, and in some cases causes a mild 3–4 (e)2 anticovariance. However, there still remains positive covariance between pixels closer than 50 pixels (or 50 clock cycles). Therefore, a background region that includes pixels with $| {\rm{\Delta }}X| \lt 50$ pixels will decrease the 1/f noise residuals. We find that a tighter background region of 7–30 pixels optimally lowers the 1/f noise for the simulated time series.

Appendix D: Dark Current Variations

In Section 4.1, we use a long dark integration with 108 groups to simulate a time series of 54 integrations with two groups each by subtracting sequential pairs of reads. This simulation does not take into account the effect of resets that would occur at the start of every integration in a real exposure that has two groups and 54 integrations. In reality, the resets can affect time series because the dark current is nonlinear with time. Here, we quantify the magnitude of this effect and how it could affect exoplanet time series.

Figure 17 shows the dark current measured from CV3 dark exposures composed of 108 groups up one integration in RAPID mode at a temperature of 37.0 K. In the beginning of an integration, the slope is about 0.011 DN s−1 but increases to 0.019 DN s−1 at around 150 s following detector reset. Thus, a simulation that breaks up a long integration (108 groups) into many shorter integrations will have dark current variation that would not be present for a long exposure that has 54 integrations with two groups each. We expect the dark current to come out with background subtraction. However, we calculate the magnitude of the dark current to see if any residuals in the background subtraction could affect the time series. The difference in signal between the early and late dark currents is 0.008 DN s−1, so the total DN difference in a 21.5 s integration is 397 DN or 107 ppm in our archetype grism wavelength bin that is 165 spectral pixels and 14 spatial pixels. Therefore, the dark current difference is only expected to be 107 ppm before subtraction and much smaller after background subtraction.

Figure 17.

Figure 17. The dark current for the NIRCam detectors is nonlinear. After about 150 s, the slope of the dark current changes, so a long integration from a dark is not equivalent to a series of shorter integrations. However, the dark current is a small signal, so the difference in the dark current is only about ∼100 ppm of the photon signal of a bright source before doing the background subtraction, and much smaller afterwards.

Standard image High-resolution image

Footnotes

Please wait… references are loading.
10.3847/1538-3881/abb811