Flux calibration involves (1) fitting a slope to the 24Hz voltage detector samples and (2) converting this to a flux. This is, of course, complicated by the presence of non-linearities in the system, glitches, and how well the conversion of voltage/sec to fluxes is known. Parts 1 and 2 are handled by Derive-SPD and AA separately.
At constant illumination the output of the SWS detectors can be approximated as a voltage changing linearly with time;
![]() |
(7.1) |
The increase of this voltage (i.e. the slope S) is dependent on the radiation falling onto the detector, the physical quantity of interest. In Derive-SPD a slope is derived from the 24 Hz data for each reset interval. See section 8.4 for a discussion of this and the errors on it.
In normal data frames all samples in a reset interval are used. For a 1 second
integration this time is 17/24 seconds - the first 7 samples are thrown away
as being affected by the reset, leaving 17 samples that can be used. For a two
second integration the time is (17+23)/24 seconds, as 1 sample is thrown away
in the last second due to the reset pulse. For an integration lasting K
seconds the effective integration time is
seconds.
The accuracy is directly estimated from the fit residuals which allow the computation of the standard deviation of the derived photo-current. Obviously, the accuracy depends not only on the intensity of the source (I), but also on how well the ramps have been previously linearized and therefore on the measurement error of the RC time constants. A statistical weight is computed which is inversely proportional to the error on I and proportional to the number of measurements between two detector resets. This weight will be used by Auto-Analysis to compute the average photo-current for each ramp. It is expected that this error will dominate all previously described ones.
If within a reset period a glitch (or any
other anomaly) is detected, processing of the reset integration is stopped.
Processing is subsequently continued after the glitch until a reset pulse (or
another glitch) is detected. If glitches have occurred within a reset interval
the slopes S of the different parts of the reset interval are averaged
together (weighted by the standard deviation
of those slopes).
Currently the SWS flux calibration as performed in AA rests on the
assumption that the measured current slope
(in
V/sec) is a linear combination of source flux
(in Jy),
instrumental gain
(
V/sec/Jy) and dark current D(t)
(
V/sec);
Note that in this equation it is implicitly assumed that all memory effects (see section 5.8) can be neglected or have been removed. A full treatment of these effects would result in some sort of convolution integral for the right hand side of eqn. 7.2.
Following this equation the actual source flux
is
reconstructed by first subtracting the dark current from the measured
slopes, and subsequently dividing them by the instrumental gain.
The instrumental gain is split into several (hopefully) orthogonal components;
Here G(t) contains all the gain variations occurring on the timescale of the observation. G(t) is derived from the observation itself, from up-down scan data - see section 4.6.1. In principle G(t) should be unity, in practice it will vary around unity during an observation, and in OLP V6 it is set to 1.
The factor G0 is used to account for long term (i.e. different SWS observations) variations in the responsivity of the instrument. It is determined by comparing the instrument response when the internal calibrator is switched on to the expected value for that response based on calibration observations.
The conversion from V/sec to Jy is contained in
.
It is taken
from a calibration table (one for each AOT band) which in turn is derived from
special calibration observations (section 8.3.6).