1. Introduction
The need for post-launch Cal/Val of satellite sensors is well-established. For the primary level 1 observations, post-launch calibration against independent references external to the spacecraft, ensures that unforeseen changes and changes that cannot be corrected by any on-board calibration systems, are accounted for. Post-launch validation supports the accurate delivery of level 1 and higher-level products, confirming expected performance including processors and any retrieval algorithms. However, cost, complexity and robustness of establishing mission specific exercises, has led to efforts of space agencies to improve coordination and generalisation of methods and infrastructure, through bodies like CEOS. This coordination not only reduces cost, but also leads to improved harmonisation and interoperability between sensors. Earth Observation data without proper calibration and uncertainty assessment has little value for most applications, because if the data cannot be trusted, no reliable information can be derived from it.
In parallel, driven in part by the critical needs of climate, there has been a significant move towards a more coordinated and comprehensive assessment and reporting of the quality, biases and residual uncertainty in the observations made by different satellite sensors. More recently, this need has been enhanced with the emergence of new commercial satellite operators ensuring users have a transparent and fair means to judge adequacy of data products meeting their needs.
All this led to the endorsement by The Group on Earth Observations (GEO) of a CEOS WGCV and WMO GSICS led community initiative to create a Quality Assurance Framework for Earth Observation (QA4EO) [
1]. QA4EO provides a set of principles and associated guidance including specifically tailored tools [
2] to encourage the provision of internationally consistent quality indicators on their delivered data. QA4EO guidelines state that ‘it is critical that data and derived products are easily accessible in an open manner and have an associated indicator of quality traceable to reference standards (preferably SI) so users can assess suitability for their applications i.e., ‘fitness for purpose’.
The QA4EO principles, and means to establish and demonstrate them, were tailored from those in common usage across many commercial and scientific sectors, ultimately derived from the best practices of the international metrology community. The latter is charged by intergovernmental treaties with ensuring international consistency and long-term reliability of measurements to support trade and legislation, through the international system of units (SI) [
3]. The robustness of the SI and its dissemination over century-long timescales is underpinned by adherence to clear principles:
- -
Unequivocal, documented, consensus definition of ‘references’ and the
mises-en-pratique [
4] to realise them.
- -
-
Documented evidence of the degree of traceability (residual bias) and associated uncertainty of a measurement relative to the ‘reference’ through:
- ∘
regular, openly reported comparison with independent peers (where there is not independent representative measurement with lower uncertainties), or
- ∘
comparison (validation/calibration) against an independent representative measurement of the same measurement with similar or lower uncertainties than that being evidenced.
- ∘
detailed comprehensive assessment of uncertainty of the measurement and its traceability to the reference that is peer reviewed alongside comparison results.
One key element of QA4EO relates to post-launch Cal/Val of satellite-delivered data through an independently derived data set that can be correlated to that of the satellite observation. Of course, for this process to be robust, the independent data set itself must also be fully characterised in a manner consistent with QA4EO principles: i.e, that it has documented evidence of its level of consistency with a suitable reference (nominal ‘truth’) within an associated uncertainty. The uncertainty, associated with the comparison process itself, which almost always involves a transformation of the ‘reference’ measurand to be comparable/representative of that observed by the satellite, must also be characterised and be sufficiently small. In practice, the ‘reference’ should be community agreed and ideally tied to SI. The need to evidence these core principles as a differentiator lies as the basis for creating the label Fiducial Reference Measurement (FRM).
2. Post-Launch Calibration and Validation (Cal/Val)
In a generic manner, a post-launch Cal/Val program for an EO mission is typically composed of a set of different complementary activities bringing elements that need to be combined in order to produce consolidated and confident validation results to encompass the full scope of a sensors observations. Note that the use of the word ‘comparison’ represents the situation when the nature of the products been compared are similar and of a similar uncertainty. ‘Validation’ is used where the link to the satellite product is more tenuous and/or the associated uncertainty of the non-satellite measurement is larger and cannot be considered similar, as in the later bullets. ‘Calibration’ is a special case of comparison where the uncertainty of the reference and the comparison process are sufficiently small that this allows the satellite sensor’s measurement to be assessed against it and potentially adjusted as part of the comparison process. Generically, the different components of a validation programme are the following:
Comparison against tailored and accurate Fiducial Reference Measurements (FRM) that provide few points but low uncertainty/high confidence;
Comparisons against general in-situ data that can provide more data over a wider set of conditions, but are generally individually less accurate;
Comparison against other sources, including: inter-satellite comparison, satellites own Level 3 (i.e. geo-referenced product of merged data derived from the sensor itself, climatologies, etc.;
Comparison against models: re-analysis data, data assimilation, integrated model analyses, forward modelling, etc.;
Validation using generic Level 3 data (i.e. merged data): statistical comparison between Level 3 products from a variety of independent sensors (determining: mean, median, standard deviation, bias, RMS…. for selected zones, transects, latitudinal bands, seasonal trends…) for a cross-validation of higher-level products,
Validation using monitoring tools: statistics, trend, systematic quality control, etc.
Comparison/calibration between operational satellites (e.g. using GSICS best practices)
All these components are important and often necessary, but the first component (comparison to FRM) is of particular importance because it gives a reference properly characterised and traceable to a standard on which the validation results can be anchored and an uncertainty assessed. If of sufficient uncertainty, the FRM can also be used to provide calibration information that can be used to update the sensor performance coefficients during operation.
As indicated above, satellite post-launch Cal/Val has traditionally relied on a wide range of data sets collated by different teams and based on, for example, in-situ and aircraft campaigns, (semi)-permanently deployed instrumentation, comparisons with other satellite sensors, opportunistic measurements (e.g., from commercial shipping and aviation) and modelled natural phenomena (e.g., Rayleigh scattering of the atmosphere over dark ocean surfaces).
To date these different datasets have varying levels of adherence to the critical QA4EO principles, and while all provide information, it is hard for satellite operators to distinguish high quality data optimised for satellites from those that need more careful treatment. Consistency between datasets for comparison with a single satellite sensor is generally not sufficient to ensure the quality of all comparisons, for all sensors and all conditions. The creation and adherence to the FRM label is designed to address this issue.
3. Fiducial Reference Measurements (FRM)
3.1. Need for FRM
The terminology ‘Fiducial Reference Measurement’, with its acronym ‘FRM’, was first elaborated and used in the framework of the Sentinel-3 Validation Team (S3VT) meeting in 2013 [
5]. It was particularly relevant in the context of Copernicus for distinguishing between in situ measurements that are managed by the Copernicus in-situ component [
6], which are also used for other purposes in their own right and the Cal/Val measurements managed by the Copernicus Space Component that are used for satellite-specific Calibration or Validation purposes. As result, the term FRM was initially defined as “The suite of independent tailored and fully characterised measurements that provide the maximum Return On Investment (ROI) for a satellite mission by delivering, to users, the required confidence in data products, in the form of independent validation results and satellite measurement uncertainty estimation, over the entire end-to-end duration of a satellite mission.” [
5].
Figure 1 illustrates this principle: If we have no FRM then we cannot really use the mission for any quantitative purpose as we have no idea how accurate the data products are; if we have many FRM this is great scientifically (statistical significance, geographic/condition coverage, robust network…) but incurs additional costs, so with a reducing ROI. There is a balance between these two extremes to deliver a satellite mission with a known product quality that is “fit for purpose”.
In an attempt to create a more robust validation framework and encourage, where appropriate, teams to specifically tailor their observations to address the needs of satellites, as well as helping funding bodies distinguish and allocate resources; ESA expanded the concept to other applications and technology domains and actively encouraged the creation and utilization of the label of ‘Fiducial Reference Measurements, FRM in the international arena through bodies such as CEOS. In creating the FRM label and associated criteria, see below, the objective has been to create a distinction and special case of observations that are optimised to meet the needs of the satellite community.
Note, that the use of FRM labels for particular data sets does not imply that observations not classified as FRM are of inferior quality. The FRM label is used where specific requirements (section 3.2) are met for a specific satellite measurand/product and, in particular, where datasets are suitable for a defined class of satellite observations.
These principles subsequently formed the basis for an ongoing series of projects funded by European organisations, EU, ESA, EUMETSAT to explore, establish and evidence FRM quality observations and in many cases, networks of sites, around the globe that adhere to the FRM principles for a wide range of observation and sensor types spanning surface temperature and reflectance, atmospheric composition, sea and ice level height etc.
3.2. What is a FRM
Fiducial Reference Measurements (FRM) are:
a suite of independent, fully characterised, and traceable (to a community agreed reference, ideally SI) measurements, tailored specifically to address the calibration/validation needs of a class of satellite borne sensor and that follow the guidelines outlined by the GEO/CEOS Quality Assurance framework for Earth Observation (QA4EO) [
1].
In many cases, these FRM may be a subset of ‘in-situ’ measurements, originally designed to make observations in their own right at a specific geo-location, both individually and/or as part of a network, but they could also include measurements from a campaign, from airborne, from seaborne platforms or even from another satellite providing it meets the FRM criteria. The newly emerging SITSats (SI Traceable Satellites) [
7] are a good example of a potential satellite-based FRM. These FRMs provide the maximum Return On Investment (ROI) for a satellite mission by delivering, to satellite operators and their users, the required confidence in data products, in the form of independent validation results and satellite measurement uncertainty estimation, over the entire end-to-end duration of a satellite mission.
The defining mandatory characteristics for FRM are:
FRM measurements should have documented evidence of their traceability (including assessment of any bias and its associated uncertainty) to a community agreed reference ideally tied to the International System of units, SI, (e.g. via a comparison ‘round robin’ or other with peers and/or a metrology institute together with regular pre-and post- deployment calibration of instruments). This should be carried out using SI-traceable ‘metrology’ standards and/or community recognised best practices, for both instrumentation and observations;
FRM measurements are independent from the satellite geophysical retrieval process;
A comprehensive uncertainty budget for all FRM instruments, and for the derived (processed) measurements (including any transformation of the measurand to match the measurand of the satellite product), is available and maintained;
FRM measurement protocols, procedures and community-wide quality management practices (measurement, processing, archive, documents, etc.) are defined, published and adhered to by FRM instrument deployments;
FRM datasets, including metadata and reports documenting processing, are accessible to other researchers allowing independent verification of processing systems;
FRM datasets are required to determine the on-orbit uncertainty characteristics of satellite geophysical measurements via independent validation activities and thus representativeness and the satellite to FRM comparison process needs to be documented and the uncertainty assessed. Note for any individual satellite sensor the exact sampling and elements of the comparison process may differ, even within a generic sensor class, but the documentation and evidence to support the uncertainty analysis must be presented in a manner that can be readily interpreted by a user. The data and associated uncertainty information should be available in a timely manner and in a form that is readily utilizable by a satellite operator.
The uncertainty of the FRM measurements, including the comparison process, must be commensurate with the requirements of the class of satellite sensor they are specified to support.
FRM datasets are designed to apply to a class of satellite missions. They are not mission specific.
These principles describe a generic approach which can be adapted to suit the requirements of specific satellite measurements and/or validation methodology but simplistically they should all, where practical seek to do the following:
Validation/measurement instruments are characterised and calibrated in laboratories where the reference instrument is characterised and traceable to standards (from metrology institutes);
Measurements protocols are established and agreed within the community.
Field in-situ measurements (on sea, land, etc.) follow protocols and uncertainty budgets are derived.
Analysis and intercomparison between measurements are carried out.
Open access websites/data portals and publications to disseminate results and best practices.
The first bullet of the mandatory characteristics is the most critical and relates to evidencing the metrological quality and rigour of the FRM measurements. It is thus important to be clear on the definition of the terminology. Here traceability and all other related terms such as uncertainty are as defined in the ISO guide on vocabulary for international metrology [
8].
Traceability is property of a measurement result whereby the result can be related to a reference through a documented unbroken chain of calibrations, each contributing to the measurement uncertainty.
Note that it is not sufficient to simply purchase an instrument, even with a valid calibration certificate that robustly can evidence traceability to SI through some independent means, but that the use of that instrument in making the measurement can also be fully evidenced. The latter usually involves detailed documentation of process and some form of independent comparison or calibration.
FRM, whilst ideally developed to meet as wide a range of sensors and applications as possible, will always be limited to a particular scope. It is important that the scope (parameter and type of satellite) for what the FRM is applicable is clearly defined and also that any procedures/tailoring of the data needed to match a particular sensor is also well described, for example an FRM for top-of-atmosphere radiance for a high-resolution optical imager, will need to define the spectral and spatial resolution of the FRM data and how that should be convolved to a satellite sensor, including the assessment of the uncertainty associated with that convolution. In many situations the specific details of how this may be done, would be on a case-by-case basis, but the limitations and considerations should be transparent to the user.
3.3. FRM in practice
One of the first projects that could be considered FRM (although the terminology did not exist at the time) was dedicated to Sea Surface Temperature (SST) validation; it was performed in Miami hosted by Miami University more than 20 years ago. The objective of the campaign was to calibrate and compare infrared radiometers used in the validation of the different sea and land surface temperature products derived from satellites. The objective included an assessment of the relative performance of each instrument as well as ensuring that surface measurements used in satellite product validation were traceable to SI units at the highest possible confidence level. This inter-comparison exercise ensured a coherence between the validation measurements used in various international Cal/Val programs and provided and increased confidence in the results used for validation of satellite-derived surface temperatures.
This first inter-comparison of infrared radiometers was held at RSMAS Miami during March 1998. This comparison involved several high-quality radiometers and some off-the-shelf devices. NIST provided their standard blackbody target [
9] for calibration of each radiometer. The first exercise was followed by a second in 2001 prior the launch of Aqua Modis and Envisat and the third one took place in May 2009, under the auspices of the CEOS WGCV IVOS, prior to launch of Suomi-NPP. This third comparison followed the newly established QA4EO guidelines, with protocols developed to ensure a robust traceability to SI (via NIST and NPL) and establish the degree of equivalence between participants. The results were then published [
10,
11].
Following the creation of the FRM label and criteria, an enhanced version of the previous ‘Miami series’ became FRM4STS – Fiducial Reference Measurement for Surface Temperature Satellites [
12]. Started in 2014, the aim was to establish and maintain SI traceability of global FRM for satellite derived surface temperature product validation (water and land) and to and help develop a case for their long-term sustainability. It required:
Comparisons to ensure consistency between measurement teams with predefined protocols;
Accessible common descriptions and evaluation of uncertainties;
Robust links to SI;
Experiments to evaluate sources of bias/uncertainty under differing operational conditions;
Guidance for prospective validation teams on efforts needed to establish and maintain their measurement instrumentation in the laboratory and on campaign;
Traceability of their instruments and also on their use in the laboratory and field.
The CEOS plenary endorsed the project in 2014 with the objective to facilitate international harmonisation.
With the success of FRM4STS [
12] as example, the European Space Agency developed a series of projects, following the same logic, aiming at improving the validation and qualifying the validation measurements as fiducial in various domains (see e.g. [
13] in the context of optical remote sensing applied for the generation of marine climate-quality products).
Table 1 lists (not exhaustive) some ESA projects with the relevant parameters and satellite missions and in most cases their websites provide access to the results and documentation.
In addition to these FRM labelled projects, there are initiatives such as CEOS WGCV RadCalNet [
14], and also Hypernets [
15] that can both justify the FRM label not only for an individual site but also a network of sites. In both these two examples, each individual site contributing to the network must adhere to FRM criteria to provide data to a central processing hub. However, in addition, the hub serving as the interface to the community, delivering commonly processed data, is itself also of FRM quality, with documented procedures, uncertainties and comparisons to underpin the evidence. There will be many more examples of networks and/or data sets that meet the FRM criteria and so some effort is being established to provide an endorsement and cataloguing process.
Although the majority of FRM data sets are likely to be sub-orbital in origin this is not a prerequisite requirement. Targets such as the moon using models such as GIRO [
16] and LIME [
17] (where the uncertainties and methods are well-established and documented) can be considered FRM for specific applications, particularly stability monitoring. Similarly, the future SITSat missions such as TRUTHS [
18] and Clarreo [
19] will provide Top of Atmosphere (ToA) and in the case of TRUTHS Bottom Of Atmosphere (BoA) FRM data of a quality to not only validate but also calibrate other satellite sensors.
3.4. Achieving and evidencing FRM status
The FRM label is now becoming recognised internationally, including within the recently published GCOS implementation plan [
20] and many data providers are seeking to claim that they are FRM compliant. Similarly, satellite operators are requesting that validation data is from an FRM compliant source. However, whilst the FRM principles are clear, they still necessarily leave room for interpretation and there is no real authority to stop the label being used in an unjustified or misleading manner even by those using it with the best intentions.
CEOS WGCV has thus recognised the need to establish a means to enable Cal/Val data providers to robustly evidence that they are FRM compliant or, at least, their progress towards being fully FRM compliant to allow their users (satellite operators and data users) to assess and weight the utility of the Cal/Val data in validating the performance of a particular satellite. To this end they have added the prefix CEOS-FRM allowing some ability to control the useage of the label. Recognising that the number of potential CEOS-FRM data providers may become very large, it is necessary that any ‘endorsement’ process needs to be time efficient and low cost but also that it maintains integrity. It also needs to be flexible to address different levels of maturity in the methodologies and products that are being validated as well as technology domains.
It is critical that the characteristics and performance of any CEOS-FRM can be unequivocally considered consistent and trustworthy (within their self-declared metrics) and readily interpretable so that any information derived from it can be reliably utilised by a user. This is not to say that FRM for any particular measurand or application must all meet the same level of performance or indeed necessarily fully meet all the CEOS-FRM criteria (at least whilst developing a capability) but that any potential user can readily identify and assess suitability for their application. Ensuring that the CEOS-FRM classification has meaning, requires not only a set of common criteria but also obliges some form of independent assessment against them, or at least the potential for such assessment. This could be achieved through the creation of a formal process implemented though a body such as ISO where independent assessment/audit against a set of criteria is undertaken by an approved body funded by the entity seeking accreditation.
Whilst aspects of what is entailed within meeting FRM compliance could benefit from and indeed may be most easily achieved through use of services adhering to ISO standards e.g. ISO 17025 for calibration of instruments, it is not anticipated that the formality and inherent cost associated with such an endorsement process will be necessary or even practical at this time.
The CEOS WGCV framework takes a more pragmatic approach; relying on self-assessment and transparency/accessibility of evidence against a set of criteria where the evidence it then subject to peer review through a board of experts led by CEOS WGCV.
In order to be flexible, to maximise inclusivity and to encourage the development and evolution of CEOS-FRM from new or existing teams, compliance with criteria will be based on a gradation scaling rather than a simple pass/fail. The degree of compliance and associated gradation can then be presented in a Maturity Matrix model, similar to that of EDAP (Earthnet Data Assessment Project) [
21], to allow intended users of the FRM to assess suitability for their particular application and indeed funders to decide on where and what aspects to focus any investment. The matrix model provides a visual ‘simple’ assessment of the state of any FRM for all given criteria making visible where it is mature and where evolution and effort needs to be expended.
The CEOS-FRM Maturity Matrix provides a high-level colour-coded summary of the characteristics of the FRM measurement under analysis against specific criteria. Although the criteria are intended to be generic, how they are interpreted and assessed in detail may vary with the nature of the FRM and sensor class. It is intended that this same concept can be used for an individual measurement/method at a geo-located site e.g., a test site, as well as at a network of such similar sites. The matrix contains a column for each category/criteria and cells for each aspect (subsection of category). Grades are indicated by the colour of the respective grid cell, which are defined in the Key (the grade criteria). The CEOS-FRM Maturity Matrix can be further subdivided into two parts as follows:
An example of such a maturity matrix is provide below for a fictitious FRM provider; note that the exact labels for each category in the MM is subject to change at the time of publication due to ongoing review and should be considered indicative.
Table 2.
Suggested format of a maturity matrix for an FRM product for a fictitious FRM provider.
Table 2.
Suggested format of a maturity matrix for an FRM product for a fictitious FRM provider.
Self-assessment |
Independent assessor
|
Nature of FRM |
FRM Instrumentation
|
Operations/ sampling |
Data |
Metrology |
Verification |
Descriptor |
Instrument Documentation |
Automation level |
Data completeness |
Uncertainty Characterisation |
Guidelines adherence |
Location/ availability of FRM |
Evidence of traceable calibration |
Measurand sampling / representativeness |
Availability and Usability |
Traceability Documentation |
Utilisation/ Feedback |
Range of sensors |
Maintenance plan |
‘Processing’ algorithms/ software |
Data Format |
Comparison/calibration of FRM |
Metrology verification |
Complementary observations |
Operator expertise |
Guidelines on transformation to satellite Pixel |
Ancillary Data
|
Adequacy for intended class of sensors |
Independent Verification |
FRM CLASSIFICATION |
A B C D (to be selected) |
Table 3.
Grading criteria for the maturity matrix in Table 2.
Table 3.
Grading criteria for the maturity matrix in Table 2.
Grade |
Not Assessed |
Not Assessable |
Basic |
Good |
Excellent |
Ideal |
Not Public
|
In addition to this broad-based summary, an overall classification of the degree of compliance will be provided based on meeting specific gradations for particular criteria. At present there are four classifications:
Class A – Where the FRM fully meets all the criteria necessary to be considered a CEOS-FRM for a particular class of sensor. It should achieve a grade of ‘Ideal’ in the ‘guidance criteria’ in the verification section of the MM and green (at least ‘excellent’) for all other verification categories where these have been carried out.
Class B – Where the FRM meets many of the key criteria and has a path towards meeting the Class A status in the near term. It should achieve at least ‘Excellent’ in the guidance criteria in the verification section of the MM and green (at least ‘excellent’) for all other verification categories where these have been carried out. Ideally it should have a roadmap/timeline of activities being undertaken to reach the higher class.
Class C – Meets or has some clear path towards achieving the criteria needed to reach a higher class and provides some demonstrable clear value to the validation of a class of satellite sensors. It should achieve at least ‘Good’ in the guidance criteria in the verification section of the MM and at least ‘good’ for all other verification categories where these have been carried out. Ideally it should have a roadmap/timeline of activities being undertaken to reach the higher class
Class D - Is a relatively basic adherence to the CEOS-FRM criteria but where there is a strategy and aspiration to progress towards a higher class. This can be considered an entry level class for those starting out on developing an FRM. It should achieve at least ‘Basic’ in the guidance criteria in the verification section of the MM and at least ‘Good’ for all other verification categories where these have been carried out. FRM owners/developers must have a roadmap/timeline of activities being undertaken to reach the higher class.
4. Conclusions
In this paper we define criteria and describe examples of a ‘new’ category, at least in terms of formalized labels, of Earth observation data expressly designed and characterised to meet the Cal/Val needs of the world’s satellite sensors in a coordinated cost-efficient manner. These FRM data sets are typically derived as a subset of a broad-based set of similar data which has been tailored so that it can provide data representative of that measured by a particular class of satellite sensors and that has well-defined and documented estimates of its uncertainty to international references such that it can be unequivocally relied upon for satellite Cal/Val.
FRM, and in particular CEOS-FRM, will provide the means to facilitate confidence and harmonisation in not only the data delivered from national and international space agency missions but also ‘new space’ commercial providers seeking to evidence the quality of their data. The criteria needed to fully achieve FRM status are detailed as is the outline of a new endorsement process currently being developed for implementation by CEOS. It is anticipated that many post-launch Cal/Val data providers will seek to achieve full CEOS-FRM status and, through the different classes allow an evolutionary development with evidence of progress towards CEOS-FRM to avoid barriers to uptake and encourage the development of new FRMs.
Acknowledgments
The FRM principles formed the basis for an on-going series of projects funded by European organisations: EU, ESA, EUMETSAT. We acknowledge the support from these organisations that led to a “joint vision by ESA and EUMETSAT for Fiducial Reference Measurements” that was presented at European Commission – ESA – EUMETSAT Trilateral meeting in June 5th, 2020. The authors would also like to thank Emma Woolliams for useful discussion and comments on the drafting of this paper.
References
- Quality Assurance Framework for Earth Observation – the website: https://qa4eo.org/.
- Quality Assurance Framework for Earth Observation – QA4EO Tools: https://qa4eo.org/tools.php.
- The International System of Units (SI): https://www.bipm.org/en/measurement-units.
- Mises-en-pratique https://www.bipm.org/en/publications/mises-en-pratique.
- Donlon, C.; Goryl, P. Fiducial Reference Measurements (FRM) for Sentinel-3. In Proceedings of the Sentinel-3 Validation Team (S3VT) Meeting, ESA/ESRIN, Frascati, Italy, 26–29 November 2013. [Google Scholar]
- The Copernicus in-situ component website: https://insitu.copernicus.eu/.
- Boesch, H; Brindley, H; Carminati, F; Fox, N; Helder, D; Hewison, T; Houtz, D; Hunt, S; Kopp, G; Mlynczak, M; Pagano, T S; Revercomb, H; Richard, E; Rosenkranz, P; Shea, Y; Simis, S; Smith, D; Stone, T C; Sun, W; Xiong, J; Wielicki, B; Yang, H; Ye, X (2022) SI-traceable space-based climate observation system: a CEOS and GSICS workshop, National Physical Laboratory, London, UK, 9-11 Sept 2019. [CrossRef]
- International Vocabulary of Metrology Fourth Edition – Committee Draft (VIM4 CD), 11 January 2021 (https://www.bipm.org/documents/20126/54295284/VIM4_CD_210111c.pdf).
- Joel B. Fowler, A Third Generation Water Bath Blackbody Source, J. Res. Natl. Inst. Stand. Technol. 5, 591–599 (1995).
- Rice, J. P., J. J. Butler, B. C. Johnson, P. J. Minnett, K. A. Maillet, T. J. Nightingale, S. J. Hook, A. Abtahi, C. J. Donlon, and I. J. Barton, 2004: The Miami2001 Infrared Radiometer Calibration and Intercomparison: 1. Laboratory Characterization of Blackbody Targets. Journal of Atmospheric and Oceanic Technology, 21, 258-267.Author 1, A.; Author 2, B. Title of the chapter. In Book Title, 2nd ed.; Editor 1, A., Editor 2, B., Eds.; Publisher: Publisher Location, Country, 2007; Volume 3, pp. 154–196.
- E. Theocharous and N. P. Fox, CEOS comparison of IR brightness temperature measurements in support of satellite validation. Part II: Laboratory comparison of the brightness temperature of blackbodies. https://eprintspublications.npl.co.uk/4759/1/OP4.pdf.
- Fiducial Reference Measurements for satellite temperature product Validation – website: https://www.frm4sts.org/.
- Zibordi, G.; Donlon, C. and Albert, P. “Optical Radiometry for Ocean Climate Measurements.” Elsevier - Hardback ISBN: 9780124170117 eBook ISBN: 9780124169944. 1st edition – Nov. 2014.
- Bouvet, M.; Thome, K.; Berthelot, B.; Bialek, A.; Czapla-Myers, J.; Fox, N.P.; Goryl, P.; Henry, P.; Ma, L.; Marcq, S.; et al. RadCalNet: A Radiometric Calibration Network for Earth Observing Imagers Operating in the Visible to Shortwave Infrared Spectral Range. Remote Sens. 2019, 11, 2401. [Google Scholar] [CrossRef]
- Goyens C., De Vis P., and, Hunt S. (2021) Automated Generation of Hyperspectral Fiducial Reference Measurements of Water and Land Surface Reflectance for the Hypernets Networks. 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Belgium, Brussels, 11-16 July 2021. [CrossRef]
- Lunar Calibration Algorithm Working Area – website: http://gsics.atmos.umd.edu/bin/view/Development/LunarWorkArea.
- Lunar Irradiance Model ESA: LIME website (https://calvalportal.ceos.org/lime).
- Nigel Fox and Paul Green, ”Traceable Radiometry Underpinning Terrestrial- and Helio-Studies (TRUTHS): An Element of a Space-Based Climate and Calibration Observatory,” Remote Sensing, Published: 27 July 2020. [CrossRef]
- Bruce A. Wielicki, D. F. Young, M. G. Mlynczak, K. J. Thome, S. Leroy, J. Corliss, J. G. Anderson, C. O. Ao, R. Bantges, F. Best, K. Bowman, H. Brindley, J. J. Butler, W. Collins, J. A. Dykema, D. R. Doelling, D. R. Feldman, N. Fox, X. Huang, R. Holz, Y. Huang, Z. Jin, D. Jennings, D. G. Johnson, K. Jucks, S. Kato, D. B. Kirk-Davidoff, R. Knuteson, G. Kopp, D. P. Kratz, X. Liu, C. Lukashin, A. J. Mannucci, N. Phojanamongkolkij, P. Pilewskie, V. Ramaswamy, H. Revercomb, J. Rice, Y. Roberts, C. M. Roithmayr, F. Rose, S. Sandford, E. L. Shirley, W. L. Smith, B. Soden, P. W. Speth, W. Sun, P. C. Taylor, D. Tobin, X. Xiong, ”Achieving climate change absolute accuracy in orbit,” BAMS (Bulletin of the American Meteorological Society), Volume 94, Issue No 10, pp: 1519-1539, Published: 1 October 2013. [CrossRef]
- The 2022 GCOS Implementation Plan (GCOS-244) (https://library.wmo.int/index.php?lvl=notice_display&id=22134).
- The Earthnet Data Assessment Project (EDAP+) website (https://earth.esa.int/eogateway/activities/edap).
|
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).