Test and Evaluation Trends and Costs For Aircraft and Guided Weapons
Test and Evaluation Trends and Costs For Aircraft and Guided Weapons
Test and Evaluation Trends and Costs For Aircraft and Guided Weapons
Support RAND
Purchase this document
Browse Books & Publications
Make a charitable contribution
Test and evaluation trends and costs for aircraft and guided weapons / Bernard Fox ...
[et al.].
p. cm.
“MG-109.”
Includes bibliographical references.
ISBN 0-8330-3540-1 (pbk. : alk. paper)
1. Airplanes, Military—United States—Testing. 2. Antiaircraft missiles—
United States—Testing. 3. Airplanes, Military—United States—Costs. 4. Antiaircraft
missiles—United States—Costs. 5. Air-to-surface missiles—Testing. 6. Air-to-surface
missiles—Costs. 7. United States. Air Force—Weapons systems—Testing.
8. United States. Air Force—Weapons systems—Costs. I. Fox, Bernard, 1951–
UG1243.T47 2004
358.4'18—dc22
2004005294
iii
iv Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Preface...............................................................................................................................................iii
Figures ..............................................................................................................................................xi
Tables............................................................................................................................................. xiii
Summary ........................................................................................................................................xv
Acknowledgments....................................................................................................................xxv
Abbreviations ........................................................................................................................... xxix
CHAPTER ONE
Background: The Cost of Testing ...................................................................................1
Changes in the Test Process ...................................................................................................5
T&E Constituencies...................................................................................................................6
Estimating the Cost of T&E..................................................................................................7
Study Objectives...........................................................................................................................9
Study Limitations ........................................................................................................................9
Organization of the Report.................................................................................................. 11
CHAPTER TWO
The T&E Process .................................................................................................................... 13
Types of Testing ........................................................................................................................ 16
Developmental Test and Evaluation ......................................................................... 16
Operational Test and Evaluation ................................................................................ 18
Multiservice Operational Test and Evaluation .................................................... 20
Qualification Operational Test and Evaluation .................................................. 20
Live-Fire Test and Evaluation ...................................................................................... 20
Test and Evaluation Process Steps.................................................................................... 21
vii
viii Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
CHAPTER THREE
Trends in Test and Evaluation ....................................................................................... 31
Acquisition Reform.................................................................................................................. 31
ORD Flexibility................................................................................................................... 32
Total System Performance Responsibility .............................................................. 34
Commercial Insertions (COTS/NDI) ..................................................................... 36
Modeling and Simulation..................................................................................................... 39
Testing of Software-Intensive Systems ........................................................................... 44
Combined and Multiservice Testing .............................................................................. 46
Contractor Versus Government Test Facilities.......................................................... 48
Live-Fire Testing ....................................................................................................................... 50
Warranties .................................................................................................................................... 51
Cost Trends ................................................................................................................................. 51
CHAPTER FOUR
Test and Evaluation Cost Data ...................................................................................... 59
Cost Element Definitions ..................................................................................................... 59
System Test and Evaluation .......................................................................................... 59
Developmental Test and Evaluation ......................................................................... 60
Contents ix
CHAPTER FIVE
Estimating Test and Evaluation Costs....................................................................... 75
Approaches to Estimating Test and Evaluation Costs ........................................... 75
A Priori Expectations of Aircraft ST&E Relationships.......................................... 80
Results of Aircraft ST&E Analysis ................................................................................... 82
Total Contractor ST&E.................................................................................................. 87
Contractor Ground Testing .......................................................................................... 89
Static and Fatigue Testing .............................................................................................. 90
Flight Testing........................................................................................................................ 91
Other Contractor Test Costs ........................................................................................ 95
A Priori Expectations for Guided-Weapon ST&E Relationships .................... 96
Results of Weapon ST&E Analysis ................................................................................. 99
Developing a T&E Cost Estimate................................................................................. 102
CHAPTER SIX
Conclusions and Recommendations........................................................................ 107
APPENDIX
A. Aircraft Program Descriptions ........................................................................... 111
B. Missile Program Descriptions ............................................................................ 145
C. System Test and Evaluation Work Breakdown Structure ................. 179
xi
Tables
xiii
xiv Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
xv
xvi Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
The tasking for this study arose from two concerns. Some pro-
gram managers have proposed test programs of greatly reduced scope
and duration, citing such initiatives as increased use of modeling and
simulation to reduce the amount of expensive “open air” testing.
Other rationales for reduced test schedules and budgets include using
lower-risk designs, combining government and contractor testing,
using nondevelopmental item (NDI) and commercial-off-the-shelf
(COTS) approaches, and applying total system performance respon-
sibility (TSPR) contracting. Acquisition decisionmakers needed to
know whether these approaches can achieve the projected savings.
The second concern was that members of the cost analysis
community, particularly those outside of the program offices and test
organizations, were not confident that the data and relationships they
were using to estimate the costs of testing for a program or to cross
check such estimates reflected the current T&E environment. Since
some of their tools were based on development programs that were
15 to 30 years old, validation against current and evolving T&E
approaches became a priority.
Although the original intention was for this study to focus on
fixed-wing aircraft, the Air Force Cost Analysis Agency (AFCAA)
asked RAND Corporation to include a cross section of tactical mis-
siles and guided munitions. Because many of the programs of interest
were joint Air Force–Navy development efforts and because the Navy
cost community had similar requirements, the Assistant Secretary of
the Navy for Research, Development, and Acquisition (ASN RDA)
agreed and directed the appropriate Navy program executive officers
and test activities to support the project.
The project scope involved the following four tasks:
• analyzing the nature of current T&E costs for aircraft, tactical
missile, and guided munition systems and the trends likely to
affect these costs in the immediate future
• identifying key cost drivers
• collecting, normalizing, and documenting representative data
• developing a set of practical, documented methodologies for
making high-level T&E estimates.
Summary xvii
Acquisition Reform
Acquisition reform initiatives are a diverse array of ideas, processes,
and practices designed to streamline the DoD acquisition process,
reducing either cost or schedule, or improving technology. A previous
RAND report (Lorell, 2001) addressed the general effects of
acquisition reform on cost estimating.
One of the acquisition reform initiatives that report discusses is
TSPR, which transfers certain T&E responsibilities from DoD to the
contractors. Although the data to support cost savings tend to be
anecdotal, it is apparent that it will shift costs from government
activities to contractor activities and must be recognized as such in
future cost estimates. Our interviews suggest that TSPR must be well
planned to avoid two test-related problems: Test data may not be
available to DoD for other, postdevelopment uses, and cross-platform
integration might not be adequately coordinated, especially in guided
weapon development. DoD must have the foresight to ensure that it
can use the system development and demonstration test data to
design modifications or to qualify additional platform-and-weapon
configurations. In addition, to maintain compatibility, DoD will
have to ensure careful coordination of cross-platform integration
issues, particularly with guided-weapon development and modifica-
tion, with other systems.
It is too early to assess the outcome of recent innovative test
management approaches that give the contractor broad latitude in
developing and executing the developmental test program. Another
innovative approach, relying on non-DoD tests and certifications of
Summary xxi
Live-Fire Testing
Although live-fire testing can be a contentious issue during early
planning for system development, our interviews did not highlight
major concerns at the program level, as long as the requirements were
known in advance and planned for accordingly. Because data were
Summary xxiii
Warranties
None of our interviews indicated that warranties significantly
changed the T&E process or costs.
Acknowledgments
xxv
xxvi Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
xxix
xxx Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
LO low observability
LRIP low-rate initial production
M&S modeling and simulation
Mk. mark
MNS mission needs statement
MOE measure of effectiveness
MOT&E multiservice operational test and evaluation
MRTFB Major Range and Test Facility Base
NAVAIR Naval Air Systems Command
NAWC-AD Naval Air Warfare Center–Aircraft Division
NAWC-PM Naval Air Warfare Center–Pt. Mugu
NDI nondevelopmental item
OAR open-air range
OPEVAL operational evaluation
OPTEVFOR Operational Test and Evaluation Force
ORD operational requirements document
OSD Office of the Secretary of Defense
OT operational testing
OTA Operational Test Agency
OT&E operational test and evaluation
P3I preplanned product improvement
PD/RR program definition/risk reduction
PEP Producibility Enhancement Program
PID program introduction document
PRTV production representative test vehicle
QOT&E qualification operational test and evaluation
QT&E qualification test and evaluation
RCS radar cross section
RDT&E research, development, test, and evaluation
Abbreviations xxxiii
1
2 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Figure 1.1
Aircraft System T&E Share of Total FSD and EMD Costs
30
Government
Contractor
25
Percentage of total cost
20
15
10
0
F/A-18E/F F-22* B-2 B-1 CMUP C-17 T-45 V-22* Avg.
BLK D
NOTE: * Program in progress
RAND MG109-1.1
Figure 1.2
Guided-Weapon System T&E Share of Total FSD and EMD Cost
25
Government
Contractor
20
Percentage of total cost
15
10
0
AMRAAM AMRAAM AIM-9X JASSM* SFW JDAM JSOW Avg.
FSD P3I Baseline
Figure 1.3
DT and OT Share of Total Aircraft FSD and EMD Duration
200
DT start to OT end
MSII to DT start
150
Test duration (months)
100
50
0
F-18A/B F-18E/F F-22* B-2 B-1 C-17 V-22 T-45 Avg.
CMUP EMD*
NOTE: * Program in progress
RAND MG109-1.3
Figure 1.4
DT and OT Share of Total Guided-Weapon FSD and EMD Duration
90
DT start to OT end
80
MSII to DT start
70
Test duration (months)
60
50
40
30
20
10
0
AMRAAM AIM-9X SFW JDAM JSOW Avg.
Baseline
RAND MG109-1.4
4 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
The T&E process has evolved with the complexity and cost of the
systems being developed and with the priorities and practices of
defense acquisition management. This evolution of the T&E process
and its effect on the development cost of the systems under test are
the subject of this monograph.
The tasking for this study arose from two concerns. The first
was the desire of senior acquisition managers to get a satisfactory
answer to a seemingly straightforward question: “Are we budgeting an
appropriate amount for T&E in this program?” With the pervasive
emphasis on acquisition streamlining, commercial practices, cost-as-
an-independent-variable (CAIV) approaches, and a variety of lean
initiatives, there was a concern that the current T&E process might
be a holdover from when contractors built hardware to detailed gov-
ernment specifications and standards and when redundant, stove-
piped processes increased both cost and schedule.
On the other hand, program managers, under intense afforda-
bility pressures, were reexamining every aspect of their programs and
looking for potential cost savings. As a result, some program manag-
ers were proposing to reduce the scope and duration of test programs
greatly, citing such initiatives as increased use of modeling and simu-
lation (M&S) to reduce the amount of expensive “open air” testing.
Other rationales for reduced test schedules and budgets include using
lower-risk designs, combining government and contractor testing,
using nondevelopmental item (NDI) and commercial-off-the-shelf
(COTS) approaches, and applying total system performance respon-
sibility (TSPR) contracting.
The second concern was that members of the cost analysis
community, particularly those outside of the program offices and test
organizations, were not confident that the data and relationships they
were using to provide early estimates of the costs of testing for a pro-
gram or to cross check such estimates reflected current trends in the
current T&E environment. Since some of their tools were based on
development programs that were 15 to 30 years old, validation
against current and evolving T&E approaches was a priority.
Background: The Cost of Testing 5
T&E Constituencies
Different constituencies bring diverse perspectives to T&E (see Table
1.1). Depending on the program phase and how the data are to be
used, T&E can be either a learning process or the final exam. Opti-
mizing T&E for each constituency would tend to drive T&E priori-
ties and objectives in different directions. For example, system
designers tend to view T&E as an integral part of the development
process. Depending on the type of test, designers may view it as an
experimental confirmation of the engineering design approaches or a
realistic exercise of a complex system to discover incompatibilities or
integration problems. In most cases, designers and engineers
approach T&E as a learning process; the more rigorous the test, the
more you learn, but the more likely you are to fail. This also tends to
be the approach in commercial product testing because “in-service”
product failures can adversely affect future sales.
Table 1.1
T&E Perspectives
Study Objectives
The Office of the Assistant Secretary of the Air Force (Acquisition)
(SAF/AQ) tasked RAND Project AIR FORCE to analyze the major
categories of acquisition support costs (commonly referred to as
“below-the-line costs”) to improve the Air Force cost-analysis com-
munity’s estimating approaches and tools. Discussions with the pro-
ject technical monitor at the Air Force Cost Analysis Agency
(AFCAA) indicated that the most pressing need was to examine sys-
tem T&E costs. Although the original intention was to focus on
fixed-wing aircraft, AFCAA asked RAND to include a cross section of
tactical missiles and guided munitions as well. Since many of these
programs were joint Air Force–Navy development efforts and since
the Navy cost community had similar requirements, SAF/AQ
requested Navy support and participation through the Office of the
Assistant Secretary of the Navy for Research, Development, and
Acquisition (ASN [RDA]). ASN (RDA) agreed and directed the
appropriate Navy program executive officers and test activities to
support the project.
The project involved the following four tasks:
• analyzing the nature of current aircraft, tactical missile and
guided munition system T&E costs and trends likely to affect
them in the immediate future
• identifying key cost drivers
• collecting, normalizing, and documenting representative data
• developing a set of practical and documented cost-estimating
methodologies.
Study Limitations
To interpret the results of this study correctly, certain limitations and
constraints should be kept in mind. First, the study includes only
those costs typically paid for by the program offices (which the test
organizations often refer to as reimbursable costs). These costs could be
considered the “price to the customer” for test services. Specifically,
10 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
we did not collect or analyze cost data on the internal costs or opera-
tions of the DoD test infrastructure.2 The program-funded test costs
are the ones that are included in weapon system estimates prepared
for service and Office of the Secretary of Defense (OSD) reviews and
that are included in program budgets. Direct appropriations cover the
costs of test organizations’ facilities and infrastructure rather than the
“variable” costs of testing. A variety of studies and panels have been
examining the broader issues of test infrastructure.
We limited the programs we analyzed to recent Air Force–Navy
fixed-wing aircraft, tactical missiles, and guided munitions. We
focused on programs that had completed development within the
past ten years or, in a few cases, slightly earlier, if test costs and pro-
grammatic data for those programs were not readily available to DoD
cost analysts.3 Older data from existing reports was used for trend
analysis and, where appropriate, to augment more-recent data in
developing relationships.
In an effort to consistently define testing to be included in the
study, we also limited ourselves to the testing categorized as system
T&E and to that associated with RDT&E-funded development
programs. Thus, the study results do not capture subsystem testing
(such as for a landing gear or an individual avionics component),
which would normally be included in what the prime contractor pays
the supplier for the subsystem. Cost-collection systems do not
normally identify such costs separately. However, the study includes
prototype testing if the activities were of sufficient scope and would
have otherwise been included in the EMD phase (now called system
development and demonstration [SDD]).4 On the other hand,
various types of postproduction follow-on testing and production
acceptance testing were not included. We also did not collect data on
_____________
2 These are costs often referred to as direct budget authority, since they are paid for using funds appropri-
ated for T&E in the DoD budget.
3 One of the purposes of the study was to provide more-current cost, technical, and programmatic data
to the cost community. In a few cases, the data we collected were slightly older than our nominal 10
years but were not generally available within the cost organizations and thus would be a useful resource.
4 T&E costs from the demonstration and validation (DEM/VAL) phases of the F-16 and AV-8B were
included in the FSD/EMD totals because of their scope.
Background: The Cost of Testing 11
13
14 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Figure 2.1
T&E Phases Within the Acquisition Process
Milestone A B C FRP
System
enhance-
ments
Development
testing
the DoD acquisition directives should be consulted for the most cur-
rent process information.
Types of Testing
Figure 2.2
System Maturity, Test Processes, and Resources
Report results
s
e ss Post-test analysis n ge
oc ra
s
pr
t
ie
Conduct test en
lit
st
ci
Te nm
fa
Pre-test analysis ro
st
vi
te
n
-e
em
Determine objectives n s
pe ie
st
lit
bs
sy
O i
tie n la
ac
d
Fielded system (P3I/mod)
f
le
ac io
L
al
s
IT
t
st
t f ra
H
Integrated system
In
ili
en eg
n
tio
m nt
Component
i
ul
Subsystem
su tem
Concept
System
sim
s
re
Sy
s
d
ce
an
ur
ea
g
so
M
e lin
re
od
E
T&
M
System maturity
RAND MG109-2.2
and therefore more complex, article. Each step of this generic T&E
process is described below.4
Conduct Test
Prior to conducting the test, the team must select specific test points
from the test matrix. In some cases, this may not be a simple progres-
sion from basic to more challenging tests because of such external fac-
tors as concurrent testing, availability of a properly configured test
article, suitable environmental conditions, and required supporting
assets. In addition, opportunities for achieving multiple test points
during a test event require careful planning and coordination to
maximize the productivity of test resources. Efficient and timely exe-
cution of a test program becomes a complex balancing act requiring
coordination (and often flexibility) among all participants.
Report Results
Test results are reported in a variety of ways, depending on the sig-
nificance of the test and the intended recipients. At the end of DT
and OT, formal reports are submitted and retained to provide a
record of test program execution and results.
If a valid trial reveals a discrepancy between predicted results
and actual system performance, the prediction algorithms are reana-
lyzed and adjusted as necessary. If a test article deficiency is found, its
cause(s) must be isolated, which may require further testing. Once
the cause(s) is (are) identified, a design change may be made.
As the type of article being tested progresses from component to
subsystem to system, design changes can become increasingly com-
plex and costly because they can affect the operation of other compo-
nents or performance in other operating regimes. As a result,
extensive retests (regression testing) may be required to ensure per-
formance results from previous tests have not been compromised.
This situation leads the program manager to a dilemma: Is it better to
reduce risk through extensive component and subsystem testing, with
known increases in cost, or to save costs through reduced testing,
with the potential for much greater disruption and cost if perform-
ance deficiencies are discovered later in development. Obviously, the
optimal balance between testing and risk depends on the maturity
and complexity of the system and on the criticality of potential per-
formance shortfalls.
Analysis of Alternatives
The analysis of alternatives (AoA) is a comparison of alternative
approaches for meeting the mission needs described in the MNS (or
in some cases, the capstone requirements document). AoAs are per-
formed by representatives of the user community, although represen-
tatives of the program manager may participate. An AoA is required
before Milestone B (or Milestone C, if there is no Milestone B). For
the test community, the AoA provides COIs and MOEs to be used in
OT.
Test Resources
The T&E process requires a variety of resources, including contractor
and government test facilities, test ranges, manpower, training, flying
time, support equipment, threat systems, M&S, instrumentation,
communications, range equipment and facilities, data protection and
security systems, munitions, and targets. While many of these needs
are relatively straightforward to estimate for a given test, the timely
availability of appropriate infrastructure to support testing is critical
to the success of any test program. Test facilities may be required
from any of the following general categories:
• M&S
• measurement facilities
• system integration laboratories
• hardware-in-the-loop facilities
• installed system test facilities (ISTFs)
• open-air ranges.
The T&E Process 27
Figure 2.3
Number of Trials by Various Test Methodologies
Number of trials
Digital modeling
measurement
Precision
Int. lab.
testing
OAR
HITL
System maturity
RAND MG109-2.3
Hardware-in-the-Loop Facilities
Hardware-in-the-loop testing involves exercising actual system or
subsystem hardware in a controlled environment using simulated
inputs. Hardware-in-the-loop facilities can simulate the systems and
threats with which the test article must interact. They are most often
part of contractor facilities, although the government maintains them
as well.
Open-Air Ranges
Open-air ranges are used for flight testing aircraft and for a variety of
guided-weapon tests, from captive carry through live fire. These are
the most resource-intensive test facilities, and nearly all are govern-
ment owned. Table 2.1 lists the primary Air Force and Navy aircraft
and missile open air ranges.
Table 2.1
Primary Air Force and Navy Open Air Ranges for Aircraft and Missile Testing
Range Location
Air Force Flight Test Center Edwards AFB, California
Air Force 46th Test Wing Eglin AFB, Florida
Holloman AFB, New Mexico
Air Force Air Warfare Center Nellis AFB, Nevada
Air Force Utah Test and Training Range Hill AFB, Utah
Naval Air Warfare Center—Aircraft Division Patuxent River, Maryland
Naval Air Warfare Center—Weapons Division China Lake, California
Pt. Mugu, California
CHAPTER THREE
Acquisition Reform
In defense procurement, acquisition reform generally refers to a broad
emphasis on eliminating activities that do not add value and ensuring
that the remaining activities are as cost effective as possible. A previ-
31
32 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
ous RAND report (Lorell, 2001) addresses the general subject of the
effects of acquisition reform on cost estimating. In that report, the
authors provide a taxonomy of acquisition reform measures:
• reducing regulatory and oversight compliance costs
• adopting commercial-like program structures and management
• using multiyear procurement.
In T&E, all acquisition reform initiatives fall into the first two
categories, reducing compliance costs and commercial-like practices.
Using the taxonomy described in Lorell (2001), the acquisition
reform principles that specifically apply to T&E are
• requirements reform (ORD flexibility)
• contractor design flexibility and configuration control (TSPR)
• commercial insertion (COTS/NDI).
In addition, the ACTD approach, which uses mature technolo-
gies to demonstrate new operational capabilities in abbreviated dem-
onstrations, often manifests all these principles. 1 ACTDs are not con-
sidered acquisition programs.
The following subsections discuss each of these approaches to
acquisition reform in T&E.
ORD Flexibility
The ORD represents the set of mission requirements a new weapon
system must meet. It describes the performance that the new system
is expected to provide to the operating forces and, throughout the
development process, is the authoritative guide for the functional
capability the system must attain. The ORD describes these
capabilities in terms of key performance parameters, which both pro-
vide guidance for the design team and are the basis for the test pro-
gram.
The ORD development process has, in the past, been criticized
as being a “wish list,” derived primarily from user desires with rela-
_____________
1 OT personnel often conduct or observe ACTD and attempt to provide an assessment of potential mil-
itary utility.
Trends in Test and Evaluation 33
the performance thresholds and objectives are clear and can be veri-
fied by testing.
A good example of the importance of the ORD in T&E was
provided when DOT&E judged that the Predator unmanned aerial
vehicle was not operationally effective or suitable, although it was
deployed in Afghanistan and operated with notable success there. The
DOT&E evaluation was based on the fact that the system had to be
operated well below the requirements set forth in its ORD, the stan-
dard against which the tests had to be conducted. If the ORD lacks
“crawl, walk, run” performance requirements, system developers and
testers must use ultimate performance thresholds as their guidance,
even if lower levels of performance would provide useful capability.
These phased requirements are expected to become the norm with
the DoD emphasis on evolutionary acquisition.
the timing and nature of testing. This may include determining the
test facilities to be used, the number of test articles, and the amount
of testing. Of course, the test IPT normally contributes to these deci-
sions, since the government must ultimately evaluate the results to
determine specification compliance. However, if government
representatives request additional tests for risk mitigation or evalua-
tion in a particular facility and if the contractor disagrees, the gov-
ernment may have to provide additional funding through a contract
modification.
Second, unless the contract specifies otherwise, TSPR may mean
that certain contractor-developed data will not be available to the
government test community. For example, the contractor may model
airflow as part of the design process. The resulting data might be use-
ful to the government for simulating other events, such as weapon
separation. But if certification of the weapons in question are not a
part of the original development program, the government may have
to recreate the data through its own testing or may have to purchase
the data from the contractor separately.
Third, representatives from several programs noted that contrac-
tors having configuration control occasionally made modifications
that affected interfaces with other systems. Thus, government per-
sonnel needed to track contractor-initiated changes closely for effects
on related systems.
Finally, TSPR highlights the importance of close cooperation
and frequent interaction with government test experts, especially the
OT community, so that contractor personnel understand user
requirements and operating procedures. Both contractors and gov-
ernment program personnel noted the importance of early identifica-
tion of potential problems. This is particularly important when
documented requirements or specifications have been reduced to fos-
ter contractor innovation. Authoritative government input early in
the process can save costly redesign and prevent potential contract
disputes. Another benefit of frequent interaction is that it improves
the familiarity of government personnel with the system, preparing
them for evaluating test data.
36 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
have limited control over some parts of their cost structure. In addi-
tion, although test ranges are, in theory, involved in test planning for
all major programs, there is a significant amount of schedule “churn.”
For example, at one range, half the actual testing for the year report-
edly was not included in the original schedule used for budgeting.
Despite these difficulties, both government and contractor program
personnel said significant unanticipated changes in range use rates
were fairly uncommon. When they did occur, the result was usually
reduction or deferral of the planned testing.
Given the occasional schedule turmoil, test article or range
equipment problems, weather delays, and limitations on range capac-
ity, surprisingly few programs complained of getting “bumped”
because of higher-priority users. This could have been due, in part, to
the fact that most of the programs in the study were relatively large or
high priority. We did hear of range availability delays of up to six
months for one FMS program.
Live-Fire Testing
Full-scale system-level live-fire testing (or a formal waiver certifying
that it would be unreasonably expensive or impractical and describing
proposed approaches for assessing vulnerability, survivability, and
lethality) has been a congressionally mandated requirement since
November 1986. Such testing has therefore been a part of all covered
programs since then. However, we found that the costs were often
combined with other DT activities.
Although the requirements for live-fire testing, especially for
modifications of previously fielded systems, are often hotly debated at
the service headquarters level, most program personnel seemed to
consider the testing to be a “fact of life” requirement and planned
accordingly. This could be because live-fire test plans and any waiver
requests must be submitted and approved early in the program
(before Milestone II or B), and are therefore integral to all subsequent
test planning. In the case of modification programs, the cost of live-
fire testing varies depending on whether the modifications are likely
Trends in Test and Evaluation 51
Warranties
None of the government or industry test personnel interviewed indi-
cated that warranties significantly changed the T&E process or costs.
The T&E scope was set independently of warranty considerations.
Cost Trends
In the preceding subsections, we examined trends in various aspects
of the T&E process. A key question remains: Have these initiatives,
in the aggregate, had the effect of reducing the overall cost of T&E?
Although it is difficult, if not impossible, to quantify the cost
effects of each of these in isolation, we did attempt to determine the
overall direction of T&E costs since the early 1970s. Although test
programs clearly differ in content, both individually and by system
type, it is at least apparent that costs are not trending downward. We
also examined the possibility that test durations might be shorter.
Figure 3.1 shows that, if anything, they are lengthening.
Have all the changes to the T&E process described above really
been ineffective? based on the interviews and some circumstantial
evidence, it appears that any net savings due to increased efficiencies
in testing are being offset by other trends.
Improvements in testing efficiency and effectiveness are difficult
to isolate and quantify from the available data. Nearly all the test per-
sonnel in government and industry commented on the increased pro-
ductivity made possible by advances in M&S data collection systems
and approaches and by the reduction of redundant testing between
contractor and government DT activities and between the DT and
OT test communities. While it is difficult to assess productivity
52 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Figure 3.1
Flight-Test Duration of Aircraft Development Programs Over Time
90
80
Flight-test duration (months)
70
60
50
40
30
20
10
0
F-14 F-15 F-16 F/A- AV- F/A- F-22* B-1A B-1B B-2 C-5A C-17 V-22 V-22 T-45
18A/B 8B 18E/F FSD EMD*
NOTES: Measured from first flight to end of DT. By comparison, the Boeing 777 flight-
test program lasted only 11 months, during which nine test aircraft provided 69
aircraft months of testing and logged approximately 7,000 flight hours. Of course,
commercial aircraft do not require testing in the range of mission requirements and
flight regimes typical of military aircraft.
* Program in progress.
RAND MG109-3.1
Figure 3.2
Growth of Flight-Test Data Collected as Indicated by Number of Sensors or
Measurement Points
12,000
Digital
10,000 Analog
8,000
Measurands
6,000
4,000
2,000
0
F-15A/B F/A-18 A/B F/A-18 E/F
SOURCE: Boeing.
NOTE: Measurands are the number of sensors or measurement points.
RAND MG109-3.2
Table 3.1
How Advanced Capabilities Affect Testing
Feature Requires
Reduced-signature design Maintaining several test aircraft in low observability
(LO) configuration
Obtaining and scheduling unique LO test assets
Managing security
Internal weapon carriage Additional wind-tunnel characterization of the flow
field with the bay open
An additional flight-test configuration (doors open)
for performance and flying qualities
Sustained supersonic cruise Reducing test time with chase aircraft
Adding tanker support
Increasing the use of supersonic test airspace
Trends in Test and Evaluation 55
Table 3.1—continued
Feature Requires
Thrust vectoring Special ground-test fixtures to control vectored
exhaust gases
Multiaxis force and moment instrumentation for
measuring thrust
Ground and in-flight performance testing
Expanded flying and handling quality testing
Failure modes and effects testing, particularly with
respect to asymmetric actuation
Integrated avionics Additional EMI and EMC testing
Comprehensive ground and air testing of integrated
system modes
Collecting data on system timelines and their effects
on system performance
Sensor fusion High-density, multispectral, integrated, enhanced-
fidelity target and threat simulation
Comprehensive integrated ground-test facilities
Highly integrated, wide- Multiple threat and target simulators with high
field-of-regard sensors update rates operating concurrently and having a
large field of view
Tailored countermeasures A target platform with a representative signature
Air and ground threats that appropriately stimulate
the system to determine countermeasure effec-
tiveness
Integrated maintenance First-of-kind DT/OT evaluations and assessments (for
information system and software-intensive, paperless systems)
technical order data
AFMSS/mission support A higher level of integration in the testing activity,
element because of the higher level of integration of these
systems in the weapon system
SOURCE: F-22 TEMP, 1999.
Table 3.2
Comparison of F/A-18 Fatigue Testing
for the JSF are to certify more than twice as many weapons as the
F/A-18E/F. Current projections show that the JSF will have roughly
three times the 2.1 million lines of code projected for the F-22 air-
vehicle software.
Overlaid on the increasing complexity and scope of test pro-
grams is the increase in the fees for using DoD test facilities. As
Figure 3.3 shows, user fees pay a significant proportion of the cost of
operating test facilities. This is consistent with the fee-for-service shift
many DoD activities have made. While this does not change DoD’s
overall cost, spreading fixed costs among fewer users increases the
costs for individual programs, unless infrastructure costs can be
reduced proportionately.
Trends in Test and Evaluation 57
Figure 3.3
Comparison of Institutional and User Funding for Major Air Force and Navy
Aircraft and Weapon Test Centers
250
Institution
User
200
Funding in FY 1999 ($M)
150
100
50
0
AF Flight NAWC Air Armament NAWC China
Test Center Patuxent River Center Lake/Pt. Mugu
SOURCE: Defense Science Board (1999).
RAND MG109-3.3
CHAPTER FOUR
59
60 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table 4.1
Representative Contractor DT&E Activities for Aircraft and Guided Weapons
Mock-Ups
Mock-ups encompass the design engineering and production of sys-
tem or subsystem mock ups that have special contractual or engi-
neering significance or that are not required solely for conducting
either DT&E or OT&E.
The reported costs for mock-ups have been only a small part of
total ST&E for recent programs, and the relative costs of mock-ups
and the share of total ST&E costs appears to be declining over time.
This decline may be due to advances in computer modeling that
reduce the need for elaborate physical mock-ups. It may also be that
most mock-up costs tend to be reported under DT&E rather than
separately.
the testing phase and are not allocated to a specific phase of testing. It
includes, for example, spares, repair parts, repair of reparables, ware-
housing and distribution of spares and repair parts, test and support
equipment, test bed vehicles, drones, surveillance aircraft, contractor
technical support, etc. We found that the content of T&E support
varies considerably across programs.
Test Facilities
Test facilities here are the special facilities required to perform the DT
necessary to prove the design and reliability of the system or sub-
system. These facilities include white rooms and test chambers but
exclude brick-and-mortar facilities identified as industrial.
In general, program offices fund only the test facilities that are
unique to their program. Once a facility is available, succeeding pro-
grams may arrange to use the facility and thus pay some of the vari-
able costs. It can be difficult to predict the program office’s share of
such costs because the required test facilities may be available as part
of the existing infrastructure; may be funded by direct test infrastruc-
ture funding; or may require funding from the program office, either
alone or in conjunction with other users. It is noteworthy that the
B-2 and F-22—programs with unprecedented levels of sophisticated
avionics, LO technology, and high levels of security—dedicated con-
siderable portions of their T&E budgets to paying for special facilities
that were not available as part of the existing test infrastructure.
Contractor Costs
Contractors report their costs to the government using a variety of
standard reports, primarily the cost performance report (CPR) and
variations of the contractor cost data report (CCDR). The source of
the cost data for these reports is the contractor’s accounting system.
Since the cost account structure of most contractors’ accounting sys-
tems will not match the approved program WBS, the costs are allo-
cated or grouped into the government-approved program WBS for
reporting purposes (providing a crosswalk between the accounting sys-
tem and the cost report).
The DoD has used this well-defined and accepted WBS struc-
ture, and standardized cost reports tied to it, for decades. This has
resulted in a large body of historical contractor costs that are intended
to be comparable within a given commodity type. The degree of
comparability depends on how well the crosswalk has been done,
which itself depends on the degree of difference between the contrac-
tor’s account structure and the program WBS and on the amount of
effort that has been devoted to reconciling them. Because of the long-
standing requirement for cost reporting and the routine use of the
data, for both management of current programs and analysis of future
programs, the system generally produces a useful high-level picture of
program costs. The fidelity and/or consistency of cost reporting tends
to decrease at lower levels of the WBS, for smaller programs, with the
extensive use of integrated product (or process) teams, and when
commercial practices are used as a justification for reducing or elimi-
nating contractor cost reporting.
Government Costs
In contrast to the long-standing and consistent method for capturing
and reporting contractor costs associated with weapon systems, the
64 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
and must adjust the rates it charges customers so that the costs they
incur and the funding balance.1 The test facilities expend consider-
able effort avoiding unplanned rate increases because of the obvious
problems for users’ test budgets. Although there are differences in the
way Navy and Air Force test facilities are funded, the funding policies
and practices at the MRTFB facilities we visited were similar because
they are all governed by the same DoD financial regulations. It
should be noted that these funding practices apply to DoD facilities
and not to the National Aeronautics and Space Administration or
private facilities that DoD occasionally uses for testing.
In summary, the user organization pays for the costs of labor
and materials related to the tests conducted at the facility and pays a
proportional share of other test-related costs. The user’s rates are
affected by how much other business is at the facility that year.
The funding practices for OT vary more significantly between
the Navy and Air Force. For the Air Force, AFOTEC has its own
program element and pays for significant items, including range costs,
rental of equipment, contractor support, special supplies and equip-
ment for data reduction, civilian pay, per diem and travel, supplies
and equipment, and pretest planning. For the Navy, OPTEVFOR
pays for its own personnel and travel; the system program manager
pays for everything else. After OPTEVFOR develops a program OT
budget, the Navy program manager sends funds to OPTEVFOR for
its contracting office to arrange for range time, assets, additional con-
tractor support, etc. This difference in funding, although a relatively
small part of the total T&E costs for a weapon system, would tend to
make OT costs for a Navy program somewhat higher than for a
similar Air Force program, all else being equal.
The process of estimating government test costs for a program
begins when the program office identifies the need for testing. In the
Air Force, the SPO prepares a program introduction document (PID)
that identifies the system that needs to be tested, the test services and
_____________
1 We attempted to determine whether there had been significant shifts of costs to users to compensate
for shortfalls in institutional funding at the test centers but, because of the lack of consistent data over
time, could draw no conclusions.
66 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
test equipment required, the starting and ending dates, and other
related information. The program office sends the PID to the organi-
zation that will conduct the testing. The test organization responds to
the PID with a statement of capability (SOC). The SOC confirms
that the test organization has the resources to perform the testing and
contains a summary of proposed test events and resources, a schedule,
and costs. The SOC also provides an assessment of technical, sched-
ule, cost, and programmatic risks. The program office consents to the
terms in the SOC by providing funding and written confirmation to
the test organization.
The test organizations we visited track budgeted funding, obli-
gations, and expenditures, generally by organization or function.
However, in contrast with the tracking of contractor costs, there was
no consistent product or output or related WBS for government costs
across programs, so comparison across programs below the level of
total government costs is not currently practical.
The Navy’s process is not as formal as that of the Air Force.
Financial obligations and expenditures are typically tracked and
recorded in the program offices that originate the funds and in the
test activities. However, the Naval Air Systems Command (NAVAIR)
cost analysis group’s attempts to extract T&E cost data for this study
from the Navy’s Standard Accounting and Reporting System
(STARS) were generally unsuccessful.
As in the Air Force, the lack of a standardized WBS for govern-
ment costs results in a wide disparity in the level of detail of cost data.
For example, in some Navy programs, STARS could identify only
total funds by appropriation by fiscal year. Below the appropriation
level, identifying the amount spent on system T&E was impossible.
A more significant difficulty is that many government organiza-
tions dispose of or archive the cost data, so that it is often unavailable
a few years after expenditure. During budget execution, most pro-
grams can track their expenditures by activity and specific tasks.
However, after the funds have been expended, this information does
not appear to be systematically retained for future analysis. In some
cases it is archived but, it is, for all practical purposes, unavailable. In
other cases, it is simply disposed of. The notable exception was the
Test and Evaluation Cost Data 67
Earned Value Cost Analysis System in the 46th Test Wing at Eglin
Air Force Base, which is used for both management and analysis.
Costs for OT were, however, available for both services from their
OT agencies.
Table 4.3
Guided-Weapon Cost Data Sources
• The B-1B airframe was derived from that of the B-1A, so the
B-1B test program had less ground testing than would a normal
development program. For CER development, we combined the
costs for both test programs.
• The F-22 program was in progress as of this writing and is
included for information only. It was not used in any quan-
titative analysis because the content and cost of the remainder of
the flight-test program were uncertain. The costs shown were
developed from the estimate at completion (EAC) for flight test
from the November 2001 CPR that the program office
provided, plus the EAC for all other test elements from the
March 2001 CCDR.
• The proportion of ST&E costs relative to the total development
program is much smaller for the F-22 program than it is for any
other aircraft in the data set. There are at least two probable
reasons. One is that the program was in progress as of this
writing, and the final test cost may be considerably higher than
the current estimate at completion. The other possibility is
related to the F-22 program’s IPT organization and time-
charging practices. Contractor personnel who are primarily
assigned to a nontest IPT but who work on test tasks charge
their time to their original (nontest) IPT. This understates the
effort expended on testing relative to other programs.
• As Chapter Three discusses, the T-45 test program was executed
primarily by the contractor, with short periods of dedicated
government testing. This minimal level of government
involvement contrasts with the other programs for which we
have government costs, in which the government had a larger,
continuous presence and a more-active role. The government
test costs for the T-45 program are a much lower proportion of
total DT costs than for other programs in the data set.
Furthermore, the source of these costs is the Navy’s STARS,
which may not fully reflect all government test costs.
• The V-22 EMD test program was in progress as of this writing
and is included for information only. The content and cost of
the remainder of the program are uncertain. We used the EAC
70 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
adjust the few programs into which we had sufficient insight and
leave the others alone, we chose to leave the costs for all programs as
reported.
weapon ST&E ranged from great detail for the AMRAAM programs
to a single cost number for other programs. Government costs varied
similarly.
In light of these realities of the nature of the cost data, we
• identified cost trends over time through discussions with experi-
enced government and contractor test personnel and examined
cost trends at the total ST&E level by including legacy programs
• identified drivers of system test cost at a high level, consistent
with the quality of the data and diversity of programs, and
aggregated data accordingly
• aggregated data so that costs could be explained by variables
typically available to estimators at the beginning of a develop-
ment program.
We grouped aircraft contractor T&E costs into ground test,
flight test, other test costs, and a subset of ground-test costs consisting
of static and fatigue test costs. These groupings allowed identification
of differences in the cost of each group by aircraft type and by charac-
teristics of the test program. When they were available, government
T&E costs for aircraft had to be combined into a single figure
because they were generally provided to us in an aggregation that did
not allow consistent identification of tasks.
Contractor and government costs for guided weapons were each
combined into a single figure for analytic purposes. This was done for
two reasons. First, in contrast to the case with aircraft, the majority of
T&E costs for most guided-weapon programs are for government
activities. However, the nature of the T&E effort done by the
government varies from program to program, so only a total of both
costs consistently represents the entire test effort. Second, aggregation
at this level is more consistent with the objective of developing CERs
for analysts with limited knowledge of the program at the beginning
of a development effort.
After aggregating the cost data as described above, we attempted
to generate CERs for each aggregation. CERs were developed for air-
craft contractor ST&E costs both in total, as well as ground, flight,
and other tests. A separate CER was also developed for static and
Test and Evaluation Cost Data 73
75
76 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
The data and methodologies in this monograph support the first two
approaches.
For estimating by analogy, the test program descriptions and the
limited-access supplement can be used to develop either a cross-check
or a primary estimate. If several programs are found to be similar to
the one being estimated, the analyst can use the data to develop cus-
tom CERs for the case being estimated. For those without access to
the supplement, Table 5.1 gives average values for the costs of com-
mon T&E elements and other characteristics of potential interest for
four representative multiengine fighter programs. Unfortunately there
were not enough consistently categorized data to develop similar
tables for other types of aircraft or for guided weapons.
For estimating by CERs or factors, we derived quantitative rela-
tionships when sufficient data were available and satisfactory relation-
ships could be developed. The aircraft and missile databases contain
diverse sets of programs of varying types, cost, and complexity. Obvi-
ously, the CERs can be most useful when the estimated program is
similar to those in the database.
Estimating by detailed buildup is normally done by test person-
nel who have knowledge of the test activities to be estimated and the
resources that should be required. These estimates are generally
developed by the test activity, program office, and system contractor.
In general, we recommend that estimators outside of test organi-
zations use a combination of the analogy and CER approaches to
estimate a test program. Because each test program is inherently
unique, CERs cannot reflect all the peculiarities of a given test pro-
gram but can provide an appropriate benchmark for similar pro-
grams. Carefully chosen analogies may be the best approach when
similar historical programs can be found. Even when there are signifi-
cant differences, useful comparisons to the historical data can often be
made. At a minimum, such comparisons can provide either a cross-
check or a starting point for a more in-depth assessment.
Other metrics, such as flight-test hours (Figure 5.1), average
flight-test hours per test aircraft month (Figure 5.2), and number of
guided-weapon launches per month (Figure 5.3), can also be helpful
Estimating Test and Evaluation Costs 77
Table 5.1
Average Test Costs for Representative Multiengine Fighters
(contractor system T&E FY 2001 $M)
Coefficient %
Standard of of
a
Average Deviation Variation ST&E
Nonrecurring development 2,595.8 494.7 0.19
Nonrecurring development
(less ST&E) 1,815.1 451.2 0.25
System T&E 795.9 59.0 0.07 100
Wind tunnel test program 37.8 14.3 0.38 5
Static articles and tests 57.6 31.3 0.54 7
Fatigue articles and tests 42.6 6.5 0.15 5
Drop and accelerated loads tests 24.7 1.5 0.06 3
Air vehicle subsystem test program 55.8 14.7 0.26 7
Avionics integration tests 43.4 13.5 0.31 5
Armament and weapon integration
tests 7.4 4.0 0.54 1
Mockups 20.3 11.5 0.57 3
Other ground 11.9 8.4 0.71 1
Ground test subtotal a 267.4 20.5 0.08 34
Contractor flight tests 367.8 33.4 0.09 46
Other test subtotal 160.8 54.2 0.34 20
Figure 5.1
Total DT Flight Hours
Fighter/attack Bomber Cargo/trainer
7,000
6,000
5,000
Flight time (hours)
4,000
3,000
2,000
1,000
0
F-14 F-15 F-16 F/A- AV- F/A- F-22* B-1A B-1B B-2 C-5A C-17 V-22 V-22 T-45
18A/B 8B 18E/F FSD EMD*
NOTE: * Program in progress.
RAND MG109-5.1
Figure 5.2
DT Flight Hours per Aircraft Month
30
25
Flight time (hours)
20
15
10
0
F-14 F-15 F-16 F/A- AV- F/A- F-22* B-1A B-1B B-2 C-5A C-17
18A/B 8B 18E/F
NOTE: * Program in progress.
RAND MG109-5.2
Estimating Test and Evaluation Costs 79
Figure 5.3
Guided DT Launches per Month in Order of Contract Award Date
Air-to-air Air-to-ground/cruise
1.4
(94)
1.2
Launches per month (number)
1.0
0.8 (13)
(27) (27)
0.6
(20)
0.4 (10)
(8)
0.2 (12)
(6)
0.0
AMR- AMR- AMR- AIM- AMR- HARM JDAM SLAM- JAS-
AAM AAM AAM 9X AAM 1978 1995 ER 1995 SM*
1981 Ph 1 Ph 2 1996 Ph 3 1998
1991 1994 1998
Number of DT guided launches
NOTE: * Program in progress.
RAND MG109-5.3
study. Limiting the data set to recent programs would not have pro-
vided sufficient data for meaningful CERs. Expanding the data set to
include older programs seemed reasonable because the relationships
between our selected cost drivers and the contractor T&E costs of
legacy aircraft programs appear to be generally consistent with more
recent data. Unfortunately, while government costs were collected for
the programs included in this study, they were generally not available
for the legacy aircraft programs. As a result, the aircraft CERs include
contractor costs only. The cost behavior of weapon programs also
appeared roughly consistent over time, so our data set included
selected legacy programs.
Table 5.2
Summary of Aircraft Test Program Characteristics
Average
Flight
Flight Hours per First
Aircraft Hours Aircraft Empty Flight
a
Months Duration (DT) Month Weight Year
F-14A 250 45 2,685 10.7 36,825 1970
F-15A/B 253 59 2,856 11.3 26,250 1972
F-16A/B 156 25 2,581 16.5 13,371 1976
F/A-18A/B 305 40 4,922 16.2 22,351 1978
AV-8B 147 37 2,038 13.9 12,931 1981
F/A-18E/F 244 41 4,620 18.9 30,564 1995
F-22 N/A N/A N/A N/A 31,670 1997
B-1A+B 264 131 3,425 13.0 181,400 1974/
1984
B-2 310 95 5,032 16.2 153,700 1989
C-5A 108 61 3,145 29.2 320,085 1968
C-17 169 39 4,104 24.3 269,696 1991
V-22 FSD N/A 43 763.6 N/A 31,886 1989
V-22 EMD N/A N/A N/A N/A 33,140 1997
T-45 N/A 64 1,932 N/A 9,394 1988
B-1 CMUP 32 11 712 22.3 181,400 1997
a
Months from first flight to end of DT.
The correlation matrix for the variables is shown at Table 5.4 and
uses the same abbreviations.
The CERs presented below as “preferred” best met our selection
criteria. In addition, we developed alternative CERs using different
independent variables, formulations, or a subset of the available data
for use in estimating situations when these constructions may be
more appropriate.4 All the CERs have F-statistics for the equation
and t-statistics for the independent variable that are significant at the
95-percent confidence level,5 unless otherwise noted. They express
contractor costs in constant FY 2001 million dollars.
Table 5.3
Definitions of Aircraft Variables
_____________
4 As previously noted, the lack of government T&E cost data for legacy programs precluded develop-
ment of CERs for government costs.
5 F- and t-statistics are measures of the significance of the coefficients of the entire equation and the
individual coefficients, respectively. Adjusted r2 is a measure of the variance “explained” by the selected
equation, adjusted for the degrees of freedom. The coefficient of variation is the estimated standard error
of the regression equation divided by the mean of the dependent variable.
84 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table 5.3—continued
ST/F a (Static and fatigue) Contractor’s T&E costs for static FY 2001 $M
and fatigue testing; includes the cost of building
the test articles and conducting the tests
T1 (T1 cost) Theoretical first unit (T1) cost for the flight- FY 2001 $M
test units built during the development programb
WEc (Weight empty) Total weight of the aircraft Pounds
structure and its subsystems, avionics, and engine
AC Mo (Aircraft months) Total time each flight-test aircraft Months
is available for flight testing during DT
EMD DUR (EMD duration) Elapsed time from development Months
contract award to end of DT
DUR (Flight test duration) Elapsed time from first flight Months
to end of DT
FLT HRS (Flight hours) Accumulated flying time during DT Hours
for all flight-test aircraft
F/A (Fighter/attack) Dummy variable used in regression 1 = fighter/
analysis to distinguish among different classes, in attack
this case, between fighter or attack aircraft and aircraft
other aircraft 0=
non–fighter/
attack
aircraft
CGO (Cargo) Dummy variable used in regression analysis 1 = cargo
to distinguish among different classes, in this case, aircraft
between cargo and other aircraft 0=
non–cargo
aircraft
a
Subtotals.
b
Calculated by dividing the total recurring air vehicle cost in development by the
number of FSD and EMD units, assuming an 80-percent cost improvement curve, using
the curve to calculate an algebraic lot midpoint, and backing up the curve from that
point to the T 1 cost:
Lot midpoint formulae can be found in various cost estimating or economics texts.
c
According to MIL-STD-1374 (DoD, 1977), weight empty is more inclusive than either
structure or airframe unit weight.
The data set for the CERs contains the following aircraft devel-
opment programs: AV-8B, B-1A and B-1B (combined), B-2, C-5A,
C-17, F-14, F-15, F-16, F/A-18A/B, F/A-18E/F, and T-45. The
Estimating Test and Evaluation Costs 85
NR DEV 100
STE 93 100
GND 92 97 100
FLT 63 80 64 100
GND + FLT 91 100 97 81 100
OTHER 95 99 95 77 97 100
ST/F 61 63 91 14 67 50 100
T1 96 95 98 58 94 95 94 100
WE 40 42 58 8 47 33 99 51 100
AC Mo 48 53 36 75 52 54 –35 36 –35 100
EMD DUR 74 85 78 81 86 81 52 73 47 42 100
DUR 54 76 67 82 77 71 40 58 35 42 94 100
FLT HRS 63 58 58 46 59 55 28 58 28 61 35 16 100
F/A –51 –52 –61 –24 –54 –47 –71 –58 –74 10 –73 –63 –11 100
CGO –11 –22 –3 –51 –19 –26 54 –3 58 –62 1 –7 –19 –67 100
Estimating Test and Evaluation Costs 87
CER CER
(Preferred) (Alternative) Database
Contractor ST&E • • •
Ground tests • • •
Static and fatigue test • •
Other ground test •
Flight tests • • •
Other tests • •
Government DT&E •
Government OT&E •
Figure 5.4
Total Contractor ST&E Cost CER
Predicted contractor ST&E costs
Table 5.7
Total ST&E CER—Alternative 1
Table 5.8
Total ST&E CER—Alternative 2
Table 5.9
Ground-Test CER—Preferred
Table 5.10
Ground-Test CER—Alternative
Figure 5.6
Static and Fatigue Test Costs
Predicted contractor static and fatigue
test costs
Table 5.11
Static and Fatigue Test CER—Preferred
Flight Testing
The CER in Figure 5.7 and Table 5.12 provided the best predictive
value for contractor flight-test costs, capturing both the fixed (dura-
92 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table 5.12
Flight-Test CER—Preferred
Figure 5.8
Percentage of Government Costs in Flight Test Have Been Increasing
100
90
80
Percentage of flight-test costs
70 65%
60
52%
50
40
30 27%
20 17%
10
0
F-14* F/A-18A/B* F/A-18E/F F-22**
* Total government costs shown.
** In progress.
RAND MG109-5.8
_____________
6 In these two cases, the government costs shown represent total program office spending on the pro-
gram during the flight-test period and are thus an upper bound on what the program office could have
spent on testing alone.
Estimating Test and Evaluation Costs 95
Table 5.15
Other Contractor Test CER—Preferred
of the rest of the test program. The CER estimates other test costs as
a function of the estimated costs of ground and flight test.
Table 5.16
Summary Characteristics of Missiles and Guided Munitions
Type of DT Guided
Type of Devel. or Live DT
Weapon Program Launches Months FSD/EMD Start
Phoenix Air-to-air
missile New 64 43 December 1962
AMRAAM Air-to-air
missile New 86 44 December 1981
AMRAAM Air-to-air
Ph 1 missile Mod 12 24 March 1991
AMRAAM Air-to-air
Ph 2 missile Mod 13 89 June 1994
AIM-9X Air-to-air
missile Mod 20 33 December 1996
AMRAAM Air-to-air
Ph 3 missile Mod 8 18 December 1998
HARM Air-to-
ground
missile New 23 18 August 1978
IIR Maverick Air-to-
ground
missile Mod 52 21 October 1978
Harpoon Cruise
missile New 33 15 June 1971
SLAM-ER Cruise
missile Mod 8 15 March 1995
JASSMa Cruise
missile New 10 19 November 1998
SFW Unpowered
munition New 25 39 November 1985
JSOW Baseline Unpowered
munition New 24 N/A June 1992
JDAM Unpowered
munition New 170 N/A October 1995
WCMD Unpowered
munition New 61 20 January 1997
a
In progress; not used for CER development.
Table 5.17
Variable Definitions for Missiles
Table 5.18
Missile Variables Correlation Matrix (10 Programs)
Figure 5.10
Missile Total ST&E CER (Excludes Guided Munitions)
Predicted missile ST&E costs
Table 5.19
Missile Total ST&E CER—Preferred
Table 5.20
Missile Total ST&E CER—Alternative
detailed data are given, major test activities can also be compared
directly. When adequate data are available, regression analysis can be
used to develop tailored CERs at the total level or for selected por-
tions of the test program. If sufficient comparable data are not avail-
able to derive meaningful statistical relationships, it may be useful to
use averages and ranges for several analogous programs or values from
a single comparable program, adjusted for differences in the program
being estimated.
As always when dealing with small data sets, cross checks should
be used to confirm estimates. Alternative metrics that can be useful to
develop cross checks include the following:
• cost per flight hour (aircraft)
• cost per shot (guided weapons)
• cost per test month.
CHAPTER SIX
107
108 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Mission
The B-1B is a long-range supersonic bomber, originally designed to
penetrate sophisticated air defenses. 1 The CMUP enhances the
B-1B’s capability to deliver modern conventional weapons. Modi-
fications are being incorporated in four combined hardware and
software block upgrades. This description summarizes only the
CMUP Block D upgrade test program.
System Description
The Block D upgrade to the B-1B consists of the following hardware
and software modifications:
• Global Positioning System (GPS) capability for navigation,
offensive avionics, and weapon delivery
• jam-resistant very high and ultrahigh frequency radios
• MIL-STD-1760 interface to the Multipurpose Rotary Launcher
to accommodate advanced conventional munitions
• JDAM integration
• offensive system and ancillary software sustainment upgrades
_____________
1 Information on the B-1B came from SPO input, Air Force Flight Test Center (2000) for
DT/IOT&E, and AFOTEC (1998) for dedicated OT&E.
111
112 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Programmatics
Boeing North American, the system prime contractor, was awarded
the contract to integrate the B-1 CMUP modifications. Program-
matic milestones included the following:
• EMD contract award: March 1995 (Contract F33657-94-C-
0001)
• critical design review (CDR): May 1996
• production approval: July 1997
• production contract award: July 1997 (Contract F33657-97-C-
0004).
Test Program
The combined DT&E/IOT&E focused on verifying the GPS
incorporation, sustainment software upgrades, reincorporation of
ground moving target indication and tracking, defensive system soft-
ware upgrades, and JDAM integration.
Initial avionics testing was conducted at the contractor’s avionics
and system integration laboratories, the Avionics Integrated Support
Facility at Tinker AFB, and the Integrated Facility for Avionics Sys-
tem Test at Edwards AFB. The avionics software modifications were
accomplished in 18 months because software anomalies were identi-
fied during ground and flight tests.
Initial JDAM testing took place at Arnold Engineering Devel-
opment Center, followed by ground testing; captive-carriage, safe-
separation, environmental conditions, safety, and electromagnetic
interference (EMI) and electromagnetic compatibility (EMC) testing;
and actual flight tests at Edwards AFB. Approximately six JDAM
separation test vehicles and 23 guided test vehicles were dropped. In
addition to testing the performance of the Block D changes, the test
program also verified correction of previously identified system-level
deficiencies. Full-up live-fire testing of the B-1B was waived, but
component-level vulnerability testing was performed. There were
5,411 hours of major ground testing.
Aircraft Program Descriptions 113
Table A.1 summarizes the aircraft sorties made for the various
types of testing. Table A.2 provides data for the participating aircraft.
The three aircraft put in a total of 34 aircraft months of service.
AFOTEC conducted an operational assessment, combined
DT&E/IOT&E, and a one-month dedicated IOT&E. Approxi-
mately nine sorties each were used for combined testing and dedi-
cated IOT&E.
EMD began in January 1995, with the first test flight in March
1997. DT ended in September 1998; two aircraft made 97 flights,
flying for 660 hours. On average, the flight-test program achieved
20.6 flight hours per aircraft month, using engine-running crew
changes to maximize aircraft utilization.
Dedicated OT&E began in August 1998 and ended in Septem-
ber 1998. The three aircraft mentioned in Table A.2 provided six air-
Table A.1
B-1B Flight Testing
Test Flight
Aircraft Sorties Time
Event (no.) (no.) (hrs.)
Table A.2
The B-1B Test Aircraft
craft months and flew a total of nine flights and 53 flight hours. Air-
craft 85-0082 was the Block D Kit Proof aircraft.
B-2A Spirit
Mission
The B-2 bomber combines LO, large payload, and long range to
deliver conventional or nuclear munitions. 2 The B-2 program began
during the early 1980s with the objective of penetrating sophisticated
air defenses and attacking high-value and heavily defended targets.
The design was modified for low-altitude operations during FSD.
Additional design and test efforts have been required to integrate pre-
cision conventional munitions, including the GPS-Aided Targeting
System, GAM, and JDAM.
System Description
The B-2 has a two-person crew and is powered by four General Elec-
tric F-118-GE-100 engines rated at 17,300 pounds of thrust. The
significant features include the following:
• reduced electromagnetic, infrared (IR), acoustic, visual, and
radar signatures
• extensive use of composite structures in the airframe
• fabrication, assembly, and finishing of parts to high tolerances to
achieve stealth
• a blended flying-wing shape
• two internal weapon bays
• 44,000-pound payload
• designed to carry general purpose bombs from 500 to 4,700
pounds, mines, GAMs, JDAM, and nuclear bombs3
• engine inlets and exhaust shaped for radar and IR stealth
_____________
2 Information on the B -2 came from CTF Highlight Summary Flight Test Production (undated, after
June 1997); T&E Master Plan for the Sustainment Phase of the B-2A Advanced Technology Bomber,
March 23, 1999; briefing slides from and discussions with the B-2 SPO.
3 A JSOW capability was added post-baseline.
Aircraft Program Descriptions 115
Programmatics
A cost-plus-incentive-fee FSD contract was awarded to Northrop in
November 1981. This contract specified delivery of two ground-test
articles and six flight-test vehicles. Boeing and Vought teamed with
Northrop to design and build the airframe. Hughes Aircraft Com-
pany was the radar contractor, and General Electric Aircraft Engine
Group developed and built the engine. After flight testing, the EMD
aircraft were updated to the baseline (Block 30) configuration for use
as operational assets. There was no full-rate production. Program-
matic milestones included the following:
• aircraft preliminary design review: November 1982 and May
19844
• CDR: December 1985
• first engine delivery: December 1986
• LRIP contract: November 1987
• first flight: July 1989
• delivery of the durability test article: September 1989
• delivery of the static test article: January 1990
• first production aircraft delivery: December 1993
• end of EMD (baseline): March 1998.
Test Program
FSD/EMD began in November 1981, with first flight in July 1989
and DT ending in March 1998. The six test aircraft made 1,013
flights, flying 5,197 hours for a total of 310 aircraft months.
Dedicated IOT&E began in October 1993 and ended in June
1997. The six aircraft made 11 flights, flying 94.7 hours for these
tests.
_____________
4 The second review was for the low-altitude redesign.
116 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table A.3
B-2 Flight-Test Program
Test Flight
Event Dates Aircraft Sorties Hours
Table A.4
The B-2 Test Aircraft
C-17 Globemaster
Mission
The C-17 provides worldwide airlift for U.S. and allied combat
forces, equipment, and supplies.5 It can deliver passengers and out-
size, oversize, or bulk cargo over intercontinental distances without
refueling. The aircraft can land at conventional or austere airfields or
can make deliveries by airdrop. The capability for rapid, in-flight
reconfiguration allows the C-17 to transition easily among its mission
modes.
System Description
The C-17 is a four-engine turbofan transport aircraft. Its engines are
modified versions of an earlier commercial airline engine (PW-2040).
Significant features include the following
• A supercritical wing design and winglets reduce drag and
increase fuel efficiency and range.
• The aircraft can be refueled in flight.
• An externally blown flap configuration, direct lift-control spoil-
ers, and a high-impact landing gear system allow the aircraft to
use small, austere airfields.
• A forward and upward thrust-reverser system provides backup
capability, reduces the aircraft’s ramp-space requirements, and
minimizes interference with ground operations.
• The airdrop system is fully automated.
• A single loadmaster can operate the cargo door, ramp, and cargo
restraint systems, and off-loading equipment does not require
special handling equipment.
• The flight-control system is electronic (quad-redundant, fly-by-
wire).
• The two-person cockpit has multifunction displays.
_____________
5 Information on the C-17 came from T&E Master Plan for the C-17 Weapon System, dated August
1999; briefing slides and discussions with C-17 SPO; C-17 Program Office Estimate, dated June 1993;
C-17 Flight Test Progress Report for Month Ending December 31, 1994.
Aircraft Program Descriptions 119
Programmatics
A fixed-price-incentive-fee full-scale development contract with two
production options was awarded to McDonnell Douglas Aircraft
Company in December 1985. The contract was restructured in Janu-
ary 1988. LRIP was approved in January 1989. The lot III produc-
tion contract was awarded in July 1991, and the first flight was in
September 1991.
Test Program
Several developmental problems affected the test program. Software
integration was more complex than originally anticipated and
increased the amount of test effort, particularly for regression testing.
In an attempt to identify and solve problems on the ground,
McDonnell Douglas (now Boeing) established an avionics integration
laboratory and a flight hardware simulator in Long Beach, California.
Tables A.5 and A.6 summarize the flight-test program and aircraft
used, respectively.
FSD/EMD began in December 1985; first flight was in Sep-
tember 1991, and DT ended in December 1994. Six test aircraft flew
a total of 1,134 sorties and 4,096 flight hours. The test program used
a total of 169 aircraft months. Although the original plan had been
for 69 aircraft months and 2,277 flight hours (33.0 flight hours per
aircraft per month), the realized rate was 24.3 flight hours per aircraft
per month.
During static testing, the wing broke at 136 percent load and
had to be redesigned and retrofitted to production aircraft. The
original design included a hydromechanical flight-control system, but
wind-tunnel testing identified problems with it. As a result the sys-
120 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table A.5
C-17 Flight-Test Program
Test Flight % of
Aircraft Sorties Time Flight
Event (no.) (no.) (hrs.) Time
Table A.6
The C-17 Aircraft Tested
Test Aircraft
Aircraft Entered Testing Exited Testing Months
Missions
The F/A-18E/F is a carrier-based multimission strike fighter derived
from the F/A-18C/D. The F/A-18E/F’s range, payload, and surviv-
122 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
System Description
The F/A-18E is a single-seat and F/A-18F is a two-seat combat-
capable trainer. Every part of the F/A-18E/F structure was redesigned
from its predecessor. In general, the structure was enlarged and
strengthened, the part count was reduced, and the use of materials
and tooling were changed. Ninety-six percent of the airframe unit
weight is unique to the E/F. The structure’s material composition
features more titanium and composites and less aluminum than its
predecessor. Changes from the F/A-18C/D include the following:
• The gross landing weight increased by 10,000 pounds.
• Redesigning the forward fuselage increased its strength and
decreased the part count.
• The center-aft fuselage was lengthened 34 inches to increase fuel
capacity.
• Wing area increased 100 ft2 , and the wingspan increased by
more than 4 ft, also increasing internal fuel capacity.
• The areas of the control surfaces, horizontal tail surfaces, and
leading-edge extension increased.
• Unitizing reduced the part count by 42 percent from that of the
C/D and reduced manufacturing costs.
• The new configuration can carry an additional 3,600 pounds of
fuel internally and 3,100 pounds externally.
• The aircraft has two additional hard points for weapons.
• Incorporating such low-cost stealth features as saw-toothed
doors and panels, realigned joints and edges, and angled anten-
nas reduced the RCS.
_____________
6 The test program description came from TEMP No. 0201-04 Rev. B; the dates, number of flights,
and aircraft months came from the EMD flight log that the program office provided; the system descrip-
tion comes from various sources.
Aircraft Program Descriptions 123
• Although the E/F and the F/A-18C/D have many common sub-
systems, the E/F has more powerful actuators to accommodate
larger control surfaces.
• Over 90 percent of the avionics are common with the F/A-18C.
• General Electric developed the F414-GE-400 turbofan engine,
rated at approximately 22,000 pounds thrust, for the new air-
craft. Two engines power the E/F.
Programmatics
The Navy awarded a cost-plus-incentive-fee contract for airframe
EMD to McDonnell Douglas (now Boeing) in June 1992. Northrop
is the major airframe subcontractor and is responsible for the center
and aft fuselage, vertical tail, and several subsystems. Three ground-
test articles—static, drop, and fatigue—were built, and seven flight-
test vehicles were built and flight tested in EMD.
• EMD contract award: June 1992
• CDR: June 1994
• first flight: November 1995
• fatigue testing completed: July 1998
• flight testing completed: April 1999.
Test Program
Two phases of studies and testing preceded EMD. The configuration
study phase (1988 to 1991) used approximately 600 hours of wind-
tunnel tests of a 12-percent scale model to define the current aero-
dynamic configuration and engine performance requirements. Dur-
ing the pre-EMD phase (1991 and 1992), an additional 2,000 hours
of wind-tunnel tests on models helped optimize the configuration
and reduce design risks before EMD.
DT in EMD was conducted in several phases. During DT-IIA
(November 1995 to November 1996), the static test article was used
for initial wing-bending tests; the drop-test article was used for a
series of landing-gear tests at increasing sink rates; the manned flight
simulator was used for aircrew training and other tests; and flight
testing focused on envelope expansion.
124 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table A.8
The F/A-18E/F Test Aircraft
Test Aircraft
Aircraft Enter Testing Exit Testing Months
F/A-22 Raptor
Mission
The F/A-22’s primary mission is air superiority, with a secondary air-
to-ground mission when equipped with JDAM.8 Its combination of
sensors, displays, weapons, and LO is designed to provide first-look,
first-kill capability in all tactical environments. It will eventually
replace the F-15 in the air superiority role. Its designation was
changed from F-22 to F/A-22 in September 2002 to recognize its
dual role.
System Description
The F/A-22 Raptor is a twin-engine, single-seat, LO, all-weather
fighter and attack aircraft. The following are some of its key features:
• Advanced turbofan engines, which allow sustained supersonic
cruise without afterburners, and thrust vectoring provide
enhanced performance.
• Reduced radar and IR signatures and internal weapon carriage
provide LO.
• The advanced integrated avionics include
– fusion of radar, EW, and communications, navigation, and
identification sensor outputs
– long-range, active and passive, offensive and defensive sensors
to improve detection and tracking
_____________
8 Information on the F-22 came from the F-22 T&E Master Plan, Version 1.0, dated September 2000;
Draft F-22 CARD, dated April 1999; briefing slides and discussions with F-22 SPO; From the
Advanced Tactical Fighter to the F-22 Raptor, ANSER, March 24, 1998.
Aircraft Program Descriptions 127
Programmatics
The Advanced Tactical Fighter entered its DEM/VAL phase in
October 1986 and completed it in August 1991. This phase pro-
duced the YF-22 (Lockheed, Boeing, and General Dynamics) and
YF-23 (Northrop and McDonnell Douglas) prototypes to demon-
strate airframe and engine design approaches.
In August 1991, the F-22 entered EMD, with Lockheed Martin
selected as the system contractor and Pratt & Whitney as the engine
contractor. The EMD contract specifies nine EMD aircraft, two
ground-test articles, and a full scale pole model for RCS testing. In
1996, the Air Force Acquisition Executive, concerned about cost
growth trends on the program, chartered a joint estimating team
(JET) to assess the F-22’s costs and schedule. The JET recommended
delaying the transition to production and adding 12 months for
completing avionics development. The restructured program dropped
the preproduction verification aircraft. Congress imposed a cost cap
on development and production that was based on the JET
projections, as well as exit criteria the program had to meet before
Congress would authorize the transition to production.
The EMD aircraft first flew in September 1997. In May 1998, a
contract for two production-representative test vehicles (PRTVs) and
a first lot of six production aircraft was awarded. A separate “program
support” contract for contractor tasks not directly identifiable to a
128 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Test Program
The F/A-22 test program is representative of the most complex pro-
duction aircraft testing because it combines a new, advanced-design
airframe, engine, avionics, and LO features in a single, highly inte-
grated system. This translates into additional testing complexity (see
Table A.9 for a summary).
The test programs during DEM/VAL focused on allowing the
two competing contractor teams to demonstrate key technologies and
risk reduction. It was not a competitive “fly-off.” The government
did not specify the testing, and the test results were not a deciding
factor in the EMD source selection. The contractors, with govern-
ment participation, executed a short but intense set of flight demon-
strations.
One carryover from the DEM/VAL phase is the use of a highly
modified Boeing 757 flying test bed to test and troubleshoot F/A-22
avionics and software before its installation on the F/A-22. The test
bed has an F/A-22 forward fuselage grafted onto its nose and a wing
for mounting sensors attached to the upper fuselage, above the cock-
pit.
The program also used the Variable Stability In-Flight Simula-
tor Test Aircraft, an F-16 specially configured to mimic the flying
characteristics of the F/A-22. This aircraft was used to verify the flight
control laws to be used in the F/A-22 flight control system.
The F/A-22 CTF consists of DT&E and OT&E test personnel
from the contractors, the Air Force Flight Test Center, AFOTEC,
the F/A-22 SPO, and Air Combat Command. Of the 770 personnel
on the CTF in late 2001, 480, or 62 percent, were government (mili-
Aircraft Program Descriptions 129
Table A.9
Special Requirements for F/A-22 Testing
Table A.9—continued
The test program has been rebaselined several times since the
beginning of EMD. To date, approximately 122 test months have
been lost primarily because of late aircraft deliveries for testing. Table
A.10 presents RAND’s projections for the total F/A-22 flight-test
program, based on the flights and flight hours through June 2001
plus the SPO’s estimate of additional flights and flight hours based on
the June 2001 replan.
Table A.11 presents the aircraft calendar months in flight test
(after initial airworthiness testing). Delivery dates for aircraft that had
yet to be delivered are as of January 2002.
Table A.12 reports the DEM/VAL and EMD schedules, along
with sorties, aircraft, and rates, for the F/A-22 planned flight test pro-
gram.
Table A.10
F/A-22 Flight Test; June 2001 Replan
Flight
Test Aircraft Sorties Time
Event (no.) (no.) (hrs)
Table A.11
The F/A-22 Test Aircraft
Aircraft Calendar
Test Enter Exit Months to Complete
Aircraft EMD Testing EMD Testing a EMD Testing
Table A.11—continued
Aircraft Calendar
Test Enter Exit Months to Complete
Aircraft EMD Testing EMD Testing a EMD Testing
Table A.12
F/A-22 Planned Flight-Test Program
Flight
First Aircraft Sorties Aircraft Time
Flight (no.) (no.) Months (hrs) Ratea
DEM/VAL
October 1986–
January 1991
(YF-22) August 1990 2 74 5b 92a 18.4
September
(YF-23) 1990 2 50 5b 65 13.0
EMDc
August 1991— September
January 2004 1997 9 1,659 234 3,680 10.7
a
In flight hours per aircraft month.
b
Approximate.
c
SPO projections for completing all EMD testing.
Aircraft Program Descriptions 133
Mission
The Undergraduate Jet Flight Training System (UJFT) provides
intermediate and advanced strike pilot training using an integrated
ground and flight training system.9 The flight training includes air-
craft familiarization, basic instruments, airway navigation, air-to-
ground and simulated air-to-air weapon delivery, aerial combat
maneuvering, carrier qualifications, low-level navigation, formation
flying, and tactical maneuvering. The system also supports an instruc-
tor training course. The T-45 Training System was developed to
replace both the T-2C and TA-4J.
System Description
The T-45 Training System (T45TS) consists of the T-45 Goshawk
aircraft; an aircraft simulator suite for both instrument and visual
flight training; flight training–related academic materials, including
training courses, equipment, and course materials for UJFT and
instructor training; a computer-based training integration system; and
contractor logistics support.10
The T-45 Goshawk is a tandem-seat single-engine carrier-
capable jet aircraft derived from the existing BAe Hawk. The aircraft
includes the Navy aircrew common ejection seat, the standard atti-
tude heading reference system, an onboard oxygen generating system,
carrier operations capability, and a training weapon delivery capabil-
ity. The T-45 is fully contractor supported, including all levels of
maintenance and logistics.
Programmatics
The T-45 entered EMD in 1984. The initial OT (OT-IIA) in
November 1988 identified major deficiencies in aircraft handling
_____________
9 Information on the T-45 and T-45 Cockpit 21 came from the Test and Evaluation Master Plan No.
786 for the Naval Undergraduate Flight Training System (T45TS) (Revisions 5 and 6), briefing slides
and discussions with the program office.
10 Note that, for this study, we addressed only the aircraft portion of the system.
134 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Test Program
The T-45 was originally planned as a firm-fixed-price demonstration
program that would require only relatively minor modifications. It
was the first modern land-based aircraft to be modified for carrier
capability. However, much of the structure of the Hawk had to be
changed to make it carrier-suitable.
As a result, the program evolved into a major development
effort. Because it used an existing airframe, little M&S was originally
planned; this increased the flight hours necessary to test changes to
the airframe. Because of the performance shortfalls and redesign,
normal aeronautical and engine simulations eventually had to be
developed anyway.
Relatively few government test engineers were assigned to the
program. The Naval Air Warfare Center’s Aircraft Division (NAWC-
AD) did most of the government DT. High-angle-of-attack testing
was done at the Air Force Flight Test Center. Most of the contrac-
tor’s testing was done at its facility in Yuma, Arizona. One test air-
craft was lost in a Class A mishap, resulting in a 12–18 month slip.
Live-fire testing was not required. COMOPTEVFOR conducted the
operational assessments and testing.
All OT periods (and associated flight hours) are government
only, as reported in the T-45 TEMP (see Table A.13). The contrac-
tor flew hundreds of additional flight hours during DT, which the
TEMP does not include. From the beginning of DT to February 1,
1994, which encompasses all DT testing and the first OPEVAL
period, the total contractor and government flight testing was 1,880
flights and 1,932 flight hours. Table A.14 summarizes the informa-
tion in Table A.13.
Aircraft Program Descriptions 135
Table A.13
T-45 Flight Testing
Test Flight
Aircraft Sorties Time
Event (no.) (no.) (hrs) Remarks
Table A.14
T-45 Testing Summary
Flight
Test Time
Event Aircraft Sorties (hrs) Remarks
T-45 Cockpit-21
Mission
The T-45 Training System provides intermediate and advanced strike
pilot training using an integrated ground- and flight-training system.
Undergraduate jet pilot training (UJPT) includes aircraft familiariza-
tion, basic instruments, airway navigation, air-to-ground and simu-
lated air-to-air weapon delivery, aerial combat maneuvering, carrier
qualification, low-level navigation, formation, and tactical maneu-
vering. Also, the system supports an instructor training course. The
T-45TS was developed to replace both the T-2C and TA-4J.
System Description
The T-45 Training System (T45TS) consists of the T-45 aircraft; an
aircraft simulator suite for both instrument and visual flight training;
academic materials, including training courses, equipment, and
course materials for UJPT and instructor training; a computer-based
training integration system; and contractor logistics support. 11 The
_____________
11 Note that, for this study, we addressed only the aircraft portion of the system.
Aircraft Program Descriptions 137
first 83 aircraft delivered to the Navy were T-45As with analog cock-
pits. Subsequent deliveries were T-45Cs, which incorporate a digital
cockpit known as Cockpit-21. To provide an early introduction to
digital displays similar to those used in current fleet tactical aircraft,
the Cockpit-21 avionics upgrade replaces current conventional pri-
mary flight instruments with two multifunctional displays in each
cockpit.
Programmatics
Prototype testing of the Cockpit-21 upgrade began in March 1994,
with the first flight the same month. Approval for fleet introduction
of the T-45C was recommended in December 1998, following OT-
IIIB earlier that year.
Test Program
Contractor and government flight testing ran from March 1994 to
September 1998 and involved up to four aircraft (see Table A.15).
Table A.15
T-45C Flight Testing
Test Flight
Aircraft Sorties Time
Event (no.) (no). (hrs) Remarks
Table A.15—continued
Test Flight
Aircraft Sorties Time
Event (no.) (no). (hrs) Remarks
These aircraft flew 846 sorties and flew for 1,199.4 hours. Following
OT-IIIB, the T-45C aircraft was determined to be operationally
effective and operationally suitable.
V-22 Osprey
Mission
The V-22 weapon system is a multiservice, multimission vertical and
short takeoff and landing aircraft.12 The Marine Corps MV-22’s pri-
mary mission is amphibious assault. The Air Force CV-22’s primary
mission is long-range infiltration, exfiltration, and resupply of Special
Operations Forces. Secondary missions are land assault, medical
evacuation, fleet logistics support, and special warfare. The MV-22
will replace the CH-46E and CH-53A/D in the Marine Corps inven-
tory. The CV-22 will replace the MH-53J and MH-60G and will
supplement the MC-130 in the Air Force inventory.
_____________
12 We derived the information in this section from a meeting with the V-22 Program Office July 24,
2001, and from subsequent comments and input; numerous issues of Bell-Boeing’s Tiltrotor Times ; and
the draft V-22 Osprey TEMP No. M960 Rev. B.
Aircraft Program Descriptions 139
System Description
The V-22 is a tilt-rotor aircraft with rotating engine nacelles mounted
on each wingtip, enabling vertical and short takeoff and landing. The
nacelles rotate to the horizontal position for cruising at high speed.
The tilt-rotor design combines the vertical flight capabilities of a heli-
copter with the speed and range of a turboprop airplane and permits
aerial refueling and worldwide self-deployment. Two Rolls Royce
T406-AD-400 turboshaft engines drive two 38-ft. diameter prop-
rotors. The proprotors are connected to each other by an inter-
connecting shaft, which maintains proprotor synchronization and
provides single-engine power to both proprotors in the event of
engine failure. A triply redundant digital fly-by-wire system controls
engines and flight controls. The airframe is primarily graphite epoxy
composite. An integrated EW defensive suite that includes a radar
warning receiver, a missile warning set, and a countermeasures dis-
pensing system will be installed.
The Air Force CV-22 and Marine Corps MV-22 share the same
basic airframe design. The CV-22 is configured for Special
Operations Forces’ infiltration and exfiltration operations. Unique
CV-22 items include a terrain following and terrain avoidance radar
system, additional fuel tanks to increase range, an additional crew
position, an expanded communications suite, and the addition of a
defensive systems suite to enhance survivability during penetration
missions.
Programmatics
V-22 development and testing have taken place in two distinct pro-
grams. An FSD contract was awarded to a Bell-Boeing joint venture
team on May 2, 1986. Six flight-test articles were planned, and five
were completed. First flight of the V-22 in FSD was in March 1989.
The Secretary of Defense cancelled the program on April 19, 1989,
but Congress and the services continued to fund the test program
incrementally until early FY 1993.
Approximately 820 hours of flight testing were completed. Two
aircraft were lost: Aircraft number five crashed on its first flight in
June 1991 because of a problem with its gyroscopic sensors, and air-
140 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
craft number four crashed on July 20, 1992, because a section of the
drive shaft failed. On October 22, 1992, the FSD airframe contract
was officially terminated, and a letter contract for EMD was awarded
on the same day.
The purpose of the V-22 EMD program is to design a lighter,
lower cost aircraft than the FSD design and to build four production
articles (aircraft numbers 7–10). An EMD engine contract was
awarded in December 1992. DT and OT continued during EMD
using the three remaining FSD aircraft, supplemented by EMD air-
craft as they were built.
Test Program
During FSD (1986 to 1992), the V-22 T&E program concentrated
on engineering and integration testing performed by the contractor.
NAWC-AD conducted three periods of formal DT, with the partici-
pation of the operational test community in integrated test team
activities. These tests provided early insight into the development
effort.
The EMD flight-test program began by using the FSD V-22
aircraft for design support, risk reduction, and envelope expansion. A
CTF, consisting of Navy, Air Force, Bell, and Boeing personnel, con-
ducted the EMD test program, with the exception of dedicated
operational test events. Testing of four EMD aircraft began in FY
1997, following the first flight of aircraft number 7 in February 1997.
Integrated testing (IT) and OT used a combination of the FSD and
EMD V-22 aircraft during initial assessment (IT-IIA/B/C and OT-
IIA/B/C). OT-IID and subsequent tests used the EMD configuration
aircraft.
The Navy conducted DT&E of the MV-22 at Patuxent Naval
Air Station, Maryland. A CTF stationed at Edwards AFB, California,
conducted the DT&E for the CV-22.
The Secretary of Defense certified a waiver of full system-level
live-fire testing. A comprehensive series of ballistic tests (582 shots
over 16 years) of critical components, major assemblies, and aircraft
structures was conducted. According to DOT&E, live-fire testing was
treated as “an integral part of the design process, not merely as a
Aircraft Program Descriptions 141
Table A.16
V-22 Flight Testing
Flight
Test Sorties Time
Event Aircraft (no.) (hrs) Remarks
Table A.16—continued
Flight
Test Sorties Time
Event Aircraft (no.) (hrs) Remarks
Table A.17
V-22 Testing Summary
AIM-9X Sidewinder
Mission
The AIM-9 Sidewinder is a heat seeking air-to-air missile designed for
short-range engagements.1 It complements the medium-range
AMRAAM in the fighter aircraft arsenal. The AIM-9X is a major
modification of the AIM-9, which the U.S. Naval Weapons Center at
China Lake, California, first developed in the 1950s. The missile has
had several variants and is used on all U.S. fighter aircraft, including
the F-15, F-16, and F/A-18, and will be employed on the F/A-18E/F,
F/A-22, and JSF. Many allied nations also use the Sidewinder, and
the Soviets and their allies copied and used its early design.
System Description
The AIM-9M, the AIM-9X’s immediate predecessor, can engage tar-
gets from all aspects; its IR counter-countermeasures and background
discrimination capability are better than those of its own predeces-
sors; and it has a reduced-smoke rocket motor. Deliveries of the M
model began in 1983.
The AIM-9X has improved counter-countermeasures, acqui-
sition range, off-boresight capability, background discrimination,
_____________
1 We derived the information in this section from the AIM-9X Selected Acquisition Report, December
31, 1999; an interview with the AIM-9X Joint Program Office (JPO), May 24, 2001; AIM-9X JPO
(2001a) and (2001b); Sidewinder Missile AIM-9X CARD, Final Update, July 7, 2000, PMA-259; Boe
and Miller (undated).
145
146 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Programmatics
The AIM-9X is a joint Navy–Air Force program, with the Navy as
lead service. AIM-9X is a CAIV program, trading off performance
and cost to achieve a best-value solution. A two-year concept-
Missile Program Descriptions 147
Test Program
Table B.1 summarizes the overall test program (as of June 2002),
which is described in greater detail in the paragraphs below.
Table B.1
AIM-9X Testing
Guided
Event Launches Remarks
DEM/VAL Two competitors
December 1994–
June 1996
DT-IIA 0 More than 50 captive-carry
January 1997– sorties
August 1998 Ground launch of
preprogrammed test vehicle
DT-IIB/C 9 More than 300 captive-carry
September 1998– sorties
August 2001 16 SCTV launches
DT-IID (TECHEVAL) 3 More than 75 captive-carry
March–December 2001 sorties
OT-IIA 5
September 1999–
July 2000
DT Assist 4 Concurrent with TECHEVAL
August–
December 2001
OT-IIB (OPEVAL) 22
August 2002–
May 2003
Missile Program Descriptions 149
Mission
AMRAAM is an all-weather, radar-guided, air-to-air missile that
replaced the AIM-7 Sparrow. 2 The F-15, F-16, and F/A-18 currently
carry this missile worldwide, and the F/A-18E/F, F-22, and JSF will
carry it for use against air threats in a variety of weather and elec-
tronic combat conditions. The U.S. Marine Corps Complementary
Low-Altitude Weapon System will use AMRAAM in a surface launch
role. FMS platforms include the German F-4F, the Swedish Gripen
and Viggen, the United Kingdom Tornado and Sea Harrier, and the
Norwegian Advanced Surface-to-Air Missile System (a ground-
launched version of AMRAAM). AMRAAM is capable of intercept-
ing maneuvering and all-aspect targets within and beyond visual
range and allows a single-launch aircraft to engage multiple targets
with multiple missiles simultaneously.
System Description
The AMRAAM weapon system includes the missile, launcher, the
fire-control system, supporting aircraft avionics, and aircrew inter-
faces.
AMRAAM can be either rail or ejector launched and interfaces
electrically with the platforms via the launch stations. The rail
launcher permits firing from wing-mounted stations and is compati-
ble with both AMRAAM and AIM-9. The ejector launcher, which
permits firing from F-15 and F/A-18 fuselage stations, is compatible
with both AMRAAM and AIM-7.
_____________
2 We derived the information in this section from Mayer (1993), information from the AMRAAM JPO
(AAC/YAF), October 17, 2001 through November 18, 2002; Advanced Medium Range Air-to-Air
Missile TEMP, March 2002.
Missile Program Descriptions 153
Programmatics
The AMRAAM acquisition strategy involved a two-contractor
DEM/VAL phase with downselection to a single FSD contractor. A
leader-follower approach was implemented during FSD to promote
competition by the fourth production lot. The AMRAAM began
FSD in December 1981 with a fixed-price-incentive contract award
to Hughes Aircraft Company. In July 1982, Raytheon won the sec-
ond source contract. Milestone II approval was granted in September
1982. At OSD direction, the AMRAAM Producibility Enhancement
Program (APREP) was initiated to lower the production cost of the
missile by identifying and incrementally redesigning high-cost com-
ponents. Congress imposed caps on both FSD and procurement cost.
Milestone IIIA was approved in June 1987 and full-rate production
(Milestone IIIB) in April 1991.
The AIM-120B missile is the result of the APREP. Improve-
ments include a new digital processor, field reprogrammable memory,
and insertion of large-scale integrated circuit and very-large-scale
integrated circuit electronic components. The AIM-120B was intro-
duced late in lot 6.
154 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Test Program
Table B.2 summarizes the overall test program, which we discuss in
more detail in the following paragraphs.
Table B.2
AMRAAM Testing
Live
Event Firings Remarks
DEM/VAL 7 One STV, three CTVs, and 3 GTVs were fired.
November 1979–
December 1981
FSD (DT&E/IOT&E) 100 Ninety AAVIs, 4 AAVs with warheads, and
October 1986–June 1990 six SCTVs were fired.
AF FOT&E (1) 37 The AFOTEC report supported full-rate
February 1990–May 1993 production.
Initial Navy OT-IIIA 6 COMOPTEVFOR concluded that the missile
January–July 1991 was potentially operationally effective and
suitable.
Navy OT-IIIB (OPEVAL) 29 COMOPTEVFOR concluded that the missile
September 1991–May 1994 was partially operationally effective and
suitable and supported IOC on the F/A-18.
AF FOT&E (2) 39 Used lot 4, 5, 6, and 7 missiles to evaluate
June 1993–March 1996 AIM-120A software in electronic attack,
chaff, and multiple-target scenarios.
Testing also included an initial evaluation
of the AIM-120B production software and
a 7,800 hour CCRP. At least one ACE (~39)
was flown for each launch.
P3I Phase 1 DT&E 12 Two SCTVs, one JTV, and nine AAVIs were
October 1993–October fired, and 36 development and preflight
1995 ACE missions were flown.
156 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table B.2—continued
Live
Event Firings Remarks
AF FOT&E (3A) 26 Verified the operational effectiveness and
August 1996–August 1999 suitability of the AIM-120B/C-3 hardware
and software; verified the correction of
deficiencies; and completed deferred
OT&E. Testing include 26 AAVI flights,
3,712 hours of captive-carry testing for
CCRP, and 25 ACE missions.
P3I Phase 2 DT&E 6 Included Phase 3 risk-reduction testing.
June 1994–November 2001 Testing included 54 ACE flights; digital and
hardware-in-the-loop simulations; and
launches of one SCTV, 13 AAVIs, and one
AAV (with warhead).
FDE (3B) 21 Testing included 16 AAVIs from lots 12 and
September 2000– a 13, five captive-carry reliability vehicles
(reconfigured lot 8 missiles), and a CCRP.
Ten ACE missions supported the launches.
P3I Phase 3 DT&E 13 Testing included eight AAVIs and one
February 2002–August reliability test vehicle, seven ground and
2003 five flight ITV missions, one IMV, eight
prelaunch ACE missions, and 30 ACE
missions to support software
development.
a
Ongoing at reporting time, with an estimated completion date of August 2002.
Mission
JASSM is a precision standoff weapon for attacking heavily defended,
high-priority targets.5 JASSM can be launched beyond the range of
enemy air defenses and can strike fixed or relocatable targets.
_____________
5 We derived the information in this section from the Lockheed Martin JASSM Web site
http://www.jassm.com/; an interview with the JASSM Program Office on May 30, 2001, and
subsequent communications.
Missile Program Descriptions 161
System Description
JASSM is 168 inches long. Its major manufactured parts include a
composite shell, fore and aft covers, tail, wings, fuel tanks, engine,
and warhead. The vehicle has an LO design, and each missile is tested
for its radio frequency signature. The tail and wings fold next to the
body before deployment to reduce space. The 1,000-pound class
warhead can penetrate hard targets, works with an impact or delayed
fuze, and is compliant with Navy requirements for insensitive muni-
tions. The missile is guided by a jam-resistant GPS and ring laser gyro
inertial navigation system (INS) with an IR seeker and pattern
matching autonomous target recognition system that provides aim-
point detection, tracking, and strike in the terminal phase. The con-
trol unit uses commercial electronics.
Programmatics
JASSM is the successor to the terminated Triservice Air-to-Surface
Standoff Missile (TASSM) program. Two 24-month program defini-
tion/risk-reduction (PD/RR) cost-plus-fixed-fee contracts totaling
$237.4 million were awarded to McDonnell Douglas and Lockheed
on June 17, 1996. The PD/RR phase was completed in October
1998.
After the open competition in the PD/RR phase, Lockheed was
selected as EMD contractor in April 1998, and a planned 54-month
EMD program began in November 1998. The Defense Acquisition
Board granted approval for LRIP in December 2001.
Test Program
Table B.3 summarizes the tests conducted for this program. The
paragraphs below describe these in greater detail.
JASSM is an acquisition reform program with no government-
directed DT. The contractor is responsible for planning and execut-
ing DT. The program progresses directly from contractor DT&E to
OT&E. DOT&E and AFOTEC were involved during the request
for proposal stage of the program to lay the foundation for data col-
lection during contractor DT&E. Lockheed’s test program (using
162 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table B.3
JASSM Testing
Guided Launch
Event Launches Platform Remarks
DT-1 0 F-16 Inert round, seeker guides to
January 2001 target
DT-2 1 F-16 Destroyed air defense target
April 2001
DT-3 1 B-52 Destroyed concrete bunker
May 2001
DT-4 1 B-52 Fuse did not arm, did not
July 2001 detonate
DT-5 1 F-16 Arming problem, did not
September 2001 detonate
DT-6 1 F-16 Destroyed hardened bunker
November 2001
DT-7 1 B-52 Destroyed hardened bunker
December 2001
DT-8 1 F-16 Penetrated hardened bunker
April 2002
DT-9 0 F-16 State of the art anti-jam GPS
July 2002 feature
DT-10 0 F-16 Successful navigation in intense
September 2002 jamming environment
OT-1a 1 F-16
April 2002
OT-2a 0 F-16 Two inert rounds
June 2002
OT-3a 1 B-52
August 2002
OT-4a 2 B-52
September 2002
OT-5a 2 B-52
January–March 2003
OT-6a 1 B-52 One live, one inert launch
March–April 2003
a
As of reporting time, these were planned events.
Mission
JDAM is an accurate, all-weather low-cost guidance kit for current
inventory 2,000- (Mark 84/BLU-109) and 1,000-pound (Mark-
83/BLU-110) bombs.6 JDAM provides highly accurate weapon
delivery in any “flyable” weather. After release, JDAM can receive
updates from GPS satellites to help guide the bomb to the target.
JDAM is a bomb-on-coordinates system that navigates from release
to the planned target coordinates. JDAM provides the user with a
variety of targeting options, such as preplanned targeting using the
Air Force Mission Support System (AFMSS) or the Navy’s Tactical
Automated Mission Planning System (TAMPS), sensor targeting, or
in-flight retargeting using onboard sensors and manual data entry.
JDAM also allows multiple target engagements on a single-pass deliv-
ery.
System Description
The guidance kit includes an INS augmented by GPS updates and a
tail control system. Target coordinates and other guidance data are
passed to the weapon through a MIL-STD-1760 interface from the
delivery aircraft during weapon initialization. Upon release, autono-
mous guidance is initiated using INS data only. After the JDAM GPS
receiver acquires the GPS satellites, precision GPS position and
velocity data are used to refine the INS data. The guidance unit
attaches to the bomb and, through controlled tail fin movements,
directs the bomb to the target. The JDAM is to be integrated on the
B-1B, B-2, B-52H, F-14B/D, F-15E, F-16C/D, F-18C/D, F-18E/F,
_____________
6 We derived the information in this section from the February 2000 TEMP and AFOTEC reports.
Missile Program Descriptions 165
F-22, F-117A, and AV-8B aircraft. The B-52 and F/A-18C/D are the
threshold aircraft.
Programmatics
JDAM is a joint Air Force–Navy program; the Air Force is the lead
service. JDAM is an ACAT identification program under OSD over-
sight. Because JDAM was selected as a defense acquisition pilot pro-
gram, many normal procurement requirements were reduced in favor
of using “best commercial practices.” Government management,
oversight, and decision processes were also streamlined.
Four contractors were involved in pre-EMD activities. The
development program had two phases. The Phase 1 EMD effort
began in April 1994 and involved two competing contractors,
McDonnell Douglas and Martin Marietta. The primary focus of
Phase 1 EMD was to reduce manufacturing risks and the projected
average unit production price by having each competitor develop its
design through CDR. The Phase 2 EMD effort began in October
1995 with the selection of one contractor, McDonnell Douglas. The
Phase 2 EMD effort completed system development with emphasis
on development and OT. The Phase 2 EMD contract was a cost-
plus-award-fee contract valued at approximately $102 million with a
period of performance from October 1995 to February 1999. The
Phase 2 EMD contract also included minimum and economic order
quantity options for production lots 1 and 2.
Test Program
Table B.4 summarizes the test program, which we discuss in greater
detail in the following paragraphs.
A test IPT that included representatives from the joint program
office, 46th Test Wing, NAVAIR, and Boeing, the system contractor,
managed DT. Representatives from the OT activities were also
involved early in DT.
Testing in EMD Phase 1 consisted of wind-tunnel testing, digi-
tal simulation, fit checks, use of weapon simulators to test interfaces
between the aircraft and the weapon, supportability analyses, and
166 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table B.4
JDAM Testing
Units
Airframe Event Fired Remarks
F-18C/D Mark 84 Safe separation 27 STVs Wind-tunnel and safe-
1–4Q FY1996 separation testing
DT&E 48 GTVs Captive carry and release,
1Q FY1997– carrier suitability, aircraft
3Q FY1998 integration, and mission
planning
DT/OT-IIA 14 GTVs Captive carry and release
July 1998– Flew 14 hours total with a
October 1998 restricted flight envelope
because of problems
with the tail actuator
subsystem
Concurrent DT evaluated
tail actuator subsystem
fixes
OT-IIB (OPEVAL) 55 GTVs Independent phase of OT;
November 1998– tested all weapon and
March 1999 fuze combinations
Flew approximately 300
hours (188 sorties)
Verified correction of tail
actuator subsystem
problems
DT/VCD 10 GTVs
1–3Q FY2000
B-52H Mark 84 Safe separation 20 STVs Wind-tunnel and safe-
3Q FY1996 separation testing
DT&E 16 GTVs
2Q FY1997–
2Q FY1998
DT/OT-IIA 16 GTVs
4Q FY1998
OT-IIB IOT&E 40 GTVs Independent phase of OT
1Q FY1999– Tested all weapon and
4Q FY1999 fuze combinations
19 sorties
F-16 Safe separation 25 STVs Wind-tunnel and safe-
3Q FY1996 separation testing
F-16 MK84
DT&E 64 GTVs
2Q FY1997–
2Q FY1998
Missile Program Descriptions 167
Table B.4—continued
Units
Airframe Event Fired Remarks
DT/OT-IIA 2 GTVs
4Q FY1998
OT-IIB IOT&E 2 GTVs Independent phase of OT
1Q FY1999– Tested all weapon and
4Q FY1999 fuze combinations
19 sorties
NOTES: The STV is a production-representative airframe with appropriate mass prop-
erties but without a guidance control unit or tail actuator subsystem. The GTV is a
production JDAM guidance kit that may or may not have telemetry, a warhead, or a
fuze.
Andersen AFB, Guam; and Minot AFB, North Dakota, flew Air
Combat Command B-52Hs for these tests. The Navy conducted
OT-IIB at Naval Air Warfare Station China Lake, and on three
aircraft carriers from November 1998 through September 2000.
A total of 207 sorties were flown during the MOT&E phase. Of
these, B-52Hs flew 19 sorties (three captive-carriage and 16 weapon-
release sorties) and FA-18C/Ds flew 188(147 captive-carriage and 41
weapon-release sorties). All missions were planned using either the
AFMSS or the Navy TAMPS. The test team evaluated the effective-
ness of these two systems, including AFMSS and TAMPS core; the
JDAM B-52H aircraft, weapon, and electronics; and FA-18C/D
mission-planning modules.
The test agencies rated JDAM as effective but not suitable
because of a combination of demonstrated weapon system perform-
ance against user requirements and the test team’s judgment of mis-
sion accomplishment. JDAM met the user’s requirements for captive-
carriage reliability for both threshold platforms. However, JDAM did
not meet the user’s requirements for storage reliability, mission reli-
ability, and system reliability. In particular, the guidance kit system
did not meet reliability thresholds, and the thresholds for mean time
to load and mission planning time were not met.
To address unresolved and unsatisfactory issues from IOT&E, a
dedicated FOT&E was planned. System reliability will be tracked
and evaluated through FOT&E and lot acceptance testing.
Mission
JSOW is a family of kinematically efficient 1,000-pound class air-to-
surface glide weapon.7 It has LO and provides multiple kills per pass,
preplanned missions, standoff precision engagement, and launch-and-
_____________
7 We derived this information from the JSOW Selected Acquisition Report of December 31, 1998;
DOT&E Annual Report FY 1996; Program Office (PMA 201) input via email dated April 5, 2002.
Missile Program Descriptions 169
leave capability against a wide range of targets, day or night and in all
weather conditions. JSOW is used for interdiction of soft or medium
fixed, relocatable and mobile light and heavy armored targets, massed
mobile armored targets, and antipersonnel and air-to-surface threats.
System Description
The JSOW Baseline (AGM-154A) consists of an airframe, a
guidance-and-control system with INS-GPS capability, and a payload
consisting of 145 BLU-97 submunitions. The JSOW/BLU-108
(AGM-154B) is similar, but the payload consists of 6 BLU-108 sub-
munitions. The JSOW Unitary (AGM-154C) is a Navy-only variant
that adds an autonomous imaging IR seeker to the INS-GPS guid-
ance system and a BAe Broach warhead with penetrator capability.
The all-up round is 160 inches long, has a wingspan of 106 inches
when fully deployed, and weighs 1,065 pounds.
JSOW missions are normally preplanned using the Navy
TAMPS or the AFMSS. JSOW interfaces with the aircraft through a
MIL-STD 1553 data bus. The F-18C/D and E/F, F-16C/D, F-15E,
JSF, B-1B, B-2A, and B-52H can employ this weapon.
Programmatics
The JSOW is a Navy-led, joint Navy–Air Force program. A DEM/
VAL contract was awarded in June 1989. The Navy awarded the
JSOW EMD contract in June 1992 to Texas Instruments Defense
Systems and Electronics (currently Raytheon Missile Systems). The
contract option for LRIP was exercised in February 1997. The Navy
approved full-rate production for the AGM-154A in October 1998.
In October 2002, the Air Force withdrew from the JSOW B
program. The Navy completed development but deferred production.
Test Program
Table B.5 summarizes the test program, which the following para-
graphs describe in more detail. Note that the description addresses
only the JSOW Baseline.
170 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table B.5
JSOW Testing
Guided
Event Launches Remarks
DT-IIA
December 1994–March 1995 2
DT-IIB
March 1995–December 1995 10 At China Lake and Point Mugu
DT-IIC
February 1996–October 1996 10 TECHEVAL and LFT&E
USAF DTE
July 1996–August 1996 2 F-16 integration
OT IIA
May 1996–September 1996 6 Operational assessment
OT-IIB
14 OPEVAL
February 1997–June 1997
Judged operationally effective and
operationally suitable
IOT&E
July 1998 2 OT for F-16
The Naval Air Warfare Center China Lake was the lead test
activity. Other facilities used included Point Mugu, NAWC-AD
(Patuxent River), and the Air Force Air Armament Center at Eglin
AFB.
The Air Force began DT&E flight testing JSOW on the F-16 at
Eglin AFB, Florida, in March 1996. AFOTEC conducted an opera-
tional assessment in December 1996, with 46th Test Wing con-
ducting an update in April 1998 in conjunction with DT&E. The
update focused on targeting and weapon platform performance, using
the results of laboratory M&S, captive-flight, and free-flight testing.
Less-than-desirable progress in integrating the baseline JSOW with
the F-16 hindered Air Force testing. The problem was the control-
section locking pins, a subassembly of the JSOW that Texas Instru-
ments did not manufacture.
The U.S. Navy began OPEVAL testing in February 1997, after
successful DT and initial OT programs. Over the entire test program,
42 of 46 shots were successful.
Missile Program Descriptions 171
Mission
SFW delivers antiarmor munitions to neutralize a large number of
enemy fighting vehicles in massed formations with a limited number
of sorties.8 It can be launched from a variety of altitudes and weather
conditions and in countermeasure environments. Two related pro-
grams, the WCMD and JSOW, can be used with SFW to give it
greater accuracy and range, respectively.
System Description
The SFW Cluster Bomb Unit (CBU-97/B) consists of three major
parts: the tactical munitions dispenser, the BLU-108 submunition,
and the projectile. The 1,000-pound-class tactical munitions dis-
penser holds ten parachute-stabilized BLU-108 submunitions. Each
submunition holds four armor-penetrating projectiles, each of which
has an IR sensor, for a total of 40 projectiles per SFW.
After the weapon is released, the tactical munitions dispenser
opens and dispenses the ten submunitions. At a preset altitude sensed
by a radar altimeter, a rocket motor fires to spin the submunition and
initiate an ascent. The submunition then releases its four projectiles
over the target area. The projectile’s sensor detects a vehicle’s IR sig-
nature, and an explosively formed penetrator fires at the heat source.
If no target is detected after a specific time, the projectiles fire auto-
matically, causing damage to material and personnel.
SFW is compatible with the A-10, F-15E, F-16, B-1, B-2
(WCMD only), and B-52 (WCMD only).
_____________
8 We derived the information in this section from the Sensor Fuzed Weapon Test and Evaluation
Master Plan, August 1996, Area Attack Systems Program Office; Sensor Fuzed Weapon Test and
Evaluation Master Plan, August 1999, Area Attack Systems Program Office; Information from Area
Attack Systems Program Office (AAC/YH), Eglin AFB; Sensor Fuzed Weapon Selected Acquisition
Report, December 31, 1998; SFW Operational Testing Summary; Director, Defense Operational Test
and Evaluation; Number of launches from FY 1997 DOT&E Annual Report.
172 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Programmatics
SFW began FSD in 1985. The development program was restruc-
tured in June 1989 because of test failures, schedule delays, and
budget changes. The restructured program included a transition to
production. After successful live-fire testing and a successful IOT&E
from September 1990 to December 1991, OSD approved LRIP in
March 1992. Decision authority was delegated to the Air Force in
1994, and the Air Force approved Milestone III for full-rate produc-
tion in June 1996.
Two Producibility Enhancement Program (PEP) hardware
upgrades were initiated for SFW to reduce costs and improve pro-
ducibility through design improvement. The first, PEP-1, involved
electronic and mechanical changes to the projectile, including the use
of an application-specific integrated circuit that placed most of the
electronic components on a single chip. The PEP 1 contract was
awarded in early FY 1994. The PEP-2 contract was awarded in early
FY 1995, and the program was to redesign the sequencer and altime-
ter into one integrated submunition electronics unit. The PEP-2 pro-
gram was cancelled because of technical problems, and the critical
elements were integrated into the P3I program.
In May 1996 the Air Force awarded a P3I contract for SFW.
The P3I program involves three major improvements: (1) improving
performance against countermeasures, (2) altering the warhead design
to improve performance against softer targets without degrading the
current target-set performance, and (3) raising the radar altimeter
height of function to increase area coverage. The current sensor will
be upgraded from a passive IR sensor only to a dual-mode type with
passive IR and an active laser sensor. This upgrade will allow the sen-
sors to discriminate between thermal and physical profiles of targets,
enhance the sensor’s performance against cooler targets, and improve
its effectiveness against countermeasures. The improved warhead con-
sists of a modified copper liner configured to form both a central
penetrator for hard targets and multiple smaller penetrators for soft
targets.
Missile Program Descriptions 173
Test Program
Table B.6 summarizes the test program. DT included LFT&E using
actual military vehicles instrumented for the test. Producibility modi-
fications have necessitated several QT&E phases as well. FOT&E of
PEP-1 was completed in 1998. All objectives were met, and testing
results indicated that PEP-1 changes have not degraded the perform-
ance of the SFW.
Mission
SLAM-ER provides standoff all-weather precision strike from carrier-
deployed aircraft against fixed, high-value land targets and, secondar-
ily, against relocatable stationary land targets and ships. 9 The missile
Table B.6
SFW Testing
_____________
9 We derived the information in this section from an interview with SLAM-ER Program Office on July
13, 2001 and from the FY 2000 DOT&E report for SLAM-ER.
174 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
System Description
The original SLAM was based on the Harpoon antiship missile, to
which it added a GPS-aided INS for midcourse guidance, a Maverick
imaging IR sensor, and a Walleye data link for man-in-the-loop con-
trol. SLAM-ER is a major upgrade with greater range; reduced sus-
ceptibility to countermeasures; greater capability against hardened
targets; an improved guidance navigation unit; and improved user
interfaces for mission planning, launch, and control. The primary
changes from the existing SLAM included the following:
1. a modified Tomahawk Block III warhead
2. an improved data link with greater range and jam resistance
3. an improved guidance set with integrated GPS-INS, a 1760 data
bus interface, a multichannel GPS receiver
4. modified Tomahawk wings
5. an automated mission-planning system.
Programmatics
SLAM-ER continued the SLAM development approach of maximiz-
ing use of existing components. After cancellation of the TASSM
program, Boeing received an EMD contract in March 1995 to
improve identified deficiencies of the interim SLAM. The first flight
was in March 1997. LRIP I was approved April 1997, LRIP II in
April 1998, and LRIP III in August 1999. Full-rate production was
approved in May 2000, with existing SLAMs to be upgraded to
SLAM-ER configuration. The threshold platform was the F/A-
18C/D.
Test Program
Table B.7 summarizes the test program. There were five DT and
eight OT launches. LFT&E was required because of the use of a new
titanium-cased warhead. LFT&E consisted of confined volume test-
Missile Program Descriptions 175
Table B.7
SLAM-ER Testing
Guided
Event Launches Remarks
DT-1 1 First flight
March 1997 Verified basic performance, mission planning,
and maneuvering
DT-2 1 Verified terrain following, data link, pilot
October 1997 designation, and target impact
DT-3 1 Verified range for low-level launch and flyout,
December 1997 shift from IR track to designated aimpoint,
and target impact
DT-4 1 Verified range for high-altitude launch, off-
February 1998 axis launch, steep impact angle, and stop-
motion aimpoint update
DT-5 1 Verified new operational flight program,
March 1998 quick-reaction launch and midcourse
update, target ID, and autonomous lock on
a moving ship target
DT/OT-1 1 Verified land-based target-of-opportunity
June 1998 capability
DT/OT-2 1 Demonstrated the ability to attack a hardened
June 1998 aircraft shelter
DT/OT-3 1 Demonstrated the ability to attack a high-
June 1998 altitude land-based target
DT/OT-4 1
September 1998 Consisted of production verification test
OT-IIA (OPEVAL) 6 Six missiles fired in operationally realistic
December 1998– scenarios
May 1999 One prior DT/OT launch included in OT
analysis
Missile judged not suitable and not effective
VCD 1 Verified software and hardware (missile and
June 1999 AN/AWW-13 data link pod) changes
DT-II 1
October 1999
OT-IIB 4 Four missiles fired in operationally realistic
November 1999– scenarios
January 2000 Deficiencies corrected
Missile judged suitable and effective
ATA OT 1 Evaluate ATA capability
February 2002
ing, three arena tests of warhead fragmentation, and four sled tests of
warhead penetration.
176 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Mission
WCMD is an all-weather guidance kit that replaces the tail on the
SFW (CBU-97), Combined Effects Munitions (CBU-87), and the
Gator Mine System (CBU-89).10 With the WCMD tail kit attached,
these weapon designations become CBU-105, CBU-103, and CBU-
104, respectively. WCMD is not a precision capability but does
improve accuracy, depending on the quality of the navigation unit
used. It can be launched from a variety of altitudes, weather condi-
tions, and in countermeasure environments. WCMD corrects for
wind effects, ballistic dispersion, and launch transients from moderate
to high altitudes.
System Description
The WCMD kit integrates a low-cost INS, control unit, and
steerable fins to guide tactical munition dispenser weapons. It is a
bomb-on-coordinates weapon and is used by aircraft that have GPS-
quality heading, velocity, and position data. The data are passed from
the aircraft to the weapon via an MIL-STD-1760 interface. The
threshold aircraft for compatibility are the F-16 and B-52. The objec-
tive aircraft are the B-1, F-15E, A-10, F-117, and JSF.
Programmatics
WCMD is an Air Force ACAT II program using commercial prac-
tices and government streamlining as much as possible. The con-
tracting process minimized the use of military specifications. The
acquisition emphasized affordability. Contractors received a perform-
ance requirement and price requirement for the average unit pro-
curement and could trade-off performance objectives to minimize
_____________
10 We derived the information in this section from the Wind Corrected Munitions Dispenser, Test and
Evaluation Master Plan, Milestone III Revision, January 2001; the Wind Corrected Munitions
Dispenser (WCMD) Engineering and Manufacturing Development (EMD) Cost Estimate, January
2001; and discussions with the Area Attack Systems Program Office (AAC/YH).
Missile Program Descriptions 177
costs. The resulting average unit procurement cost for 40,000 units
was approximately $10,000 in FY 1994 dollars.
The development program had two phases. EMD Phase 1 con-
tracts for development with an option for pilot production were
awarded in January 1995 to Alliant Techsystems and Lockheed Mar-
tin. The contractors built production-representative tail kits, and the
Air Force held a fly-off competition. In January 1997, the Air Force
selected Lockheed to continue into Phase 2, pilot production. In fall
1997, flight testing revealed an actuator problem during high-speed
releases. The program was restructured to extend EMD with a third
LRIP lot to allow time to design and qualify a fin-locking mecha-
nism. In April 2001, WCMD was approved for full-rate production.
Test Program
Table B.8 summarizes the test program, which we describe in more
detail below.
Table B.8
WCMD Testing
Number
Guided of
Event Launches Flights Remarks
Contractor pre–fly-off tests 11 10 Light testing verified
January 1995– predictions of contractors’
October 1996 six-degrees-of-freedom
models
Fly-off 14 8
November 1996–
December 1996
DT/OT 61 45 Approximately 60 captive-
carry flights
(approximately 120–180
flight hours)
Phase 1
May 1998–
October 1998
178 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Table B.8—continued
Number
Guided of
Event Launches Flights Remarks
Phase 2
October 1999–
December 2000
IOT&E 21 6
January–
October 2000
_____________
11 The program office estimates that simulation eliminated at least 12 flights and drops.
APPENDIX C
179
180 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons
Excludes:
• all formal and informal testing up through the subsystem
level which can be associated with the hardware/software
element
• acceptance testing
H.3.3.3 Mock-ups
The design engineering and production of system or subsystem
mock-ups which have special contractual or engineering signifi-
cance, or which are not required solely for the conduct of one of
the above elements of testing.
System Test and Evaluation Work Breakdown Structure 183
Bell, Don G., Naval Weapons Center Test and Evaluation Model (for Air-
Launched Weapons), China Lake, Calif.: Naval Weapons Center,
October 1988.
The Boeing Company, “Boeing 777 Facts.” Online at http://www.boeing.
com/commercial/777family/pf/pf_facts.html (as of January 28, 2004).
Cook, Cynthia, and John C. Graser, Military Airframe Acquisition Costs:
The Effects of Lean Manufacturing, Santa Monica, Calif.: RAND
Corporation, MR-1325-AF, 2001.
Commander Operational Test and Evaluation Force, Operational Test
Director’s Guide, COMOPTEVFORINST 3960.1H, December 13,
1995.
Defense Science Board, Report of the Defense Science Board Task Force on
Test and Evaluation, Washington, D.C.: Office of the Under Secretary of
Defense for Acquisition and Technology, September 1999.
Defense Systems Management College, Test and Evaluation Management
Guide, 4th edition, Fort Belvoir, Va.: The Defense Acquisition Univer-
sity Press, November 2001.
Director, Operational Test and Evaluation, Annual Report to Congress for
Fiscal Year 2000.
DoD—see U.S. Department of Defense.
F-22 System Program Office, F-22 Test & Evaluation Master Plan, Wright-
Patterson AFB, Ohio, June 1999.
185
186 Test and Evaluation Trends and Costs for Aircraft and Guided Weapons