2015 - Melese, Richter & Solomon (Military Cost-Benefit Analysis)
2015 - Melese, Richter & Solomon (Military Cost-Benefit Analysis)
Volume 1 Volume 7
European Armaments From Defense to Development?
Collaboration International perspectives on
Policy, problems and prospects realizing the peace dividend
R. Matthews A. Markusen, S. DiGiovanna and
M. Leary
Volume 2
Military Production and Volume 8
Innovation in Spain Arms Trade and Economic
J. Molas-Gallart Development
Theory, policy, and cases in arms
Volume 3
trade offsets
Defence Science & Technology
Edited by Jürgen Brauer and
Adjusting to change
J. Paul Dunne
R. Coopey, M. Uttley and
G. Spiniardi Volume 9
Volume 4 Exploding the Myth?
The Economics of Offsets The peace dividend, regions and
Defence procurement and market adjustment
countertrade Derek Braddon
S. Martin Volume 10
Volume 5 The Economic Analysis of
The Arms Trade, Security and Terrorism
Conflict Tilman Brück
Edited by P. Levine and R. Smith
Volume 11
Volume 6 Defence Procurement and
Economic Theories of Peace Industry Policy
and War Edited by Stefan Markowski and
F. Coulomb Peter Hall
Volume 12 Volume 14
The Economics of Defence Policy Military Cost–benefit
A new perspective Analysis
Keith Hartley Theory and practice
Edited by Francois Melese,
Volume 13
Anke Richter and
The Economics of UN
Binyam Solomon
Peacekeeping
Nadège Sheehan
Celebrating 50 Years
Excellence in Education since 1965
This page intentionally left blank
Military Cost–benefit
Analysis
Theory and practice
Typeset in Times
by Sunrise Setting Ltd, Paignton, UK
Contents
PART I
Introduction and problem formulation 1
PART II
Measuring costs and future funding 111
PART III
Measuring effectiveness 195
8 Multiple objective decision-making 197
KENT D. WALL AND CAMERON A. MACKENZIE
PART IV
New approaches to military cost–benefit analysis 261
10 The role of cost-effectiveness analysis in allocating
defense resources 263
KENT D. WALL, CHARLES J. LACIVITA, AND ANKE RICHTER
PART V
Selected applications 333
13 Embedding affordability assessments in military
cost–benefit analysis: defense modernization
in Bulgaria 335
VENELIN GEORGIEV
Index 432
Figures
The editors are grateful to the series editors for valuable suggestions, and for
reviewers’ comments provided by Taylor & Francis that offered constructive guid-
ance and support for this project. We would also like to thank the publisher for
excellent editorial direction. Dr Melese is grateful to his wife, Heather, for her
patient role as a sounding board, and especially to RADM James Greene, US Navy
(retired) and his staff at the Acquisition Research Program at the Naval Postgrad-
uate School for their early support of this book. Our recently departed defense
economist colleague, Michael Intriligator, was an inspiration to us all. Finally,
we are indebted to the authors for their outstanding contributions, as well as to
current and past faculty and participants at the Defense Resources Management
Institute (DRMI) and the broader defense economics community from whom we
have learned so much over the years. The views expressed are those of the edi-
tors and authors and do not necessarily represent those of the US Department of
Defense or the Canadian Department of National Defense.
Part I
1.1 Background
Military cost–benefit analysis (CBA) offers a vital tool to help guide governments
through both stable and turbulent times. As countries struggle with the dual chal-
lenges of an uncertain defense environment and cloudy fiscal prospects, CBA
offers a unique opportunity to transform defense forces into more efficient and
effective twenty-first-century organizations.
Defense reforms typically involve politically charged debates over invest-
ments (in projects, programs, or policies) as well as contentious divestment
decisions—from base realignment and closure (BRAC) to outsourcing and asset
sales. A powerful contribution of CBA is to inform such complex and con-
tentious decisions—carefully structuring the problem and capturing relevant costs
and benefits of alternative courses of action. Lifting the veil on military CBA,
this edited volume reveals several systematic quantitative approaches to assess
defense investments (or divestments), combined with a selection of real-world
applications.
The frameworks and methods discussed in the following chapters should
appeal to anyone interested or actively involved in understanding and applying
CBA to improve national security. These valuable approaches also have broader
government-wide applications, especially in cases where it is difficult to monetize
the benefits of a public project, program, or policy.
Unprecedented government spending to counter the global financial crisis has
placed enormous pressure on public budgets. Combined with alarming demo-
graphics, many countries struggle to fulfill past promises to underwrite health care
expenditures, social security payments, government pensions, and unemployment
programs. As debt burdens grow to finance current operations, the risk of esca-
lating interest payments threatens to crowd out vital future public spending. As
the single largest discretionary item in many national budgets,1 military expen-
ditures make a tempting target. Especially vulnerable are military and civilian
compensation (pay and benefits) and the purchase and operation of equipment,
facilities, services, and supplies.
4 F. Melese, A. Richter, B. Solomon
Anticipating future spending cuts, this book explores both conventional and
unconventional approaches to contemporary defense decisions—from critical
investments in facilities, equipment, and materiel to careful vendor selection to
build, operate, and maintain those investments. Recognizing the value of sys-
tematic quantitative analysis, senior US Army leadership has “directed that any
decisions involving Army resources be supported by a CBA.”2
Faced with severe budget cuts and an uncertain threat environment, defense
officials around the world confront urgent decisions on whether to approve
specific projects (e.g. infrastructure—military housing; training, and mainte-
nance facilities) or programs (e.g. weapon systems—unmanned aerial vehicles
(UAVs), armored personnel carriers (APCs), cyber defense). Military CBA offers
a valuable set of analytical tools to increase the transparency, efficiency, and
effectiveness of critical defense decisions.
A synthesis of economics, management science, statistics, and decision theory,
military CBA is currently used in a wide range of defense applications in coun-
tries around the world: i) to shape national security strategy, ii) to set acquisition
policy, and iii) to inform critical investments in people, equipment, infrastructure,
services, and supplies. This edited volume offers a selection of carefully designed
CBA approaches, and real-world applications, intended to help public officials
identify affordable defense capabilities that effectively counter security risks in
fiscally constrained environments.
1.3 Outline
This edited volume reveals how military CBA can reduce budget pressures and
improve defense decisions that contribute to national security. The dual purpose
of CBA is to encourage more efficient and effective allocation of society’s scarce
resources to increase social welfare.23 Governments often employ CBA to rank
8 F. Melese, A. Richter, B. Solomon
(mutually exclusive) portfolios of projects or programs. The typical CBA involves
at least eight steps:
These eight basic steps of a CBA are explored throughout the chapters of this
edited volume. The book consists of seventeen chapters divided into five parts.
1.4 Conclusion
Tight budgets make for hard choices. The greater the pressure on public budgets
the greater the opportunity to apply military CBA. Today, the impact of govern-
ment deficits and debt on military spending is inescapable.35 As one of the largest
discretionary items in government budgets, military spending is an obvious tar-
get for cuts. While wise use of military power can underpin economic growth,
it is equally clear that economic strength underpins military power. Shrinking
budgets place a renewed premium on affordability. As sovereign debt challenges
squeeze national budgets, and emerging threats challenge existing security forces,
this edited volume offers a valuable set of tools and techniques to help navigate
the political landscape and meet calls to increase the transparency, efficiency, and
effectiveness of the defense sector.
Notes
1 According to the World Bank average, military expenditures globally account for 9.2
percent of government spending. http://data.worldbank.org/indicator/MS.MIL.XPND.
ZS/countries/AU-C5-C7?display=graph (accessed February 7, 2014).
2 Office of the Deputy Assistant Secretary of the Army, U.S. Army Cost–Benefit Anal-
ysis Guide, January 12, 2010, p.6. In general, the US Federal Government’s Office
of Management and Budget (OMB) Circular A-94 provides guidance for the applica-
tion of CBA across the entire Executive Branch. DoD Instruction 7041.3 “Economic
Analysis for Decision Making” provides explicit guidance for the Department of
Defense.
3 Notable among these is the concept of “consumer surplus.”
4 Note that this criterion does not require transfers to actually occur and is occasionally
debated on equity grounds (see Footnote 10).
5 Assuming policies, projects or programs are independent and there are no binding
constraints on inputs.
6 In theory, selecting projects with positive net benefits maximizes aggregate wealth
(e.g. GDP growth) which indirectly helps those that might be made worse off. More-
over, an implicit assumption is that costs imposed on some and benefits accrued to
others will tend to average out across individuals. Where interactions occur among
projects, the general rule is to choose the combination of projects that maximizes net
benefits.
7 For example, benefits of a decision to allocate scarce financial resources among major
military missions.
8 For example, see Brooks (2005) or Gartzke (2007).
9 Chapter 3 in this edited volume offers a notable exception.
10 Examples of proximate criteria or partial measures of effectiveness include speed,
operating range, weapons accuracy, and armor protection.
11 For example, see OMB Circular A-94 Guidelines and Discount Rates for Benefit-Cost
Analysis of Federal Programs published by the US Office of Management and Budget.
Note that this edited volume will continue to use the generic term “military CBA” to
refer to cases where benefits cannot be monetized. Although CEA also produces a rank-
ing, there is no explicit information about whether the highest ranked alternative would
14 F. Melese, A. Richter, B. Solomon
provide positive net social benefits. If all alternatives are mutually exclusive, and the
status quo is among the alternatives, sharing similar scale and phasing of costs and
benefits, then it is possible to apply CEA to select the most efficient policy.
12 Given the vast existing literature on building MOEs, this book instead focuses on the
careful construction of military CBAs. Although we occasionally explore the ques-
tion of developing non-monetary benefit measures, we encourage the reader to review
the extensive literature on multi-criteria decision-making for alternative approaches to
deriving such measures of effectiveness. (For example, see Keeney and Raiffa 1976;
Buede 1986; or Kirkwood 1997.)
13 “The ultimate objective of PPBS shall be to provide operational commanders-in-chief
the best mix of forces, equipment, and support attainable within fiscal constraints”
(DoD Directive 7045, 14 May 22, 1984). The basic questions of systems analysis
are twofold: i) given a fixed budget, which weapon systems are most cost-effective
and, conversely, ii) given a fixed military mission, which system(s) could generate
the desired level of effectiveness at the lowest cost? The basic ideas of PPBS were:
“the attempt to put defense program issues into a broader context and to search for
explicit measures of national need and adequacy;” “consideration of military needs
and costs together;” “explicit consideration of alternatives at the top decision level;”
“the active use of an analytical staff at the top policymaking levels;” “a plan combin-
ing both forces and costs which projected into the future the foreseeable implications
of current decisions;” and “open and explicit analysis . . . made available to all inter-
ested parties, so that they can examine the calculations, data, and assumptions and
retrace the steps leading to the conclusions” (Enthoven and Smith 2005). http://www.
rand.org/content/dam/rand/pubs/commercial_books/2010/RAND_CB403.pdf.
14 In practice, measuring contributions of various inputs towards a defense goal can be
difficult and contentious, and these desirable results only hold under the assumption
(“homotheticity”) that optimal input ratios are independent of the budget and depend
only on relative costs of each input. It is also important to recognize transaction costs
associated with the application of military CBA (or systems analysis). For example, cen-
tralization of decision-making authority under Secretary of Defense Robert McNamara
resulted in a proliferation of management systems to collect data required to evaluate the
costs and benefits of alternative projects and programs (weapon systems). Increasingly
buried in paperwork, the term “paralysis of analysis” was coined by some members of
the defense establishment (personal conversation with A. Enthoven).
15 In 1972, the Systems Analysis division evolved into the Office of Program Analysis and
Evaluation (PA&E) and later, in 2009, into the Office of Cost Assessment and Program
Evaluation (CAPE). Prior to his departure, Dr Hitch launched an OSD-sponsored educa-
tion institution to teach civilian and military managers in DoD (and partners and allies)
basic principles of PPBS and CBA. Today it is known as the Defense Resources Man-
agement Institute (DRMI) located at the Naval Postgraduate School in Monterey, Cali-
fornia. Two co-editors of this volume (Dr Melese and Dr Richter) are faculty members at
this institution which celebrated its 50th anniversary in 2015 (http://www.nps.edu/drmi/;
http://www.dtic.mil/whs/directives/corres/pdf/501035p.pdf).
16 McNamara relied heavily on systems analysis to reach several controversial weapon
decisions. He canceled the B-70 bomber, begun during the Eisenhower years as a
replacement for the B-52, stating that it was neither cost-effective nor needed, and later
he vetoed its proposed successor, the RS-70. McNamara expressed publicly his belief
that the manned bomber as a strategic weapon had no long-run future; the interconti-
nental ballistic missile was faster, less vulnerable, and less costly. Similarly, McNamara
terminated the Skybolt project late in 1962. Begun in 1959, Skybolt was conceived as
a ballistic missile with a 1,000-nautical mile range, designed for launching from B-52
bombers as a defense suppression weapon to clear the way for bombers to penetrate to
targets. McNamara decided that Skybolt was too expensive, was not accurate enough,
Introduction: theory and practice 15
and would exceed its planned development time. He claimed other systems, including
the Hound Dog missile, could do the job at less cost.
17 Our view is that a clean military CBA is a valuable starting point for political debate.
Careful analysis can constrain political attempts to turn defense spending into a jobs
program or an opportunity to redistribute income. Since there exist considerably more
efficient and effective ways to promote job growth and income distribution, if these are
the goals, then they should be stated explicitly and explored using a separate CBA. This
may prove a valuable avenue for future research.
18 An example is the case of the F-35 aircraft program. Lockheed-Martin claimed to
have “created 125,000 US-based direct and indirect jobs in 46 states” http://www.
businessweek.com/news/2014-01-22/lockheed-martin-inflating-f-35-job-growth-claims-
nonprofit-says (accessed March 21, 2014).
19 Since costs (e.g. investment expenditures) tend to occur earlier with benefits appearing
later, discount rates can also be strategically selected to make projects appear more or
less attractive. (See Chapter 17.)
20 For example, see NATO’s Building Integrity initiative at www.nato.int/cps/en/natolive/
topics_68368.htm [last accessed June 24, 2014].
21 For example, see Camm and Greenfield (2005).
22 As public officials face growing resistance to tax increases, pressure increases on
governments to work more efficiently and effectively.
23 National defense satisfies two key characteristics of a “public good.” It is: i) non-rival
and ii) non-excludable. In the former case, unlike private goods, if one person in a
geographic area is defended from foreign attack or invasion, his or her consumption is
non-rival in that others in the area can consume the same level of national security for
little or no additional cost. In the latter case, if one person is defended, others in that
same area cannot be excluded from the security benefits. This leads to a classic free-
rider problem, making it difficult to charge people for national defense. The key here
is that whereas it is generally agreed that the provision of national defense is a public
good that must be funded with taxes, the production of national defense depends on
the relative costs (including transaction costs) and benefits of public and private sector
production, which can be evaluated using military CBA.
24 Since benefits of proposed defense investments are often difficult to monetize, various
approaches have been developed to construct MOEs that capture the value or utility of
alternatives. In theory, the benefits of alternative defense investments could be mone-
tized if their contribution to security and stability encourages foreign direct investment
that contributes to economic growth and social welfare. In practice, precise linkages
between defense investments and economic growth are difficult to establish. As a conse-
quence, the benefits of most military investments are not monetized, and instead various
MOEs are constructed to conduct a CEA that is referred to in this volume as a military
CBA (e.g. see OMB Circular A-94).
25 If benefits can be monetized, then calculate the discounted sum of net benefits (benefits
minus costs) from each alternative over the specified time period, i.e. the discounted
NPV. For example, consider two alternative military projects designed to achieve the
goal of reducing the Navy’s fuel budget. Suppose there is a fixed investment budget
available, and the two mutually exclusive alternatives each require the same identically
phased investment. The first proposal is to invest in a program to convert ship propul-
sion from conventional diesel to a less expensive bio-diesel. The second proposal is to
invest in an energy conservation program at Navy installations. Since the two alterna-
tives require the same identically phased investments, the CBA can simply focus on
the stream of benefits (savings from cheaper fuel in the first case and reduction in
fuel demand in the second) that accrue from each project. Assuming a preference for
present vs. future savings, the discounted present value of each stream of savings can be
calculated to determine the winning project. (See Chapter 17.)
16 F. Melese, A. Richter, B. Solomon
26 Alternatively, a “Real Options” approach could be adopted, conducting Monte Carlo
simulations assigning probability distributions to key parameters. (For example, See
Chapters 11 and 14.)
27 The authors also warn that until new mechanisms such as CBA are adopted to address
the continual failure to fuse requirements with necessary resources, DoD will essentially
continue to create a disjointed, ineffective framework for addressing national security
concerns rather than the vital cohesive plan required to confront the changing dynamics
of modern, global warfare.
28 See Ullman and Ast (2011) and OMB Circular A-11 (2008) for discussions of AoAs.
29 In the US, AoAs are conducted in early phases (milestones) of major defense acquisi-
tions. Since they frequently occur in early development before a project is fully funded,
they rarely incorporate future funding forecasts. Instead, the budget estimate for the
program or project is generated as an output of the AoA. As major budget cuts cre-
ate funding challenges for new defense programs, “affordability” in terms of realistic
budget constraints is gaining increasing importance in AoAs.
30 A key difference between traditional AoAs and EEoA is that instead of modeling
competing vendors as points in cost-effectiveness space, EEoA solicits vendor offers
as functions of optimistic, pessimistic, and most likely funding (budget) scenarios. A
formal mathematical model of EEoA can be found in Simon and Melese (2011).
31 Following the recommended EEoA approach also provides a unique opportunity to
achieve a significant defense reform: to coordinate the Requirements Generation Sys-
tem, Defense Acquisition System, and Planning, Programming, and Budgeting System
(PPBS)—to lower costs, and improve performance and schedules.
32 For example, see Williamson and Masten (1999) or Melese et al. (2007).
33 Finding the optimal mix of forces to accomplish a military goal at the lowest possible
cost, or alternatively, for a given budget, to find the optimal mix that maximizes defense
capabilities.
34 The real options approach builds on what was previously referred to as incremental or
“spiral” development of military programs and projects.
35 The response of many governments to the global recession was to bail out banks and
businesses and to stimulate their economies. Combined with falling tax receipts, this
led to unprecedented increases in government spending. The result in many countries
transformed the financial crisis into a chronic sovereign debt crisis. Annual deficits
soared and cumulative debt loads reached unsustainable levels. Aging demographics in
some countries compounded the problem, placing impossible demands on social welfare
programs and introduced further pressure on government budgets. Combined with an
uncertain threat environment, the fiscal crisis makes a compelling case for widespread
application of military CBAs to ensure future defense decisions to produce efficient,
effective, and accountable security forces.
References
Buede, D. M. 1986. “Structuring Value Attributes.” Interfaces 16(2): 52–62.
Brooks, S. G. 2005. Producing Security: Multinatinal Corporations, Globalization, and the
Changing Calculus of Conflict. Princeton, NJ: Princeton University Press.
Camm, Frank, and Victoria A. Greenfield. 2005. How Should the Army Use Contractors on
the Battlefield: Assessing Comparative Risk in Sourcing Decisions. Santa Monica, CA:
RAND Corporation.
Dupuit, J. 1844. De La mesure de l’utilité des travaux publics. Annales des Ponts et
Chaussées, Memories et Documents 2, 8: 332–375. Translated by R. H. Barback into
English, “On the Measurement of the Utility of Public Works.” 1952. International
Economic Papers 2: 83–110.
Introduction: theory and practice 17
Enthoven, A. C., and K. W. Smith. 2005. How Much is Enough? Shaping the Defense
Program, 1961–1969. Santa Monica, CA: RAND Corporation.
Fitzsimmons, M. 2007. “Whither Capabilities-Based Planning?” Joint Forces Quarterly
44(1): 101–105.
Gartzke, E. 2007. “The Capitalist Peace.” American Journal of Political Science 51(1):
166–191.
Haveman, R. 1976. “Policy Analysis and the Congress: An Economist’s View.” Policy
Analysis 2(2): 235–250.
Hicks, J. R. 1940. “The Valuation of the Social Income.” Economica 7(26): 105–124.
Hitch, C., and R. McKean. 1960. The Economics of Defense in the Nuclear Age. London:
Oxford University Press.
Kaldor, N. 1939. “Welfare Propositions of Economics and Interpersonal Comparisons of
Utility.” Economic Journal 49(195): 549–552.
Keeney, R. L., and H. Raiffa. 1976. Decisions with Multiple Objectives: Preferences and
Value Tradeoffs (Probability and Mathematical Statistics Series). New York: John Wiley
& Sons, Inc.
Kirkwood, C. W. 1997. Strategic Decision Making: Multiobjective Decision Analysis with
Spreadsheets. Belmont, CA: Duxbury Press.
Marshall, A. 1920. Principles of Economics (8th edition). London: MacMillan & Co. Ltd.
Melese, F., R. Franck, D. Angelis, and J. Dillard. 2007. “Applying Insights from Transac-
tion Cost Economics to Improve Cost Estimates for Public Sector Purchases: The Case
of U.S. Military Acquisition.” International Public Management Journal 10(4): 357–85.
Pareto, V. 1909. Manuel d’Economie Politique. Paris: V. Giard et E. Briere. Reprinted,
Paris: M. Girard, 1927; Geneva: Librarie Droz, 1966.
Prest, A. R., and R. Turvey. 1965. “Cost–Benefit Analysis: A Survey.” The Economic
Journal 75(300): 683–735.
Simon, J., and F. Melese. 2011. “A Multiattribute Sealed-Bid Procurement Auction with
Multiple Budgets for Government Vendor Selection.” Decision Analysis 8(3): 170–79.
Ullman, D. G., and R. Ast. September 2011. “Analysis of Alternatives (AoA) Based
Decisions.” Military Operations Research Society, Phalanx 44(3): 24–36.
US Office of Management and Budget (OMB). 1992. Circular No. A-94, Guidelines and
Discount Rates for Benefit-Cost Analysis of Federal Programs. Washington, DC: Execu-
tive Office of the President. Available at www.whitehouse.gov/omb/circulars_a094/ [last
accessed September 18, 2014].
US Office of Management and Budget (OMB). 2008. Circular No. A-11, Preparation,
Submission, and Execution of the Budget. Washington, DC: Executive Office of the Pres-
ident. Available at www.whitehouse.gov/omb/circulars_a11_current_year_a11_toc [last
accessed November 11, 2014].
Williamson, O. E., and S. E. Masten (Eds). 1999. The Economics of Transaction Costs
(Elgar Critical Writings Reader). Northampton, MA: Edward Elgar Publishing.
2 Allocating national security
resources
Jacques S. Gansler and
William Lucyshyn
Most innovations fail. And companies that don’t innovate die . . . In today’s world,
where the only constant is change, the task of managing innovation is vital for
companies of every size in every industry. Innovation is vital to sustain and advance
companies’ current business; it is critical to growing new business. It is also a very
difficult process to manage.
Henry W. Chesbrough 2006, Haas School of Business
Berkeley California
2.1 Introduction
Private sector firms have accepted that they must continuously innovate to stay
competitive in today’s global environment. This innovation is not a one-time
event, but a continuous exercise. Similarly, in a world of emerging and evolv-
ing challenges and threats, ministries of defense must be willing to continuously
innovate. The constancy of change cannot be met with technology, training, or
operational concepts that are outdated and inadequate. Equally important, as pri-
vate sector firms have discovered, are innovations improving business processes.
Within the US Department of Defense (DoD), one business process needs imme-
diate attention—the one used to allocate national security resources. Military
cost–benefit analysis (CBA) has an important role to play. To fully appreciate this
imperative, one must understand and appreciate the environment that DoD finds
itself in today. The willingness of DoD to adopt innovative CBA concepts to over-
haul its resource allocation processes will directly affect the department’s capacity
to meet future threats to US and global security.
2.2 Background
During the cold war, US defense planning was, to a large degree, threat-based.
The monolithic threat posed by the Soviet Union could sensibly, if not always
accurately, be described in terms of troops, fighter aircraft, tanks, and missiles.
The threat was remarkably stable and evolved slowly, simplifying the Pentagon’s
planning challenge—as well as the cost–benefit analysis of investments to cope
with that threat.
Allocating national security resources 19
With the dissolution of the Soviet Union, the relentless influence of the cold
war upon the nation and its national security strategy vanished. Some, believ-
ing that mankind had reached the pinnacle of ideological evolution, wrote of the
end of history with the United States as the world’s lone super-power (Fukuyama
1989).
By the beginning of the 1990s, the United States found itself in the midst of a
very different war—Desert Storm. This war showcased a new manner of warfare,
involving precision strike capabilities used in conjunction with real-time informa-
tion exchange—an early version of net-centric warfare. There was still much to be
learned, however, and many additional advancements to be made. The Gulf War
offered the first glimpse of technological capabilities, and these will continue to
reshape warfare throughout the twenty-first century.
Immediately following Operation Desert Storm, powerful forces began to influ-
ence policy-makers as they anticipated a new era of change. These included the
acceleration of the information technology revolution, globalization, the poten-
tial threat of terrorism, and perhaps most importantly the desire to reap a “peace
dividend.” With no clear rival in sight, the United States began a period of
disinvestment, and it expedited efforts to downsize its military forces (Walker
2014).
Then, on September 11, 2001, the attacks on the World Trade Center and the
Pentagon highlighted a new, global threat from terrorist organizations. The events
of that day created an unprecedented urgency for innovative changes in the US
defense establishment, and they laid the foundation for adjustment to the new
global security environment. As the threats now addressed by the DoD were vastly
different from those it faced in the previous five decades, this adjustment proved
especially vital (Freier 2007).
While numerous earlier reports had highlighted the potential for transformation
of the security environment (Cha 2000), the events of 9/11 provided the catalyst
for reform. The formerly relatively stable Soviet threat was replaced by profoundly
uncertain, fragmented, and complex threats that proved far more difficult to satis-
factorily address. In addition to the threat from global terrorism, there were threats
from failed or failing states that resulted in civil wars, humanitarian catastrophes,
and regional instability. This tremendously unstable international security environ-
ment makes it extremely challenging to predict with any confidence what threats
the United States, its partners, and allies might face in the medium or even the
near-term.
Although military CBA can help DoD as it shifts its portfolio of capabilities to
meet these new challenges, the current force structure is still far from being able
to fully respond to this spectrum of new challenges. As a result, the Department is
faced with difficult investment decisions to bridge this gap, while simultaneously
responding to current operational challenges.
14
12
10
Percentage of GDP
0
19
19
19
19
19
19
20
20
20
20
20
70
75
80
85
90
95
00
05
10
15
20
Year
Dramatic Increase
$800,000
with OCO Funding
$700,000
Korean War Vietnam Reagan
War Buildup
$600,000
Pre-9/11
$500,000 Low
CY 2015 $ M
$400,000
$300,000
$200,000
$100,000
$0
20
19
19
19
19
19
20
20
20
20
20
19
19
19
19
19
19
19
19
19
19
19
19
19
17
69
72
75
78
81
84
87
90
93
96
99
02
05
08
11
14
48
51
54
57
60
63
66
quoted as saying. “With the spending declines following the Vietnam War and the
post-Reagan years, the decline in procurement was steeper than the rate of decline
in the overall defense budget” (Boessenkool 2009). In the present environment,
it is likely that increasing O&M costs for legacy systems, along with projected
rising personnel expenses, will be financed at the expense of future R&D and
procurement within DoD.
80%
70%
70%
64%
66%
Percentage of DoD budget
60%
56%
50%
49%
40%
30%
23% 22%
19%
20% 17% 16%
Figure 2.3 Resources for Defense in selected years as a percentage of total DoD budget.
Source: CBO 2010.
Note: 2009 and 2010 include supplemental funding of US$74 billion and US$130 billion.
2.7 System-of-systems
With major advances in information technology, many of DoD’s weapon systems
are now interconnected to form interdependent systems-of-systems. A system-of-
systems (SoS) is a set of individual systems that is integrated to operate optimally
as a single system. SoS development provides the DoD with the unique ability
to integrate and field a full range of assets (new and legacy) to provide required
military capabilities. Although one can conceive of developing an entire SoS from
scratch, most of DoD’s SoS efforts consist of synthesizing existing capabilities
28 Hon. J. S. Gansler, W. Lucyshyn
of legacy systems, combined with selected newly developed systems to address
additional capability needs.
Since it must pull together numerous projects and combine them into a cohe-
sive whole, the size and scale of a SoS complicates any military CBA and the
development of these integrated programs. The challenge for DoD is that, although
the systems will be integrated in this fashion, they are still developed as individual
systems.
In the traditional engineering framework, the goal of a typical development
project is to optimize the performance of a particular system (e.g. platform). Since
there are multiple systems within the overall SoS, each element is individually
optimized (to derive maximum individual performance). In this framework, it is
believed that the whole is equal to the sum of its parts.
A problem with this narrow engineering approach is that it fails to take into
account how changes in one component of the interrelated whole may affect the
performance or requirements (or costs) of other components of the entire system.
If one system relies upon a second system having a minimum threshold capabil-
ity that is not realized, the first may have to be modified to achieve the desired
effectiveness. Conversely, optimizing the performance of an individual platform
may not necessarily increase the performance of the SoS and it may, in some
cases, degrade its overall performance. Due to the size and interconnected nature
of SoS, one problem may have ripple effects throughout. Optimizing each indi-
vidual system can produce a sub-optimal result at the SoS level. The goal must be
to optimize the performance of the SoS as a whole and not individual component
systems.
Currently, a SoS program generally spans multiple organizations. As a result, a
number of authorities may be in a position to influence the decision-making pro-
cess. Managing the component systems should be done using a portfolio approach
with some flexibility to shift resources within the portfolio to meet the SoS
requirement. In this way, there would be a level of flexibility not available with
the “platform-centric” approach. Ideally, these SoSs will be able to leverage the
unique advantages of each component system and result in a capability greater
than the sum of its parts.
In the most likely future of limited resources, however, there is another problem
with optimizing individual systems versus the overall SoS: namely “affordabil-
ity” (see Chapter 4). SoS is a potentially valuable innovation that creates new
challenges for military CBAs. In a resource-constrained environment, the focus
of military CBA should be to optimize overall performance at an affordable price
(often integrating lower-cost individual elements).
2.8 Conclusion
Challenges facing US National Security begin with an increasingly complex and
uncertain environment of global threats and also include domestic budgetary con-
straints that will significantly limit the resources available to address those threats.
Coupled with top-line budgetary pressures, DoD must also deal with the chal-
lenge of the year-to-year line-item budgeting process and ongoing acquisition
challenges, all of which threaten to severely limit DoD’s capacity to support
current acquisitions. All of this must be done while balancing the need for new sys-
tems to address evolving missions, such as “irregular” operations, missile defense,
and cyber-defense. And, of course, sound acquisition decisions require accurate
forecasts of vendor costs and future defense budgets. This “witches’ brew” of
conditions calls for the wider implementation of military CBA-guided resource
allocation decisions throughout DoD. The goal is for the nation to receive the best
value for every dollar it invests in defense programs.
Until new mechanisms are adopted that explicitly incorporate military CBA to
address failures to fuse requirements with necessary resources, DoD will risk a
disjointed, ineffective approach that may not properly address national security
concerns. A comprehensive military CBA approach such as that proposed in this
volume is vital in confronting the changing dynamics of modern, global warfare.
Even if this approach is adopted, however, DoD must still engage in a contin-
uing process of reassessment and innovation in order to retain the strong bond
between resources and requirements these innovations may bring. This ability
Allocating national security resources 33
to constantly innovate will be central to remaining effective in the face of an
increasingly complex and uncertain national security environment.
Notes
1 For example, in Somalia 1992, Iraq 2003 and Libya 2011.
2 In recent updates of Social Security financial projections, the US Congressional Budget
Office (CBO) notes that the Social Security deficit will grow to be significantly larger by
2020 than previously expected (CBO 2010).
3 Social Security, Medicare, and Medicaid have reached 10.1 percent of GDP (2009) and
are projected to continue increasing as a percentage of GDP to 11.8 percent in 2020,
15.5 percent in 2040, and 21.1 percent in 2075 (CBO 2010). According to GAO, these
programs will near 30 percent of GDP by 2040.
4 Both CBO and GAO have conducted long-term studies surrounding the federal bud-
get. Both have noted the dramatic increases in entitlement spending and interest on
the national debt. GAO projections have been carried out until 2084 based upon CBO
assumptions and projections ending in 2020.
5 This current projection was diverted from the baseline after an assumption that the US
will draw down troops in Iraq, Afghanistan, and elsewhere to 60,000 by 2013. Other
projections based upon different troop levels or different spending freezes, however, place
DoD’s budget even lower as a percentage of GDP.
6 Compounding future funding challenges is the fact DoD has come to rely on “supplemen-
tal” funding which is being significantly reduced. The Administration requested US$33.0
billion in 2010 supplemental funding on top of the US$129.6 billion already provided and
a total of US$159.3 billion for its 2011 overseas contingency operations (OMB 2010).
7 The Cost Analysis Improvement Group (CAIG) launched by what is now known as
CAPE (Cost Assessment and Program Evaluation) was meant to be responsible for pro-
viding independent cost estimates for the US Secretary of Defense: see www.cape.osd.mil
[last accessed November 25, 2014].
8 Milestone B is normally the formal initiation of an acquisition program (see DoD Instruc-
tion 5000.02, November 25, 2013, http://www.acq.osd.mil/osbp/sbir/docs/500002_
interim.pdf [last accessed February 9, 2015]).
9 Chapter 10 offers an interesting theoretical framework that captures aspects of a more
general portfolio approach, while Chapter 4 explicitly addresses the issue of affordability.
References
Al-Salhy, S. and T. Arango. June 12, 2014. “Iraq Militants, Pushing South, Aim at Capital.”
New York Times: A1.
Arena, M. V., R. S. Leonard, S. E. Murray, and O. Younossi. 2006. Historical Cost Growth
of Completed Weapon System Programs (Technical Report Series). Santa Monica, CA:
RAND Corporation.
Boessenkool, A. November 11, 2009. “Analysts: U.S. Procurement Drop Coming as Costs
Mount.” Defense News.
Carden, M. J. April 8, 2010. “Army Command Sustains Logistics for Warfighters.” Fort
Hood Sentinel.
CBO. 2010. The Budget and Economic Outlook: Fiscal Years 2010 to 2020. Congressional
Budget Office.
———. 2010. Long Term Implications of the Fiscal Year 2010 Defense Budget. Congres-
sional Budget Office.
Cha, V. D. 2000. “Globalization and the Study of International Security.” Journal of Peace
Research 37(3): 391–403.
34 Hon. J. S. Gansler, W. Lucyshyn
Chesbrough, H. W. 2006. “Open Innovation: A New Paradigm for Understanding Industrial
Innovation.” In Innovation: Researching a New Paradigm (Chapter 1). Oxford: Oxford
University Press.
Chu, D. S. C., and N. Berstein. 2003. “Decisionmaking for Defense.” In New Challenges
New Tools for Defense Decisionmaking, edited by M. L. Stuart and G. F. Treverton.
Santa Monica, CA: RAND Corporation.
Frank, D. F. Summer 1997. “A Theoretical Consideration of Acquisition Reform.” Acqui-
sition Review Quarterly 4(3): 279.
Freier, N. 2007. Strategic Competition and Resistance in the 21st Century: Irregular,
Catastrophic, Traditional, and Hybrid Challenges in Context. Author via CreateSpace
Independent Publishing Platform (a DBA of On-Demand Publishing LLC, part of the
Amazon group of companies).
Fukuyama, F. Summer 1989. “The End of History.” The National Interest: 3–18.
Gates, R. M. May 3, 2010. Remarks as Delivered by Secretary of Defense Robert M. Gates
(cited May 5, 2010). Gaylord Convention Center, National Harbor, MD. Available at
www.defense.gov/speeches/speech.aspx?speechid=1460 [last accessed November 25,
2014].
Gordon, L. A. 2000. Mangerial Accounting: Concepts and Empirical Evidence (5th
edition). Columbus, OH: McGraw-Hill Higher Education.
Grossman, E. 2008. “Marine General Lays Groundwork for Unprecendented Change.”
Government Executive. May 30,2008. Available at www.govexec.com/defense/2008/
05/marine-general-lays-groundwork-for-unprecedented-change/26953/ [last accessed
November 25, 2014].
The Institute for Economics and Peace (IEP). 2012. Global Terrorism Index 2012: Captur-
ing the Impact of Terrorism for the Last Decade (IEP Report Series). New York: The
Institute for Economics and Peace.
Keehan, M. P. 2008. Planning, Programming, Budgeting, and Execution (PPBE) Proccess
(Teaching Note). Fort Belvoir, VA: Defense Acquisiton University.
Perl, Raphael. 2006. Terrorism Expert Calls for a Redefinition of National Security. US
Department of State, Office of International Information Programs, Washington, DC.
Perlo-Freeman, S. and C. Solmirano. April 2014. “Trends in World Military Expenditure,
2013.” Stockholm International Peace Institute Fact Sheet.
Popescu, I. C. March 1, 2010. “The Last QDR? What the Petagon Should Learn from
Corporations about Strategic Planning.” Armed Forces Journal: 26.
Panel on Defense Acquisition Reform. 2010. Interim Findings and Recommendations.
Washington, DC: House Armed Services Committee.
Schwirtz, M., A. Barnard, and C. J. Chivers. August 9, 2008. “Russia and Georgia Clash
Over Separatist Region.” New York Times.
Smith, R. 2005. The Utility of Force: The Art of War in the Modern World. London: Allen
Lane.
Sullivan, M. J. 2008. Defense Acquisitions: Fundamental Changes are Needed to Improve
Weapon Program Outcomes (GAO-08-1159T). Available at www.gao.gov/products/
GAO-08-1159T [last accessed November 25, 2014].
———. 2009. Many Analyses of Alternatives have not Provided a Robust Assesment of
Weapon System Options (GAO-09-665). Available at www.gao.gov/products/GAO-09-
665 [last accessed November 25, 2014].
United States Census Bureau. 2004. Available at www.census.gov [accessed November 15,
2007].
United States Government Accountability Office. 2010. The Federal Government’s Long-
Term Fiscal Outlook. GAO-10-468SP.
Allocating national security resources 35
US Office of Management and Budget (OMB). 2010. President’s Budget for FY 2011.
Washington, DC: Executive Office of the President.
Walker, D. July 15, 2014. “Trends in U.S. Military Spending.” Council on Foreign
Relations. Available at www.cfr.org/defense-budget/trends-us-military-spending/
p28855 [last accessed November 25, 2014].
Younossi, O., M. V. Arena, R. S. Leonard, C. R. Roll, Jr., A. Jain, and J. M. Sollinger. 2007.
Is Weapon System Cost Growth Increasing? A Quantitative Assessment of Completed
and Ongoing Programs. Santa Monica, CA: RAND Corporation.
3 Measuring defense output
An economics perspective
3.1 Introduction
Measuring defense output is a necessary step to successfully apply military cost–
benefit analysis (CBA) to evaluate alternative security investments. Other chapters
in this book focus on what economists call “intermediate outputs” (e.g. military
forces). This chapter offers a higher-level “macro” perspective of overall defense
output that encompasses total defense spending.
In most countries, the defense sector absorbs substantial scarce resources that
have many valuable alternative uses (schools, hospitals, etc.). Whereas defense
expenditures are well known within each country, there is no single indicator of
value (or benefit) of overall defense output. This contrasts with the valuation of
private sector outputs in market economies. In defense, the economist’s solution to
measuring output assumes output equals inputs (a convention widely used across
the public sector), or that the value of defense output is roughly equivalent to
expenditures made to produce that output.
In sharp contrast, measuring the value of market outputs is not usually regarded
as a policy problem. Market economies ‘solve’ the problem through market prices
that reflect choices of large numbers of buyers and sellers. Defense, however, dif-
fers in several key ways from private markets, which helps to explain the challenge
in measuring and valuing defense output.
An important step in applying military CBA to evaluate security investments is
to discuss output measures. This chapter examines the measurement of defense
output from an economics perspective. Specifically, the chapter identifies key
questions which need to be addressed in measuring overall defense output. These
include: What is defense output? How can it be valued? Under what conditions is
it a worthwhile investment?
Economic theory offers some policy guidelines for determining the optimal
defense output for any society. As an optimizing problem, the rule is to aim at the
socially desirable or optimal level of defense output. This is achieved by equating
additional or marginal costs of proposed defense expenditures with additional or
marginal benefits. While the economics approach is difficult to ‘operationalize’
into a set of clear, unambiguous policy guidelines, it does provide a framework
for designing valuations for defense outputs and activities.
Measuring defense output 37
This chapter is organized as follows. Section 3.2 reframes the defense output
measurement issue as an economic problem utilizing insights from public goods,
public choice, and principal-agent models. Section 3.3 presents the military pro-
duction function as an early attempt to quantify defense outputs, and reveals
some challenges in operationalizing the model. Section 3.4 discusses more recent
attempts at measuring defense outputs through the transformation of defense bud-
gets reporting. Section 3.5 analyzes both economic and non-economic benefits of
defense, while Section 3.6 surveys international experience in measuring defense
outputs. Nations surveyed include Australia, New Zealand, the United Kingdom,
the United States, and a group of European nations.
3.2.2 Markets
There are several major differences between private markets and public (defense)
markets. Private markets involve prices that reveal society’s valuation of outputs,
where these prices reflect market incentive and penalty mechanisms. Goods that
are ‘private’ rather than public are characterized by both excludability and rivalry;
large numbers of private consumers and buyers; rivalry between firms; motivation
and rewards through profits; and a capital market that imposes penalties on poor
economic performance through take-overs and the ultimate sanction, bankruptcy
(with managers often losing their jobs).
Public bureaucracies such as the armed forces lack such incentive and penalty
mechanisms, and they consequently tend to be slow to adjust to change. Often,
change in the armed forces results from budget pressures, new technology, vic-
tories and defeats, and occasionally, views of senior military leaders (Solomon
et al. 2008). In contrast to private markets, there is no market price for publicly
provided defense forces: for example, there are no market prices for submarine or
tank forces.
Although some rivalry exists between suppliers (Navy, Army, Air Force, etc.),
there is no profit motive for public suppliers, nor capital market pressures cor-
responding to takeovers and bankruptcy in private markets. Defense has another
distinctive feature reflected in the state-funding and state provision (ownership) of
its armed forces. Governments are monopsony buyers and monopoly providers of
armed forces.
This contrasts with private markets where there are large numbers of buyers
and rivalry amongst suppliers. State-owned and funded defense markets are less
likely to undertake worthwhile changes (Tisdell and Hartley 2008, Chapter 10).
There is also a unique military employment contract which differs drastically from
private sector employment contracts. The military employment contract requires
military personnel to obey commands which relate to type, duration, location, and
Measuring defense output 39
conditions of work (e.g. world-wide deployments) with significant probability of
injury and even death. Such a contract contains elements resembling indentureship
and command systems.
Each of the armed forces is a monopoly supplier of air, land or sea systems with
monopoly property rights in the air, land or sea domains. There are barriers to new
entry which prevent rival internal armed forces from offering competing products.
For example, armies often operate attack helicopters and unmanned aerial vehi-
cles (UAVs) which are rivals to close air support and surveillance provided by
air forces. Similarly, land-based aircraft operated by air forces are alternatives to
naval carrier-borne aircraft. Efficiency requires a mechanism for promoting such
competition; instead, each service guards its traditional monopoly property rights
in the air, land, or sea domain, thereby creating barriers to new entry.
This has an impact on efficiency. Specifically, is the correct amount of output
being produced? Is the correct mix of inputs being used? As monopolies with
significant barriers to entry, each of the armed forces lacks strongly competing
organizations, and hence has less incentive for efficiency improvements and for
innovation (where efficiency embraces both allocative and technical efficiency).
Allocative efficiency requires the choice of socially desirable output, and tech-
nical efficiency requires the use of least-cost methods to produce that output.
Again, problems arise in determining allocative efficiency (see a discussion below
on principal-agent models). Technical efficiency, however, can be assessed by
allowing activities traditionally undertaken ‘in-house’ by the armed forces to be
‘opened-up’ to competition from private suppliers (market testing leading to mili-
tary outsourcing). Indeed, military cost–benefit formulation of such competitions
can offer improvements in allocative efficiency (e.g. by inviting competition for
different levels of service in order to identify true marginal costs for different
levels of output or service).2
Internal defense markets lack other incentives of private markets. There are no
profit incentives to stimulate and reward military commanders to search for and
introduce productivity improvements or to identify new and profitable opportuni-
ties (for example, the role of entrepreneurs in private markets). The absence of a
capital market also means that military managers are unlikely to lose their jobs
for poor performance and there are no capital market opportunities for promot-
ing and rewarding mergers and take-overs. For example, a military commander
of a regiment cannot merge with another regiment to achieve economies of scale
and scope, nor can an Army regiment acquire Air Force or Navy transport units
where such mergers might offer both cost savings and output improvements (such
as horizontal, vertical, or conglomerate mergers).
Uncertainty dominates defense policy. Defense policy has to respond to a range
of future threats, some of which are unknown and unknowable. Assumptions are
needed about likely future allies and their responses to threats, the location of
threats, new technologies, and the time dimension of threats (e.g. today, in 10–15
years, or 30–50 years ahead where uncertainties are greatest). These uncertain-
ties mean that forces have to be capable of adapting to change, and that today’s
defense investments must be capable of meeting tomorrow’s threats. Admittedly,
40 K. Hartley, B. Solomon
the private sector also faces considerable uncertainty about future markets and new
technologies, and these unknowns extend over lengthy time horizons. Defense is
different, however, in that uncertainties are dependent upon, and determined by,
governments, nation states, and some non-state actors, rather than by the actions
of large numbers of private individuals as consumers, workers, and shareholders.
There is one further key difference between defense and private markets. Defense
aims to avoid conflict, but where conflict arises it often destroys markets and valuable
infrastructure and creates disequilibrium as resources are re-allocated to military
forces to gain strategic advantage, with consequent opportunity costs in civilian
goods and services. War involves the destruction of labor and capital. In contrast,
private markets seek the optimal mix of labor and capital to provide goods and
services through voluntary trading and exchange. Resource allocation is based on
price and profit signals that lead to “creative destruction” reflected in continuous
investment in new innovations, inventions and the output of new goods and services.
Q = f (A, K , L) (3.1)
42 K. Hartley, B. Solomon
where Q is defense output and A, K , and L are inputs of technology (A), capital
(K ), and labor (L).
While the model appears attractive, there are at least four major caveats. First,
a production function assumes factor inputs are combined to minimize costs. This
assumption is unrealistic in view of the lack of efficiency incentives in internal
defense markets: there are few rewards or penalties to achieve least-cost produc-
tion. Second, all defense inputs have to be identified and correctly valued. Third,
defense output is simply asserted without recognizing the problems of identify-
ing and measuring output, including the multi-product nature of overall defense
output. Fourth, the model simply identifies defense outputs resulting from various
inputs: there are no criteria for determining society’s preferred defense output (the
“optimal” defense output).
Two central problems with military production functions arise over inputs and
outputs. Consider the problem of identifying and valuing all relevant inputs. These
comprise technology, capital and, labor, and include the following items:
E q = f (R Dd , Z ) (3.2)
Granger causality test statistics are uninformative about the size and direc-
tion of the predicted effects and Granger causality measures incremental
predictability and not economic causality.
(Dunne and Smith 2010, 440)
This particular critique ends with the need to provide “measures of the political
and strategic determinants of military expenditures, such as threats” (Dunne and
Smith 2010, 440).
Measuring defense output 45
3.4 Assessing defense outputs: problems
and challenges
Defense outputs involve a complex set of variables concerned with security, pro-
tection, and risk management, including risks avoided, safety, peace, and stability.
Private markets routinely provide benefit measures such as sales, labor productiv-
ity, and profitability. Unlike private markets, there are no precise benefit measures
for defense output.
Defense inputs are more easily identified, measured, and valued than outputs, as
reflected in many nations’ annual input-oriented defense budgets. For economists,
questions then arise as to whether annual defense budget information provides suf-
ficient data to assess the efficiency and effectiveness of military expenditure: how
do expenditures on inputs correspond to desired defense outputs? Do defense bud-
gets provide policy-makers and politicians with the sort of data needed to conduct
military CBAs?
Typical questions include assessing the benefits and costs of alternative defense
forces; expanding (or contracting) the Army, Navy, or Air Force; substitut-
ing equipment (capital) for military personnel (labor); or substituting national
guard and reserves for regular (active) forces. Various defense budgets used by
nations include: input budgets, output budgets, management budgets, and resource
accounting budgets.
CDN$000
Personnel 10,438,096
Transportation and communications 768,058
Information 13,666
Professionnal and special services 2,982,038
Rentals 383,972
Purchased repair and maintenance 1,465,091
Utilities, materials and supplies 1,017,831
Construction and/or acquisition of land, buildings and works 500,631
Construction and/or acquisition of machinery and equipment 2,434,609
Transfer payments 181,705
Total subsidies and payments 221,561
Total gross expenditure 20,407,258
Total revenue 429,068
Total net expenditure 19,978,190
Source: Public Accounts Volume II. Receiver General of Canada (RGC 2013).
values. These and other limitations led to the development of output budgeting
(Hartley 2011).
3.4.5 Challenges
Two growing pressures make it essential to focus on the size of a nation’s defense
budget and the efficiency with which defense resources are used. The first pres-
sure is to reduce defense budgets and reallocate resources to other public spending
programmes, especially education, health, and welfare (including care for the
increasing elderly populations). Second, added pressure on defense budgets comes
from the increasing costs of defense equipment. A simple example shows the
Measuring defense output 49
importance of rising unit equipment costs which affects all nations (all figures
are for unit production costs in 2010 prices):11
Increasing unit costs and constant or falling defense budgets (in real terms)
suggest difficult defense choices cannot be avoided. The challenge in divestment
discussions as in investment decisions is to establish useful measures of defense
outputs, or in the absence of such measures, to develop proxies to enable CBA.
This chapter, having identified multiple issues and obstacles involved in
developing defense output measures; next reviews the experience of several
nations. Many chapters in this book offer valuable techniques to develop proxy
measures of intermediate outputs at lower levels of the organization in support of
military CBA.
3.5.1 Security
In principle, defense provides security, which itself is a multi-product output
embracing protection, safety, insurance, peace, economic stability, and risk avoid-
ance or reduction (Solomon et al. 2008). Further dimensions include prosperity,
individual and national freedoms, liberty, and a ‘way of life.’ These are all dif-
ficult to measure and might be influenced by factors other than defense. Also,
these aspects of security are public goods which are not marketed and include
non-marketable services involving no tangible and physical products.
Security is sometimes defined as the absence of threats or risks (Baldwin 1997;
Engerer 2011). A world of no threats or risks, however, does not and cannot
exist: the real world is characterized by continually emerging threats and risks.
50 K. Hartley, B. Solomon
Questions then arise about which threats and risks can be reduced, by whom, and
at what cost.
Recent developments have led to security referring to issues other than military
security (creating fuzzy boundaries). Individuals are faced with threats to their
lives, health, property, other assets, and their prosperity (e.g. from criminals and
terrorists, disease/pandemics and ill health, natural or man-made disasters, and
economic recessions). Threats to individuals augment the threats to nation states
(e.g. military threats from other nations and environmental problems originating
from other nations) which raises questions about which threats should be handled
privately and which publicly.
Where threats are handled publicly, which is the most appropriate and least-
cost solution? For example, military solutions are appropriate for external military
threats whilst internal police forces are more appropriate for internal threats
from criminals (e.g. physical violence to individuals involving injury, death,
and robbery). Threats to an individual’s state of health require dietary, medi-
cal, and care solutions (e.g. from doctors, nurses, and care homes). Threats to
prosperity require government macroeconomic and microeconomic policies to
promote full employment and economic growth (e.g. opportunities for educa-
tion, training, and labor mobility—although some of these activities can be funded
privately).
Technical progress and changing consumer preferences have resulted in shifts
from the public provision of security to private protection measures provided
and financed by individuals (e.g. private security guarding, camera surveillance
of property, and creation of neighborhood watch schemes providing local club
goods). Security also has important geographical dimensions.
For example, defense can be viewed as a means of protecting a nation’s property
rights over its land, sea, and air space. A nation’s defense forces, however, might
also be used to protect other nations’ citizens so that the public good becomes inter-
national which further increases the problem of obtaining and financing the optimal
amount of the international public good (including peace). Overall, security mea-
sures can be analyzed as national or international public goods, club goods, and
private goods—each with different solutions and each embracing different indus-
tries (e.g. security and defense industries: Engerer 2011). These different industries
have different customers, products, and technologies (Sempre 2011).
MoD budgets pay for the UK force elements to be ready for operations as out-
lined in the DPAs. The costs of these missions, however, are funded from the
government’s contingency reserves. Over time, the rising unit costs of defense
equipment and of volunteer military personnel will result in smaller armed forces
and reduced defense capabilities (as defined by the UK MoD).
More important would be an assessment of the costs of achieving these defense
capabilities compared with other nations providing similar capabilities (i.e. is the
UK providing its capabilities at least-cost?). Within MoD, measures of defense
training activities are used to assess performance. These include flying hours, days
spent at sea, and Army personnel data on gains to trained strength and data on
military exercises (ONS 2008).
The MoD publishes an annual performance report which offers some further
insight into its defense capabilities (HCP 992 2011).17 Useful though such infor-
mation might be, it is both qualitative and vague (i.e. “success” in Afghanistan)
and focuses on input costs which are unhelpful data for measuring output by
themselves. On force readiness, the MoD’s Performance Report admits that “Mea-
suring and aggregating readiness is complex, not least because it is based on
judgements of what is required to enable the armed forces to respond to a wide
range of potential challenges” (HCP 992 2011, 21).
The MoD also reports on where there are “critical and serious weaknesses” in
UK Forces.18 The MoD’s Performance Report also included a section on imple-
menting the 2010 Strategic Defence and Security Review which provided further
information on the UK’s defense capabilities.19
The UK’s defense capabilities output measures are an improvement on the tra-
ditional input approach but there are deficiencies at least in terms of publicly
available information. For example, the National Audit Office has reported that the
UK MoD has a good system for defining, measuring, and reporting the readiness
56 K. Hartley, B. Solomon
of its armed forces which compares well with other countries (e.g. Australia,
Denmark, and US: NAO 2005). It is recognized that 100 percent readiness is
too costly. The published data on readiness, however, refers to whether there are
serious or major weaknesses, which is useful but not very illuminating (e.g. with-
out knowing what and where such weaknesses arise and their impact on force
effectiveness). For instance, a statement that 50 percent of UK Forces had no seri-
ous or critical weaknesses suggests that the remaining 50 percent suffered serious
weaknesses, which should be a source of concern!
Moreover, these performance assessments are mostly undertaken by MoD
personnel, which could raise questions of independence and objectivity. An
NAO report on the performance of MoD in 2009–10 presented and reviewed
performance indicators (NAO 2010). The report focused on financial manage-
ment information (e.g. management of stocks and assets) and made no men-
tion of defense output measures. Though there was mention of defense output
indicators, these only included qualitative measures: ‘success on operations;’
the existence of serious and critical weaknesses in readiness; manning levels
in relation to manning balance by Service (with no data); and flying hours
achieved against targets (again, without any data). In relation to the MoD aim
of global and regional reductions in conflict, no output measure was reported
(NAO 2010).
The NAO also publishes value-for-money reports (a type of retrospective
CBA)—for example, on the multi-role tanker aircraft capability—and annual
reports on MoD’s major projects. These project reports assess major defense
projects against their contractual commitments on cost, delivery, and performance,
usually identifying cost overruns, delays, and any failures to meet performance
requirements. Such value-for-money reports are a useful addition to knowl-
edge but do not include the wider industrial and economic benefits of major
projects, nor do they provide any assessment of the “battle-winning” perfor-
mance of defense equipment (e.g. as demonstrated in conflicts such as Afghanistan
or Iraq).
Overall, the UK’s defense capabilities are useful measures of defense output but
deficiencies remain. Some indicators of force readiness are qualitative: readiness
is a variable measure depending on circumstances (readiness for what, when and
where?); capabilities are not all identified; benefit values are not attached to capa-
bilities; and the capabilities cannot be aggregated into a single measure of overall
defense output.
None of the output measures address the contribution of defense to conflict pre-
vention and its contribution to minimizing the costs of conflict, including saving
lives. In fact, MoD economists examined different approaches to capturing output
used in various parts of the MoD. As Davies et al. (2011, 399) put it:
The UK’s experience in measuring outputs in other parts of the public sector
and in the private sector is given in Appendix 3.2, offering selected insights for
defense.
• the defense of Australia against armed attack with the capability to act inde-
pendently so as not to be reliant on foreign military forces. This principal
task requires the Australian Defence Force (ADF) to control the air and sea
approaches to Australia;
• the security, stability, and cohesion of Australia’s immediate neighborhood
which is shared with Indonesia, Papua New Guinea, East Timor, New
Zealand, and the South Pacific Island states;
• an enduring strategic interest in the stability of the wider Asia-Pacific region;
and
• a strategic interest in preserving the world international order which restrains
aggression, manages other risks and threats, and addresses the security
impacts of climate change and resource scarcity.
• the capabilities needed for sea and air control around Australia;
• the ability to deploy a brigade group for combat operations for a prolonged
period of time in the primary operational environment (for shorter period
beyond that area);
• the ability to deploy a battalion group to a different area of operations in the
primary operational environment;
• the ability to maintain other forces in reserve for short-notice, limited warning
missions;
• the ability to provide tailored contributions to operations in support of
Australia’s wider strategic interests (e.g. special task forces group); and
• the ability to provide assistance to civil authorities (e.g. fisheries protection,
terrorist incidents, support for major events, emergency responses, humani-
tarian and disaster relief in Australia and its neighbors, provision of search
and rescue support, etc).
• the rise of the Asia-Pacific and the Indian Ocean rim as regions of global
strategic significance;
• the growth of military power projection capabilities of the Asia-Pacific
countries;
• the growing need for the provision of humanitarian assistance and disaster
relief following extreme events in the Asia-Pacific; and
• energy security and security issues associated with expanding offshore
resource exploitation in Australia’s North West and Northern approaches.
The ADF Force Posture Review considers how the ADF will support Australia’s
ability to respond to a range of activities including deployments on overseas
missions and operations; support of operations in Australia’s wider region; and
engagement with the countries of the Asia-Pacific and Indian Ocean rim in ways
which will help to shape security and strategic circumstances in Australia’s inter-
est. The Force Posture Review also makes recommendations on basing options for
Force 2030. There is also a Submarine Sustainment Review to assess sustainment
of Australia’s Collins Class submarines (ADF 2011).
The Australian Defence White Paper of 2013 identified the capabilities the
ADF will need in the future, reflecting the withdrawal from Afghanistan and the
Solomon Islands. It emphasized air, naval,20 special-forces, intelligence, and cyber
security. The commitment to spending 2 percent of GDP on defense became a
long-run target.
The 2013 White Paper presented a broad description of future forces but pro-
vided little detail on their capabilities. Interestingly, the 2013 White Paper referred
to productivity and the need for “defence to become more efficient and prudent
in its use of resources to remove waste and achieve better economies of scale.”
This is a politically attractive phrase, but does not explicitly address the chal-
lenge of preserving an adequate level of overall defense output and corresponding
measurement challenges (DoD 2013).
3.6.4.1 France
A new French defense policy was announced in 2008 with the aim of making
the French armed forces more flexible for rapid deployment from the Atlantic to
the Indian Ocean. France aims to provide the necessary resources to ensure the
security of its citizens, to safeguard national independence, and to consolidate the
nation’s military and diplomatic power. Under the new policy, France will be able
to project 30,000 personnel with 70 combat aircraft, one carrier group and two
naval battle groups within a six-month period for up to a year (a force capable
of dealing with one major war or crisis at a time). Nuclear deterrence remains
a key military mission, but terrorism is the most immediate threat, and there are
also public service missions. There will be reductions in the numbers of military
personnel and investment in new equipment. Some equipment is of poor quality:
for example, only 50 percent of Leclerc tanks are mission ready; refuelling aircraft
are 45 years old; and some Puma helicopters are 30 years old (Penketh 2008).
In 2010, the UK and France signed an Anglo-French Defence Treaty with the
potential for greater bilateral co-operation between their armed forces and defense
industries.
By 2014, French defense policy was adjusting to a “difficult financial situa-
tion” leading to further reductions in the numbers of military personnel, ships,
tanks, helicopters, and transport aircraft. A Joint Reaction Force was created, with
2,300 personnel and the ability to deploy up to 15,000 troops in a single over-
seas deployment. Plans were in place to cut the defense budget to 1.5 percent
of GDP.
3.6.4.2 Germany
NATO remains the centrepiece of Germany’s defense policy. The new defense
policy announced in 2011 involved major changes for Germany’s armed forces:
reductions in the defense budget; abolishment of conscription replaced by an
all-volunteer force; improvement of Germany’s expeditionary capabilities; and
closer military co-operation in Europe, especially in procurement and training
(German Ministry of Defense (GMOD) 2011). Under the new policy, Germany
plans to increase the deployment of the Bundeswehr outside Germany from the
current 7,000 to some 10,000 soldiers (but there is no statement of the extent
of geographical coverage of these expeditionary forces). There are also plans
to reduce quantities of equipment (aircraft, helicopters, and ships). In 2014,
however, Germany announced a review of its commitment to overseas military
Measuring defense output 63
deployments, including its military contribution to the international pooling and
sharing of military resources.
3.6.4.3 Italy
Despite cuts in defense spending due to Italy’s austerity programme, Italy retains
an expeditionary capability. Reports suggest the Air Force has been particularly
affected by defense cuts. There were also reports that in 2010, Italy planned to
reduce its involvement in peacekeeping missions in the Balkans and possibly in
Lebanon, concentrating instead on Afghanistan where force levels peaked at 4,000
soldiers (Nativi 2010).
Italy’s Military Policy Review of 2013 aimed to reduce the numbers of military
personnel to 150,000 and civilian personnel to 20,000 by 2024. This represents a
lengthy adjustment period. Italy also identified reductions in the numbers of senior
military personnel within the reduced total numbers (i.e. fewer admirals, generals,
colonels, and naval captains). The plan is for a defense budget of 0.84 percent of GDP.
The 2013 Military Policy Review declared that “available resources will be used
on developing systems that combine high operational efficiency, adequate cost-
effectiveness, and a development/growth margin that allows them to be integrated
into complex and net-centric systems.”
3.6.4.4 Spain
Reductions in defense spending were part of Spain’s austerity programme. The
2011 budget reflected four objectives: the safety of the troops (force protection
via operating and logistics expenditure); operational readiness; the maintenance
of weapons systems; and international operations and fulfilment of Spain’s inter-
national commitments. New tools were announced for improved oversight and
management of defense expenditures.
3.6.4.5 Sweden
A new defense policy was announced in 2009 with an emphasis on mobility and
flexibility of Sweden’s armed forces. The plan is for it to be possible for an entire
operational organization of some 50,000 people to be used within one week after
a decision on heightened alert. In contrast, today only one-third of the national
operational organization is equipped and prepared for an operation within one
year. Some defense capabilities were listed in terms of inputs: numbers of military
personnel (e.g. deployment of 1,700 people for continuous international peace-
support operations) and numbers of Gripen aircraft (100 of the C/D model). An
all-volunteer force will replace compulsory military service and there will be sub-
stantial reserve forces (e.g. four mechanized battalions). For the Army, only a
small proportion of its soldiers will be full-time. Sweden specified its area of
national interest: namely, the Baltic Sea or the northern area (Sweden Ministry
of Defence (SMOD) 2009).
64 K. Hartley, B. Solomon
3.6.5 The European Defence Agency
The European Defence Agency (EDA) publishes defense data for its member
states. These include annual financial data such as levels of defense spending and
shares of GDP for member states, equipment procurement and R&D expenditures,
spending on infrastructure and construction, defense expenditure that is out-
sourced, and expenditures on collaborative equipment programs. Input data is
also published that include numbers of military, civilian and internal security
personnel, personnel expenditures, as well as data on numbers of different types
of equipment (combat aircraft tanks and warships).
Most of the data published is for inputs rather than defense outputs, although
some EDA officials regard indicators such as quantity of military personnel as
intermediate output measures. There is, however, data which are clearly a mea-
sure of intermediate output and proxies for defense output: namely, operation and
maintenance expenditure, operational costs, average numbers of troops deployed,
and the average numbers of sustainable (land) forces (EDA 2011). Comparative
analysis of such data could be useful in revealing variations in efficiency and
effectiveness across member states.
3.7 Conclusion
This chapter has identified a set of important questions that arise in efforts to
measure defense outputs. Indeed, it has raised more questions than answers, but
this investigation contributes to further the understanding needed to address the
central research questions: What is defense output? How can it be valued? Under
what conditions is it a worthwhile investment?
In its published form, the international experience of measuring defense output
reveals some useful intermediate output measures, usually in the form of spe-
cific defense capabilities.22 These are improvements on the traditional reporting
of inputs that have typically included numbers of military personnel and equip-
ment (e.g. combat aircraft, tanks, and warships). By themselves, input measures
offer little indication of the value of overall defense capabilities such as peace,
protection, deterring conflicts, and insurance against future threats.23
A starting point in answering the central research questions is to apply CBA: to
identify the costs of defense and then ask whether defense provides at least a com-
parable level of benefits in the outputs produced. It is also important to capture
non-economic benefits in addition to measurable economic benefits in measur-
ing the overall benefits of defense spending. For example, if defense spending
66 K. Hartley, B. Solomon
costs $X billion, does it provide overall benefits of a similar value? Similar ques-
tions can be asked about the costs and benefits of conflict and peacekeeping
operations.24
Next, the CBA can focus on incremental (or marginal) changes. If defense
spending is increased or decreased by 10 percent, what are the effects on defense
outputs (benefits)? Such marginal analysis can be assessed as a whole (on overall
defense output), or by each military service (e.g. what would be the impact of a
10 percent increase or decrease in the size of the Army?).
Specifying the important questions is the first stage in any evaluation; but
who raises and answers the questions? In a democracy, elected politicians are
ultimately responsible for determining the size of military expenditures and its
allocation among each of the services. Typically, unelected agents within the
military propose many of these choices.25 This reinforces the importance of devel-
oping meaningful defense output measures to guide future military investment and
divestment decisions.
Notes
1 This chapter is based on a report prepared by Keith Hartley for the Defence Research
and Development Canada under contract number DND-10/23136. Parts have been
updated to 2014.
2 See also Williamson (2000) for more nuanced discussion.
Measuring defense output 69
3 For example, a nation’s international peacekeeping contributions might provide consid-
erable satisfaction to the country’s prime minister, senior ministers, and civil servants
able to attend international meetings at the UN and to participate in regional meetings.
The principal-agent and public choice analysis raises the general question of who gains
and who pays for these defense policies (e.g. international peacekeeping, national pro-
curement of defense equipment including offsets). Ultimately, taxpayers pay and receive
some defense benefits whilst agents consume some benefits not explicitly supported by
the majority of voters and taxpayers.
4 For example, compare today’s space satellite communications systems with the military
communications facilities in 1914 (e.g. observation balloons).
5 Skills and productivity differ between regular forces, conscript, and reserve forces.
Other complementary or substitutable labor inputs include civilian labour, contractors,
national guard and reserves, and even police forces (e.g. police forces substituted for
British Army troops in policing Northern Ireland: Ridge and Smith 1991).
6 Chapter 14 offers an interesting application of a new “real options” cost–benefit analysis
of R&D investments.
7 The varied results in this field reflect different economic and econometric models,
combinations of variables, time-periods, cross-section and time-series studies; a het-
erogeneous set of countries; and the use of data with varying degrees of reliability and
scope of coverage.
8 With output budgets, a distinction needs to be made between the budget available to
the MoD and the budget released to Parliament and the public. The published version of
the budget does not reveal all the information available to the MoD and the basis for the
choices which are reflected in the publicly released version of the budget (Davies et al.
2011, Chapter 17).
9 As an example, note that in Table 3.3 capital departmental expenditures limit (DEL) is
part of overall resource DEL and reflects investments spending (cash) which appears
in MoD’s balance sheet to be consumed over a number of years. The resource DEL
includes depreciation and cost of capital changes.
10 Formerly known as Program Activity Architecture.
11 Norman Augustine famously forecasted that with continued rising unit costs, by 2054
the entire US defense budget will purchase just one aircraft which would have to be
shared between the Air Force and Navy (the Marines would have it for one day in leap
years). He also forecasted that the UK and France would reach this position two years
earlier (Augustine 1987, 143).
12 See Chapter 9 in this book for a novel application of the statistical value of life.
13 It was not expected that both deployments would involve warfighting or that they would
be maintained for longer than six months. One might be a short warfighting deployment;
the other an enduring non-warfighting operation (SDR 1998, 23).
14 “Small scale” is defined as the UK’s deployment to Macedonia in 2001; “medium-scale”
as Afghanistan (2001); and “large scale” as operation TELIC (Iraq).
15 When part of a coalition, the optimum ratio for prolonged commitments was 3–4 ships
and 5 Army and RAF crews/units for each one deployed (Cmnd. 6269 2004).
16 This is about two-thirds of the force deployed to Iraq in 2003.
17 For example, its 2011 report focused on success in Afghanistan reflected in the costs of
operations; the costs of its force elements (e.g. a ship at an annual cost of £28 million;
a fixed wing combat aircraft at £6.5 million per year); and the direct costs of Service
personnel (£49,000 per Service personnel per year: HCP 992 2011).
18 Interestingly, MoD’s Performance Report included a section on Defence exports where
one aim is to support British industry and jobs (HCP 992 2011). Defence exports are
not an obvious output indicator for the MoD.
19 There is a NATO (NATO 2010) commitment to spend at least 2 percent of GDP on
defense; an aim of achieving savings from contract renegotiations with the defense
70 K. Hartley, B. Solomon
industry; a goal to retain a surface fleet of 19 warships; a commitment to reduce the
force of main battle tanks by 40 percent; and, lastly, a commitment to scrap the Nimrod
MRA4 fleet (at a savings of some £200 million per aircraft: HCP 992 2011).
20 The decision to replace the Collins’ class submarines was confirmed but the life of the
Collins’ class was extended by seven years.
21 Compares to “Lines of Development” in the UK and DOTMILPF (doctrine, orga-
nization, training, material, leadership and education, personnel, and facilities) in
the US.
22 None of the nations reviewed in this chapter addressed the challenges of measuring and
valuing overall defense output. The nearest to an intermediate output measure consisted
of the identification of various defense capabilities, but these were not always com-
prehensive. For example, the UK did not identify all its capabilities, including defense
of the UK homeland and its nuclear deterrent, and no valuations were provided for
the assorted capabilities. Nonetheless, the focus on defense capabilities is an improve-
ment over the traditional focus on input measures such as numbers of personnel and
equipment. The next challenge may be to assign weights to the mix of capabilities and
aggregate them into an index that represents an overall measure of defense output.
23 Nor should it be assumed that there exists a single “best” indicator: performance indi-
cators can often give unexpected and perverse results (e.g. the operation was a success
but the patient died).
24 For example, was the Iraq conflict a worthwhile investment for the US?
25 Governments could use representative samples of voters to form focus groups which
would offer their views on the size of alternative defense budgets and force structures.
These focus groups could receive advice from officials and military personnel. Focus
groups are not an ideal solution (e.g. free-rider problems remain; groups have to be
selected; and they will have their internal momentum and dynamics), but would provide
politicians with an additional mechanism for identifying voter preferences on defense
spending and policy.
References
Atkinson, Sir A. B. 2005. “Measurement of Government Output and Productivity for the
National Accounts.” In Atkinson Review: Final Report. Basingstoke, England: Palgrave
Macmillan.
Augustine, N. R. 1987. Augustine’s Laws. London: Penguin Books.
Australian Defence Force (ADF). August 25, 2011. Australian Defence Force Posture
Review Public Consultation and Submission Process. Canberra, AU: Department of
Defence.
Baldwin, D. A. 1997. “The Concept of Security.” Review of International Studies 23: 5–26.
Benoit, E. 1973. Defense and Economic Growth in Developing Countries. Boston, MA:
Lexington Books.
Berkok, U. G., and B. Solomon. 2011. “Peacekeeping, Private Benefits and Common
Agency.” In Handbook on the Economics of Conflict, edited by D. L. Braddon and K.
Hartley. Northampton, MA: Edward Elgar Publishing.
Black, R. 2011. Review of Defence Accountability Framework. Canberra, AU: Australian
Government Department of Defence.
Comptroller and Auditor General (CAG). June 2011. Central Government: Cost-
effectiveness and Improving Annual Reports. Wellington, NA: Comptroller and Auditor
General.
Cornes, R., and T. Sandler. 1996. The Theory of Externalities, Public Goods and Club
Goods (2nd edition). Cambridge: Cambridge University Press.
Measuring defense output 71
Davies, N, T. Turner, A. Gibbons, D. Jones, S. Davis, and N Bennett. 2011. “Helping
Secure the ‘Biggest Bang for the Taxpayers’ Buck:’ Defence Resource Management
in the United Kingdom.” In Handbook on the Economics of Conflict, edited by D. L.
Braddon and K. Hartley. Northampton, MA: Edward Elgar Publishing.
Defence Analytical Services and Advice (DASA). 2011. U.K. Defence Statistics Com-
pendium 2011. London: The Stationery Office Limited.
Deger, S., and R. P. Smith. 1983. “Military Expenditure and Growth in Less Developed
Countries.” Journal of Conflict Resolution 27(2): 335–353.
Department of Defence (DoD). 2009. Defending Australia in the Asia Pacific Century:
Force 2010. Canberra, AU: Department of Defence.
——. 2011. Defence Portfolio Budget Statements 2011–2012. Canberra, AU: Department
of Defence.
——. 2013. Defence White Paper 2013. Canberra, AU: Department of Defence.
—— 2014. Quadrennial Defense Review. Available at www.defense.gov/pubs/
2014_Quadrennial_Defense_Review.pdf [accessed 10 September, 2014].
Department of National Defence (DND). 2013. Department of National Defence: Report
on Plans and Priorities 2013–14. Ottawa, CA: Department of National Defence.
Dunne, P., and R. P. Smith. 2010. “Military Expenditure and Granger Causality: A Critical
Review.” Defence and Peace Economics 21(5–6): 427–441.
Engerer, H. 2011. “Security as a Public, Private or Club Good: Some Fundamental
Considerations.” Defence and Peace Economics 22(2): 135–145.
European Defence Agency (EDA). March 31, 2011. Defence Data 2009. Brussels, BE:
EDA.
German Ministry of Defense (GMOD). May 2011. Defence Policy Guidelines. Berlin, DE:
GMOD.
Hartley, K. 2010a. “Defense Economics.” In The Economics of Human Systems Integration:
Valuation of Investments in People’s Training and Education, Safety and Health, and
Work Productivity, edited by W. B. Rouse. Hobeken, NJ: Wiley.
——. 2010b. The Case for Defence. Defence and Peace Economics 21(5–6): 409–426.
——. 2011. Defence Output Measures: An Economic Perspective. Defence R&D Canada–
CORA. Contract Report: DRDC CORA CR 2011-178.
Hildebrandt, G. G. 1990. “Services and Wealth Measures of Military Capital.” Defence
Economics 1(2): 159–176.
——. 1999. “The Military Production Function.” Defence and Peace Economics 10(3):
247–272.
Hitch, C. J., and R. N. McKean. 1960. The Economics of Defense in the Nuclear Age.
London: Oxford University Press.
House of Commons (HCP) 992. 2011. Ministry of Defence Annual Report and
Accounts 2010–2011 (including Performance Report). London: The Stationery Office
Limited.
House of Commons Defence Committee. 2003. Delivering Security in a Changing World.
2003 Defence White Paper: Cmnd. 6041, Volumes I–II. London: The Stationery Office
Limited.
House of Commons Ministry of Defence. 2004. Delivering Security in a Changing World:
Future Capabilities. 2004 Defence White Paper: Cmnd. 6269. London: The Stationery
Office Limited.
Jones-Lee, M. E. 1990. “Defence Expenditure and the Economics of Safety.” Defence
Economics 1(1): 13–16.
72 K. Hartley, B. Solomon
Middleton, A., S. Bowns, K. Hartley, and J. Reid. 2006. “The Effect of Defence R&D on
Military Equipment Quality.” Defence and Peace Economics 17(2): 117–139.
Middleton, A., K. Copsey, and K. Hartley. 2011. Estimating a Production Function for
Military Power (Mimeo). Malvern, UK: Economex Limited.
Military Policy Review 2013. Available at www.difesa.it/Primo_Piano/Documents/2013/
gennaio%202013/Direttiva%20Ministeriale_ENG.pdf [accessed 10 September, 2014].
National Audit Office (NAO). 2005. Assessing and Reporting Military Readiness (HCP
72). London: The Stationery Office Limited.
——. 2010. Performance of the MoD 2009–10. Briefing to House of Commons Defence
Committee. London: The Stationery Office Limited.
National Security and Intelligence. 2010. Securing Britain in an Age of Uncertainty: The
Strategic Defence and Security Review. (Cmnd. 7498). London: The Stationery Office
Limited.
Nativi, A. November 19, 2010. “Italy Protects Defense Modernization Spending.” Aviation
Week.
New Zealand Defence Force (NZDF). June 1991. Annual Plan: G55. Wellington, NZ:
NZDF.
——. 1993. New Zealand Defence Force Annual Report 1993. Wellington, NZ: NZDF.
——. 2011. New Zealand Defence Force Statement of Intent 2011–2014. Wellington. NZ:
NZDF.
——. 2013. New Zealand Defence Force Annual Report 1993. Wellington, NZ: NZDF.
North Atlantic Treaty Organization (NATO). 2010. Financial and Economic Data Relating
to NATO Defence. Brussels, BE: NATO, Press and Media.
Office of Under Secretary of Defense (USDOD). 2011. US DoD Fiscal Year 2012 Budget
Request. Washington, DC: Department of Defense.
Office for National Statistics (ONS). 2008. Scoping Paper on the Possible Improvements
to Measurement of Defence in the U.K. National Accounts: U.K. Measurement of
Government Activity. Newport, SW: ONS.
——. 2010. Annual Report 2009–10. Newport, SW: UK Centre for Measuring Government
Activity.
Organisation for Economic Co-operation and Development (OECD). 2010. Main Science
and Technology Indicators. Paris: OECD.
Penketh, A. June 18, 2008. “The Big Question: What is the New French Defence Strategy,
and Should we Follow Suit?” The Independent Newspaper.
Receiver General of Canada (RCG). 2013. The Public Accounts of Canada Volume II.
Ottawa, CA: Government of Canada.
Ridge, M., and R. P. Smith. 1991. “U.K. Military Manpower and Substitutability.” Defence
Economics 2(4): 283–293.
Sempre, C. M. 2011. “The European Security Industry: A Research Agenda.” Defence and
Peace Economics 22(2): 245–264.
Solomon B., P. Chouinard, and L. Kerzner. October 2008. The Department of National
Defense Strategic Cost Model (Volume II: Theory and Empirics, Defence R&D). Ottowa,
CA: Canada-CORA.
Stockholm International Peace Research Institute (SIPRI). 2011. Military Expenditure
Database (Database). Solna, SE: SIPRI.
The Strategic Defence Review (Cmnd. 3999). 1998. (See also Supporting Essays). London:
The Stationery Office Limited.
Sweden Ministry of Defence (SMOD). March 19, 2009. A Functional Defence. (Press
Release by the Sweden Ministry of Defence).
Measuring defense output 73
Tisdell, C., and K. Hartley. 2008. Microeconomic Policy: A New Perspective (2nd edition).
Cheltenham, UK: Edward Elgar Publishing.
Williamson, O. E. 2000. “The New Institutional Economics: Taking Stock, Looking
Ahead.” Journal of Economic Literature 38: 595–613.
Zaller J. R., and S. Feldman. 1992. “A Simple Theory of the Survey Response: Answering
Questions versus Revealing Preferences.” American Journal of Political Science 36(3):
579–616.
4 The economic evaluation
of alternatives
Francois Melese
All participants in the acquisition system shall recognize the reality of fiscal
constraints . . . DoD components shall plan . . . based on realistic projections of
the dollars . . . likely to be available [and] the user shall address affordability
in establishing capability needs.
[emphasis added] (DoD 5000.01 2007, Enclosure 1, 5)15
Economic evaluation of alternatives 77
According to the US Defense Acquisition Guidebook, the purpose of an
affordability assessment is to demonstrate that a program’s projected funding
requirements are realistic and achievable:16
In general, the assessment should address program funding over the six-year
programming [FYDP] period, and several years beyond. The assessment
should also show how the projected funding fits within the overall DoD
Component plan.17
[emphasis added] (DoD July 7, 2006, Section 3.2.2)
The next section offers a brief description and critical evaluation of the status
quo. This includes two common decision criteria widely used in cost-effectiveness
analyses. The first is the popular “bang-for-the-buck” or benefit/cost ratio.24 The
second criterion is essentially a weighted average of cost and effectiveness, a deci-
sion rule generated by the standard MCDM approach to cost-effectiveness analysis
that underpins most contemporary AoAs (see Chapters 8 and 16, for example).
Section 4.6 illustrates the six approaches to structure an EEoA. The final section
concludes with a decision map to guide analysts and DMs in selecting which of
the six approaches is best suited to their circumstances.
[o]ur work in [uncovering] best practices has found that an executable busi-
ness case [requires] demonstrated evidence that . . . the chosen concept can be
developed and produced within existing resources [funding/budgets].
[emphasis added] (GAO September 25, 2008, 6)
1 In a military text entitled Executive Decision Making, the author offers that
“[w]hen we cannot fix cost or effectiveness, we might combine them to help
us choose between alternatives . . . If neither can be fixed . . . we can establish
a cost/effectiveness ratio” (i.e. a cost/benefit ratio) (Murray 2002, 6–3, 6–10).
2 In presenting what they claim is a “novel cost–benefit analysis” for the “com-
prehensive evaluation of competing military systems,” authors in the Acquisi-
tion Review Quarterly define a “merit function” as “a single number . . . [that]
reflects the ratio of benefits derived to dollars spent” (i.e. a benefit/cost ratio).
Economic evaluation of alternatives 81
The article asserts: “with this . . . approach, the cost-effectiveness of compet-
ing systems can be compared and “provides for objective and reliable decision
making,” where “a large system merit [benefit/cost ratio] is preferable to a
small one” (Byrns et al. 1995).
3 Similarly, in a section entitled “Comparing Costs and Benefits,” the US
Department of the Army’s Economic Analysis Manual states: “When the
results yield unequal cost and unequal benefits [. . .] in this situation all alter-
natives [. . .] may be ranked in decreasing order of their benefit/cost ratios”
(DoA February 2001, 32).28
The use of [benefit/cost] ratios usually poses no problem as long as the anal-
ysis is conducted in [a] framework . . . with the level of effectiveness or cost
fixed. However, it is common to encounter studies where this has not been
done, with the result that comparisons [are] essentially meaningless.
[emphasis added] (p. 11)
One common ‘compromise criteria’ is to pick that [alternative] which has the
highest ratio of effectiveness to cost . . . [M]aximizing this ratio is the [deci-
sion] criterion. [While] it may be a plausible criterion at first glance . . . it
allows the absolute magnitude of [effectiveness] or cost to roam at will. In
fact, the only way to know what such a ratio really means is to tighten the
constraint until either a single budget (or particular degree of effectiveness)
is specified. And at that juncture, the ratio reduces itself to the test of maxi-
mum effectiveness for a given budget (or a specified effectiveness at minimum
cost), and might better have been put that way at the outset . . . .29
(pp. 165–7)
The first two ways to structure an EEoA, as well as the fourth and fifth
approaches, follows their advice:
The test of maximum effectiveness for a given budget (or alternatively, mini-
mum cost of achieving a specified level of effectiveness) . . . seems much less
likely to mislead the unwary.30
(p. 167)
The linear, additive separable version of this value function is frequently used to
calculate a positively weighted MoE and negatively weighted cost for each alter-
native (for example, see Beil and Wein 2003; Che 1993; Clemen 1996; French
1986; Hwang and Yoon 1981; Keeney 1994; Keeney and Raiffa 1976; Kirkwood
1997; Liberatore 1987; Pinker et al. 1995; Vazsonyi 1995).
The typical decision sciences’ (MCDM) approach to an AoA38 thus
involves:
Given several alternatives, select the preferred alternative that provides the
best value, i.e. Maximize: V(MoE,COST) = w1*MoE − w2*COST.
The solicitation shall state whether all evaluation factors other than cost/price,
when combined [i.e. an MoE], are significantly more important than, approx-
imately equal to, or significantly less important than cost/price.
(GSA, DoD, NASA, March 2005, Section 15.101-1(2))
The specific weight given to cost or price shall be at least equal to all other
evaluation factors combined unless quantifiable performance measures can be
used to assess value and can be independently evaluated.
(OMB Circular A-76, B-8)
A2
Effectiveness
Cost
A2
Effectiveness
If Performance has a
A1 sufficiently greater weight
(w1 >>w2), then high cost,
high effectiveness
alternative A2 wins
Cost
In the fixed budget case, the alternatives being considered are compared on
the basis of effectiveness likely to be attainable for the specified budget level
(p. 12). The analysis attempts to determine that alternative (or feasible com-
bination . . .) which is likely to produce the highest effectiveness.
(p. 10)47
I) INTRA-PROGRAM ANALYSIS
A) Build alternatives:
1. Fixed budget approach
2. Fixed effectiveness approach
3. Economic (expansion path) approach: (Construct alternatives as
cost-output/effectiveness relations or “response functions.”)
B) Modify existing alternatives: “Level the playing field”
4. Modified budget approach: GOTO 1.
5. Modified effectiveness approach: GOTO 2.
II) INTER-PROGRAM ANALYSIS
6. Opportunity cost/benefit approach
88 F. Melese
B* Cost($)
MoE(Utils) Benefit/cost
= MoE*/$B1
Cost($)
the British call “market testing”) competitions conducted under OMB Circular
A-76, which “requires . . . a structured process for [evaluating] the most efficient
and cost-effective method of performance for commercial activities” (OMB May
29, 2003). This process, sometimes referred to as Lowest Cost Technically Accept-
able (LCTA), involves four steps: 1) develop a Statement of Work (SOW) or
Performance Work Statement (PWS) to define desired performance/effectiveness;
2) construct the Most Efficient Organization for the in-house competitor; 3) issue
an Invitation for Bid (IFB) for well-defined, routine commercial activities; and
4) compare bids or proposals, and select the “least cost.” This offers an illustration
of the fixed effectiveness approach: minimizing costs of achieving a given MoE
(e.g. defined by the SOW or PWS).51
The test of maximum effectiveness for a given budget seems much less likely
to mislead the unwary.
(p. 167)
Alternative 1
A1
MoE(Utils)
Alternative 2
A2
Effectiveness
B2 B1 Budget ($)
Source Selection Decision: A2 for pessimistic budget; A1 for optimistic budget
This approach is illustrated in Figure 4.5 with two notional forecasted future
funding/budget levels: B1 (pessimistic) and B2 (optimistic).
The expansion paths for each vendor reveal combinations of attributes each ven-
dor can offer at different budget levels (e.g. pessimistic, optimistic, and most likely).
In Figure 4.5, each vendor’s expansion paths are transformed (through the govern-
ment’s utility function) into cost-utility or cost-effectiveness response functions (A1
and A2). These response functions reveal each vendor’s proposal under different
budget scenarios and represent the most general description of “alternative vendor
proposals” in the EEoA. Given a range of likely budgets for the program, the most
effective vendor over that range of budgets can be selected by the buyer.55
A key difference between traditional AoAs and this economic (expansion
path) approach is that instead of modeling competing vendors as points in
cost-effectiveness space, EEoA solicits vendor offers as functions of optimistic,
pessimistic, and most likely funding (budget) scenarios. This approach explicitly
addresses a key concern voiced by GAO:
Whereas the first three ways to structure an EEoA assume alternatives can be
constructed or built based on budget or effectiveness considerations (endogenous
alternatives), the last three approaches assume alternatives are exogenously deter-
mined, so that DMs must evaluate pre-specified alternatives. The most interesting
cases are where an alternative (vendor proposal) costs more but offers greater
utility (MoE), while others cost less and offer less utility (MoE).
A2
Effectiveness
A1
MoE(Utils)
A1* A2
Equal MoE
Effectiveness
A1
Cost
Budget ($)
it would cost to achieve the same target level of MoE.57 In Figure 4.7, vendor 1 is
preferred since the response (A1*) minimizes the budget required for the program.
A2 B2
A1 B1
Cost($) Cost($)
programs that must be sacrificed (e.g. a budget cut in Program B shifting the deci-
sion from B2 =>B1) for funds to be released to purchase greater marginal utility
(MoE) in the program under review (e.g. boosting the budget of program A to shift
the decision from A1 =>A2). “[T]he assessment should provide details as to how
excess funding . . . demands will be accommodated by reductions in other mission
areas, or in other . . . accounts” (DoD July 7, 2006, Section 3.2.2).
Alternatively, the DM can explore how much additional utility the extra money
might generate somewhere else if the low-cost alternative (A1 in program A)
was chosen. These are tough but useful questions that break through the sub-
optimization of most traditional AoAs. As a consequence, EEoA encourages
critical communication between different layers of the organization and a seam-
less interface between the requirements generation system, the acquisition system,
and PPBS.58
The bottom line is that it is often more transparent, efficient, and effective to
develop MoEs that are independent of costs. Equally important are theroles of budget
(funding) forecasts and opportunity costs in helping structure defense investment
decisions. Structuring an Economic Evaluation of Alternatives (EEoA) using one
of the six approaches summarized in Figure 4.9 can help achieve the primary goals
of defense acquisition, to lower costs and improve performance and schedules.
Dr. F. Melese
CAN YOU BUILD
Naval Postgraduate School
ALTERNATIVES?
[email protected]
NO:
Identify/Plot MOE &
YES
Cost/Budget of each
Alternative
IS THE DESIRED
MOE Can you Modify
SIMPLE TO Alternatives?
DEFINE/MEASURE?
NO: YES:
YES Do you have Level the NO
a BUDGET? Playing Field
Appendix 4.1
A simple example helps to illustrate the danger in using benefit/cost ratios with-
out anchoring either the budget (here measured in dollars) or a specified measure
of effectiveness (MoE) (here measured in “utils”). Suppose Alternative A1 in
Figure A4.1 costs $10 million and yields an MoE of 10 utils, while Alternative A2
costs $1 billion and yields an MoE of 900 utils. Applying a benefit/cost ratio cri-
terion indicates that A1 has a bigger “bang-for-the-buck” since it returns 1 util per
million dollars, while A2 only offers 0.9 utils per million dollars. (In Figure A4.1,
the slope of any ray from the origin represents the (constant) benefit/cost ratio
anywhere along that ray: the steeper the slope, the greater the benefit/cost
ratio.)
Using benefit/cost ratios to rank alternatives, however, is dangerous in this
case since it ignores the absolute magnitude of the costs involved. Now suppose
the situation was reversed and that A2 offered a higher benefit/cost ratio than
A1. Anyone who chooses A2 in a simple ranking of alternatives strictly on the
basis of “bang-for-the-buck,” ignoring affordability, would be in for an unpleasant
surprise: a $1 billion vs. $10 million commitment!
98 F. Melese
Is A1 really superior to A2 ?
Lesson: Danger in using Benefit/cost (Bang/Buck) or
Cost/Benefit (Buck/Bang) ratios without anchoring Budget or MoE
RANKING:
MoE(Utils)
Benefit/Cost (A1) > Benefit/Cost (A2)
A2
Effectiveness
Efficient Set
A1
Marginal Benefit/Marginal Cost
“The perceived benefits of the
higher priced proposal shall
merit the additional cost…”
www.arnet.gov FAR 15.101-1(2)c
Cost
Notes
1 This study often uses the term “analysis of alternatives” (AoA) in its broad, generic sense.
Although focused on defense acquisition, the results of the study apply to any public-
sector procurement where benefits can be quantified but not monetized. It should be clear in
context whenever the term AoA references major defense acquisition programs (MDAPs)
or the acquisition of major automated information systems (MAISs).
2 For examples of the theoretical foundations of AoAs, see Clemen (1996), French
(1986), Keeney (1982, 1992, 1994), Keeney and Raiffa (1976), or Kirkwood (1997).
3 As quoted in E. Newell “Business group urges reform of Pentagon contract requirements
process,” Government Executive, www.govexec.com [last accessed July 27, 2009].
4 The US Government Accountability Office (GAO) emphasizes that a major challenge
facing the Department of Defense (DoD) is to “achieve a balanced mix of weapon
systems that are affordable” (GAO, 2009, 5). Section E1.1.4, “Cost and Affordabil-
ity” of US DoD Directive 5000.01, states: “All participants in the acquisition system
shall recognize the reality of fiscal constraints. They shall view cost as an independent
variable, and the DoD Components shall plan programs based on realistic projections of
the dollars and manpower likely to be available in future years . . . The user shall address
affordability in establishing capability needs” [The Defense Acquisition System, DoDD
5000.01 (certified current as of November 20, 2007)].
5 “Affordability means conducting a program at a cost constrained by the maximum
resources the Department can allocate for that capability” [quote from “Better Buying
Economic evaluation of alternatives 99
Power” Memorandum for Acquisition Professionals from Ashton Carter, Undersecre-
tary of Defense for Acquisition, Technology and Logistics (dated September 14, 2010)].
Available at www.acq.osd.mil/docs/USD_ATL_Guidance_Memo_September_14_2010
_FINAL.PDF [last accessed November 11, 2014].
6 Instead of modeling decision alternatives from competing vendors as points in cost-
effectiveness space, EEoA generates vendor proposals as functions of optimistic,
pessimistic, and most likely funding (resource/budget) scenarios.
7 The DoD uses the Planning, Programming, Budgeting And Execution (PPBE) process
as its principal decision support system to provide the best possible mix of forces, equip-
ment, and support within fiscal constraints. Two other major decision support systems
complement the PPBE process: a Requirements Generation System to identify military
investment opportunities and the Defense Acquisition System (DAS) to develop and
procure new weapon systems. In explicitly addressing affordability up front, EEoA pro-
vides a unique opportunity to achieve a significant acquisition reform to lower costs and
to improve performance and schedules by tightly integrating requirements generation
and defense acquisition within the PPBE.
8 Based on US strategic guidance (the National Security Strategy, National Military Strat-
egy, Quadrennial Defense Review, Strategic Planning Guidance, etc.), the Requirements
Generation System reviews existing and proposed capabilities and identifies critical capa-
bility gaps. To fill those capability gaps, senior leadership examines the full range of “doc-
trine, organization, training, materiel, leadership and education, personnel and facilities”
(DOTMLPF). (JCIDS 2007, A-1; DoD 5000.2 2008, 14) The DAS provides principles
and policies that govern major defense acquisition decisions and milestones. To ensure
transparency and accountability and to promote efficiency and effectiveness, various
instructions (e.g. FAR; DFARS; DoD Directive 5000.01; DoD Instruction 5000.02; etc.)
specify statutory and regulatory reports (e.g. AoAs) and other information requirements
for each milestone and decision point of major defense investments.
9 Major Defense Acquisition Program (MDAP) and Major Acquisition Information Sys-
tems (MAIS) proposals that emerge from the planning process enter the DAS and are
incorporated into the programming phase of PPBE.
10 Translating the budget implications of these decisions into the usual Congressional
appropriation categories [military personnel, procurement, operations and maintenance
(O&M), military construction, etc.] generates the nation’s defense budget and FYDP.
While DoD’s biennial defense budget projects funding only two years into the future, it
includes more financial detail than the six-year POMs.
11 Office of Management and Budget (OMB) Circular A-11 titled Preparation and Sub-
mission of Budget Estimates is the official guidance on the preparation and submission
of budget estimates to Congress. The Army’s acquisition guidance emphasizes “the
requirement for presenting the full funding for an acquisition program—that is the
total cost [for] a given system as reflected in the most recent FYDP [. . .] pertains to
all acquisition programs” (DoA July 15, 1999, 41).
12 Current DoD directives require that an AoA be performed at key milestone decision
points (i.e. A, B, C) for all MDAPs and MAISs. “Affordability Analysis should be con-
ducted as early as possible in a system’s life cycle so that it can inform early capability
requirements trades and the selection of alternatives to be considered during the Analy-
sis of Alternatives (AoA)” (USD (AT&L), November 25, 2013, 120). The EEoA offers
a mechanism to embed affordability assessments into AoAs.
13 “A 2008 DoD directive established nine joint capability-area portfolios, each managed
by civilian and military co-leads [. . .]. However, without [. . .] control over resources
[funding/budgets], the department is at risk [. . .] of not knowing if its systems are being
developed within available resources [funding/budgets]” (GAO 2009, 11).
14 “Typically, the last analytical section of the AoA plan deals with the planned approach
for the cost-effectiveness comparisons of the study alternatives” (DoD July 7, 2006,
100 F. Melese
Section 3.3). Note that there is no mention of “affordability,” but instead only an
ex-post cost-effectiveness tradeoff that implies a concern for affordability. Moreover,
this tradeoff occurs at the end of a process in which alternatives under consid-
eration have been developed independently of any cost/budget/funding/affordability
constraint.
15 One of the multiple cost analysis tasks prescribed in the US Air Force AoA Handbook
is to: “provide funding and affordability constraints. . .” and “describe the appropriate
CAIV [cost as an independent variable] methodology for the AoA” (US Air Force 2008,
32 and 49). Yet in its explanation of the final vendor selection process, the Hand-
book states “There is no formula for doing this; it is an art whose practice benefits
from experience” (p. 41). In sharp contrast, Figure 4.9 at the end of this chapter offers
explicit guidance to help analysts and decision-makers structure an EEoA with a focus
on affordability.
16 “Since this assessment requires a DoD Component corporate perspective, the
affordability assessment should not be prepared by the program manager nor should
it rely too heavily on the user. It requires a higher-level perspective capable of balancing
budget tradeoffs (affordability) across a set of users” (DAU, 2006, Section 3.2.2).
17 A first step in the program’s affordability assessment is to portray the projected annual
modernization funding (Research, Development, Test and Evaluation (RDT&E) plus
procurement, measured as Total Obligation Authority (TOA)) in constant dollars for
the six-year programming period and for twelve years beyond. Similar funding streams
for other acquisition programs in the same mission area also would be included. What
remains to be determined is whether this projected funding growth is realistically afford-
able relative to the DoD component’s most likely overall funding. This chapter proposes
structuring the EEoAs not only for a “most likely” budget, but also for an “optimistic”
(higher) and “pessimistic” (lower) budget.
18 These dual constrained optimization approaches represent the first two of six ways
proposed in this study to structure an EEoA. See Figure 4.9: A decision map.
19 A Senate report (Acquisition Advisory Panel 2007) states that “Awards are made on the
basis of the solicitation of factors and sub-factors by a Source Selection Official who,
using his or her discretion and independent judgment [e.g. guided by an AoA], makes
a comparative assessment of [. . .] competing proposals, trading off relative benefits and
costs” (Chapter 1, Commercial Practices, 65). The Senate Committee’s recommenda-
tion is that “Regulatory guidance [. . .] be provided in the FAR [Federal Acquisition
Regulations] to [include] a minimum weight to be given to cost/price” (p. 102). Missing
in this discussion is an explicit and realistic acknowledgement of “affordability”—the
resources, funding, or budgets available for the procurement—something only indirectly
and implicitly addressed in assigning a “weight” to cost.
20 The first and second EEoA approaches are the standard fixed budget or fixed effec-
tiveness techniques. In the former case, the recommended alternative offers the greatest
effectiveness for a fixed budget and in the latter case the lowest cost to achieve the
desired level of military effectiveness. More interesting and realistic cases involve deci-
sions among alternatives that offer somewhat greater (less) effectiveness for somewhat
higher (lower) costs.
21 This is in the spirit of the Army’s Acquisition Procedures, which explicitly states that
“Cost as an Independent Variable (CAIV) applies to all defense acquisition programs
[. . .and] treats cost as an input to, rather than an output of, the materiel requirements
and acquisition processes.” The Army guidance emphasizes “CAIV is focused on [. . .]
meeting operational requirements with a solution that is affordable [. . . and that does]
not exceed cost constraints [and to] establish CAIV-based cost objectives (development,
procurement, and sustainment costs) early in the acquisition process. Moreover, the
“RFP must [. . .] solicit from potential suppliers an approach [. . .] for meeting CAIV
objectives” (DoA July 15, 1999, 63).
Economic evaluation of alternatives 101
22 This offers an alternate approach to defense investment decisions based on explicit
funding (resource/budget/affordability) scenarios that supports the “. . . long-standing
DoD policy to seek full funding of acquisition programs . . .” [Defense Acquisi-
tion Guidebook, Chapter 3.23: “Full Funding, Defense Acquisition” University,
https://akss.dau.mil/dag/DoD5000.asp?view=functional (accessed April 18, 2009)].
23 The cost of many programs reviewed by the GAO exceeded planned funding/budget
levels (GAO July 2, 2008).
24 The inverse cost/benefit ratio is also used and sometimes calculated as a “cost-per-kill.”
For an example, see the case of the Javelin anti-tank missile in Chapter 14.
25 Fisher (1965) argues that “numerous terms [. . .] convey the same general meaning
[. . .] ‘cost–benefit analysis,’ ‘cost-effectiveness analysis,’ ‘systems analysis,’ ‘oper-
ations analysis,’ etc. Because of such terminological confusion, [. . .] all of these
terms are rejected and ‘cost-utility analysis’ is employed instead” (p. 185). Although
the terms “cost–benefit” and “cost-effectiveness” are used interchangeably here, the
assumption throughout is that neither “benefits” nor “effectiveness” can be measured
in monetary terms.
26 “[E]ffectiveness analysis . . . is built upon the hierarchy of military worth, the assumed
scenarios and threats, and the nature of the selected alternatives . . . .In many AoAs
involving combat operations [a] typical classification would consist of four levels: (1)
system performance, based on analyses of individual components of each alternative or
threat system, (2) engagement, based on analyses of the interaction of a single alterna-
tive and a single threat system, and possibly the interactions of a few alternative systems
with a few threat systems, (3) mission, based on assessments of how well alternative
systems perform military missions in the context of many-on-many engagements, and
(4) campaign, based on how well alternative systems contribute to the overall military
campaign.” (DAU, Section 3.3.3.5).
27 Ironically, if a budget scenario is specified, there is no need to take the MCDM approach
that underpins most AoAs since it is possible to adopt the EEoA approach. The EEoA
approach constructs alternatives to fit within a budget envelope, converting the problem
into a straightforward MOE maximization (see next section).
28 A fourth example involves a recent landmark RAND study on capabilities-based plan-
ning. The author falls into the same trap. In a section entitled “Choosing Among
Options in a Portfolio,” Paul Davis (2002) develops “A Notional Scorecard for Assess-
ing Alternatives in a Portfolio Framework,” where alternatives differ in both their costs
and effectiveness. Nevertheless, the decision criterion recommended by the author
to select an alternative in “[t]he last column is the ratio of effectiveness over cost”
(pp. 45–6).
29 The authors continue: “Of course, if the ratios did not alter with changes in the scale
of achievement (or cost, the higher ratio would indicate the preferred system, no matter
what the scale [. . .]. But to assume that such ratios are constant is inadmissible some of
the time and hazardous the rest” (Hitch and McKean 1967, 167).
30 The third (and most general) approach to EEoA follows another of Hitch and McKean’s
(1967) recommendations: “As a starter [. . .] several budget sizes can be assumed. If
the same [alternative] is preferred for all [. . .] budgets, that system is dominant [. . .].
If the same [alternative] is not dominant the use of several [. . .] budgets is never-
theless an essential step, because it provides vital information to the decision maker”
(p. 176).
31 An additional (necessary and sufficient) condition is a linear, separable, additive
objective function.
32 “Usually, ratios are regarded as potentially misleading because they mask important
information” (DoD July 7, 2006, Section 3.3.1).
33 “Measures of Effectiveness [. . .] provide the details that allow the proficiency of
each alternative in performing the mission tasks to be quantified [. . .]. A measure of
102 F. Melese
performance typically is a quantitative measure of a system characteristic (e.g. range,
etc.) chosen to enable calculation of one or more measures of effectiveness . . . The cost
analysis normally is performed in parallel with the operational effectiveness analysis. It
is equal in importance in the overall AoA process [. . .]. [I]ts results are later combined
with the operational effectiveness analysis to portray cost-effectiveness comparisons”
(DoD July 7, 2006, Section 3.3.1).
34 It is often implicitly asserted certain key assumptions are satisfied for this utility func-
tion to be valid, such as “additive independence” (see French 1986; Keeney and Raiffa
1976; and Keeney 1994).
35 For example, one issue is that normalization is not necessary, and worse, can be mislead-
ing. The author is aware of several applications where relative weights were assigned
to different attributes based on soliciting acceptable tradeoffs among measurable char-
acteristics from decision-makers, but then later those same weights were applied to
normalized values of the characteristics to obtain MOEs (personal correspondence with
DoD officials).
36 This corresponds to the first of the six EEoA approaches presented in this chapter—the
fixed budget approach.
37 “In the European Union, a legislative package intended to simplify and modernize exist-
ing public procurement laws was recently adopted. As before, the new law allows for
two different award criteria: lowest cost and best economic value. The new provisions
require that the procurement authority publishes ex-ante the relative weighting of each
criteria used when best economic value is the basis for the award” (see EC 2004a and
EC 2004b).
38 “An AoA is an analytical Comparison of the operational Effectiveness, suitability,
and Life-cycle Cost of Alternatives that satisfy established Capability needs” [Defense
Acquisition Guidebook, Chapter 3.3: “Analysis of Alternatives,” Defense Acquisition
University. https://akss.dau.mil/dag/DoD5000.asp?view=functional (accessed April 18,
2009)].
39 According to US Federal Acquisition Regulations (FAR), “source selection” is the deci-
sion process used in competitive, negotiated contracting to select the proposal that
offers the “best value” to the government: “In different types of acquisition, the rel-
ative importance of cost or price may vary” (General Services Administration 2005,
Section 15.101).
40 Surprisingly, the author has continued to write prolifically in this field and continued to
promote this decision criterion, apparently never taking the time to reflect back on these
key observations.
41 Note that the slope of the straight-line indifference curves that reflect the DM’s relative
preference (or tradeoffs) between effectiveness (MOE) and cost are given by the relative
weights assigned, the ratio –w2/w1.
42 The Army’s Economic Analysis (EA) Manual states that “good EA should go beyond
the decision-making process and become an integral part of developing requirements in
the PPBE process” (DoA February, 2001, 12).
43 The weight on cost in the unconstrained optimization (MCDM approach) roughly
corresponds to the Lagrangian multiplier (shadow price) of the budget constraint in
the constrained optimization (the EEoA approach). (See Chapter 10 for an alternative
interpretation.)
44 In a section describing Building a Model, Fisher (1965) comments: “Since by definition
a model is an abstraction from reality, the model must be built on a set of assump-
tions. These assumptions must be made explicit. If they are not, this is to be regarded
as a defect of the model design” (p. 190). It is easy to inadvertently conceal the impor-
tance of affordability (budget/funding) issues in the MCDM decision sciences approach
that underpins many AoAs. In sharp contrast, the EEoA approach encourages explicit
affordability (budget/funding) assumptions.
Economic evaluation of alternatives 103
45 The OMB Circular A-109 for Major Systems Acquisition mentions the goal of
“design-to-cost.” “Under the CAIV philosophy, performance and schedule are consid-
ered dependent on the funds available for a specific program.”
46 The technical model can be found in Simon and Melese (2011). The static, determin-
istic, multi-stage, constrained-optimization, microeconomic production (procurement
auction) model developed in that study underpins the third, and most general, approach
to EEoA, called the “expansion path approach.”
47 In a footnote, Fisher (1971, 10) adds: “[T]he fixed budget situation is somewhat analo-
gous to the economic theory of consumer [optimization] . . . For a given level of income
[budget/funding] the consumer is assumed to behave in such a way that he maximizes
his utility.”
48 Note that we refer to the usual deterministic “utility function” that is a conventional
term used in the economics literature. This is in contrast to the way a utility function
is typically defined in the decision sciences and operations management literature as a
stochastic function. The “value function” described in the latter literature is similar to
our “utility function,” except that costs can enter into a value function and are excluded
from our utility function since they appear as part of the budget constraint.
49 Note that in the first and second EEoA approaches, since either the budget (funding
level) or MOE (level of effectiveness) is anchored in the constrained optimization, the
benefit/cost ratio decision criterion can be used as a decision rule in the selection pro-
cess. The steeper the slope from the origin through an alternative (A1, A2, A3), the
bigger the “bang-for-the-buck.”
50 An example of this second approach to EEoA is RAND Corporation’s AoA study for
the KC-135 recapitalization program: “in this AoA, the most ‘cost-effective’ alternative
[fleet] means precisely the alternative whose effectiveness meets the aerial refueling
requirement at the lowest cost” (Kennedy et al. 2006, 7).
51 Title 10, Subtitle A, Part IV, Chapter 146, Sec. 2462 of the US Code reads: “A func-
tion of the Department of Defense . . . may not be converted . . . to performance by a
contractor unless the conversion is based on the results of a public-private competition
that . . . examines the cost of performance of the function by Department of Defense
civilian employees and the cost of performance of the function by one or more con-
tractors to demonstrate whether converting to performance by a contractor will result in
savings to the Government over the life of the contract” (January 3, 2007). This law con-
cerning public-private competitions offers another illustration of the fixed effectiveness
approach, minimizing the cost of achieving a given MOE.
52 Budget announcements are analogous to an agency exploring in order to uncover its
true reservation price for the product (given competing demands for scarce budgets).
Adoption of this approach of evaluating vendor proposals under different “reservation
prices” could eventually lead to greater use of fixed-price contracts. Note also that
although requiring vendors to offer multiple bids based on different budget scenarios
increases vendors’ bidding costs, the government’s ability to make a superior vendor
selection with the extra information should more than compensate for any incremen-
tal price increase charged by vendors or for extra costs the government might incur to
compensate bidders for their higher bidding costs.
53 The vendors’ constrained optimizations define distinct expansion paths, one for each
vendor (for example, two vendor expansion paths are illustrated in Figure 4.5). From
the Envelope Theorem, the Lagrangian multiplier in each vendor’s constrained opti-
mization problem reveals the marginal product [the extra output or attribute mix they
are capable of producing) if the military buyer (DoD) relaxes its funding constraint (i.e.
corresponding to a more optimistic budget)].
54 The buyer reveals desired attributes/characteristics of the investment to the sellers, but
not the relative importance weights, and requests a single offer from each seller based on
a pre-specified budget (affordability) constraint, and then he selects the one he prefers
104 F. Melese
among the submitted offers. “We call this procedure a ‘single-bid auction with secret
scoring rule”’ (Asker and Cantillon 2004, 1). A more general construct is presented in
Simon and Melese (2011).
55 For example, if the PPBS process ultimately results in narrowing the budget/expenditure
constraint for the program around B2, then alternative 2 (A2) is selected. Note that if the
planning process allows the optimistic budget assumption, B1, to persist, then A1 would
have been selected, but when reality struck and budget B2 was all that was available,
choosing A1 would have turned out to be highly inferior decision (see Figure 4.5). This
EEoA approach relies upon and reinforces the importance of closely coupled iterative
interactions between the requirements generation system, defense acquisition system,
and PPBS.
56 Alternatively, different valuable uses for the money saved by choosing the lower-cost
alternative could be brought into the effectiveness calculation. Some will recognize this
search for the “next best alternative use of funds” as the standard economic definition
of opportunity costs. This sets the stage for the sixth way to structure an EEoA.
57 Note that the response function (expansion path) for vendor 1 includes points A1 and A1*.
58 Fisher (1965, 182) quotes Secretary of Defense Robert McNamara: “Suppose we have
two tactical aircraft which are identical in every important measure of performance
[MOE] except one—aircraft A can fly ten miles per hour faster than Aircraft B. Thus,
if we need about 1,000 aircraft, the total additional cost would be $10 million. If
we approach this problem from the viewpoint of a given amount of resources, the
additional combat effectiveness . . . of Aircraft A would have to be weighed against
the additional combat effectiveness which the same $10 million could produce if
applied to other defense purposes—more Aircraft B, more or better aircraft munitions,
or more ships, or even more military family housing . . . This kind of determina-
tion is the heart of the planning-programming-budgeting . . . problem with the Defense
Department.”
Bibliography
Acquisition Advisory Panel. Report to the US Congress, Commercial Practices (Chap-
ter 1), Jan. 2007, pp. 31–157. Available at www.acquisition.gov/comp/aap/24102_
GSA.pdf [last accessed November 22, 2014].
Asker, J., and E. Cantillon. December 2004. Equilibrium in Scoring Auctions. FEEM
(Fondazione Eni Enrico Mattei) Working Paper No. 148.04.
Beil, D. R. and L. M. Wein. 2003. “An Inverse-Optimization-Based Auction Mechanism to
Support a Multiattribute RFQ Process.” Management Science 49(11): 1529–1545.
Byrns, Jr., E. V., E. Corban, and S. A. Ingalls. 1995. “A Novel Cost–Benefit Analysis
for Evaluation of Complex Military Systems.” Acquisition Review Quarterly, Winter:
1–20.
Carter, A. Undersecretary of Defense for Acquisition, Technology and Logistics (AT&L).
2010. Memorandum for Acquisition Professionals. Office of Secretary of Defense
(OSD), Washington DC.
Che, Y.-K. 1993. “Design Competition through Multidimensional Auctions.” RAND Jour-
nal of Economics 24(4): 668–80.
CJCS. May 2007. Joint Capabilities Integration and Development System (JCIDS) (CJCSI
3170.01F). Washington, DC: Author.
Clemen, R. T. 1996. Making Hard Decisions: An Introduction to Decision Analysis (2nd
edition). Belmont, CA: Duxbury Press.
Corner, J. L. and C. W. Kirkwood. 1991. “Decision Analysis Applications in the Operations
Research Literature, 1970–1989.” Operations Research 39(2): 206–219.
Economic evaluation of alternatives 105
Davis, P. K. 2002. Analytic Architecture for Capabilities-Based Planning, Mission-System
Analysis, and Transformation. Santa Monica, CA: National Defense Research Institute,
RAND.
Defense Acquisition University. 2006. Defense Acquisition Guidebook. Section 3.2.2:
AoA Study Plan—Affordability. Available at https://dag.dau.mil/Pages/Default.aspx
[last accessed November 18, 2014].
Defense Federal Acquisition Regulation Supplement (DFARS) and Procedures, Guidance,
and Information (PGI). Available at www.acq.osd.mil/dpap/dars/dfarspgi/current/ [last
accessed September 12, 2014].
DoA. July 15, 1999. Pamphlet 70–3. Washington, DC: US Army Headquarters.
——. February 2001. Economic Analysis Manual. Arlington, VA: US Army Cost and
Economic Analysis Center.
DoD. May 22, 1984. The Planning, Programming, and Budgeting System (PPBS) (DoD
Directive 7045.14). Washington, DC: Author.
——. April 9, 1987. Implementation of the Planning, Programming, and Budgeting System
(PPBS) (DoD Directive 7045.07). Washington, DC: Author.
——. November 7, 1995. Economic Analysis for Decision making (DoD Directive 7041.3).
Washington, DC: Author.
——. 1998. DoD Guide to Cost as an Independent Variable (DA PAM 70–3). Washington,
DC: Author.
——. July 7, 2006. Department of Defense (DoD) Defense Acquisition Guidebook. Chapter
1.1 (Integration of the DoD Decision Support System); Chapter 1.2 (Planning, Program-
ming, Budgeting and Execution Process; Office of Management and Budget (OMB)
and the PPBE Process). Available at https://acc.dau.mil/CommunityBrowser.aspx [last
accessed March 16, 2009].
DoDD. November 20, 2007. Department of Defense Directive Number 5000.01, USD
(AT&L) (SUBJECT: The Defense Acquisition System).
——. January 25, 2013. Department of Defense Directive Number 7045.14, USD (C)
(SUBJECT: The Planning, Programming, Budgeting and Execution Process).
DoDI. December 8, 2008. Department of Defense Instruction Number 5000.02, USD
(AT&L) (SUBJECT: Operation of the Defense Acquisition System).
Dyer, J. S., and R. K. Sarin. 1979. “Measurable Multiattribute Value Functions.” Operations
Research 27: 810–822.
Dyer, J. S., P. C. Fishburn, R. E. Steuer, J. Wallenius, and S. Zionts. 1992. “Multiple Crite-
ria Decision Making, Multiattribute Utility Theory: The Next Ten Years.” Management
Science 38(5): 645–654.
Edwards, W., D. von Winterfeldt, and D. Moody. 1988. “Simplicity in Decision Analy-
sis.” In Decision Making: Descriptive, Normative, and Prescriptive Interactions, edited
by D. Bell, H. Raiffa, and A. Tversky, pp. 43l–464. Cambridge, England: Cambridge
University Press.
Eger, R. J., and A. L. Wilsker. 2007. “Cost Effectiveness Analysis and Transporta-
tion: Practices, Problems and Proposals.” Public Budgeting and Finance 27(1):
104–116.
European Commission. March 2004a. Coordinating the Procurement Procedures of Enti-
ties Operating in the Water, Energy, Transport and Postal Services Sectors (Article 55 of
the European Commission Directive 2004/17/EC of the European Parliament and of the
Council of March 31, 2004).
——. March 2004b. On the Coordination of Procedures for the Award of Public Works
Contracts, Public Supply Contracts and Public Service Contracts (Article 53 and 54 of
106 F. Melese
the European Commission Directive 2004/18/EC of the European Parliament and of the
Council of March 31, 2004).
Ewing, P. L., W. Tarantino, and G. S. Parnell. 2006. “Use of Decision Analysis in the
Army Base Realignment and Closure (BRAC) 2005 Military Value Analysis.” Decision
Analysis 3(1): 33–49.
Federal Acquisition Streamlining Act of 1994 (Public Law 103–355).
Federal Acquisition Reform Act of 1995 (Public Law 104–106).
Federal Activities Inventory Reform Act of 1998 (Public Law 105–270).
Fisher, G. H. 1965. “The Role of Cost-Utility Analysis in Program Budgeting.” In Pro-
gram Budgeting, edited by D. Novick, pp. 61–78. Cambridge, MA: Harvard University
Press.
——. 1971. Cost Considerations in Systems Analysis. New York: American Elsevier
Publishing.
Fowler, B. W. and P. P. Cason. 1998. “The Cost Exchange Ratio: A New Aggregate
Measure of Cost and Operational Effectiveness.” Military Operations Research 3(4):
57–64.
French, S. 1986. Decision Theory: An Introduction to the Mathematics of Rationality (Ellis
Horwood Series in Mathematics and Its Applications). Cambridge: Ellis Horwood.
Gansler, J. S. June 2003. Moving Toward Market-Based Government: The Changing Role
of Government as the Provider (New Ways to Manage Series). Arlington, VA: IBM
Endowment for the Business of Government.
GAO. July 2, 2008. Defense Acquisitions: A Knowledge-Based Funding Approach Could
Improve Major Weapon System Program Outcomes (GAO-08-619). Washington, DC:
Author.
——. September 25, 2008. Defense Acquisitions: Fundamental Changes are Needed
to Improve Weapon Program Outcomes (GAO-08-1159T). Testimony of M. Sullivan,
Director of Acquisition and Sourcing Management before the Subcommittee on Federal
Financial Management, US Senate. Washington, DC: Author.
——. December 2008. Defense Logistics: Improved Analysis and Cost Data Needed to
Evaluate the Cost-Effectiveness of Performance-Based Logistics (GAO-09-41). Wash-
ington, DC: Author.
——. March 18, 2009. Defense Acquisitions: DoD must Prioritize its Weapon System
Acquisition and Balance them with Available Resources (GAO-09-501T). Testimony of
M. Sullivan, Director of Acquisition and Sourcing Management before the Committee
on the Budget, House of Representatives. Washington, DC: Author.
General Services Administration, Department of Defense, National Aeronautics and Space
Administration. March 2005. Federal Acquisition Regulation (FAR). Washington, DC:
Author.
Golabi, K., C. W. Kirkwood, and A. Sicherman. 1981. “Selecting a Portfolio of Solar
Energy Projects Using Multiattribute Preference Theory.” Management Science 27:
174–189.
Gorman, W. 1980. “The Demand for Related Goods: A Possible Procedure for Analysing
Quality Differentials in the Egg Market.” Review of Economic Studies 47: 843–856.
Henry, M. H. and W. C. Hogan. 1995. “Cost and Effectiveness Integration.” Military
Operations Research Society’s PHALANX, 28/1.
Hitch, C. J. and R. N. McKean. 1967. The Economics of Defense in the Nuclear Age.
Cambridge, MA: Harvard University Press.
Howard, R. 1988. “Decision Analysis: Practice and Promise.” Management Science 34(6):
679–695.
Economic evaluation of alternatives 107
Hwang, C.-L, and K. Yoon. 1981. Multiple-Attribute Decision Making: Methods and
Applications. New York: Springer-Verlag.
Keeney, R. L. 1982. “Decision Analysis: An Overview.” Operations Research 30(5):
803–38.
——. 1992. Value-Focused Thinking: A Path to Creative Decision Making. Cambridge,
MA: Harvard University Press.
——. 1994. “Using Values in Operations Research.” Operations Research 42(5): 793–813.
Keeney, R. L. and H. Raiffa. 1976. Decisions with Multiple Objectives: Preferences and
Value Tradeoffs. New York: John Wiley & Sons.
Kennedy, M., L. H. Baldwin, M. Boito, K. M. Calef, J. S. Chow, J. Cornuet, M. Eisman, C.
Fitzmartin, J. R. Gebman, E. Ghashghai, J. Hagen, T. Hamilton, G. G. Hildebrandt, Y.
Kim, R. S. Leonard, R. Lewis, E. N. Loredo, D. M. Norton, D. T. Orletsky, H. S. Perdue,
R. A. Pyles, C. R. Roll, Jr., W. Stanley, J. Stillion, F. Timson and J. Tonkinson. 2006.
Analysis of Alternatives (AoA) for KC-135 Recapitalization. Santa Monica, CA: Project
Air Force, RAND.
Kirkwood, C. W. 1997. Strategic Decision Making: Multiobjective Decision Analysis with
Spreadsheets. Belmont, CA: Duxbury Press.
Lancaster, K. J. 1966a. “A New Approach to Consumer Theory.” Journal of Political
Economy 74(2): 132–157.
——. 1966b. “Change and Innovation in the Technology of Consumption.” American
Economic Review 56(2): 14–23.
——. 1971. Consumer Demand: A New Approach. New York: Columbia University Press.
——. 1979. Variety, Equity, and Efficiency. New York: Columbia University Press.
Larsen, R. and D. Buede. 2002. “Theoretical Framework for the Continuous Early
Validation Method.” Systems Engineering 5(3): 223–241.
Liberatore, M. J. 1987. “An Extension of the Analytic Hierarchy Process for Industrial
R&D Project Selection and Resource Allocation.” IEEE Transaction on Engineering
Management 34(1): 12–18.
Lorell, M. and J. Graser. 2001. An Overview of Acquisition Reform Cost Saving Estimates.
Santa Monica, CA: RAND Project Ari Force.
Melese, F. May 17, 2007. “Outsourcing for Optimal Results: Six Ways to Structure an
Analysis of Alternatives.” In Proceedings of the 4th Annual Acquisition Research Sym-
posium (Creating Synergy for Informed Change). Monterey, CA: Naval Postgraduate
School.
Melese, F. and D. Bonsper. December 1996. “Cost Integration and Normalization Issues.”
Military Operations Research Society’s PHALANX 29(4).
Melese, F., R. Franck, D. Angelis, and J. Dillard. January 2007. “Applying Insights from
Transaction Cost Economics to Improve Cost Estimates for Public Sector Purchases:
The Case of US Military Acquisition.” International Public Management Journal 10(4):
357–385.
Melese, F., J. Lowe, and M. Stroup. 1997. “Integrating Cost and Effectiveness: An Eco-
nomic Perspective.” Military Operations Research Society’s PHALANX 30(3): 13–17.
Michael, R. T. and G. S. Becker. 1973. “On the New Theory of Consumer Behavior.”
Swedish Journal of Economics 75(4): 378–396.
Murray, C. H. 2002. Executive Decision Making (6th edition). Newport, RI: National
Security Decision Making Department, US Naval War College.
Office of Aerospace Studies. 2008. Analysis of Alternatives (AoA) Handbook. Air Force
Materiel Command, Kirtland AFB, New Mexico. Available at http://www.ndia.org/
Divisions/Divisions/SystemsEngineering/Documents/Forms/AllItems.aspx?RootFolder
108 F. Melese
=https%3a%2f%2f2.zoppoz.workers.dev%3a443%2fhttp%2fwww%2endia%2eorg%2fDivisions%2fDivisions%2fSystemsEngine-
ering%2fDocuments%2fCommittees%2fMission%20Analysis%20Committee%2fSupp-
ort%20Documentation&FolderCTID=0x0120003F6EF07406A5714EB4933D54BF88
1105 [last accessed November 19, 2014].
OMB. April 5, 1976. Major Systems Acquisitions (OMB Circular A-109). Washington, DC:
Author.
——. October 29, 1992. Guidelines and Discount Rates for Benefit-Cost Analysis of
Federal Programs (OMB Circular A-94). Washington, DC: Author.
——. May 29, 2003. Performance of Commercial Activities (Public-Private Competitions)
(OMB Circular A-76, Revised). Washington, DC: Author.
Parkes, D. C. and J. Kalagnanam. 2005. “Models for Iterative Multiattribute Procurement
Auctions.” Management Science 51(3): 435–451.
Parnell, G. S. 2006. “Value-Focused Thinking Using Multiple Objective Decision Analy-
sis.” In Methods for Conducting Military Operational Analysis: Best Practices in Use
Throughout the Department of Defense, edited by A Loerch and L. Rainey, Chapter 19.
Alexandria, VA: Military Operations Research Society.
Pinker, A., A. H. Samuel, and R. Batcher. December 1995. “On Measures of Effectiveness.”
Military Operations Research Society’s PHALANX, 28/4.
Quade, E. S. 1989. Analysis for Public Decisions (3rd edition). Englewood Cliffs, NJ:
Prentice Hall.
Ramesh, R. and S. Zionts. 1997. “Multiple Criteria Decision Making.” In Encyclopedia
of Operations Research and Management Science, edited by S. Gass, and C. Harris,
pp. 538–543. Boston, MA: Kluwer.
Retchless, T., B. Golden, and E. Wasil. 2007. “Ranking US Army Generals of the 20th
Century: A Group Decision-Making Application of the Analytic Hierarchy Process.”
Interfaces 37(2): 163–175.
Ruefli, T. 1971. “PPBS—An Analytical Approach.” In Studies in Budgeting, edited by R.
Byrne. Amsterdam: North Holland.
——. 1974. “Analytic Models of Resource Allocation in Hierachical Multi-Level Systems.”
Socio-Economics Planning Science 8(6): 353–363.
Saaty. T. L. 1977. “A Scaling Method for Priorities in Hierarchical Structures.” Journal of
Mathematical Psychology 15(3): 234–281.
——. 1980. The Analytic Hierarchy Process. New York: McGraw-Hill.
Simon, J., and F. Melese. 2011. “A Multi-Attribute Sealed-Bid Procurement Auction
with Multiple Budgets for Government Vendor Selection.” Decision Analysis 8(3):
170–179.
Smith, J. E. and D. von Winterfeldt. 2004. “Decision Analysis in Management Science.”
Management Science 50(5): 561–574.
Stigler, G. J. 1945. “The Cost of Subsistence.” Journal of Farm Economics 27: 303–314.
Theil, H. 1952. “Qualities, Prices and Budget Enquiries.” Review of Economic Studies 19:
129–147.
Thomas, C. and R. Sheldon. 1999. “The Knee of the Curve.” Military Operations Research
4(2): 17–24.
US Code. January 2007. Title 10—Armed Forces, Subtitle A, Part IV, Chapter
146, Section 2462. US Government Printing Office. Available at www.gpo.gov/
fdsys/granule/USCODE-2006-title10/USCODE-2006-title10-subtitleA-partIV-chap146
[last accessed November 19, 2014].
US Congress. 1997. Clinger-Cohen Act—Information Technology Management Reform Act
of 1996 (Public Law 104–106). Washington, DC: Chief Information Officers Council.
Economic evaluation of alternatives 109
USD (AT&L). November 20, 2007. The Defense Acquisition System (DoD Directive
5000.01). Washington, DC: Author.
——. November 25, 2013. Operation of the Defense Acquisition System (DoD Instruc-
tion 5000.02). Washington, DC: Author. Available at: http://www.acq.osd.mil/osbp/
sbir/docs/500002_interim.pdf [last accessed February 9, 2015].
Vazsonyi, A. December/January 1995. “Multicriteria Decision Making.” Decision Line
26(1): 13–14.
Willard, D. 1998. “Cost-Effectiveness, CAIV, and the Knee of the Curve.” Military
Operations Research Society’s PHALANX 31(3).
Winterfeldt, D. and W. Edwards. 1986. Decision Analysis and Behavioral Research.
Cambridge, England: Cambridge University Press.
Zahedi, F. 1986. “The Analytic Hierarchy Process—A Survey of the Method and its
Applications.” Interfaces 16: 96–108.
Zionts, S. 1980. “Methods for Solving Management Problems Involving Multiple Objec-
tives.” In MCDM: Theory and Applications, edited by G. Fandal, and T. Gaul,
pp. 540–558. New York: Springer-Verlag.
This page intentionally left blank
Part II
5.1 Introduction
Military cost–benefit analysis (CBA) requires careful consideration of future
costs. Cost estimating is the process of collecting and analyzing historical data and
applying quantitative models, techniques, tools, and databases to predict the future
cost of an item, product, program or task. The term “cost analysis” is broadly used
to include not only the process of estimating (measuring) the cost of a project but
also the process of discovering, understanding, modeling and evaluating the rel-
evant information necessary to estimate the cost as well as the cost uncertainty
and risk.
Some might make a distinction between the cost estimating techniques and
analysis used for commercial off-the-shelf items (where a competitive market
determines the price), and products developed specifically for defense (where
there is no obvious market price). When comparing total life-cycle costs of alter-
natives, the techniques and practices presented in this chapter apply equally to
both cases. The only distinction is whether the acquisition price is determined by
the market or through contract negotiations.
It is important to note that cost analysis and estimating, as applied to finan-
cial management throughout government and industry, provides for the structured
collection, analysis, and presentation of life-cycle cost (LCC) data to assist in
decision-making for weapon systems capabilities throughout their useful life.
Examples of the components of LCC analysis conducted during each phase are:
There are three main applications of cost estimating in the Planning, Program-
ming, Budgeting, and Execution System (PPBES) process: (1) preparation and
justification of budgets, (2) making choices among alternatives, and (3) source
selection (contracting). Descriptions of these applications are given below.
114 D. I. Angelis, D. Nussbaum
• Preparation and justification of budgets. Cost estimates are performed within
all phases of the acquisition process to determine the budget amounts required
to fund the program (or “funding requirements”) throughout its life-cycle.
• Making choices among alternatives. This is further subdivided into the
following areas:
◦ acquisition cost forecasting;
◦ analysis of alternatives (AoA).
Cost estimates for the AoA requires forecasts of time-phased cash flows that
occur over the life of a system. This forms the basis for estimates of the total LCC
of a system. The LCC estimates are usually developed using constant dollars,1
and cost comparisons of alternatives are based on the present value2 of the total
LCC.
Cost estimating for budgeting assumes that the choice has been made and we are
interested in the near-term implementation cost of the chosen alternative. Budget
estimates must include the effects of expected inflation over the budget years and
are therefore stated in current (or nominal) dollars.
While the planning phase of PPBES is conducted with some resource con-
straints, it is not an intensive consumer of cost estimating support. The execution
phase is primarily focused on actual costs rather than estimated costs and the
difference between the two can be a source of much consternation, and is
one of the main reasons credible and accurate cost estimates are so important.
Additionally, the execution phase often uses the earned value management sys-
tem (EVMS), which has its roots in cost estimates, both to identify planned
vs actual cost variances, retrospectively, as well as to forecast cost variances,
prospectively.3
Develop
Develop Develop
Develop
Develop Develop
Develop
Develop
Develop Develop
Develop
Develop Conduct a Present Update the
Define the Develop the Develop Develop
Develop Document
'-------"' Conduct risk and estimate to estimate to
estimate's estimating the
sensitivity uncertainty management reflect actual
purpose plan Develop the point estimate
Obtain analysis for approval costs/changes
estimate and compare
the
data it to an independent
•
cost estimate
Aircraft
Level 1
System
Air Peculiar
Training Level 2
Vehicle Support
Equipment
Aircraft
Intermediate
Consumable Mat/ Sustaining
POL/Energy Other Maintenance
Repair Parts Support
Depot Indirect
Maintenance
Maintenance Support
Contractor
Operational
Support
Support
GROSS ESTIMATES
DETAILED ESTIMATES
PARAMETrC System l
EXTRAPOLATION
FROM ACTUALS
Development
Demonstration I
&l
ANALOGY ENGINEERING
The Director, CAPE is the principal staff assistant to the Secretary of Defense
for Cost Assessment and Program Evaluation. The Director’s principal
responsibilities include:
• analyzing and evaluating plans, programs, and budgets in relation to
US defense objectives, projected threats, allied contributions, estimated
costs, and resource constraints;
• reviewing, analyzing, and evaluating programs, including classified pro-
grams, for executing approved policies;
• providing leadership in developing and promoting improved analyti-
cal tools and methods for analyzing national security planning and the
allocation of resources;
• ensuring that the costs of DoD programs, including classified programs,
are presented accurately and completely;
• assessing effects of DoD spending on the US economy, and evaluating
alternative policies to ensure that DoD programs can be implemented
efficiently.
Each service has, as part of its Assistant Secretary for Financial Management
and Comptrollership, its own service cost center, as follows:
• In the Army Material Command there are cost estimating organizations within:
◦ Communications and Electronics Command (CECOM);
◦ Aviation and Missile Command (AMCOM);
◦ Tank and Automotive Command (TACOM).
• In the Air Force Systems Command, there are cost estimating organizations
within:
◦ Space and Missile Command (SMC);
◦ Aeronautical Systems Center (ASC);
◦ Electronics Systems Center (ESC).
• Within the Navy and Marine Corps, there are cost estimating organizations
within:
◦ Naval Air Systems Command (NAVAIR);
◦ Naval Sea Systems Command (NAVSEA);
◦ Space and Naval Warfare Systems Command (SPAWAR);
◦ Marine Corps Systems Command (MCSC).
What are known as Budget Item Justification sheets in the OSD budget can also
provide useful information on cost and schedule for major weapon systems, albeit
at an aggregated level.
A critical step when using historical data is to first “normalize” the data before
using it to estimate future costs. Appendix 5.1 offers a comprehensive overview
of data challenges and the importance of data normalization in cost estimates.
These approaches rely on statistical properties and logical relationships, and are
largely based on historical data. One view of these three approaches is often heard
at cost estimating conferences, such as the one put on annually by the International
Society of Cost estimating and Analysis (ICEAA, formerly called the Society of
Cost Estimating and Analysis (SCEA)):
The relationships between these variables and cost are frequently determined
using a statistical technique, such as ordinary least squares (OLS) regression.
The flowchart in Figure 5.5 shows the steps taken in developing a CER.
An example, taken from the Constructive Cost Model (COCOMO)15 used to
estimate the cost of developing software, is:
MM = 2.4(KDSI)∗∗ 1.05
Define Estimating
Collect
“Hypothesis”
“Relationship”
Data Evaluate and
Analyze Data
Normalize Data
for Candidate
Relationships
Perform Statistical
(Regression)
Analysis
Test
Relationships
Select Cost
Estimating
Relationship
must be considered to apply the method correctly. The user (i.e. cost estimator)
must also understand design tradeoffs and the current state of technology.
A third disadvantage is that the system must be well defined, i.e. there is little
allowance for unknown factors. For example, a component’s cost must be esti-
mated even though that component might represent a first-of-a-kind technology.
Finally, the user of the bottom-up approach must have access to, or maintain
an extensive and detailed database of, development, production, operating, and
support costs for the particular technology.
Table 5.1 summarizes the advantages and disadvantages of each approach. A
fourth approach, “expert judgment,” is included, which can be used when histori-
cal data is not available. Note, however, that there are no statistics associated with
an estimate based upon this method.
Once these critical questions have been answered, the greater challenge of risk
management is to address and control the following three issues articulated by
Haimes (1991):
We will focus on the first set of questions when assessing cost risk. For cost
estimates, we pose the following three questions:
For a cost estimate, the answer to the first question is: the actual cost of the pro-
gram exceeds the program budget (a cost overrun). Note that while it is certainly
possible that the actual cost of the program is less than the budget, this is not a
“risk” since it is a desirable outcome. In addition, there are a number of reasons
why cost overruns are much more likely than cost underruns, as data and analysis
from RAND has shown (see Figure 5.6). The next section reviews these reasons
in examining biases in cost estimating.
128 D. I. Angelis, D. Nussbaum
16
14
12
10
Frequency
0
0.75–1.00 1.00–1.25 1.25–1.50 1.50–1.75 1.75–2.00 2.00–2.25 2.25–2.50
CGF range
Figure 5.6 Distribution of cost growth (difference between estimate and actual costs).
Source: Arena et al. (2006).
The answer to the second question (what is the probability of a cost overrun?)
depends on how much money is programmed in the budget. The real question is:
how much money should we put in the budget? We could start with the mean esti-
mate for the cost of the project. If we budget the mean and the cost distribution
is approximately normal, the risk of exceeding the budget (for any particular pro-
gram) is about 50 percent. In other words, the risk of a cost overrun is 50 percent,
not a very attractive number if you are the program manager.
If we budget less than the mean for each program in a portfolio of programs,
there is a higher risk of individual program cost overruns (not necessarily for the
total portfolio), but it permits funding more programs in a portfolio. Of course,
funding a program at such a high level of risk may imply to program managers
that a cost overrun is expected. It also reduces the amount of funds available for
contract negotiations, management reserves, and performance incentives.
We can reduce the risk of cost overruns if we budget an amount greater than the
mean. This will lower the risk of an individual program exceeding its budget, but
it may leave dollars on the table if actual costs are less than the amount budgeted.
Of course, this less risky approach means fewer programs can be funded in the
portfolio. It also provides less incentive to program managers to control costs, but
it increases the funds available for management reserves, contract negotiations,
and performance incentives.
The third question (what are the consequences of a cost overrun?) may provide
some insight into why defense programs are more often than not underfunded
(resulting in cost overruns). What are the consequences of a cost overrun in a
defense program? Cost overruns tend to result in either a reduction in the number
of units purchased (for example, cost overruns in the B-2 program forced the US
Cost analysis 129
Air Force to cut the number of aircraft procured from 132 to 21) or an extension
of the schedule, so that less units are procured each year. Understanding produc-
tion costs is essential to evaluating the impact of such actions. Often, cutting the
quantity of units produced increases the unit cost, resulting in less capability, as
fewer units can be purchased for a given budget.
Funding shortfalls may also lead to a reduction in scope or capabilities and
sometimes even program cancellations, but such outcomes are less frequent. Much
more likely is that one program will be cut to fund another (higher priority)
program to maintain a robust portfolio of capabilities. While some of these conse-
quences, such as cancellation of the program, are highly undesirable, historically
they have been rare and as such present less risk. As a result, the US DoD has been
willing to accept the risk of cost overruns, perhaps judging that the consequences
are manageable and thus not very risky.
It is common for CBA to focus on a comparison of the capabilities (effective-
ness) and total cost of each alternative. It is also useful to compare the cost risk of
different alternatives. To do this, we need the cost distribution for each alternative.
In some cases, one alternative dominates the other, meaning that regardless of the
funding (budget) level, the risk of a cost overrun is always lower for the dominant
alternative. In other cases, the comparison is not as clear. Consider the alternatives
A and B shown in Figure 5.7.
Alternative A has a lower cost risk up to a certain funding level (US$1,770),
and alternative B has a lower risk beyond that level. Figure 5.8 illustrates how risk
curves can be used to highlight this information and help decision-makers make
better choices.
4.5
4.0
3.5
Alternative A
3.0
Alternative B
Probability
2.5
2.0
1.5
1.0
0.5
0.0
0
2000
4000
6000
8000
10000
12000
14000
dollars
1.0
0.8
Alternative A
Alternative B
0.6
Probability
0.4
0.2
0.0
0
2000
4000
6000
8000
10000
12000
14000
dollars
5.10 Summary
In this summary section, we provide recommendations on future research direc-
tions. In many disciplines, research is driven by both demand-side and supply-side
132 D. I. Angelis, D. Nussbaum
considerations. The demand-side represents the research which is done in response
to what users of the discipline are demanding, while the supply-side represents
those new techniques driven by researchers internal to the profession.
In cost estimating, the impetus for research is driven almost exclusively from
the demand-side, which encompasses changes in various ministries of defense
and military acquisition strategies. From this perspective, the most important
developments in the US are:
There are also some internal challenges within the professional community to
make sure that the next generation of cost estimators, with the right academic
background, are properly identified and developed in order to provide a better
level of service to the next generation of defense leadership.
Finally, there is the ever-present challenge of data availability. In this regard,
there have been some major improvements in data collection in the US over
the past 15 years, such as the development and maintenance of extensive data
collections like VAMOSC17 and DCARC18 (see Appendix 5.1). More recently,
OSD/CAPE has been working to consolidate all DoD cost data sources into the
Cost Assessment Data Enterprise, a unified initiative to collect, organize and use
cost data across the cost community. These and similar databases must be contin-
uously updated and maintained to support future cost estimates that are a critical
input in military CBA.
Bibliography
Angelis, Diana I., J. Dillard, R. Franck, and F. Melese. 2007. Applying Insights from Trans-
action Cost Economics (TCE) to Improve DoD Cost Estimation. Acquisition Research
Sponsored Report Series, NPS-GSBPP-07-008, Naval Postgraduate School.
Arena, Mark V., R. S. Leonard, S. E. Murray, and O. Younossi. 2006. Historical Cost
Growth of Completed Weapon System Programs. Santa Monica, CA: Rand Corporation.
Defense Acquisition University (DAU). Weapon Systems Acquisition Reform Act of 2009
(WSARA 09), DTM 09-027.
Fisher, Gene H. 2007. Cost Considerations in Systems Analysis. Santa Monica, CA: Rand
Corporation.
Cost analysis 135
Garvey, Paul R. 2000. Probability Methods for Cost Uncertainty Analysis. New York:
Marcel Dekker.
Government Accountability Office. 2009. Cost Estimating and Assessment Guide. Avail-
able at www.gao.gov/new.items/d071134sp.pdf
Haimes, Y. Y. 1991. Total Risk Management. Risk Analysis 11(2): 169–171.
Kaplan, S. and B.J. Garrick. 1981. On The Quantitative Definition of Risk. Risk Analysis
1: 11–27.
McNicol, D. L. 2005. Cost Growth in Major Weapon Procurement Programs (2nd edition,
IDA Paper No. P-3832). Alexandria, VA: Institute for Defense Analyses.
Melese, F., R. Franck, D. Angelis, and J. Dillard. 2007. Applying Transaction Cost
to Improve Cost Estimates for Public Sector Purchases: The Case of US Military
Acquisition. International Public Management Journal 10(4): 357–385.
Office of the Secretary of Defense. 1992. Cost Analysis Improvement Group Operating and
Support Cost-Estimating Guide. Available from www.dtic.mil/pae.
Shrull, D., ed. 1998. The Cost Analysis Improvement Group: A History. McLean, VA:
Logistics Management Institute (LMI).
Stewart, R. D. 1991. Cost Estimating (2nd edition). New York: John Wiley and Sons, Inc.,
Chapter 4.
US Air Force. 2007. Cost Risk and Uncertainty Analysis Handbook. Available at https://acc.
dau.mil/CommunityBrowser.aspx?id=316093
US Department of Defense. 2005. Work Breakdown Structures for Defense Materiel Items
(MIL-HDBK-881A). Available at www.everyspec.com.
6 Advances in cost estimating
A demonstration of advanced machine
learning techniques for cost estimation
Bohdan L. Kaluzny
6.1 Introduction
6.1.1 Background
The International Cost Estimating and Analysis Association (ICEAA) Cost Esti-
mating Handbook of Knowledge (CEBoK)1 categorizes different costing tech-
niques commonly applied for defense cost estimation. These include costing by
analogy, costing by parametric approaches, engineering build-up approaches, and
extrapolation from actuals. Figure 6.1 shows when each of these costing tech-
niques are commonly applied relative to a program’s life cycle. The figure shows
the approximate proportional usage of cost techniques by phase. At the beginning
of a program there is more emphasis on using analogies and parametrics. As the
program matures, it becomes more defined, additional data is collected, and the
estimates get more detailed. A challenge to cost analysts is developing good cost
estimates during early design phases using costing by analogy and parametrics.
Cost estimation by analogy is typically accomplished by forecasting the cost of
a new system based on the historical cost of similar or analogous systems. This
requires a reasonable correlation between the new and historical system. The cost
of the historical system is adjusted by undertaking a technical evaluation of the
differences between the systems, deducting the cost of components that are not
comparable to the new design and adding estimated costs of the new components.
Usually subject matter experts are required to make a subjective evaluation of
the differences between the new system of interest and the historical system, and
subjectively chosen complexity factors are often used to adjust the analogous sys-
tem’s cost to produce an estimate. This subjectivity is a disadvantage of traditional
analogy methods.
Advances in cost estimating 137
Program Life
Program Life
Program Life Program Life
Program Life
6.1.3 Outline
The remainder of this chapter is structured as follows. Section 6.2 provides a very
brief description of the dataset used to exemplify the machine learning algorithms
discussed herein. Section 6.3 presents a machine learning algorithm for costing by
analogy and Section 6.4 exhibits the application of M5 model trees to parametric
cost estimation. Section 6.5 provides a combined discussion. Finally, the chapter
concludes in Section 6.6.
Advances in cost estimating 139
6.2 Data
Data mining approaches are data intensive. Both the quality and quantity of the
data drive the quality of the outputs of machine learning algorithms. While the
appropriate level of data quality and quantity will be case-specific, to illustrate
the application of machine learning algorithms a database of 59 naval ships com-
piled by the NATO RTO SAS-076 working group is considered. This dataset
is subsequently referred to as the SAS-076 ship dataset. The database captures
over 100 descriptive, technical, and cost attributes for 59 naval ships from 18
classes (of ships), from seven nations, commissioned between 1950 and 2010.
The normalized costs also span multiple orders of magnitude.
The SAS-076 ship dataset contains military and civilian auxiliary (coast guard
or similar) vessels that were judged to be more or less similar to Landing Platform
Dock (LPD) ships. A benefit of employing machine learning algorithms is that
they allow for, or can excel on, a greater variability in the input dataset—variability
that could be questioned when using traditional costing by analogy or parametric
approaches. The fidelity of costing by analogy and parametric approaches also
depends on the size of the input dataset. In this case, the availability of ship cost
data limited the size of the SAS-076 ship dataset. While the decision as to which
ships were or were not included in the dataset was subjective, it was primarily
driven by the availability of cost data.
Using machine learning terminology, each ship is an instance and together the
ships listed in the SAS-076 dataset form a set of instances for input to a machine
learning scheme. Instances are things that are to be classified, associated or clus-
tered. Each instance is characterized by the values of a set of predetermined
features or attributes. The matrix of instances versus attributes forms the input
dataset (e.g. the SAS-076 ship dataset). The value of an attribute can be numeric,
nominal, ordinal, interval or ratio. Numeric attributes measure numbers (real or
integers) and are not necessarily continuous. Nominal, or categorical, attributes
take on values from a prespecified, finite set of possibilities. These can take the
form of distinct symbols or even words like “yes” or “red.” Ordinal quantities are
nominal quantities with a rank order (e.g. fast > slow > very slow). Interval quan-
tities are ordinal but also measured in fixed and equal widths (e.g. temperature in
degrees Celsius). Ratio quantities are relative; specified by an inherent zero-point
(e.g. the distance between two objects).
Descriptive, technical, and cost data were gathered for each of the ships and
broken down into categories, as shown in Table 6.1. Attribute units are expressed
by either nominal values (e.g. “fixed pitch,” “controlled pitch,” “yes,” “no”), or
by numerical units such as meters, megawatts, knots, hours, nautical miles, etc.
Unknown or missing data is permissible and is denoted by empty or “?” entries.
I Description 6
II Construction 8
III Dimensions 5
IV Performance 8
V Propulsion 9
VI Electrical power generation 3
VII Lift capacity 35
VIII Flight deck 19
IX Armament 13
X Countermeasures 5
XI Radars/sonars/sensors 13
XII Combat data systems 1
XIII Weapons control systems 1
XIV Other capabilities 7
XV Cost data 3
Total 136
analogy approaches) in quantifying the cost of the technical and other differences
between the historical system and the new system. The approach also considers
multiple analogous systems rather than just one.
For numeric attributes, dijk is normalized to lie in the [0, 1] range with dijk = 1 indi-
cating that ships i and j lie at opposite ends of the observed spectrum for attribute
k (e.g. shortest and longest length ships), and dijk = 0 indicating that the ships are
the same with respect to attribute k. For nominal attributes, dijk is binary—set to 0
if ship i and j are the same with respect to attribute k, and 1 if they are not.
A variety of distance metrics can be used to calculate similarity of two ships
based on the attribute distances dijk . Using a simple Euclidean distance metric, the
aggregate distance between two ships i and j , is expressed as
di j = di2j k , (6.2)
k∈A
Cj 1
C̃i = · 1 (6.3)
d2
j =i i j di2j
j =i
is the estimated cost of ship i . Figure 6.2 plots the actual ship costs, Ci , versus the
costs predicted, C̃i , by the analogy method using hierarchical clustering based on
a simple distance metric. The solid line running at 45 degrees represents perfect
prediction. The measures of worth of the analogy method using hierarchical clus-
tering analysis (simple distance metric) for predicting ship costs are R = 0.48 and
R 2 = 0.23. The mean absolute percentage error is 49 percent indicating that this
raw approach does poorly in learning the known ship costs and would likely yield
poor cost predictions.
An assumption in the above approach is that all attributes are of equal impor-
tance; the (normalized) differences or similarities for each attribute contribute
equally to the measure of similarity between ships. To potentially improve the
142 B. L. Kaluzny
700
600
Predicted Cost
500
400
300
200
100
0
0 100 200 300 400 500 600 700
Actual Cost
Figure 6.2 Hierarchical clustering with simple distance function: correlation plot of
actual versus predicted ship costs.
Using a weighted Euclidean distance metric, the aggregate distance between two
ships i and j , is expressed as
d̂i j = (wk · di j k )2 , (6.5)
k∈A
where k∈A wk = 1 and wk ≥ 0 for all k. As before, let C j be the known cost of
ship j , then
Cj 1
Ĉi = · 1 (6.6)
2
j =i d̂i j d̂i2j
j =i
is the estimated cost of ship i using weighted attributes. The optimal allocation of
weights is determined by minimizing the prediction error for the known ships,
2
minimize Ci − Ĉi . (6.7)
i
A2 = 0.204 × length
+ 0.196 × beam width
+ 0.183 × vehicle space
+ 0.18 × no. of expeditionary fighting vehicles
+ 0.165 × 1 if has a well deck, otherwise 0
+ 0.165 × width of the well deck
+ 0.164 × length of the well deck
+ 0.159 × no. of large personnel landing craft
(6.8)
+ 0.156 × no. of Chinook helicopters supported
+ 0.155 × full load displacement
+ 0.153 × no. of combat data systems
+ 0.150 × light load displacement
+ 0.144 × well deck capacity
+ 0.143 × no. of elevators
+ 0.142 × vehicle fuel capacity
etc.
A1 17 17 0
A2 12 29 0.452
A3 11 41 0
A4 9 49 0
A5 8 57 0.334
A6 7 64 0
A7 6 69 0
A8 5 74 0
A9 4 78 0
A10 3 81 0.214
A11 3 85
A12 3 88
A13 3 90
A14 2 92
A15 2 94
A16 1 95
Only macro-attributes A2, A5, and A10 have non-zero weights. This is a typi-
cal extreme output of mathematical optimization software. There may exist other
optimal solutions with other non-zero macro-attributes weights.
The hierarchical cluster analysis using the weighted distance function on the top
ten macro-attributes determined by PCA (effectively the three macro-attributes
assigned non-zero weight) is visualized in Figure 6.3. The figure illustrates the
resulting dendrogram indicating the relative grouping of ships. As expected, ships
within the same class (but different rank) are closely grouped together. The
weighted distance function is used to determine the distances of each ship to a
particular ship. For example, plugging in the macro-attributes weights of Table 6.2
into Equation (6.5) results in the list of computed distances to the Rotterdam LPD
presented in Table 6.3.
Figure 6.4 plots the actual ship costs, Ci , versus the costs predicted, Ĉi , by the
analogy method using hierarchical clustering based on a weighted distance met-
ric (the figure is based on Table 6.3 and Equation (6.6)). The solid line running
at 45 degrees represents perfect prediction. The measures of worth of the analogy
method via hierarchical clustering analysis (weighted distance matrix) for predict-
ing ship costs are R = 0.93 and R 2 = 0.86. The latter coefficient of determination
indicates that 86 percent of the total variation in the ship costs can be explained
by an average cost of the known ships weighted by an optimized distance met-
ric. The mean absolute percent error is 16 percent, a significant improvement
over the hierarchical clustering based on the simple distance metric that yielded a
49 percent error.
Hermitage
Alamo
Monticello
‘Spiegel Grove’
‘Point Defiance’
‘Fort Snelling’
‘Plymouth Rock’
Thomaston
‘Gunston Hall’
Comstock
‘Fort McHenry’
Germantown
‘Whidbey Island’
Rushmore
Ashland
Tortuga
‘Fort Fisher’
‘Mount Vernon’
Pensacola
Portland
Anchorage
Denver
Juneau
Duluth
Cleveland
Dubuque
Ogden
Austin
Ponce
Trenton
Nashville
Coronado
Shreveport
‘La Salle’
Vancouver
Raleigh
‘Pearl Harbour’
‘Oak Hill’
‘Carter Hall’
‘Harpers Ferry’
Ocean
BPC3
Mistral
Tonnerre
Svalbard
‘Johan de Witt’
Preserver
Protecteur
Oden
Atle
Carlskrona
Siroco
‘Cardigan Bay’
‘Mounts Bay’
‘Lyme Bay’
‘Largs Bay’
Bulwark
Albion
Rotterdam
Figure 6.3 Dendrogram illustrating the arrangement of the clusters produced by the hierarchical clustering of ships (weighted distance function).
146 B. L. Kaluzny
Table 6.3 Weighted distance of the Rotterdam LPD to ships in the SAS-076 dataset
700
600
Predicted Cost
500
400
300
200
100
0
0 100 200 300 400 500 600 700
Actual Cost
Figure 6.4 Weighted hierarchical clustering: correlation plot of actual versus predicted
ship costs.
The standard deviation, indicating the variation that the predictions have from
the actual costs, is computed as
n
(yi − ŷi )2
i=1
. (6.9)
n
Advances in cost estimating 147
To summarize, as an analogy costing approach, hierarchical agglomerative clus-
ter analysis, principal component analysis, and nonlinear optimization were used
to calculate a matrix of distances among the systems of a dataset. These distances
can then be used to predict the cost of new systems. The fidelity of the estimation
model is very dependent the dataset. It is a “top down” approach applicable in
early design phases of the procurement cycle.
• it can capture major portions of an estimate quickly and with limited infor-
mation (which is often all that is available in early phases of the procurement
cycle);
• the CER is objective—it is based on consistent and quantitative inputs;
• once the CER is established it is easy to perform sensitivity analyses (deter-
mine the change in cost subject to changes in the independent input variables);
and
• CERs established using regression analyses include standard tests of validity,
including a coefficient of correlation indicating the strength of association
between the independent variables and the dependent variable in the CER.
x2
LM3 LM2 x2 > 2
4
Yes No
3
x1 > 2.5 x1 < 4
LM1 Yes No Yes No
2
LM6 LM3 LM4
x2 < 3.5 x2 < 1
1 LM4 Yes No Yes No
LM5 LM1 LM2 LM5 LM6
1 2 3 4 5 6 x1
Y (output)
6.4.1.1 Discussion
Regression trees are more accurate than straightforward linear regression, but the
trees are often cumbersome and difficult to interpret (Witten and Frank 2005).
Model trees are more sophisticated than regression trees as they approximate con-
tinuous functions by linear “patches”—M5 model trees are analogous to piecewise
linear functions. M5 model trees have an advantage over regression trees with
respect to compactness and prediction accuracy due to the ability of model trees
to exploit local linearity in the data. Regression trees will not predict values lying
outside the range of learned training cases, while M5 model trees can extrapolate.
The M5 model tree algorithm is optimized both to learn known cases and predict
unknown cases. The trees are also smaller, easier to understand, and their average
error values on the training data are lower.
M5 model trees tackle the difficulties in considering a multitude of cost-driving
factors; they determine the right set of independent variables (ship attributes) by
construction. The effect of the M5 system is somewhat similar to the use of indica-
tor variables in standard linear regression analysis, with the key difference being
that appropriate indicators are identified and included based on an optimization
algorithm.
M5 model trees have been shown to excel even when limited data is available
(Wang and Witten 1997) or learn efficiently from large datasets. They can handle
datasets which include systems with notable differences, missing data, and noise
(as is the case for the SAS-076 ship dataset). The decision tree can branch on any
variable type: nominal (e.g. military versus non-military) or numeric (e.g. tonnage
less than 15,000 or greater than 15,000).
The use of M5 model trees for numeric prediction has increased since com-
prehensive descriptions, implementations, and refinements of Quinlan’s method
became available (Wang and Witten 1997; Malerba et al. 2004; Torgo 2000, 2002;
Dobra 2002). Recently Chen (2006) and Sharma and Litoriya (2012) discussed
the benefits of the system for estimating the cost of software development.
There are many other types of prediction methods that employ data mining tech-
niques. For example, neural networks are commonly used for predicting numeric
quantities, but suffer from the disadvantage of producing opaque structures that
150 B. L. Kaluzny
mask the nature of the solution. Their arbitrary internal representation means that
there can be variations between networks trained on the same data. By comparison,
the M5 system is transparent and the model tree construction is repeatable. The
final tree and linear regression models are straightforward and clear.
# of LCAC
< 2 or ? ≥2
0 ≥ 1 or ? ≥6 ≤5
≤1 ≥2 ≤1 ≥2
LM5 LM6
LM1 LM2
LM3 LM4
LM5 LM6
LM7
Log(Cost) = 8.4544
− 0.0202 × rank in class
+ 0.0012 × length (m)
+ 0.0001 × crew size
+ 0.0216 × no. of LCAC
+ 0.0155 × no. of elevators
ship increases as the length, number of elevators, crew size, number of LCAC sup-
ported, or number of torpedo decoy systems increase(s). The regression models
also predict a shipbuilding learning curve as the cost of constructing a ship
decreases as a function of the ship’s rank in class. However, the negative coeffi-
cient preceding the number of ship propeller shafts in the regression models LM1,
LM2, and LM3 is counter-intuitive. It seems unlikely that a ship will cost less if
152 B. L. Kaluzny
Table 6.5 Statistics of attributes used in the M5 model tree linear regression models
Rank 1 3 3.6 12
Length 103.7 173.8 170.2 203.4
Range (total distance) 6,000 8,000 9,933 30,000
Crew size 15 352 321 491
No. of propeller shafts 0 2 1.7 4
No. of LCAC 0 2 1.9 4
No. of elevators 0 0 0.6 3
No. of helicopters supported 0 6 5 18
No. of torpedo decoy systems 0 0 0.5 1
the number of propeller shafts is increased. The SAS-076 ship data explains this
anomaly: all but one of the ships captured in the SAS-076 ship dataset have at
most two propeller shafts; only Sweden’s Atle icebreaker has four. The Atle is
also the lowest costing ship. Since there is no constraint on keeping the coeffi-
cients positive, the linear regression model LM2 uses the propeller shaft attribute
to model Atle’s low cost. The combination of Atle’s low cost and unique propeller
shaft attribute drives the negative coefficient sign. The smoothing process of the
M5 model tree construction spreads this feature to the adjacent linear regression
models LM1 and LM3. The M5 model can be potentially adjusted in an attempt to
remove such anomalies by disabling the particular attribute (number of propeller
shafts); however, there is no guarantee that the regenerated model will not sub-
stitute this attribute with another, also allocated a negative coefficient. Similarly,
removing the instance (e.g. Atle) from the dataset provides no guarantees. Rather
than subjectively diminishing the dataset, anomalies are noted and discussed as
part of the results.
Figure 6.7 plots the actual ship costs versus the costs predicted by the M5 model
tree. The worth of a regression-based model is most commonly measured by the
coefficient of correlation, the quantity that gives the quality of a least squares
fitting to the original data. Let yi be the actual cost of ship i and ŷi the predicted
cost (using the M5 model tree), then the coefficient of correlation, R, of the M5
system is calculated as
n yi ŷi − yi ŷi
R=
2
2 2 , (6.10)
n yi −
2
yi n ŷi − ŷi
where n = 59, the number of ships in the dataset, and each of the sums are over
these 59 ships. By design R can range from −1 to +1 with R = 1 indicating
a perfect positive linear correlation. A correlation greater than 0.8 is generally
described as strong (Ryan 1997). The value R 2 , known as the coefficient of deter-
mination, is a measure of how well the regression line represents the data—it
is a measure determining how certain one can be in making predictions from
Advances in cost estimating 153
700
600
Predicted Cost
500
400
300
200
100
0
0 100 200 300 400 500 600 700
Actual Cost
Figure 6.7 M5 model tree: correlation plot of actual versus predicted ship costs.
the model. R 2 is also the ratio of the explained variation (by the model) to the
total variation of the dataset. The measures of worth of the M5 model tree for
predicting ship costs are a strong R = 0.9646 and R 2 = 0.9304. The coefficient
of determination indicates that 93 percent of the total variation in the ship costs
can be explained by the linear relationships described by the M5 model tree lin-
ear regression equations. The remaining 7 percent of the total variation remains
unexplained.
The mean absolute percentage error quantifies the amount by which the esti-
mated cost differs from the actual cost. The mean absolute percentage error is
computed as follows:
1 |yi − ŷi |
i=59
. (6.11)
59 i=1 yi
The mean absolute percentage error for the M5 model tree applied to the SAS-076
ship dataset is 11 percent. Mean absolute percentage errors specific to the
individual M5 model tree linear regression models are shown in Table 6.6. In
addition to absolute percentage errors, the standard deviation is also indicative of
a model’s accuracy. For the purpose of determining the standard deviation of a
M5 model tree output prediction, the standard deviation over all the training data
is more reflective of the M5 system than just the standard deviation of the train-
ing cases reaching the particular leaf node used for the prediction. By M5 model
tree construction, each of the training cases influence the structure of the final
model tree.
Figure 6.8 plots the residuals for attributes used in the M5 model tree linear
regression models, providing a visual confirmation that the residuals are randomly
154 B. L. Kaluzny
Table 6.6 Mean absolute percentage errors of known instances per individual M5 model
tree linear model
scattered (indicating that a linear fit is suitable) for all attributes except for the
rank in class. The residuals for the rank in class attribute indicate more accurate
predictions the higher the ship’s rank in class. This indicates that a linear fit may
not be the most suitable. Typically the construction of lower-ranked ships of the
same class benefit from labor learning curves and economies of scale. Learning
curves are nonlinear, taking the general form of Cr = a · r b , where Cr is the cost of
unit (of rank) r , and a and b are fitted parameters (see Goldberg and Touw (2003)
for more information). Nonlinear regression may be more suitable to model the
relationship of ship cost to the rank in class attribute, however this is outside the
capabilities of the M5 model tree algorithm.
The results presented in Table 6.4 show that linear regression model LM5 con-
tributes the greatest to the standard deviation. LM5 models the nine highest costing
ships. Figure 6.7 reveals that in eight of these cases, LM5 underestimates the actual
cost. By the piecewise linear M5 model tree construction, LM5 is influenced by
the adjacent linear regression models LM4 and LM6. To evaluate the degree of
this influence, a separate multiple linear regression model was fitted to the ships
reaching the LM5 leaf using the same five ship attributes as LM5. The resulting
CER is as follows:
(Predicted)
(Predicted)
(Predicted)
–0.10 –0.10 –0.10
Log(Actual)–Log
Log(Actual)–Log
Log(Actual)–Log
–0.15 –0.15 –0.15
2 4 6 8 10 12 100 120 140 160 180 200 5000 10000 15000 20000 25000 30000
Rank Length Range (total distance)
(Predicted)
(Predicted)
(Predicted)
–0.10 –0.10 –0.10
Log(Actual)–Log
Log(Actual)–Log
Log(Actual)–Log
–0.15 –0.15 –0.15
100 200 300 400 500 0 1 2 3 4 0 1 2 3 4
Crew size Number of propeller shafts Number of LCAC
(Predicted)
(Predicted)
(Predicted)
–0.10 –0.10 –0.10
Log(Actual)–Log
Log(Actual)–Log
Log(Actual)–Log
–0.15 –0.15 –0.15
0.0 0.5 1.0 1.5 2.0 2.5 3.0 0 5 10 15 0.0 0.2 0.4 0.6 0.8 1.0
Number of elevators Number of helicopters supported Number of torpedo decoy systems
Figure 6.8 Plots of residuals for attributes used in the M5 model tree linear regression models.
156 B. L. Kaluzny
Figure 6.9 plots the actual ship costs versus the costs predicted by the M5 model
tree when the smoothing procedure is disabled. The R value is 0.985 (R 2 = 0.97).
The unsmoothed M5 model tree does better in learning the costs of the most
expensive known ships, but likely at the expense of prediction accuracy of unseen
cases.
with an R value of 0.75 (R 2 = 0.56). One typically tests the hypothesis that the
linear regression fit is a better fit than fitting to just the mean length of the ships.
Define the total variation to be the variance when a model is fit to just the mean
length of the ships. Residual variation is defined as the variance when the lin-
ear model is fit. The F-statistic is the ratio of the model and residual variance
and represents the proportional increase in error of fitting a mean model versus
the linear model. The p-value is the probability of rejecting the null hypothesis
that the linear fit is equal to the mean fit alone, when it is in fact true. A sig-
nificant p-value (e.g. less than 5 percent) implies that the linear fit is a better
fit than the mean alone. In this case the F-statistic is 74.2 and the p-value is
6.7 × 10−12.
Applying multiple linear regressions with a greedy attribute selection method
(step through the attributes removing the one with the smallest standardized
700
600
Predicted Cost
500
400
300
200
100
0
0 100 200 300 400 500 600 700
Actual Cost
Figure 6.9 Unsmoothed M5 model tree: correlation plot of actual versus predicted
ship costs.
Advances in cost estimating 157
coefficient until no improvement is observed in the estimate of the error given
by the Akaike (1980) information criterion), yields the CER
with an R value of 0.91 (R 2 = 0.83). Table 6.7 lists the F-statistics and p-values
associated with each of the model attributes. The rank in class and number of
propeller shaft attributes have a p-value near 10 percent, indicating that the fitted
linear model should not be claimed to be a better fit than just using the mean values
for these two attributes.
While the CER produced by applying simple linear regression is straightfor-
ward to understand, the negative coefficient signs of the multiple linear regression
CER makes its interpretation non-trivial.
6.5 Discussion
Figure 6.4 shows that the analogy method using hierarchical clustering based
on a weighted distance metric underestimates the cost of seven of the eight
most expensive ships. This was also a characteristic of the parametric estimation
presented in the preceding section. In the latter it was conjectured that the under-
estimation was likely a result of the smoothing and pruning functions of the M5
model tree algorithm, designed to optimize the predictive capability of the method.
However, the underestimation of expensive ships in both methods is potentially an
indication that the attributes (and their values) of the SAS-076 ship dataset do
not provide enough information to help distinguish the highest costing ships from
their peers.
Multiple cost estimations are a good thing, especially during early design and
developmental phases of a system. Multiple independent cost estimates are ideal,
and strongly endorsed as best practice (see NATO RTO SAS-054 Task Group
(2007) report). The two machine learning approaches described in this chapter
are independent of each other; however, given their reliance on a common source
of data (in this instance) the independence of the cost estimates they produce can
be argued. Given multiple estimates, decision-makers must exercise sound judg-
ment to assess and select a final figure. Properties of the machine learning methods
discussed can provide guidance:
• The M5 model tree algorithms are optimized both to learn known cases
and predict unknown cases. The attribute weights used in the hierarchical
clustering method are optimized to learn the known cases.
• The hierarchical clustering approach uses PCA to reduce the dimensionality
of the attribute space. Due to computational limitations, the weight optimiza-
tion method could only be applied on the top ten macro-attributes, accounting
for 80 percent of the original dataset’s variability. In comparison, the M5
model tree algorithm is computationally superior as it efficiently learns from
large datasets.
• The M5 model tree results in a better correlation measure, lower mean abso-
lute percentage error, and smaller standard deviation in estimating the known
cases.
Table 6.8 synthesizes the predictions and compares properties of the M5 model
tree and hierarchical clustering methods.
Given this information, decision-makers could be advised that estimates gener-
ated by the M5 model tree should be considered to be the primary estimates and
the hierarchical clustering estimates as secondary estimates.
6.6 Conclusion
Two novel approaches to cost estimation using machine learning algorithms are
demonstrated. Both approaches incorporate a multitude of cost-driving factors and
Advances in cost estimating 159
allow for a greater variability in the input dataset—variability that could be cur-
tailed when using traditional approaches. As with other parametric and analogy
approaches, the fidelity of the estimation models are very dependent the dataset,
especially if the size of the dataset is small. Both are “top down” approaches
applicable in early design phases of the procurement cycle.
As an analogy costing approach, hierarchical agglomerative cluster analysis,
principal component analysis, and nonlinear optimization were used to calculate
a matrix of distances among the systems of a dataset. These distances can then be
used to predict the cost of new systems.
The parametric approach combined features of decision trees with linear regres-
sion models to both classify similar systems (based on attributes) and build
piecewise multivariate linear regression models. The attributes of a new system
can then be used to trace down the tree and as input to the resulting regression
models which output a cost prediction.
Acknowledgements
Portions of this chapter reproduce work originally presented in the NATO RTO
SAS-076 Task Group (2012) report, Journal of Cost Analysis and Parametrics
article by Kaluzny et al. (2011), and Defence Research and Development Canada
report by Kaluzny and Shaw (2011). The author is indebted to Sorin Barbici,
Göran Berg, Renzo Chiomento, Dimitrios Derpanis, Ulf Jonsson, R.H.A. David
Shaw, Marcel Smit, and Franck Ramaroson for their contributions.
Notes
1 See www.iceaaaonline.org [last accessed November 20, 2014].
2 The linear models are also tested to see if eliminating a set of attributes reduces the
estimated error (a measure of accuracy of the model on unseen cases).
3 Data mining software in Java, available at www.cs.waikato.ac.nz/ml/weka/ [last accessed
November 20, 2014].
4 The program’s default parameters for the M5 model tree algorithm were used.
References
Akaike, H. 1980. “Likelihood and the Bayes Procedure.” Trabajos de Estadística y de
Investigaciín Operativa 31: 143–166.
Chen, Z. 2006. Reduced-Parameter Modeling for Cost Estimation Models. PhD thesis,
Faculty of the Graduate School, University of Southern California.
Dobra, A. 2002. “Secret: A Scalable Linear Regression Tree Algorithm.” In Proceedings of
the Eighth ACM SIGKDD International Conference on Knowledge Discovery and Data
Mining. New York: ACM Press, pp. 481–487.
Flynn, B. and Garvey, P. 2011. “Weapon Systems Acquisition Reform Act (WSARA) and
the Enhanced Scenario-Based Method (ESBM) for Cost Risk Analysis.” In Proceedings
of 44th Annual Department of Defense Cost Analysis Symposium Williamsburg, Virginia.
Goldberg, M. and Touw, A. 2003. Statistical Methods for Learning Curves and Cost Analy-
sis. Topics in Operations Research Series. Linthicum MD, USA: Institute for Operational
Research and Management Sciences.
160 B. L. Kaluzny
Jolliffe, I. 2002. Principal Component Analysis. New York: Springer.
Kaluzny, B., Barbici, S., Berg, G., Chiomento, R., Derpanis, D., Jonsson, U., Shaw, R.,
Smit, M., and Ramaroson, F. 2011. “An Application of Data Mining Algorithms for
Shipbuilding Cost Estimation.” Journal of Cost Analysis and Parametrics 4: 2–30.
Kaluzny, B. and Shaw, R. 2011. A Parametric Approach to Cost Estimation During Options
Analysis. Technical Memorandum DRDC CORA TM 2011-069, Defence Research and
Development Canada.
Malerba, D., Esposito, F., Ceci, M., and Appice, A. 2004. “Top-Down Induction of Model
Trees with Regression and Splitting Nodes.” IEEE Transactions on Pattern Analysis and
Machine Intelligence 26(5): 612–625.
Miroyannis, A. 2006. Estimation of Ship Construction Costs. PhD thesis, Department of
Mechanical Engineering, Massachusetts Institute of Technology.
NATO RTO SAS-054 Task Group 2007. Methods and Models for Life Cycle Costing.
Technical Report RTO-TR-SAS-054, NATO Research & Technology Organization.
NATO RTO SAS-076 Task Group 2012. NATO Independent Cost Estimating and the Role
of Life Cycle Cost Analysis in Managing the Defence Enterprise. Technical Report
NATO RTO-TR-SAS-076 AC/323)SAS-076)TP/420, NATO Research & Technology
Organization.
Quinlan, J. 1992. “Learning with Continuous Classes.” In Proceedings AI’92, edited by A.
Adams and L. Sterling. Singapore: World Scientific, pp. 343–348.
Ryan, T. 1997. Modern Regression Methods (2nd edition). New York: John Wiley & Sons,
Inc.
Sharma, N. and Litoriya, R. 2012. “Incorporating Data Mining Techniques on Software
Cost Estimation: Validation and Improvement.” International Journal of Emerging
Technology and Advanced Engineering 2(3): 301–309.
Torgo, L. 2000. Inductive Learning of Tree-based Regression Models. PhD thesis, Univer-
sidade do Porto.
Torgo, L. 2002. “Computationally Efficient Linear Regression Trees.” In Proceedings
of International Federation of Classification Societies (IFCS) 2002: Classification,
Clustering and Data Analysis: Recent Advances and Applications.
Wang, Y. and Witten, I. 1997. “Inducing Model Trees for Continuous Classes.” In
Proceeding of the 9th European Conference on Machine Learning Poster Papers,
pp. 128–137.
Witten, I. and Frank, E. 2005. Data Mining: Practical Machine Learning Tools and
Techniques (2nd edition). Massachusetts: Morgan Kaufmann.
7 Facing future funding realities
Forecasting budgets beyond the
future year defense plan
7.1 Introduction
Introducing future funding realities is a critical component of any military cost–
benefit analysis (CBA). Building a national defense force for future generations,
planning missions that will be assigned to military forces in the coming decades,
developing a capital equipment program to meet anticipated future operational
requirements, and building the infrastructure and logistics system to support
them are by their very nature long-term tasks. It is significant, therefore, that in
defense departments around the world one of the biggest challenges is to plan
defense budgets over the medium to long-term.2 Indeed, “despite ferociously
complex planning processes, coherence between strategic guidance, capability
development and the budget remains elusive” (Thompson 2007, 3).
Forecasting future funding envelopes for defense programs is a daunting task.
The challenge of allocating defense budgets among labor (military and civilian
personnel) and capital (weapon systems, platforms and infrastructure) inputs is
further complicated by three additional factors. First, the long-term planning hori-
zon necessary in defense is in stark contrast to the relatively short-term planning
outlook of most national governments, a consequence of the four to five year
election cycles that exist in most modern democracies. Second, defense directly
competes for discretionary funding with publicly popular programs such as health-
care and infrastructure spending. Third, defense faces external shocks in the form
of dramatic changes in the international strategic environment, such as the end of
the cold war in 1989, and the rise of multiple inter-state and intra-state conflicts
throughout the world after September 11, 2001. Although some of these prob-
lems were anticipated by analysts, it is still difficult to predict the scale, scope and
impact of those changes (Dorff 2005).
Given these challenges, no one strategic management approach or forecasting
model can provide a completely robust framework for all nations that face differ-
ent external threats, economic growth trajectories, and mutual defense alliances.
One key element required to enhance defense resources management is the devel-
opment of a more comprehensive approach to forecast and manage budgets within
defense departments.
162 R. Fetterly, B. Solomon
In this chapter three forecasting approaches and two management strategies are
presented to help develop future funding lines (i.e. budget envelopes) for defense
investments. The development of funding lines would allow defense planners to
conduct more rigorous cost–benefit and tradeoff analyses using reasonable proxies
of future budget constraints. While most multi-year planning exercises tend to be
fiscally informed, future cost constraints are often not anchored by a reasonable
approximation of a funding line.
The risk is that a capability acquired in the absence of careful budget esti-
mates may impact other programs if funds are “borrowed” from those programs to
fund the new program that exceeded the actual funding constraint. Alternatively,
unanticipated budget cuts can seriously impact program performance sacrificed to
accommodate the smaller budget.
Recognizing these challenges, a United States Government Accountability
Office (GAO) report on cost growth in US defense acquisition stated, “the goal
is to get a better match between budgeted funds and costs in order that the true
impact of investment decisions is known” (GAO 2005). The major finding of a
Center for Strategic and International Studies (CSIS) study adds that “resource
constraints require a more integrated approach in defense” (CSIS 2005).
The rest of the chapter is structured as follows. The first section presents stylized
facts about defense and fiscal frameworks. The second section examines manage-
ment of the defense budget while the third section discusses defense cost drivers.
The fourth section introduces various forecasting models and contrasts their uses
to the challenges of optimal defense resources management. The last section con-
cludes the chapter and highlights some future directions for research in forecasting
defense budgets beyond the future year defense plan (FYDP).3
150
Other Depts DND (GDP Defl) 93–94 = 100 DND (DSI Defl)
140
130
120
110
100
90
80
70
93–94 95–96 97–98 99–00 01–02 03–04 05–06 07–08 09–10
22
Pre-Budget 2010
Post-Budget 2010
20
18
16
14
12
10
2005–2006 2008–2009 2011–2012 2014–2015 2017–2018
7.4.1 Personnel
Defense establishments are first and foremost people-driven organizations. As
a consequence, personnel costs in a defense budget are often both the most
significant and the least flexible expenditure category. The cost of military
172 R. Fetterly, B. Solomon
personnel and departmental civilians has an overwhelming impact on other
expenditure categories and drives much of the remaining defense spending.
Consequently, the preliminary and possibly most significant step in defense
decision-making is the determination of the required number of military
personnel. Forecasted Army, Navy, Air Force, and Marine Corps personnel, to a
large extent, guide the force planning process. The requirement for military bases,
quantities of military equipment, training facilities, as well as O&M expenses all
flow from decisions regarding personnel strengths. Government costs for military
personnel includes direct pay, environmental and operational allowances, and the
employer share of programs such as pensions and unemployment insurance.
7.4.2 Capital
The acquisition of capital equipment in defense is generally the most visible aspect
of defense spending, because of the substantial annual allotment of taxpayer dol-
lars and the expectations of industry to obtain contracts related to those planned
expenditures. To maintain a viable and effective military capability, it is essen-
tial to have a well-conceived capital equipment replacement plan that is supported
by a stable long-term financial commitment. And in order to develop a coherent
and effective plan for the acquisition of new and modernized equipment there are
always tradeoffs that must be made. Future tradeoffs include:
In terms of goods and services procured and personnel employed, defense depart-
ments spend a large percentage of the overall federal operating budget and capital
funds. This is significant because these categories of expenditure are highly visible,
generate employment, and to a certain extent are discretionary; accordingly these
expenditures are particularly vulnerable to spending cuts during budget crises.
60
Forercast Lower80% Upper80%
55
50
€000
45
40
35
30
2008 2009 2010 2011 2012 2013 2014 2015 2016 2017 2018 2019
Figure 7.3 Belgian defense expenditures forecast based on an ARIMA model (000
euros).
W = f (C, S, Z n ). (7.1)
I = pc C + pm M (7.2)
where I is aggregate income often proxied by GDP; M and C are real military
expenditures and consumption, and the price for each are denoted by pm and
pc respectively. Note that alternate welfare function specifications will use mil-
itary and non-military expenditures and prices in the budget constraint model. For
detailed discussions on the empirical estimation of demand for military spending
models see Hartley and Sandler (1990) and Smith (1995).6
The other key constraint in the welfare maximization problem is the security
function. Security can be represented as
S = f (M1 , . . . , Mn , Z s ). (7.3)
S = f (I, p, Q, Z n , Z s ) (7.4)
Thus the security of a nation depends on its military forces and other countries’
forces and other strategic variables that may affect the security environment.
Looking at the ME of other countries, allies’ military expenditure may be seen
as “spillin”, thus
n
Q = Mit + M j t − Mit
j =1
MEt = β0 + β1 ( pm / pc ) + β2 I + β3 T h + β4 Q + β5 Z + εt . (7.5)
Recall that the ME of other nations elicit varied security postures depending
on whether one is an ally or a threat. The reaction is captured as either Th
(threat) or Q (“spillin”). Aspects that are unique to a nation, such as external con-
flict, new security policy or other exogenous effects are summarized as strategic
variables (Z ).
Facing future funding realities 177
The empirical work on demand for military spending falls largely into two
groups. The first group, popularized by the important work of Murdoch and San-
dler (1982), supports a well-specified theoretical model that generates the military
budget forecasting equation. The second group proposes an approach where the
estimating equation is determined by the best fit to the data, based on rigorous
statistical testing. Chowdhury (1991) is one example of the latter.
Recent studies have departed from these two groups in favor of a compromise
strategy. This strategy, adopted in this chapter, starts with a model based on a
general theory of the demand for military spending, followed by rigorous statis-
tical testing to investigate the relative importance of strategic and other economic
factors for the country(ies) under investigation. Smith (1989) and most recently
Dunne et al. (2003) and Solomon (2005) employ such a strategy.7
For this application of model #2, we use Belgian and Canadian ME for the
period 1955 to 2007 from SIPRI annual publications. To convert US and other
NATO members’ military spending data into constant dollars the IMF database
was consulted. Note that the GDP deflator of each country was used. All figures are
in US dollars. Belgium, Denmark, France, Germany, Italy, Luxembourg, Nether-
lands, Norway, and the United Kingdom were included in the NATO Europe
sample. This variable (NATO_Europe) proxies the potential spill-ins from allied
contribution. If the sign on this allied spill-in (Q) variable is negative, this sug-
gests that the nation (Belgium or Canada) is a free rider. However, Canada and
Belgium articulate their defense posture within the context of a multilateral secu-
rity arrangement and thus a positive response to allies’ ME is expected. To reflect
the choices faced by decision-makers between socio-economic and security needs,
government expenditures were obtained from the World Bank (removing ME from
the aggregate). A negative association is theoretically predicted.
Various databases were consulted to identify potential enemies, militarized
inter-state disputes and peacekeeping related expenditures and missions. Threat
variables include USSR (sample period 1952–1990 and Russia for the period
1991–2006) on which data were obtained from the SIPRI (various years) year
book, and Soviet/Russia Nuclear Equivalent Megatons and Intercontinental Ballis-
tic Missiles (data from Jim Finan, Royal Military College, Canada). The latter was
found to be a better proxy for the Soviet threat since the smaller NATO nations’
primary concern was strategic capability and the potential nuclear fallout.
Data on foreign aid were obtained from the Organization for Economic Coop-
eration and Development (OECD). In addition various dummy variables were
used to proxy various country-specific security policies. These were proxied
according to the dates of the introduction or suspension of government docu-
ments that articulated the nations’ security strategy. NATO doctrine changed from
Mutual Assured Destruction (1952–1969 = 1, 0 elsewhere) to Flexible Response
(1970–2001 = 1, 0 elsewhere; 1970 was used as the period of shift based on
statistical properties) and was also used as a possible determinant of military
spending.
Other standard variables from social welfare maximization, such as income
(GDP) and the relative price between military and civilian good are anticipated
178 R. Fetterly, B. Solomon
to be significant and show positive and negative effects respectively. This is
because defense is considered a normal good, and also reflects the tendency
for governments to economize on military spending when it becomes relatively
expensive. Note, however, that this variable is only estimated for Canada, for
which military price data is readily available. In addition, public finance the-
ory suggests that military spending may crowd out other government programs;
as such, a significant and negative coefficient for non-military expenditures is
anticipated.
Preliminary tests were conducted on the specification of the variables. Levels as
opposed to shares of the variables were used, as these better explain the nature of
the demand for ME (see also Hartley and Sandler 1995).8 The estimated long-run
coefficients for Belgium and Canada are presented in Table 7.3.
Examining results for Canada, country income, and US and European allies,
both contribute positively to military spending. Surprisingly the threat from inter-
continental nuclear attack is negative; this is discussed later in this section and
turns out to be somewhat consistent with the theoretical prediction of the model.
The price variable is not significant for Canada, indicating that rising military
prices have little or no impact on government ME (resource allocation) decisions.
On the other hand, the significance of the income variable for Canada suggests
that income constraints have important long-run impacts.
The long-run coefficients for Belgium are mostly not significant, and only the
threat variable proxied by nuclear arsenals was significant. This suggests that for
Belgium, short-run effects tend to govern defense spending.
Table 7.4 presents the short-run effects through the error correction representa-
tion of the demand model;9 see also Appendix 7.2.
It is important to point out that the short-term dynamics of the demand for
military spending models in Canada are governed mostly by a complementary
reaction to allied military spending, and budgetary inertia. The latter is evident for
all the nations considered, and especially when observed from the perspective of
peace support operations and defense procurement.
Peacekeeping and other military expeditionary operations, to be effective as a
military response to the threat of international instability, are by necessity medium
Notes
∗∗∗
1% significance level ∗∗ 5% and ∗ 10%; n.a. = not applicable; n.s. = not significant.
Sample period 1955–2007.
Policy = Defense policy changes (dummy variables); NICM = Nuclear Intercontinental Missile
(USSR/Russia); NT = NICM plus surface; US = US Military Expenditures.
Facing future funding realities 179
Table 7.4 Error correction representation
Notes
Key: See Equation (7.5) for variable descriptions, and also Table 7.3.
∗
10%, ∗∗ 5%, ∗∗∗ 1% significance.
denotes the first difference of the variable; X −1 implies X −1 − X −2 .
13
CANME2 Forecast
12
11
$B (nominal Canadian $)
10
4
19 te
19 6
19 8
19 0
19 2
19 4
19 6
19 8
19 0
19 2
19 4
19 6
19 8
19 0
19 2
19 4
19 6
19 8
19 0
19 2
19 4
19 6
20 8
20 0
20 2
20 4
06
20 8
20 0
20 2
20 4
16
5
5
6
6
6
6
6
7
7
7
7
7
8
8
8
8
8
9
9
9
9
9
0
0
0
0
1
1
1
a
20
D
Figure 7.4 Canadian defense expenditures forecast based on constant GDP growth
(2 percent).
of 0.22. Given the long-run estimate of GDP inflation at 1.5 percent, the forecast
for nominal defense growth is around 1.9 percent.
Personnel
Military (officers and enlisted civilian)
Reserve or other non-regular
Inflation_Pay
Operations & maintenance
People related
Equipment
Infrastructure
Inflation_O&M
Capital
Equipment
Infrastructure
Technology
Procurement policy
Defense inflation
Statutory and transfers
Personnel related
Other transfers
Exogenous
Federal budgets
Defense White Papers
Other government initiatives
Grand total
Personnel costs can be further disaggregated into military and civilian personnel
costs,
where ( pm Pmil ) is number of military personnel times the wage rate, and similarly
civilian personnel costs are ( pci Pciv ). Equipment costs are represented as
1 below the required goal, how long will it take to achieve the desired level given
the attrition rates in basic training and regular attrition in the existing force?
2 above the required level, how would the reduction be achieved (e.g. to reduce
recruitment until regular attrition reduces the number, or design an accelerated
reduction program)?
Finally, O&M costs can be forecast using projected inflation rates. Note that
O&M expenditures depend on the number of personnel (P), infrastructure (I ),
and weapon stocks (E):10
The model is applied to Canada. Baseline personnel numbers and expected cap-
ital acquisitions are taken from Canada’s departmental plans and priorities (DND
2009). As shown earlier in Figure 7.2, the Federal Budget has reduced its own
20-year projection for defense, and this information is coupled with the fact that
personnel numbers will hold steady.
Personnel-related costs are projected 20 years out (Figure 7.5). Projections for
capital will depend on the ability of DND to re-scope and re-phase projects given
the absorption challenges discussed previously. Specifically a report to Parliament
indicated DND did not have the capacity to absorb the announced increases to
its budget, and the result was a funding lapse of approximately $300 million
(1.6 percent of the total budget; DND 2009). Projections on personnel, readi-
ness, infrastructure and equipment are constrained by the CFDS 20-year funding
ceiling but year to year growth is based on the 1.5 percent and then 2 percent
increase expected by DND (DoF 2010). Figure 7.6 shows overall future funding
projections for the defense budget by calibrating policy, technology and inflation
projections.11
It should be reiterated that the funding line (budget envelope) projected here
may or may not equal derived demand since the funding line will always be con-
strained by legislated personnel numbers, approved inflation funding, and force
acquisition and elimination decisions based on CBP management decisions.
On the demand side the attribution rules generate, in input-output model par-
lance, technology coefficients that are updated annually. Long-term projections
remain fixed on these updated technology coefficients. To the extent the mod-
eller receives a priori information about a new in-service support arrangement
184 R. Fetterly, B. Solomon
14.00
Pers Cost Pers Cost +Inflation
12.00
10.00
$ Billions
8.00
6.00
4.00
2.00
–
2008–09
2009–10
2010–11
2011–12
2012–13
2013–14
2014–15
2015–16
2016–17
2017–18
2018–19
2019–20
2020–21
2021–22
2022–23
2023–24
2024–25
2025–26
2026–27
2027–28
Figure 7.5 Personnel growth projections with and without inflation adjustment.
Source: Initial Personnel Numbers from DND (2009). Inflation projection based on DND (2008).
35
Demand
30
25
$ Billions
20
15
10
0
2008–09
2009–10
2010–11
2011–12
2012–13
2013–14
2014–15
2015–16
2016–17
2017–18
2018–19
2019–20
2020–21
2021–22
2022–23
2023–24
2024–25
2025–26
2026–27
2027–28
or training that may affect factors such as the attribution of training, material
and infrastructure support, these can be incorporated in long-term projections.
For example, if the future security environment and capability planning favors
an emphasis on the Arctic, then one example of “early warning” data is in-service
support arrangements for Arctic patrol capability.
Facing future funding realities 185
While this final model is closely linked to the strategic management strategy
we outlined in preceding sections, each of the forecasting models discussed here
have their merits depending on the maturity and availability of strategic and
management information systems.
1.2
ACF
1
0.8
Autocorrelation
0.6
0.4
0.2
0
1 2 3 4 5 6 7 8 9 10 11 12
–0.2
Lags
-
1.5
1
0.5 I
I I •I -I •
-
0
I I -
Autocorrelation
,
-o.s 1
" n 1 '>
-1
-1.5
-2
-2.5
-3
• PACF
-3.5
Lags
as a “random walk” model can be estimated by linear methods. Whether the tenta-
tive model is appropriate for forecasting is assessed through a series of diagnostic
checks or tests known in Box–Jenkins parlance as necessity, invertibility, and
sufficiency tests. Each parameter included in the model should be statistically sig-
nificant (necessary) and each factor must be invertible. In addition, the residuals
from the estimated models should be random (model sufficiency).
The test for necessity is performed by examining the T-ratios for the indi-
vidual parameter estimates. Parameters with non-significant coefficients may be
deleted from the model in order to have a parsimonious model. Invertibility is
Facing future funding realities 187
determined by extracting the roots from each factor in the model. All the roots
must lie outside of the unit circle. If one of the factors is non-invertible, then the
model must be adjusted. The appropriate adjustment is dictated by the type of the
factor that is non-invertible. For example, a non-invertible autoregressive factor
usually indicates under-differencing, while a non-invertible moving average fac-
tor may indicate over-differencing. A non-invertible moving average factor could
also represent the presence of a deterministic factor. Since the model structure is
not really clearcut, the overall model must be considered when adjusting for non-
invertibility (Box and Jenkins 1970). The residuals are then tested for white noise
by studying the autocorrelation and partial autocorrelation of the residuals. Fur-
thermore, a test statistics Q or “portmanteau test” is performed on the residuals
autocorrelations of all lags. If the model is mis-specified or inappropriately fitted,
the Q test tends to be inflated (Box and Jenkins 1970; Kennedy 1998).
The general multiplicative Box–Jenkins model as articulated in its standard
form is (see Box and Jenkins 1970)
where s denotes periodicity of the seasonal component (4 for quarterly and in our
case annual which is 1); B denotes the backward operator, i.e.
B Z t = Z t−1
(B )Z t = (B s )Z t−s ;
s
(t) (t)
Z t (l) = b0 (t) f0 (l) + b1 f 1 (l) + · · · + bm−1 fm−1
for l > n − m. Note also that f represents functions of the lead time l and includes
polynomials, exponentials, etc. The general autoregressive operator ψ(B) defined
above, which determines the mathematical form of the forecasts function, i.e. the
nature of f . In other words, it determines whether the forecasting function is to
be a polynomial, a mixture of sines and cosines, a mixture of exponentials or
some combinations of these functions. As indicated in Box and Jenkins (1970),
the above is called the “eventual forecast function”, and since n > m it provides
the forecasts only for lead times l > n − m.
Outlier detection
Most time series based forecasting software packages include a routine that
automatically detects the presence of outliers (isolated events such as strikes,
earthquakes, etc.) or “shocks,” and whether these shocks are transitory or induce a
level shift. Often such events are modelled as “dummy” variables with a series of
zeroes and one(s) for the time period(s) of the isolated event. The automatic out-
lier detection algorithm is obviously better than the theory-based method for cases
when the modeller does not have a priori knowledge of when the event may have
occurred. In addition, if the outlier causes a structural shift (such as new institu-
tional legislation, defense White Papers, etc.) and if the impact of the event takes
a longer time lag to affect behaviour, then the modeller risks biasing the effect of
the outlier by choosing the day the event occurred as the break in the series.
On the other hand, theory-free methods have the “potential for finding spurious
significance” during the testing of the hypothesis (AFS 2000). However, since many
of these events may or may not be identifiable to the modeller, a theory-free detection
algorithm is often desirable. The algorithm begins by first fitting a univariate ARIMA
model to the series divided by a series of regressions at each time period to test the
hypothesis that there is an intervention. After identifying an outlier the residuals are
modified and the test will resume until all outliers are uncovered.
Some forecasters (see the 1990 AFS manual) argue that fitting a univari-
ate ARIMA model does not necessarily imply that the series is homogeneous
(the error term of the series is random about a constant mean). In such situations,
the identification of the outliers become dubious due to the recursive process. The
suggested solution is the testing of a more rigorous specification, i.e., the mean of
the errors must be near zero for all time sections (AFS 2000). As an alternative,
an endogenous determination of the timing of structural breaks can be more infor-
mative. Lanne et al. (2002) developed an Augmented Dickey–Fuller (ADF) type
test on such a method by considering shift functions of a general nonlinear form
that is used to add to the deterministic term. Lütkepohl and Saikkonen (2002) also
argue that a shift may occur in various forms ranging from a shift that is spread
Facing future funding realities 189
out over a number of periods to a smooth transition to a new level. This method is
employed in the next section to supplement dummy variable selection.
k
θ (L, p)yt = α1 + α2 Tt + βi (L, qi )x it + εt (A7.2.1)
i=1
n
n
n
M E t = β0 + αM E t−i + χ( pm / pc )t−i + φQ t−i
i=1 i=1 i=1
n
n
n
+ γ T h t−i + ϕIt−i + ηM E t−i + εt .
i=1 i=1 i=1
190 R. Fetterly, B. Solomon
A variable addition test where the lagged values of the level variables repre-
sented as
One can use the variable addition test to fit a more parsimonious model of demand
for military expenditures. For example, if by dropping a variable x from the model
it ceases to be cointegrated, then we can infer that x has a significant effect on
the demand for military spending in the long run. However, if the reverse holds,
then one can remove the variable. The use of the F-test to eliminate variables is
similar to one employed by Beenstock (1998) to test significance of variables for
cointegration in Israel’s demand model. Specifically, if the null hypothesis testing
for long-run relations is sensitive to the removal of the variable in question, then
one can safely assume that the variable in question is likely a long-run forcing
variable explaining military expenditure.
Notes
1 The views expressed in this chapter are the views of the authors and do not necessar-
ily reflect the views or policies of the Canadian Armed Forces or the Department of
National Defence.
2 National defense is a complex, high consequence of error, capital-intensive, knowledge-
dependent, national security instrument. Categorization of defense in this manner results
from the diversity of tasks assigned to the defense department and the military forces, the
elevated level of technology employed within these organizations, the aggregate size of
the budget, and the considerable annual capital investment in weapon systems. Indeed,
management of the defense budget “is a highly constrained exercise in pricing the exe-
cutability of programs within the parameters of affordability and political feasibility”
(Jones 1991, 20).
3 While most of the examples used in this chapter have a Canadian flavor, selected NATO
member nations are also discussed where data and information are available.
4 There is considerable debate on this issue. While Solomon (2003) makes the case for
defense inflation for Canada as does Kirkpatrick (2008) for the United Kingdom, a
special issue of the Royal United Services Institute (RUSI) dedicated a special issue on
the subject (RUSI 2009) that needs to be considered to assess whether a separate index
of inflation is required in defense costing.
5 The addition of population may seem peculiar given the fact that defense is a pub-
lic good; however, consumption probably is not and as such, per capita consumption
is assumed to matter for welfare. In other words, demand for defense spending may
increase if it has a high-income elasticity and increase in population reduces the tax cost
faced by the median voter (Dunne et al. 2003).
6 Most empirical studies in the literature, citing practical and conceptual difficulties asso-
ciated with military price indices, make the assumption that the relative price differential
between the military and civilian price is constant. Solomon (2003, 2005) discusses the
arguments for and against this view as well as the implications for empirical work.
7 The conventional econometric method for estimating time series models is based on
cointegration. (See Appendix 7.2.)
8 The log-transformation of the variables was preferred, based on non-nested tests. As
such, these results are presented.
9 The ARDL (autoregressive distributed lag) estimates and the lag order were selected
using the Schwarz Bayesian Criterion. The error correction term enters the equation with
Facing future funding realities 191
the expected negative sign and is significant at the 1 percent and 5 percent significance
for both countries considered. The results from the model in general do not point to any
econometric problems, as the model successfully passes all the diagnostic tests.
10 If the information system or costing model allows further disaggregation then personnel
can be limited only to operational and support personnel.
11 Initial calibrations are based on the official data from DND Report on Plans and Pri-
orities (FY 2008–09) Section 3 Supplementary Tables A&B as well as Table 13 on all
approved major capital projects.
References
Amara, J. 2008. “NATO Defense Expenditures: Common Goals Or Diverging Interests? A
Structural Analysis.” Defence and Peace Economics 19(6): 449–469.
Automatic Forecasting. Systems-AFS. 2000, 1990 and 1986. User’s Manual for Autobox
Software. Hatboro: Pennsylvania.
Beenstock, M. 1998. “Country Survey XI: Defence and the Economy in Israel.” Defence
and Peace Economics 9(3): 171–222.
Berkok, U. 2005. “Specialization in Defence Forces.” Defence and Peace Economics 16(3):
191–204.
Box, G. E. P. and D. R. Cox. 1964. “An Analysis of Transformations.” Journal of the Royal
Statistical Society Series B 26: 211–243.
Box, G. E. P. and G. Jenkins. 1970. Time Series Analysis: Forecasting and Control (rev. 1st
edn). San Francisco: Holdenday.
Center for Strategic and International Studies. 2005. European Defense Integra-
tion: Bridging the Gap between Strategy and Capabilities (Washington: Cen-
ter for Strategic and International Studies). Available at www.csis.org/media/
csis/pubs/0510_eurodefensereport.pdf [accessed April 12, 2007].
Chouinard, P. and I. Wood. 2007. The Department of National Defence Strategic Cost
Model: Development. DND Centre for Operational Research & Analysis, Ottawa, p. 2.
Chowdhury, A. R. 1991. “A Causal Analysis of Defence Spending and Economic Growth.”
Journal of Conflict Resolution 35: 80–97.
Congressional Budget Office. 2010. Long-Term Implications of the Fiscal Year
2010 Defense Budget (CBO, Washington). Available at www.cbo.gov/ftpdocs/
108xx/doc10852/01-25-FYDP.pdf [accessed July 23, 2010].
Department of Finance. 2010. The Federal Budget 2010. Ottawa: Public Works and
Government Services Canada.
Department of National Defence. 2007. Investing in the Future: A Ten Year Strategic Invest-
ment Plan Framework for the Department of National Defence. Ottawa: Department of
National Defence.
———. 2008. Canada First Defence Strategy. Ottawa: Department of National Defence.
———. 2009. Report on Plans and Priorities. Ottawa: Department of National Defence.
Dodaro, Gene L. 2010. Maximizing DOD’s Potential to Face New Fiscal Challenges and
Strengthen Interagency Partnerships. Government Accountability Office, Washington,
p. 2. Available at www.gao.gov/cghome/d10359cg.pdf [accessed July 5, 2010].
Dorff, Robert H. 2005. “Failed States After 9/11: What Did We Know and What Have We
Learned?” International Studies Perspectives 6(1): 20–34.
Dunne J. P., E. Nikolaidou and N. Mylonidis. 2003. “The Demand for Military Spending in
the Peripheral Economies of Europe.” Defence and Peace Economics 14(6): 447–460.
GAO. 1994. Future Years Defense Program: Optimistic Estimates Lead to Billions
in Over-programming. GAO, Washington. Available at www.legistorm.com/showFile/
192 R. Fetterly, B. Solomon
L2xzX3Njb3JlL2dhby9wZGYvMTk5NC83/ful24551.pdf [accessed July 28, 2010].
———. 2005. Defense Acquisitions: Improved Management Practices Could Help Min-
imize Cost Growth in Navy Shipbuilding Programs GAO, Washington). Available at
www.gao.gov/new.items/d05183.pdf [accessed July 4, 2010].
———. 2010. The Federal Government’s Long-Term Fiscal Outlook: January 2010.
GAO, Washington. Available at www.gao.gov/new.items/d10468sp.pdf [accessed July
19, 2010].
Hartley, K. and Sandler, T. (Eds). 1990. The Economics of Defence Spending: An
International Survey. London: Routledge.
———. 1995. Handbook of Defence Economics, Volume 1 Amsterdam: North Holland.
Hilton, B. and A. Vu. 1991. “The McGuire Model and the Economics of the NATO
Alliance.” Defence Economics 2(1): 105–121.
International Monetary Fund (IMF). Various years. World Economic Outlook Database.
Jones, L. R. 1991. “Policy Development, Planning, and Resource Allocation in the
Department of Defense.” Public Budgeting and Finance 11(3): 15–27.
Jones, L. R. and Jerry L. McCaffery. 2005. “Reform of the Planning, Programming, Bud-
geting System, and Management Control in the US Department of Defense: Insights
from Budget Theory.” Public Budgeting & Finance 25(3): 1–19.
Jordan, Leland G. 2000. “Systematic Fiscal Optimism in Defense Planning.” Acquisition
Review Quarterly (Winter): 47–62.
Kennedy, P 1998. A Guide to Econometrics (4th edition). Cambridge: MIT Press.
Kirkpatrick, D. 1995. “The Rising Unit Costs of Defence Equipment – the Reasons and the
Results.” Defence and Peace Economics 6(4): 263–288.
———. 2004. “Trends in the Costs of Weapon Systems and the Consequences.” Defence
and Peace Economics 15(3): 259–273.
———. 2008. “Is Defence Inflation Really as High as Claimed?” RUSI Defence Systems
11(2): 66–71.
Krepinevich, Andrew F., and Barry D. Watts. 2009. Regaining Strategic Competence.
Center for Strategic and Budgetary Assessments, Washington, p. vii. Available at www.
csbaonline.org/4Publications/PubLibrary/R.20090901.Regaining_Strategi/
R.20090901.Regaining_Strategi.pdf [accessed August 21, 2010].
Lanne, M., H. Lütkepohl, and P. Saikkonen. 2002. “Comparison of Unit Root Tests for
Time Series with Level Shifts.” Journal of Time Series Analysis 23: 667–685.
Lütkepohl, H. and P. Saikkonen. 2002. “Testing for a Unit Root in a Time Series with a
Level Shift at Unknown Time.” Econometric Theory 18: 313–348.
Melese, F. 1997. “Gain-Sharing, Success-Sharing and Cost-Based Transfer Pricing: A New
Budgeting Approach for the Department of Defense.” Military Operations Research
3(1): 23–50.
Melese, F., J. Blandin, and S. O’Keefe. 2004. “A New Management Model for Government:
Integrating Activity Based Costing (ABC), the Balanced Scorecard (BSC), and Total
Quality Management (TQM) with the Planning, Programming and Budgeting System
(PPBS).” International Public Management Review 5(2): 103–131.
Murdoch, James C. and Todd Sandler. 1982. “A Theoretical and Empirical Analysis of
NATO.” Journal of Conflict Resolution 26(2): 237–263.
Nelson, Daniel S. 2001. “Beyond Defence Planning.” Defence Studies 1(3): 25–36.
Pesaran, M. H. and B. Pesaran. 1997. Working with Microfit 4.0: Interactive Econometric
Analysis. Oxford: Oxford University Press.
Pesaran, M. H. and Yongcheol Shin. 1999. “An Autoregressive Distributed Lag Modelling
Approach to Cointegration Analysis.” In Econometrics and Economic Theory in the 20th
Facing future funding realities 193
Century, edited by S Strom. The Ragnar Frisch Centennial Symposium, Chapter 11.
Cambridge: Cambridge University Press.
Pesaran, M. H., Y. Shin, and R. J. Smith. 1996. “Testing for the Existence of a Long-
Run Relationship.” DAE Working Papers Amalgamated Series, No.9622, University of
Cambridge.
———. 2001. “Bounds Testing Approaches to the Analysis of Level Relationships.”
Journal of Applied Econometrics 16: 289–326.
Pugh P. G. 2007. Source Book of Defence Equipment Costs. London: Pugh.
Royal United Services Institute (RUSI). 2009. “Defence Inflation: Myth or Reality.”
Defence Systems 12(1): 12–21.
SIPRI Yearbooks. Various years. Armaments, Disarmament and International Security.
Stockholm International Peace Research Institute. Available from http://www.sipri.
org/yearbook [last accessed February 9, 2015].
Smith, R. 1989. “Models of Military Expenditure.” Journal of Applied Econometrics 4(4):
345–359.
———. 1995. “The Demand For Military Expenditure.” In Handbook of Defence Eco-
nomics, Volume 1, edited by K. Hartley and T. Sandler. Amsterdam: North Holland.
Solomon, B. 2003. “Defence Specific Inflation: A Canadian Perspective.” Defence and
Peace Economics 14(1): 19–36.
———. 2005. “The Demand For Military Expenditures in Canada.” Defence and Peace
Economics 16(3): 171–189.
———. 2009. “The Defence Industrial Base in Canada”, In The Public Management of
Defence in Canada, edited by Craig Stone. Toronto: Breakout Educational Network.
Solomon, B., P. Chouinard, and L. Kerzner. 2008. The Department of National Defence
Strategic Cost Model: Volume 2 – Theory and Empirics. DRDC-CORA TR2008-03,
Ottawa.
Stone, J. C. and B. Solomon. 2005. “Canadian Defence Policy and Spending.” Defence and
Peace Economics 16(3): 145–169.
Thompson, M. 2007. Improving Defence Management. Canberra: Australian Strategic
Policy Institute.
Treddenick, J. M. 2000. “Modelling Defense Budget Allocations: An Application to
Canada.” In The Economics of Regional Security: NATO, The Mediterranean, and South-
ern Africa, edited by Jurgen Brauer and Keith Hartley, pp. 43–70. Amsterdam: Harwood
Academic Publishers.
World Bank 2010. Global Economic Prospects Summer 2010: Fiscal Headwinds and
Recovery. World Bank, Washington, p. 2. Available at http://siteresources.worldbank.
org/INTGEP2010/Resources/FullReport-GEPSummer2010.pdf [accessed July 5, 2010].
This page intentionally left blank
Part III
Measuring effectiveness
This page intentionally left blank
8 Multiple objective decision-making
Kent D. Wall and Cameron A. MacKenzie
8.1 Introduction
Imagine the the following decision problem for the fictitious country of Drmecia,
a small nation with many security problems and limited funds for addressing these
problems:
The Army of Drmecia operates a ground based air-search early warning radar
base located near the capital city, Sloat. The radar exhibits low availability
because of reliability and maintainability problems. This exposes the country
to a surprise air attack from its enemy, Madland. The Army obtained pro-
curement funds two years ago and installed another radar of the same type
as a back-up, or standby, radar. Unfortunately, the second radar unit also has
exhibited low availability and too often both radars are off-line. The com-
mander of Sloat Radar Base is now requesting funds for a third radar, but the
Army Chief of Staff is very concerned. He knows that the defense budget is
under increasing pressure and funds spent on another radar may have to come
from some other Army project. In addition, two new radar models are now
available for procurement at higher cost than the existing radars. It may be
better to purchase one of these than another radar of the existing model.
The Chief of Staff knows he must have a strong case if he is to go to the
Minister of Defense and ask for additional funds. He wants to know which
radar is best. He wants to know what he is getting for the extra money spent.
He wants to know what is the most cost-effective course of action.
The director of analysis for the Army of Drmecia knows that making a
cost-effective decision requires two things: (1) a cost analysis and (2) an
effectiveness analysis. The analysis of cost is a familiar problem and he has a
staff working on it. The effectiveness part of the problem, however, is a con-
cern. What is effectiveness in this situation? How can effectiveness be defined
and quantified? How can the cost analysis be integrated with effectiveness
analysis so that a cost-effective solution can be found?
The goal of this chapter is to develop a method to think about and quantify
effectiveness in the public sector, specifically defense. Effectiveness can best be
198 K. D. Wall, C. A. MacKenzie
measured in the public sector by developing a framework for solving decision
problems with multiple objectives. The framework will provide you with a practi-
cal tool for quantitative investigation of all factors that may influence a decision,
and you will be able to determine why one alternative is more effective than others.
This analytical ability is very important because many real-life decision problems
involve more than a single issue of concern. This holds true for personal-life deci-
sions, private sector business decisions, and public sector government resource
allocation decisions.
Examples of personal-life decision problems with multiple objectives are plenti-
ful: selecting a new automobile; choosing from among several employment offers;
or deciding between surgery or medication to correct a serious medical problem. In
the private sector, maximizing profit is often the sole objective for a business, but
other objectives may be considered, such as maximizing market share, maximizing
share price performance, and minimizing environmental damage.
Public sector decision-making almost always involves multiple objectives. State
and local government budget decisions are evaluated, at least, in terms of their
impacts on education programs, transportation infrastructure, public safety, and
social welfare. The situation is more complex at the Federal level because national
defense issues enter the picture. For example, consider the ways in which the
US Department of Defense (DoD) evaluates budget proposals. Top-level decision-
makers consider the effects of a budget proposal on (1) the existing force structure,
(2) the speed of modernization, (3) the state of readiness, and (4) the level of risk,
among other factors. Other national defense objectives include the ability to deter
aggression, project force around the world, and defeat an enemy in combat regardless
of its location or military capability. This multidimensional character permeates all
levels of decision-making within a planning, programming, budgeting and execution
system (PPBES). Acquisition decisions, training and doctrine policy changes, and
base reorganization and consolidation choices all have multiple objectives (Buede
and Bresnick 1992; Parnell et al. 1998; Ewing et al. 2006).
Government decisions in general, and defense resource allocation decisions
in particular, have an added evaluation challenge. Outcomes are difficult, if not
impossible, to represent in monetary terms. First, benefit cannot be expressed
in terms of profit. Unlike the private sector, the public sector is not profit moti-
vated and this single monetary measure of benefit is not relevant. Second, market
mechanisms often do not exist for “pricing out” the many benefits derived from
public sector decisions. Thus, it is not possible to convert all the benefits into
monetary terms and conduct a cost–benefit analysis (CBA). In national defense,
benefits are often characterized in terms such as deterrence, enhanced security,
and increased combat capability (Dyer et al. 1998; Parnell et al. 2001). No mar-
kets exist that generate a price per unit of deterrence or a unit increase in national
military security.
Solving these decision problems requires a structured systematic approach that
aids the discovery of all the relevant objectives and makes it easy to work with
objectives expressed in many different units of measure. It also must allow the
decision-maker to account for the relative importance of the objectives and the
Multiple objective decision-making 199
importance of marginal changes within individual objective functions. Following
such an approach can allow decision-makers to perform cost-effectiveness analy-
sis (CEA) to compare programs and alternatives.
{x 1 ( j ), x 2( j ), x 3 ( j ), . . . , x M ( j )}.
These numbers become the attributes for the j th alternative, and each of the M
numbers represents how well the j th alternative meets each one of the decision-
maker’s M objectives. Evaluating an alternative means determining the numerical
value of each of the measures. These serve as the raw data upon which the decision
is made.
The key to solving the decision problem is to explain how this set of M num-
bers is viewed in the mind of the decision-maker. This is done by constructing
a collective measure of value that reflects how much the decision-maker prefers
one collection of attributes vis-à-vis another collection of attributes. For example,
how does the decision-maker value the set of attributes for the first alternative,
{x i (1)}, compared to the set of attributes for the second alternative, {x i (2)}, where
1 ≤ i ≤ M? Is the the set {x i (1)} more “attractive” than the set {x i (2)}?
It is challenging for a decision-maker to compare alternatives if each alternative
has several attributes (i.e., if M is greater than 3 or 4). Developing a function
that combines the M attributes into a single number provides a method for the
decision-maker to compare alternatives based on the attributes that describe each
alternative. This single number is called the Measure of Effectiveness (MoE).
8.2.2 Effectiveness
The function that measures the effectiveness of the j th alternative is called a
value function, v( j ). This function incorporates the preferences of the decision-
maker and converts each collection of attributes as represented by the set {x i ( j )}
into a number that represents the attractiveness or desirability of the collection in
the mind of the decision-maker. It measures the extent to which the j th alterna-
tive helps the decision-maker pursue all the objectives while taking into account
the relative importance of each. Effectiveness is a number. It quantifies “how far
we go” towards achieving our goals as measured by the value of the objectives.
The objectives are not all of equal importance, however, and v( j ) also takes into
account the relative importance of each objective.
Multiple objective decision-making 201
Table 8.1 Variable definitions
Variable Definition
0 vera II 0 bjective
To Maximize Effectiveness
To Max. Availability
I
To Max. P e rformanc e ToM in . Complexity
To Max.
Inter-
I
To Max To Max.
Effective
~
To Min To Max.
ECCM Cognitive Ease-of-Use
Operability Range Load
0.9
0.8
., 0.7
::I
~., 0.6
> 0.5
~
:i
E 0.4
::I
(J
0.3
0.2
0.1
0
0 100 200 300 400 500 600 700 800 900 1000
Range(km)
1 − e−a[xi −xmin ]
vi (x i ) = (8.1)
K
,............, ..----
~
!-A"""
v
0.9
.,
0.8
0.7
v-
::I
0.6 ~
~.,
> 0.5 /
~
:;
E 0.4
r/-
::I
(.)
0.3 /
0.2 ~~-
0.1 I
0 J
0 100 200 300 400 500 600 700 800 900 1000
Range(km)
1 − e−a[xmax −xi ]
vi (x i ) = (8.2)
K
where K is the same as before. Figure 8.6 depicts the results of fitting the cumu-
lative value function for radio weight in Excel using the Solver add-in tool where
a = 0.21, x min = 0, and x max = 20. Once again, closer fit can be obtained by using
more terms in the exponent as described in Appendix 8.1.
A linear function can also be used to approximate the exponential function for
either the more-is-better case or the less-is-better case. The linear function can use
the same cumulative value function as described previously, or just two values can
be assessed from the decision-maker. The two values necessary are a number cor-
responding to “not enough performance” and a number corresponding to “good
enough performance.” If more is better, the number corresponding to not enough
performance is “too little” or x too little . If less is better, the number corresponding
Multiple objective decision-making 211
I I
I I I I I
0.9
I I I I I I
0.8
.,
I I I I I I I
0.7
::I
~.,
I I I I I I I
0.6
I I I I I I I I
> 0.5
i
'S
I I I I I I I I
E 0.4
::I
(J
I I I I I I I I
0.3
I I I I I I I I I
0.2
I I I I I I I I I
0.1
0
0 2 6 8 10 12 14 16 18 20
Weight(kg)
...........,
0.9 N KJ
0.8 r-.....
., 0.7 ~
"
::I
~., 0.6
> ~
"n'
0.5
i
'S
E 0.4
::I
(J
0.3
0.2
0.1
\
0 \
0 2 6 8 10 12 14 16 18 20
Weight(kg)
to not enough performance is “too much” or x too much . In both cases, the number
corresponding to good enough performance is “ideal” or x ideal . For example, in
the more-is-better example of radar range, 0 km can represent not enough per-
formance or too little and 600 km can represent good enough performance or the
ideal. We choose 600 km rather than 1,000 km as the ideal because the assessed
value at 600 km is 0.963 and the increase in value from 600 to 1,000 km is only
0.037. In the less-is-better example of the radio’s weight, we choose 20 kg as too
212 K. D. Wall, C. A. MacKenzie
much and 8 kg as ideal. Eight kilograms is chosen as ideal because the value at
8 kg is 0.986 and the increase in value from 8 to 0 kg is only 0.014. The linear
value function is defined for more is better in which [x too little ≤ x i ≤ x ideal ]:
x i − x too little
vi (x i ) = , (8.3a)
x ideal − x too little
x i − x too much
vi (x i ) = . (8.3b)
x ideal − x too much
w1 = w A (8.4a)
w2 = w P · w I (8.4b)
w3 = w P · w E (8.4c)
w4 = w P · w R (8.4d)
w5 = wC · w L (8.4e)
w6 = wC · wU . (8.4f)
Since each group of local weights sum to one, we know the global weights as
calculated above will also sum to one:
w1 + w2 + w3 + w4 + w5 + w6 = w A + w P + wC = 1.
w A = w P = wC = 1/3.
This also means that all the component objectives for performance are equally
important:
w I = w E = w R = 1/3.
Finally, suppose the decision-maker says that maximizing ease-of-use and mini-
mizing cognitive load are equally important, so then we know
wU = w L = 1/2.
Range = 1/1,
Interoperability = 1/2,
ECCM = 1/3,
which sum to 11/6. Dividing the reciprocal ranks by their sum gives three numbers
that satisfy the summation condition and represent a valid set of weights: w R =
6/11, w I = 3/11, and w E = 2/11.
Both the rank sum and rank reciprocal methods only require rank order infor-
mation from the decision-maker. The rank sum method returns weights that are
less dispersed than the rank reciprocal method, and less importance is placed on
the first objective in the rank sum method.
v( j ) = w1 · v1 (x 1 ( j )) + w2 · v2 (x 2 ( j ))
M
+ w3 · v3 (x 3 ( j )) + · · · = wi · vi (x i ( j )). (8.5)
i=1
The MoE is calculated as the weighted sum of all the individual value functions.
Nothing of relevance is left out, and everything that enters the computation does
so according to the preferences of the decision-maker (Keeney and Raiffa 1976;
Kirkwood 1997).
Evaluating the overall effectiveness of an alternative requires computing its
v( j ). The result allows us to order all the alternatives from best to worst according
to their MoE. The “best” alternative is now the alternative with the largest MoE.
While this provides us a way to find “the” answer, we have a far more powerful
tool at hand.
We can use Equation (8.5) to investigate why we get the answers we do. For
example, what makes the best alternative so desirable? What individual effective-
ness measures contribute most to its desirability? An alternative may be the best
because of very high effectiveness in only one single measure. This may tell us
that we could be “putting all our eggs in one basket.” If this alternative is still
under development, uncertainties about its development may make this alternative
risky, and identifying the second best alternative may be important.
We can also assess the sensitivity of the answer to the importance weights. We
can find out how much the weights must change to give us a different answer.
For example, would a change of only 1 percent in one of the higher level weights
change the ordering of the alternatives? How about a change of 10 percent? This is
of practical significance because the decision-maker assigns weights subjectively,
and this always involves a lack of precision.
Most important of all is the ability to assess the effect of uncertainty in the
future condition. The decision-maker’s preferences are a function of the future
condition. If the decision-maker believes the future condition will change, then the
preferences, value function and weights may change. The effects of uncertainties
in the problem formulation can be readily evaluated once we have our model.
218 K. D. Wall, C. A. MacKenzie
We illustrate each of these situations using the Sloat Radar example. Suppose
there are four alternatives: (1) “do nothing” (keep the two existing radars at Sloat
which is labeled as Sloat 2); (2) purchase a third radar for Sloat of the same type
(which is labeled as Sloat 3); (3) purchase the new SkyRay radar; and (4) pur-
chase the new Sweeper radar. The latter two alternatives are new, with better range,
availability, interoperability and ECCM. These come at the cost of higher cogni-
tive load and less ease-of-use. The costs of procurement for the two new radars
are also higher than purchasing an existing radar. The data on which each of these
four alternatives can be assessed are depicted in Table 8.2.
1.000
0.900
0.800
0.700
0.600
=> 0.500
0.400
0.300
0.200
0.100
0.000
Sloat 2 Sloat 3 SkyRay Sweeper
Multiplying each of the global weights by the each of the values and adding
them together calculates an MoE for each radar, as pictured in Figure 8.7. SkyRay
is the most effective alternative and we know why. It has the highest value for
interoperability and range. Sweeper is second best because it has the second high-
est value for range. All alternatives possess approximately the same availability
rating. Interoperability and range are the attributes that are most important in
discriminating among these alternatives.
1.0
b.
b.
-~ ~
0.9 ·u
b.
...---:::::::: -~~-
b.
0.8
-~ ~
e>
0.7
... :::----
~
0.6
1.0
~ :t
0.8
~
-v·· ~
~
b.
e>
b.
0.7
~
0.6
o-----
0.5
0.45 0.50 0.55 0.60 0.65
1.000
0.900
0.800
0.700
0.600
e> 0.500
0.400
0.300
0.200
0.100
0000
Sloat 2 Sloat 3 SkyRay Sw eeper
1.000
0.900
0.800
0.700
0.600
G 0.500
>
0400
0.300
0.200
0.100
0.000
Sloat 2 Sloat 3 Sky Ray Sweeper
depicted in Figure 8.11. Sweeper is now the most effective alternative for this
future condition because it is the only radar with ECCM capability.
8.6 Cost-effectiveness
Computing the MoE for each alternative allows the decision-maker to order the
alternatives from most effective to least effective. This is only half the story,
though, because cost also matters. Ultimately the decision-maker will have to inte-
grate cost information with effectiveness and engage in cost-effectiveness analysis
(CEA). This section presents the conceptual elements used in the analysis and
outlines the process followed in the analysis.
1.00
0
0.90 SkyRay
0.80 0
0 0 Sweeper
Sloat2 Sloat3
0.70
e>
0.60
0.50
0.40
0.30
$0 $10 $20 $30 $40 $50 $60 $70 $80 $90
alternative, the exact meaning of which requires understanding the many ways the
solution can be interpreted.
<fJ 4
•7
r----------... 6
c
3
u
4
~
"'
>- 4
System Cost
<fJ
i;'l r •7
. •s
r----------...i 6
c
·"u
Q)
• 3
~ ~---------- . . 4
E
Q)
"'
>-
(f)
4
System Cost
I
I
I
max System Cost
I
·~
·~I e 6
- - - - - - ~;L - . - - ~------------ v(i) m;o
I
I
I
·~ I
I
I
I
"'"'
Ql
c: -------• 5 •7
--------, 3
Ql
.2:
t3
~ I
w I
E I
Ql I
Ui
>-
2• 1
(f) I
I
I
I
I
S y stem Cost
where W E and WC are the importance weights for effectiveness and cost,
respectively. Upper case letters distinguish these weights from those used in
Multiple objective decision-making 227
defining v( j ). W E represents the importance to the decision-maker of maximizing
effectiveness and WC represents importance to the decision-maker of minimiz-
ing cost. The weights must satisfy 0 ≤ W E ≤ 1, 0 ≤ WC ≤ 1, and W E + WC = 1.
The relative importance of these two conflicting objectives is captured by the
ratio WC /W E .
Equation (8.6) represents the preferences of a rational decision-maker. When
confronted by a choice between two alternatives equal in effectiveness, the
decision-maker will choose the less expensive alternative because vc (c( j ))
decreases as the cost c( j ) increases. Similarly, when confronted by a choice
between two alternatives equal in cost, the decision-maker will choose the alter-
native with the greater MoE because v( j ) is greater for the alternative that is more
effective.
The overall cost-effectiveness function V ( j ) defines a preference structure that
allows us to develop valuable insights. We can see this graphically by rearranging
the cost-effectiveness function and drawing a straight line to represent V ( j ) on
the cost-effectiveness graph. Express v( j ) as a function of V ( j ) and c( j ):
1 WC
v( j ) = V ( j) − · vc (c( j )).
WE WE
1 WC
v( j ) = V ( j) − · [1 − k · c( j )]
WE WE
1 WC WC
= V ( j) − + · k · c( j ).
WE WE WE
This is the equation for a straight line when we interpret c( j ) as the independent
variable and v( j ) as the dependent variable. The intercept for the line is given
by V ( j )/W E − WC /W E and the slope is given by k · WC /W E . This line repre-
sents all combinations of cost and effectiveness corresponding to a given level
of overall cost-effectiveness, V ( j ). As V ( j ) increases (for example, V ∗∗ > V ∗ in
Figure 8.17), this “isoquant” line shifts up and to the left. As we move closer to
the northwest corner of the cost-effectiveness plot, we move to greater levels of
overall cost-effectiveness.
Suppose the decision-maker is much more interested in minimizing cost
than maximizing effectiveness. This decision-maker would select weights where
WC W E . The slope of the lines representing constant overall cost-effectiveness
would be very steep. This situation is depicted in Figure 8.18. Alternative 3 is
the best in this situation because it lies on the highest achievable isoquant of
cost-effectiveness. A decision-maker who places more emphasis on maximiz-
ing effectiveness would choose weights that result in less steep lines of constant
228 K. D. Wall, C. A. MacKenzie
Patient
!113 !113
!113 Patient
Patient
Patient
!113
System Cost
/ / / / /
/ / / / /
/ / / / /
/ / / / /
/ / / / /
/ / / / /
"'"'c
/ / / / /
/ / / // • 7 /
Q)
Q) /
/
/
/ • s// / /
/
>
t5 /
/
/
/
!113 /
/
/
/
/
/
/
/
/
:Jg / / / / /
w / / / / /
/ / / / /
E / / / / /
Q)
/ / / /
1i5
>- / • 2 / / /
(/) / / / /
/ / / /
/ / / /
/ / / /
/ / / /
System Cost
Figure 8.18 WC W E .
overall cost effectiveness, which could yield a picture like Figure 8.19. Now alter-
native 5 is the most cost-effective. Finally, consider a decision-maker who places
considerable importance on the maximization of effectiveness, which implies that
W E WC . The slope of the lines is very small resulting in nearly flat isoquants of
cost-effectiveness as in Figure 8.20. Alternative 7 is now the most cost-effective
alternative.
The above three cases illustrate the importance of the efficient set, and the
most cost-effective alternative is selected from the set of efficient alternatives. The
most cost-effective depends on the decision-maker’s preferences for cost reduction
versus effectiveness maximization. The answer to the decision problem requires
elicitation of decision-maker preferences—we do not know what we mean by
“cost-effectiveness” until we incorporate decision-maker preferences over cost
and effectiveness.
Multiple objective decision-making 229
------------------------------
(/)
S y stem Cost
Figure 8.20 W E WC .
1.0 ,---------------------
//-/y----/-
//~_ _/_/_/~--/-/~/c----
//~/,---/-/~
/ ----/-/'/~---
//_.---/-/c
/ r--/-/'
/
0.9 S,weepef /
0.8
// o/ //"" / ,"b /" / " ,., / "
S,weepef
/ / Sl9af"2 "/ "stoat 3/" /" "/ "/
0.7
/// ,"" ,./' //" ,.,., /// ,-/' //"
G
> ,-"" ,/'/ ,/'/ ,"" ,-"" ,/'/ ,"" ,-/'
0.6 // // ," // ," /// // ,"
/ / / // ," / / /
/// ,"" /// // /" ," //" //"
/ / / ," ," / / /
0.5
~--~--~----~--~--~~-.~---.------.------.------,-----~
/ /
/
," /
1.0
0.9 S,weepef /
0.8
S,weepef /
0.7 "stoat
Sl9af"2 "stoat ,.,.,
G
>
0.6
0.5
0.4
0.3
$0 $10 $20 $30 $40 $50 $60 $70 $80 $90
System Discounted Life-Cycle Cost ($M)
and has greater interoperability. Framing the questions in this manner can help the
decision-maker understand precisely what additional capability he is getting for
an increased cost.
The more formal manner of answering which alternative is the most cost-
effective requires elicitation of the final set of importance weights, WC and W E .
Multiple objective decision-making 231
Suppose the decision-maker believes cost and effectiveness are equally important.
These preferences translates to WC /W E = 1.0 resulting in the picture given in
Figure 8.21. The decision-maker would select the “do nothing” alternative because
at least one isoquant line lies to the right of Sloat 2 and to the left of Sloat 3 and
SkyRay.
If the decision-maker believes maximizing effectiveness is twice as important as
minimizing cost, WC /W E = 0.5, which results in the picture given in Figure 8.22.
Sloat 2 is still most cost-effective because at least one isoquant line separates
Sloat 2 from Sloat 3 and SkyRay. If the decision-maker believes maximizing
effectiveness is four times more important than cost reduction, WC /W E = 0.25.
These preferences give the picture shown in Figure 8.23. SkyRay is the most cost-
effective because at least one isoquant line lies below SkyRay and above Sloat 2
and Sloat 3.
This method also can serve to find the ratio between WC and W E which would
make the decision-maker indifferent between two alternatives. For example, when
WC /W E = 1/3 we obtain the situation in Figure 8.24. Here Sloat 2 and SkyRay
are almost of equal cost-effectiveness. The analysis of cost-effectiveness afforded
by the model reduces the management question to one of asking the decision-
maker: “Do you feel that effectiveness is more than three times as important as
cost?” If the answer is yes, SkyRay should be selected. If the answer is no then
do nothing (Sloat 2) is the best alternative. The model helps to focus attention on
critical information.
Now that we have found the critical value of WC /W E 1/3, it is of interest
to consider the effects of uncertainty in the future condition. Two other plan-
ning scenarios besides the baseline were considered during the MoE discussion.
1 . 0.---------------
~-~-=---------------
----
~-----------------
----
~-~--.
--- D ---- ---
--- -----
-
------
0.9
------ --- --- --- ----- -------sk;Ray --- --- ---
----- --- ------- --- --------D ----
0.8
-------D
----- 0 ------
--- ---
$90
0.9 Kind
0.8
Kind
0.7 Kind Kind
G
> ~~~
0.6
0.5
0.4
$90
$90 $10 $20 $30 $40 $50 $60 $70 $80 $90
System Discounted Life-Cycle Cost ($M)
If the supersonic attack scenario is highly likely, then we use different weights
in the definition of effectiveness. The v( j ) values change for all alternatives and
Figure 8.25 shows that SkyRay is the most cost-effective if WC /W E ≤ 0.57. The
decision-maker should be asked: “Is effectiveness more than 1.75 times as impor-
tant than cost?” If he answers yes, SkyRay is the best alternative. The ECM
planning scenario changes the importance weights in effectiveness, and the result-
ing cost-effectiveness is shown in Figure 8.26. If WC /W E = 1/3, Sweeper is the
best alternative. Under this scenario, Sweeper remains the best alternative as long
as the decision-maker values effectiveness at least 1.4 times as much as cost, or
WC /W E ≤ 0.74.
8.7 Summary
Multiple objective decision problems arise very frequently. Their solution requires
the decision-maker to first determine what really matters and list the issues or
consequences of concern. This process of discovery should be pursued through
the construction of an objectives hierarchy. It is the single most important step
towards a solution. Without doing this we run the risk of not knowing “what is the
real problem” and of not asking “the right question.”
The lowest levels of the hierarchy define the individual measures of effective-
ness. These are the natural, constructed or proxy measurement scales by which
we begin to quantify the effectiveness of an alternative. The proper integration of
these measures into a single MoE requires quantification of decision-maker pref-
erences. First we need to know decision-maker preferences over marginal changes
in the individual effectiveness measures—the answer to “How much is enough?”
Multiple objective decision-making 233
1.0
0.9
,....skyRa~
0.8
0.7
,....skyRa~
e>
0.6
,....skyRa~ ,....skyRa~
0.5
0.4
0.3
$0 $10 $20 $30 $40 $50 $60 $70 $80 $90
System Discounted Life-Cycle Cost ($M)
1.0 /
/
/
/
/
/ /
/
/
/
/ / / / /
/ / / / /
/ / / / /
/ / / / / /
/ / / / / /
/ / / / / /
0.9 /
/
/
/
/
/
,."""' p "' ,."'
/
/
/
/
/ / / / / /
/ / /
/
/
/
/
/
/
/
/
/
/'
/
;>w'e..ep~r / /
/
/
/
/
/
/
0.8 /
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/ / / / /
/ / / / / / /
/ / / / / / /
/ / / / / / /
/ / / / / / / /
/ / / / / / / /
/ / / / / / / /
0.7 /
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/ / / / / / / /
/ / / / / / /
e>
/
,."'
/ / / / / / / /
/
/
/
/
/
/
,." ,.,. /
/
/
/
/
/
/
/ / / / / / /
/ / / / / / / / /
0.6 /
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/ / / /
/ / / / / / / / /
/ / / / / / / / /
/ / / / / / / / /
/ / / / / / / / / /
/ / / / / / / / / /
/ / / / / / / / / /
0.5 /
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
/
,-0
/ / / / / / / / /
/ / / / / / / / /
/ / / / / / / / /
/ / / / / / / / /
,....skyRa~.........
/ / / / / / / / /
/ / / /
/
/
/ /
/
/
/ / / / / /
/
/
/
/
/
, / ~16at-2 ' / / /
/
/
/
Sloa~ )- '
/
/
/
/
/
/
/
/
/
/
/
/
/ / / / / / / / /
0.3
$0 $10 $20 $30 $40 $50 $60 $70 $80 $90
System Discounted Life-Cycle Cost($M)
These changes can represent increases (when more is better) or decreases (when
less is better). This information allows us to convert the individual effectiveness
measures into individual value measures on a 0–1 scale. Second we need to know
decision-maker preferences across the individual measures—the answer to “How
234 K. D. Wall, C. A. MacKenzie
important is it?” This information allows us to combine the individual scaling
functions using a weighted sum defined by the importance weights.
Next, we combine the resulting MoE with the discounted life-cycle cost. This
provides all the information needed to conduct the analysis of cost-effectiveness.
Viewing things in a two-dimensional framework expedites thinking about the
solution by allowing visual inspection and making use of humans’ ability at
pattern recognition. In this framework, five solution concepts apply: (1) the supe-
rior solution, (2) the efficient solution, (3) the satisficing solution, (4) marginal
reasoning, and (5) weighting cost versus effectiveness. The first concept should
always be sought, and a superior solution (if one exists) is the best alternative
regardless of the preferences for cost reduction versus effectiveness maximiza-
tion. The efficient solution and the satisficing solution often yield non-unique
answers and gives the decision-maker flexibility over which alternative to select.
The set of efficient solutions is intrinsically important because because the most
cost-effective alternative will come from this set. Selecting the most cost-effective
alternative requires eliciting additional decision-maker preferences. We must
know the decision-maker’s preferences over cost minimization and effectiveness
maximization before determining the most cost-effective alternative.
1 − e−azi
vi (z i ) =
K
0.9
"'"'
""
0.8
""""
0.7
0.6
.,
""""
:::l 0.5
~
0.4
""""
0.3
0.2
""
0.1
0
$0 $20 $40 $60 $80 $100 $120
z i = x i − x min
z i = x max − x i
zi
vi (z i ) = .
x max − x min
2
1 − e−azi −b1 zi
vi (z i ) = ,
K1
or a cubic term,
2 3
1 − e−azi −b1 zi −b2 zi
vi (z i ) = ,
K2
2
where K 1 = 1 − e−a[xmax −xmin ]−b1 [xmax −xmin ] and K2 = 1 −
2 3
e−a[xmax −xmin ]−b1 [xmax −xmin ] −b2 [xmax −xmin ] . The quadratic form with a = 0 allows
value functions with an “S-shape” to be represented. The cubic form with a = b = 0
permits us to incorporate value functions that appear almost like “step” functions.
Higher-order terms can be specified but in practice are almost never needed.
The additional terms permit a wider range of preference behavior to be modeled
but are hardly ever worth the increased complexity.
References
Buede, Dennis M., and Terry A. Bresnick. 1992. “Applications of Decision Analysis to the
Military Systems Acquisition Process.” Interfaces 22:110–25.
Clemen, Robert T., and Terence Reilly. 2001. Making Hard Decisions with DecisionTools.
Mason, OH: South-Western.
236 K. D. Wall, C. A. MacKenzie
Dyer, James S., Thomas Edmunds, John C. Butler, and Jianmin Jia. 1998. “A Multiat-
tribute Utility Analysis of Alternatives for the Disposition of Surplus Weapons-Grade
Plutonium.” Operations Research 46:749–62.
Ewing Jr., Paul L., William Tarantino, and Gregory S. Parnell. 2006. “Use of Decision Anal-
ysis in the Army Base Realignment and Closure (BRAC) 2005 Military Value Analysis.”
Decision Analysis 3:33–49.
Hammond, John S., Ralph L. Keeney, and Howard Raiffa. 1999. Smart Choices: A Practical
Guide to Making Better Life Decisions. New York: Broadway Books.
Keeney, Ralph L. 1996. Value-Focused Thinking: A Path to Creative Decisionmaking.
Cambridge, MA: Harvard University Press.
Keeney, Ralph L., and Howard Raiffa. 1976. Decisions with Multiple Objectives: Prefer-
ences and Value Tradeoffs. New York: John Wiley & Sons.
Kirkwood, Craig W. 1997. Strategic Decision Making: Multiobjective Decision Analysis
with Spreadsheets. Belmont, CA: Brooks/Cole.
Parnell, Gregory S., Harry W. Conley, Jack A. Jackson, Lee J. Lehmkuhl, and John M.
Andrew. 1998. “Foundations 2025: A Value Model for Evaluating Future Air and Space
Forces. Management Science 44:1336–50.
Parnell, Gregory S., Ronald E. Metzger, Jason Merrick, and Richard Eilers. 2001. “Mul-
tiobjective Decision Analysis of Theater Missile Defense Architectures.” Systems
Engineering 4:24–34.
Stillwell, William G., David A. Seaver, and Ward Edwards. 1981. “A Comparison of Weight
Approximation Techniques in Multiattribute Utility Decision Making.” Organizational
Behavior and Human Performance 28:62–77.
von Winterfeldt, Detlof, and Ward Edwards. 1986. Decision Analysis and Behavioral
Research. Cambridge, UK: Cambridge University Press.
9 A new approach to evaluate
safety and force protection
investments
The value of a statistical life
Thomas J. Kniesner, John D. Leeth,
and Ryan S. Sullivan
9.1 Introduction
A fundamental tenet of economics is that public decisions should be evaluated in
terms of benefits and costs. Such decisions include laws, rules, regulations, policies,
and investments, and all are aimed at improving civilian safety and reducing military
casualties. Economic welfare is only increased if the benefits of safety improve-
ments exceed their costs. Whereas calculating monetary costs of safety investments
is relatively well established—determined through engineering or accounting stud-
ies (see Chapter 5, for example)—monetizing benefits can be highly controversial.
As we explain in this chapter, assessing the benefit of fewer fatalities and injuries
requires both an accurate forecast of fatalities (and injuries) prevented or eliminated
and an estimate of the monetary value of lives saved or injuries avoided.
Some contend that no monetary value can be placed on human life. The associ-
ated implication is that life is infinitely precious, so any action to improve safety
regardless of the cost is worthwhile. The dilemma facing government officials can
be stated as follows: although everyone believes saving lives is moral, societal
resources are limited, so actions to save lives in one area come at the expense of
other worthwhile goals, including the possibility of saving lives elsewhere. This
chapter reveals how pricing lives can allow government officials to determine if
the resources devoted to improving safety in one area might be better allocated to
other areas or other programs. We urge governments with limited budgets to con-
sider using the value of a statistical life (VSL) approach described in this chapter in
any cost–benefit analysis (CBA) of alternative force protection investments. This
will not only ensure resources are allocated to their highest valued use, but more
importantly, may result in more lives saved.
Any military that operates as if human life had infinite value would soon be
out of business. The focal message of our chapter is that hard choices must be
made because complete safety is impossible. Approving every force protection
investment (in armaments, technology or training) to reduce casualties or injuries
would rapidly exhaust military budgets.
In reality, most of us do not behave as if our life has infinite value. Many make
conscious decisions to drive small cars, knowing large cars are significantly safer,
because small cars are cheaper to own and operate. We routinely engage in risky
238 T. J. Kniesner, J. D. Leeth, R. S. Sullivan
activities (skiing, biking, roller-blading, etc.), because the perceived benefits com-
pensate for the risks. People smoke, drink alcoholic beverages, and eat unhealthy
food, even though health hazards are well known, presumably because the imme-
diate pleasure outweighs any possible long-run consequences. In short, people
routinely make cost–benefit decisions that influence their own safety, and those
decisions do not reflect an infinite value of life.
This chapter examines how economists evaluate the costs and benefits of safety
improvements, and then provides a contemporary case study of a major recent
force protection investment. The study examines the cost-effectiveness of adding
armor protection to tactical wheeled vehicles (TWVs) to protect soldiers. We
demonstrate how the conclusions depend on the measure called the “value of a
statistical life,” or VSL.
A common approach to value the benefits of greater safety is to observe actual
tradeoffs people make between safety and other valuable job or product charac-
teristics (Viscusi and Aldy 2003). Consider two identical jobs that only differ in
terms of safety: one job is completely safe, but in the other job there is a 1 in
100,000 (or 0.00001) chance of a workplace injury resulting in death. The com-
pletely safe job pays US$80 less per year than the job with the slight chance of a
workplace fatality.
This scenario offers two complementary perspectives. Workers in the safe job
essentially sacrifice US$80 annually in the form of a lower wage to eliminate
a 0.00001 chance of injury on the job that could result in death over the next
year. Alternately, workers in the dangerous job demand a wage premium of
US$80 extra annually to accept a 0.00001 increase in the chance of a deadly
injury.
One possible interpretation is that if 100,000 workers were each willing to pay
US$80 per year (a total of US$8 million) to avoid a 0.00001 probability of injury,
then one life per year would be saved on average. Collectively, the US$8 million
is the value workers implicitly place on that one life and implicitly on their own
individual lives. Because the actual life that is saved is unknown, but in a sense
is drawn randomly from the 100,000 workers, economists refer to the resulting
figure (US$8 million) as the VSL.
The VSL in this case can be obtained by dividing the wage gain in the riskier
occupation (US$80) by the higher probability of death (0.00001), resulting in
a dollar figure (US$80/0.00001 = US$8 million) standardized to one life saved.
Although the monetary value is expressed as the amount per life, it is derived by
observing the actual amount of money workers sacrifice to avoid a slightly higher
chance of death, and so might be better thought of as the value of mortality risk
reduction (US Environmental Protection Agency (EPA) 2010b).1
After describing the theory behind VSLs, this chapter briefly summarizes vari-
ous approaches used to generate VSL estimates and provides a broad overview of
empirical results from the literature. VSLs can vary by income and demographic
characteristics (age, gender, race), so that no single estimate is appropriate in all
situations. In fact, the range of VSL estimates reported in the literature turns
out to be sizeable, anywhere from US$0 to US$40 million. However, careful
Value of a statistical life 239
empirical modeling and improved risk information reduce the likely relevant range
of VSLs in the United States to between US$4 million and US$10 million. For
example, US Government agencies such as the EPA and US Department of Trans-
portation (DOT) generally evaluate safety improvements using a VSL of around
US$8 million.2
The chapter concludes with a force protection case study that examines the cost-
effectiveness of two large-scale vehicle replacement programs recently imple-
mented by the US military to increase vehicle armor protection for troops in
day-to-day operations. The principal aim of the replacement programs was to
reduce fatality and injury risks faced by US military personnel.
This case study illustrates a common problem when evaluating safety or force
protection investments. Estimates of the benefits of risk reduction are based on
engineering studies that do not consider changes in behavior that might mitigate
or completely counteract the safety investment. For example, requiring drivers
to wear seatbelts might reduce deaths or injuries in an accident, but could also
encourage those drivers to drive faster or more recklessly, thus increasing the num-
ber of accidents. The behavioral response places drivers at greater risk, as well
as increasing risks to other drivers, passengers, and pedestrians (Peltzman 1975).
Simply estimating the lower probability of death or injury from using a seatbelt
in a specific accident scenario will tend to overestimate the benefits of requiring
seatbelt usage if behavioral responses that increase the total number of accidents
are not included.
The case study presented in this chapter shows that the move from a lightly
armored vehicle to a moderately armored vehicle reduced infantry fatalities at
a relatively low cost of US$1 million to US$2 million per life saved for some
military units. Given an estimated VSL in the range of US$4 million to US$10 mil-
lion, such a force protection investment easily passes the cost-effectiveness test.
Interestingly, switching to a more costly, highly armored third alternative did
not appreciably reduce fatalities in any of the different military units examined,
despite engineering studies that demonstrated the new vehicle was clearly safer
against specific risks such as improvised explosive devices (IEDS).
An explanation for the surprising lack of life-saving benefits observed from
the added armor protection in the most costly vehicle may come from offset-
ting behavior of our own troops or from actions by the enemy in response to
the change. Both may have contributed to counteract the added force protec-
tion investment of additional armor, ultimately reflecting diminishing returns that
essentially neutralized the value of the most costly safety investment. This result
points to the importance of properly framing a cost–benefit study, especially
when lives are at stake, and the value of VSL to help guide force protection
investments.
π=0 π = πhigh
Wage Wage Sh
whigh
S0
w0 Dh
D0
Labor Labor
Wage
W (π)
Military
pay
P*
Supply with
a draft
Observed difference
in quantity supplied
m
n
ln(w) = c + αi πi + β j X j + ε, (9A.1.1)
i=1 j =1
Value of a statistical life 253
where ln(w) is the natural logarithm of wage; πi are measures of workplace risk;
X i are demographic variables (such as education, race, marital status, and union
membership) and job characteristics (such as wage replacement under workers’
compensation insurance, and industry, occupation, or geographic location indi-
cators); ε is an error term; and c, αi , and β j are parameters to be estimated.
Many studies also include interaction terms between workplace risk and various
X i to determine variation in risk compensation by type of worker (union/nonunion,
white/black, native/immigrant, young/old) or by level of income replacement from
workers’ compensation insurance (Viscusi and Aldy 2003).
m
n
ln( p) = c + αi E i + β j C j + ε, (9A.2.1)
i=1 j =1
where ln( p) is the natural logarithm of price; the E i are environmental hazards;
and C j are structure characteristics, neighborhood, and other control variables
that affect price (for instance, number of rooms, number of bathrooms, lot size,
neighborhood median income, and distance to central city) (Gayer et al. 2000).
3
fatalitiesit = αcjf × q j it + βcf xit + εitf , (9A.3.1)
j =1
where q1it , q2it , and q3it represent the quantities of each of three types of vehicles
possessed by unit i in month t; xit is a vector of control variables that might include
other vehicle quantities, troop characteristics, or fixed effects for month, province,
month by province, or unit; εitf is random error; and the coefficients are allowed
to vary by the unit’s classification c as infantry, armored or cavalry, administrative
254 T. J. Kniesner, J. D. Leeth, R. S. Sullivan
and support, or other. Let vehicle types one, two, and three be defined as Type 1,
Type 2, and Type 3 TWVs.
Because the focus of the Rohlfs and Sullivan (2013a) study is the effects of
replacing one vehicle type with another, it is convenient to rearrange the terms in
Equation (9A.3.1) to obtain the following specification:
Equation
3
(9A.3.2) serves as their main regression specification, with
j =1 q j it , q2it , q3it , and varying formulations of xit as the regressors. The differ-
ences (αc2f − αc1f ) and (αc3f − αc1f ) measure the effects of replacing Type 1 TWV
with a Type 2 or Type 3 TWV.
Rohlfs and Sullivan (2013a) next focus on the costs related to the vehicles by
estimating changes in expenditures by unit as a function of the different vehicle
types and their usage rates. Their expenditure modeling technique follows.
To relate effects on fatalities to dollar costs, let p0 j denote the purchase price
(including shipping cost), and let p1 j denote the additional cost per mile driven
for a type j vehicle. Let mileage j it denote the miles driven of type j vehicles by
unit i in month t, and define expenditureit as unit i ’s month t expense for all three
vehicle types
3
1
expenditureit = p0 j × q j it + p1 j × mileage j it (9A.3.3)
j =1
36
expenditureit = (αc2
e
− αc1
e
) ∗ q2it + (αc3
e
− αc1
e
) ∗ q3it
3
+ αc1
e
∗ q j it + βce xit + εite (9A.3.4)
j =1
where (αc2
e
− αc1e
) and (αc3
e
− αc1
e
) represent the monthly costs of replacing a Type 1
TWV with a Type 2 or Type 3 TWV. Expressing the costs of changing vehicle type
through Equation (9A.3.4) helps to ensure that the same factors are held constant
and the same sets of observations are compared to one another for the cost as
for the fatalities estimation. The mileage and expenditure data are constructed
from the SDC sample and use the same sampling weights as the vehicle counts;
consequently, any measurement error in vehicle quantities that affects the fatalities
regressions does not affect the coefficients in Equation (9A.3.4).
Value of a statistical life 255
The cost per life saved from replacing a Type 1 with a Type 2 TWV is
f f
−(αc2e
− αc1
e
)/(αc2 − αc1 ), which expressed verbally is negative one times the ratio
of the difference between the coefficient for Type 2 TWVs and the coefficient for
Type 1 TWVs from Equation (9A.3.4), the expenditures equation, divided by the
corresponding difference in coefficients from Equation (9A.3.3), the fatalities equa-
tion. The cost per life saved from replacing a Type 2 TWV with a Type 3 TWV
is −(αc3e
− αc2
e
)/(αc3f − αc2f ), which is computed as negative one times the differ-
ence between the coefficient for Type 3 TWVs and the coefficient for Type 2 TWVs
from Equation (9A.3.4)divided by the corresponding difference in coefficients from
Equation (9A.3.3). Other dependent variables considered in the study are combat-
related injuries and miles driven by different vehicles. Rohlfs and Sullivan calculate
standard errors for the ratios by estimating Equations (9A.3.3)and (9A.3.4) together
using seemingly unrelated regression and applying the delta method.
Measuring effects on injuries helps to identify a benefit of changing vehicle
type not included in the cost per life saved calculations. To interpret the effects on
injuries better, Rohlfs and Sullivan (2013a) compute the ratios of injuries reduced
per life saved as (αc2i
− αc1
i
)/(αc2f − αc1f ) and (αc3
i
− αc2
i
)/(αc3f − αc2f ), where the i
superscript denotes coefficients from the injury equation. Measuring effects on the
mileage variables helps to determine the extent to which the replacement policies
led units to alter their behavior.
Notes
1 Rarely do you see two jobs that are otherwise identical except for the chance of a work-
place fatality. Economists derive VSL estimates econometrically using large datasets
that allow them to control for other factors affecting wages or product prices to generate
the ceteris paribus impact of risk. It is worth noting, though, that sometimes risks are
explicitly priced for workers by employers—for example, with so-called hazardous duty
pay in the military.
2 Implementing a safety improvement that costs more than the benefits generated can
result in more lives being lost because the funding could be better spent on alter-
native safety improvements. Kniesner and Leeth (2004), for instance, examine safety
inspections of underground coal mines by the Mine Safety and Health Administration
(MSHA) and find that inspections fail the cost-benefit test. They show that eliminat-
ing one safety inspection per mine per year (MSHA inspects underground coal mines
quarterly) and moving the funding to other activities such as heart disease screening
or purchasing defibrillators, would save 39,700 more miners’ lives. Formally, the cost-
benefit test shows whether resources are allocated efficiently among competing uses,
but many times a competing use can be an alternative means of improving safety and
saving lives.
3 The “all else equal” requirement means that one might not always see a simple positive
relationship between an occupation’s pay and its level of risk. Loggers face greater risk
on the job than financial analysts, but financial analysts earn more than loggers. The
additional education required to become a financial analyst outweighs the added risk
of logging. Economists attempt to control for other factors affecting wage rates using
large datasets and multiple regression techniques. Two popular reality TV shows, The
History Channel’s Ice Road Truckers and the Discovery Channel’s Deadliest Catch,
provide nice anecdotal examples of the relationship between pay and risk holding all
else equal. The two shows follow workers who face the most extreme hazards in their
256 T. J. Kniesner, J. D. Leeth, R. S. Sullivan
professions (driving across frozen lakes and rivers in the Arctic and fishing for Alaskan
king crabs in the Bering Sea) and document both the dangers they face and the very high
monetary rewards they receive for facing the dangers.
4 To use an analogy, it is as if the worker is being given the choice between a dollar with
certainty or a gamble where there is a 1 − π chance of a dollar payoff and a π chance
of a 0 payoff. Regardless of risk preferences, rational workers would always choose the
dollar payoff with certainty.
5 The wage gap sorts firms and workers into the two submarkets. The firm on the margin is
indifferent between offering safe or hazardous employment. For the so-called marginal
firm, the costs of the safety programs necessary to eliminate job hazards just equal
the benefits of greater safety, the drop in wage. Likewise, the worker on the margin
is indifferent between the two types of jobs: the utility gain from the higher wage just
offsets the utility loss from the greater likelihood of a drop in income, medical expenses,
and pain and suffering from an injury or illness.
6 See Kniesner and Leeth (1995, Chapter 3) for a formal derivation.
7 Viscusi uses the CFOI data to determine mortality risk by first grouping fatalities by
2-digit industry (72 industries) and then by 1-digit occupation (10 occupations) to gen-
erate a total of 720 industry-occupation cells. He then calculates the frequency of a
workplace death by dividing the average number of fatalities by the average number
of employees within each industry-occupation cell, using data from the period 1992 to
1997. By averaging over six years, Viscusi minimizes possible errors in characterizing
risk resulting from the random fluctuation of workplace fatalities over time.
8 As is fairly typical, he excludes from his sample agricultural workers, part-time work-
ers, workers younger than 18, and workers older than 65. Viscusi also does not include
interactions between fatal injury risk and other control variables (see discussion of
Equation (9A.1.1) in Appendix 9.1).
9 Although the VSL function depends on the values of the right-hand side in Equation
(9A.1.1) in Appendix 9.1, most commonly considered is the mean VSL. With a fatality
risk measure of deaths per 100,000 workers and a work year of 2,000 hours, the value of
a statistical life is VSL = α × exp(ln(w)) × 100,000 × 2,000. If α is small then it is the
case that α × exp(ln(w)) ≈ αw. (See Appendix 9.1.) All dollar figures presented have
been adjusted for inflation to 2010 using the CPI-U.
10 As part of a Remedial Investigation of a Superfund site, the EPA releases an assess-
ment of the site’s risks across the various affected areas, including the elevated risk of
contracting cancer (Gayer et al. 2000).
11 The range of values arises from differences in empirical specifications underlying the
estimates of costs and risk. Rohlfs argues credit constraints prevented many young men
from attending college during the period, causing the enrollment data to understate the
cost of the draft and the VSL of draft-age men.
12 One example of such an approach is Vassanadumrongdee and Matsuoka (2005). They
determine the VSL in Bangkok, Thailand using a survey instrument that describes the
current level of air pollution in the city as resulting in annual deaths of 430 in 1,000,000.
Individuals are then told that they can reduce their chance of dying from air pollution
by 30 in 1,000,000 (or 60 in 1,000,000) if they take an annual health exam that detects
respiratory system impairments. Each individual is randomly assigned one of four prices
and asked if they would be willing to pay the amount for the annual checkup. Based on
the hypothetical price that would need to be paid for the checkup and the reduction
in the chance of premature death from air pollution from taking the health exam, they
determine a VSL in Bangkok of about US$1 million.
13 In terms of Figure 9.2, market studies estimate the market wage equation W (π),
whereas contingent valuation studies estimate the preferences underlying the wage
equation. In equilibrium, the slope of the market wage function equals the marginal
benefit to the worker from a slightly safer job. The equality of the two values means it
Value of a statistical life 257
is unnecessary to estimate underlying preferences to determine appropriate VSLs, but
for more substantial improvements in safety market-determined values will overstate
people’s willingness to pay for greater safety.
14 Contingent valuation surveys are also more challenging to construct than other surveys
because many people have difficulty properly understanding complex probability con-
cepts and are unaccustomed to the notion of trading income for expanded safety. For
example, people generally understand the simple difference in payoffs between betting
on a race horse with 4:1 winning odds versus betting on a horse with 2:1 winning odds.
They might have a harder time, however, understanding whether or not they would sup-
port a government policy that reduces mercury levels in the local water supply and thus
decreases the risk of death for each constituent by 0.00001 at a cost of US$50 per tax-
payer. If respondents do not fully understand the probability concepts in surveys then
the VSL estimates from the survey may be inaccurate.
15 For further discussion see Carson et al. (2001) and Hausman (2012).
16 VSL estimates vary depending on both the set of control variables included in the esti-
mating wage or price equation and the underlying population examined (Viscusi 2013).
In the United States, VSL estimates are higher for union workers than for non-union
workers, higher for whites than blacks, and higher for women than men (Viscusi and
Aldy 2003; Viscusi 2003; Leeth and Ruser 2003). VSLs for native workers are roughly
the same size as for immigrant workers, except for non-English speaking immigrants
from Mexico who appear to earn little compensation for bearing very high levels of
workplace risk (Hersch and Viscusi 2010). VSLs rise as people age, at least until the
mid-40s, and then gradually decline (Kniesner et al. 2006; Aldy and Viscusi 2008). Not
surprisingly given that safety is generally considered to be a normal good, VSLs are
larger for higher income groups within a country at a point in time, within a country
over time as the country’s income increases, and within developed countries than within
less developed countries (Mrozek and Taylor 2002; Viscusi and Aldy 2003; Costa and
Kahn 2004; Bellavance et al. 2009; Kniesner et al. 2010). VSLs can vary across pop-
ulations because of differences in risk preferences. Groups more willing to take risks
will locate further to the right along the market wage function, and if the wage function
is concave from below as shown in Figure 9.2, more risk-tolerant groups will have a
smaller reduction in wage for a given increase in safety and a lower VSL. Alternatively,
some populations may have a lower VSL because they are less careful than others and
their lower safety productivity causes the wage function they face to lie below and to
be flatter (resulting in a lower VSL) than the wage function faced by others. The lower
ability of some workers to produce safety makes it desirable for employers to offer them
smaller wage premiums for accepting risk. Discrimination may also result in two sepa-
rate market wage functions, with the disadvantaged group facing not only lower wages
at every level of risk but also smaller increases in wages for given changes in risk than
for the advantaged group.
17 A difficulty with meta-analyses is that they include results from studies with known
problems, which “imparts biases of unknown magnitude and direction” (Viscusi 2009,
p. 118).
18 The authors devote particular attention to measurement errors, which have been noted in
Black and Kniesner (2003), Ashenfelter and Greenstone (2004), and Ashenfelter (2006).
Although they do not have information on subjective risk beliefs, they use very detailed
data on objective risk measures and consider the possibility that workers are driven by
risk expectations. Published industry risk beliefs are strongly correlated with subjective
risk values, and they follow the standard practice of matching to workers in the sample
an objective risk measure. Where Kniesner et al. (2012) differ from most previous stud-
ies is the pertinence of the risk data to the worker’s particular job, and theirs is the first
study to account for the variation of the more pertinent risk level within the context of a
panel data study. Their work also distinguishes job movers from job stayers. They find
258 T. J. Kniesner, J. D. Leeth, R. S. Sullivan
that most of the variation in risk and most of the evidence of positive VSLs stems from
people changing jobs across occupations or industries rather than from variation in risk
levels over time in a given job setting.
19 Rohlfs and Sullivan (2013a) used a variety of publicly available datasets from the
Department of Defense (DoD) in their empirical research. They obtained vehicle data
from the Theater A portion of US Army Materiel Systems Analysis Activity’s Sam-
ple Data Collection (US AMSAA, SDC, 2010) and cost information from cost experts
within the Army Material Command (AMC) and Cost Analysis and Program Evalua-
tion (OSD-CAPE) departments in DoD. Some of the data used in Rohlfs and Sullivan
(2013a) are considered For Official Use Only (FOUO). For another author to use the
data requires that person to have an official defense-related government purpose. Rohlfs
and Sullivan have made arrangements with the US Naval Postgraduate School (NPS) to
sponsor potential replicators of their project, so that replicators could obtain access to
the data.
20 All prices presented in the replacement vehicle case study are in 2010 dollars.
21 This is most likely an underestimate of the actual cost of the program because it does
not include many costs such as maintenance and fuel. It should also be noted that the
US$50 billion figure is for all Services, not just the Army, which was the focus of the
study by Rohlfs and Sullivan (2013a).
22 The three cost estimates use 484 miles per month as a constant rate of usage, which is
the rate observed for the average vehicle in the SDC data.
23 See Tables I–IV in Rohlfs and Sullivan (2013a) for details on their findings.
24 Exceptions, however, could be made for particularly combat-intensive units to receive
heavily armored Type 3 TWVs depending on the combat environment and their specific
capability requirements.
References
Aldy, J. E., and W. K. Viscusi. 2008. “Adjusting the Value of a Statistical Life for Age and
Cohort Effects.” Review of Economics and Statistics 90: 573–581.
Ashenfelter, O. 2006. “Measuring the Value of a Statistical Life: Problems and Prospects.”
Economic Journal 116: C10–C23.
Ashenfelter, O., and M. Greenstone. 2004. “Estimating the Value of a Statistical Life: The
Importance of Omitted Variables and Publication Bias.” American Economic Association
Papers and Proceedings 94: 454–460.
Bellavance, Francois, Georges Dionne, and Martin Lebeau. 2009. “The Value of a Statis-
tical Life: a Meta-Analysis with a Mixed Effects Regression Model.” Journal of Health
Economics 28.2: 444–464.
Black, D. A., and T. J. Kniesner. 2003. “On the Measurement of Job Risk in Hedonic Wage
Models.” Journal of Risk and Uncertainty 27: 205–220.
Brown, J. N., and H. S. Rosen. 1982. “On the Estimation of Structural Hedonic Price
Models.” Econometrica 50: 765–768.
Carson, R. T., N. Flores, and N. F. Meade. 2001. “Contingent Valuation: Controversies and
Evidence.” Environmental and Resource Economics 19: 173–210.
Costa, D. L., and M. E. Kahn. 2004. “Changes in the Value of Life, 1940–1980.” Journal
of Risk and Uncertainty 29: 159–180.
Epple, D. 1987. “Hedonic Prices and Implicit Markets: Estimating Supply and Demand
Functions for Differentiated Products.” Journal of Political Economy 95: 59–80.
Gayer, T., J. T. Hamilton, and W. K. Viscusi. 2000. “Private Values of Risk Tradeoffs
at Superfund Sites: Housing Market Evidence on Learning about Risk.” Review of
Economics and Statistics 82: 439–451.
Value of a statistical life 259
Hausman, J. 2012. “Contingent Valuation: From Dubious to Hopeless.” Journal of Eco-
nomic Perspectives 26: 43–56.
Hersch, J., and W. K. Viscusi. 2010. “Immigrant Status and the Value of Statistical Life.”
Journal of Human Resources 45: 749–771.
Kahn, S., and K. Lang. 1988. “Efficient Estimation of Structural Hedonic Systems.”
International Economic Review 29: 157–166.
Kniesner, T. J., and J. D. Leeth. 1995. Simulating Workplace Safety Policy. Boston, MA:
Kluwer Academic Publishers.
——. 2004. “Data Mining Mining Data: MSHA Enforcement Efforts, Underground Coal
Mine Safety, and New Health Policy Implications.” Journal of Risk and Uncertainty 29:
83–111.
Kniesner, Thomas J., W. Kip Viscusi, and James P. Ziliak. 2006. “Life-Cycle Consumption
and the Age-Adjusted Value of Life.” Contributions in Economic Analysis & Policy 5.1.
——. 2010. “Policy Relevant Heterogeneity in the Value of Statistical Life: New Evidence
from Panel Data Quantile Regressions.” Journal of Risk and Uncertainty 40: 15–31.
Kniesner, T. J., W. K. Viscusi, C. Woock, and J. P. Ziliak. 2012. “The Value of a Statistical
Life: Evidence from Panel Data.” Review of Economics and Statistics 94: 74–87.
Leeth, J. D., and J. Ruser. 2003. “Compensating Wage Differentials for Fatal and Nonfatal
Injury Risk by Gender and Race.” Journal of Risk and Uncertainty 27: 257–277.
Mrozek, J. R., and L. O. Taylor. 2002. “What Determines the Value of Life? A Meta-
Analysis.” Journal of Policy Analysis and Management 21: 253–270.
Peltzman, S. 1975. “The Effects of Automobile Safety Regulation.” Journal of Political
Economy 83: 677–725.
Office of Management and Budget. 1992. Circular No. A-94 Revised. The White House.
October 29. Available at www.whitehouse.gov/omb/circulars_a094 [last accessed
November 18, 2013].
Rohlfs, C. 2012. “The Economic Cost of Conscription and an Upper Bound on the Value
of a Statistical Life: Hedonic Estimates from Two Margins of Response to the Vietnam
Draft.” Journal of Benefit-Cost Analysis 3: 1–37.
Rohlfs, C., and R. Sullivan. 2013a. “The Cost-Effectiveness of Armored Tactical Wheeled
Vehicles for Overseas US Army Operations.” Defence and Peace Economics 24: 293–316.
Rohlfs, C., and R. Sullivan. 2013b. “A Comment on Evaluating the Cost-Effectiveness of
Armored Tactical Wheeled Vehicles.” Defence and Peace Economics 24: 485–494.
US Army Material Systems Analysis Activity. 2010. Theater A Sample Data Collection.
US Defense Manpower Data Center. 2008. Unit Master File. Data Request System number
23242.
US Environmental Protection Agency. 2010a. Guidelines for Preparing Economic
Analyses. Available at http://yosemite.epa.gov/ee/epa/eerm.nsf/vwAN/EE-0568-50.pdf/
US$file/EE-0568-50.pdf [last accessed November 18, 2013].
——. 2010b. Valuing Mortality Risk Reductions for Environmental Policy: A
White Paper. Available at http://yosemite.epa.gov/ee/epa/eerm.nsf/vwAN/EE-0563-
1.pdf/US$file/EE-0563-1.pdf [last accessed November 18, 2013].
US Department of Transportation. 2008. Revised Departmental Guidance 2013: Treat-
ment of the Value of Preventing Fatalities and Injuries in Preparing Economic Analyses.
Available at www.dot.gov/sites/dot.dev/files/docs/VSL%20Guidance%202013.pdf [last
accessed November 18, 2013].
Vassanadumrongdee, S., and S. Matsuoka. 2005. “Risk Perceptions and Value of a Statis-
tical Life for Air Pollution and Traffic Accidents: Evidence from Bangkok, Thailand.”
Journal of Risk and Uncertainty 30: 261–287.
260 T. J. Kniesner, J. D. Leeth, R. S. Sullivan
Viscusi, W. K. 2003. ‘Racial Differences in Labor Market Values of a Statistical Life’.
Journal of Risk and Uncertainty 27: 239–256.
——. 2004. “The Value of Life: Estimates with Risks by Occupation and Industry.”
Economic Inquiry 42: 29–48.
——. 2009. “The Devaluation of Life.” Regulation and Governance 3: 103–127.
——. 2013. “Estimating the Value of a Statistical Life Using Census of Fatal
Occupational Injuries (CFOI) Data.” Monthly Labor Review, October. Avail-
able at http://www.bls.gov/opub/mlr/2013/article/using-data-from-the-census-of-fatal-
occupational-injuries-to-estimate-the-1.htm [last accessed November 18, 2013].
Viscusi, W. K., and J. Aldy. 2003. “The Value of a Statistical Life: A Critical Review of
Market Estimates throughout the World.” Journal of Risk and Uncertainty 27: 5–76.
Part IV
New approaches to
military cost–benefit
analysis
This page intentionally left blank
10 The role of cost-effectiveness
analysis in allocating
defense resources
Kent D. Wall, Charles J. LaCivita, and
Anke Richter
10.1 Introduction
Defense resource allocation is a multi-level resource allocation process with multi-
ple levels within the organization making budget requests and granting approvals.
Not only are there multiple levels within a defense organization, but the resource
allocation process may iterate between the levels. For example, in the US PPBES
system, each of the services presents its desired set of capabilities (weapons
systems, personnel, new acquisitions, maintenance and readiness requirements,
training, research and development, etc.) requiring a certain portion of the total
Department of Defense (DoD) budget to the Office of the Secretary of Defense
(OSD). OSD is then responsible for reviewing these submissions, combining the
requested needs and desires into an overall proposal that will become the total
requested defense budget and which will be included in the President’s budget
request to Congress. Given system incentives, the requests always exceed the
expected available budget and OSD must decide which mix of capabilities will
provide the strongest defense force for the nation within the budgetary limitations.
As they decide to allocate different levels of funds to the military services, the
services have the possibility to adjust their submission packages of capabilities.
Each service strives for an optimal allocation among capabilities within its organi-
zation given a specific budget allocation. OSD strives for an optimal allocation of
budget among the services to attain what it views as the best mix of capabilities.
The process iterates among these two decision-makers (and two decision-making
levels) until the services and OSD are in agreement on the capabilities that will be
provided at a given allocation of the budget. (For an explanation and documenta-
tion of the PPBS process, please refer to the websites and documents provided at
the end of the references.) Making the process more complex, different types and
amounts of information are available to these two levels of decision-makers. Oper-
ations research, especially cost-effectiveness analysis (CEA), can provide valuable
insight to this iterative programming process.
While the preceding paragraph emphasizes the interactions seen at the top level
of the PPBES system, similar situations arise throughout the defense organiza-
tion. At lower levels, there is competition among capability providers in a service
to be included in the service’s budget proposal to OSD; for example, the armored
264 K. D. Wall, C. J. LaCivita, A. Richter
division and the infantry division in the Army; the infantry division and the field
artillery in the Army; the surface navy and aircraft carriers in the Navy; or infor-
mation operations and the surface navy in the Navy. At the higher level, this occurs
when the DoD budget is presented to Congress as part of the overall budget for
the country.
The situation is also seen outside of defense. Excellent examples are found
in multi-national aid programs. For example, the majority of HIV/AIDS inter-
vention programs implemented in Africa are funded by external sources (e.g.
donor governments; UN agencies; the Global Fund to Fight AIDS, Tubercu-
losis, and Malaria; and the World Bank). Each of these sources chooses to
allocate funding to individual countries based on their proposed interventions.
The individual countries then allocate these funds to local organizations and pro-
grams to conduct these interventions. Finally, the individual organization tasked
with implementing the interventions may make further allocation decisions to
spend monies on prevention and treatment efforts with individual populations
or geographical regions. Lasry et al. (2007) provides more information for these
applications.
Normative or prescriptive theory tells us that decision-makers ultimately must
choose among alternatives based on their benefits and their costs. For decision-
makers in the public sector benefits are most often expressed in terms of effective-
ness because of the difficulties in “pricing out” the “products” associated with
each alternative. National defense and homeland security are prime examples.
There are no markets for placing monetary value on the “products” associated
with an aircraft carrier battle group, national missile defense system or counter
terrorism policy. The decision-maker somehow must integrate information on
cost and effectiveness and decision problems are focused on striking a balance
between these two attributes. Cost-benefit analysis (CBA) is replaced by CEA
and examination of tradeoffs between effectiveness and cost becomes a crucial
task for the decision-maker. Unfortunately the integration of cost and effective-
ness is not straightforward and there are many views on how it should be done
(see Willard 1998; Fowler and Cason 1998; Thomas and Sheldon 1999; and the
excellent literature review by Ruefli 1974). This issue is even further compounded
and complicated in a multi-level decision process.
Mathematical programming techniques can provide vital insight into these
types of optimization problems. For any optimization problem (maximize effec-
tiveness subject to budget constraints) it is possible to develop the dual problem
formulation (minimize the shadow prices—the value of each constraint—subject
to desired levels of effectiveness). With basic convexity assumptions, the opti-
mal solution of the primal problem is an optimal solution of the dual problem as
well. This relationship has been extensively explored in the operations research
literature (see for example Bazaraa and Shetty 1979; Hillier and Lieberman 1990;
Luenberger 1973; Solow 1984) and it has been shown how the problem can be
solved either by working with the primal formulation or the dual formulation.
Frequently, one formulation is much easier to solve than the other. This is indeed
the case when considering the multi-level resource allocation process as it can
Defense resources management 265
take advantage of (rather than be hindered by) the different information available
at each decision-making level.
This chapter explores these two potential solution techniques in a multi-level
resource allocation problem. The goal of the resource allocation is to maximize
the overall effectiveness in a public sector organization through the allocation of a
single limited resource (the budget). We begin with a formulation of the resource
allocation decision problem in its primal form and explore practical aspects of its
solution, paying special attention to information requirements. The next section
presents the dual formulation of the problem and its solution. The practical imple-
mentation of the dual solution and its information requirements are compared with
those of the primal solution. The third section considers the resource allocation
problem from the perspective of CEA. Here we show that a decision-maker who
claims to be motivated by cost-effectiveness concerns is actually implementing
the dual solution. The desire to find “the best tradeoff between cost and effec-
tiveness” is, in reality, a desire to solve the resource allocation problem via the
dual formulation. Two illustrative examples are provided, one involving contin-
uous alternatives and one involving discrete alternatives, before we end with a
summary of our main points.
DM0
subject to
x i = B, (10.2)
∂ V (x∗ )
− λ = 0; i = 1, 2, . . ., N, (10.3)
∂ xi
where x∗ is the vector of optimal budget allocations and λ is the Lagrange mul-
tiplier associated with the budget, i.e. the marginal effectiveness of another unit
of budget. The standard prescriptive result is clear. The first-level decision-maker
should select an allocation such that marginal effectiveness across all second-level
units is equal.
The main issue with this formulation is the amount and type of information that
is available at the different levels of an organization. The “all-knowing” and “all-
seeing” decision-maker necessary to formulate and solve Equation (10.3) does
not exist. At the first level, the decision-maker has a strategic view (a joint view)
over all of the lower levels, the capabilities they produce, and how the entire
organization and portfolio of capabilities interact and contribute to the overall
effectiveness of the defense force. This level of decision-maker is less knowl-
edgeable of the operational details; exactly how this capability is best produced.
That information resides with the second-level decision-makers (in our example,
the military services). They have all of the operational knowledge, which is why,
when given a budget allocation, they will make a further allocation that enables
them to best produce the needed capabilities. The second-level decision-makers
will typically have less of a strategic/joint view. These different levels of informa-
tion can be captured in the mathematical formulation of the problem presented in
Equations (10.1)–(10.3).
Let yi represent the units of output produced by each second-level decision-
maker if it is given an allocation x i . This function, yi (x i ) is known only to the
second-level decision-makers. The first-level decision-maker knows the x i and can
observe the yi but does not know the details of the production function, i.e. the
first-level decision-maker does not know how yi would change with a slightly
Defense resources management 267
larger or slightly smaller allocation, x i . The first-level decision-maker assesses and
values the yi and their relative strategic contribution to the overall effectiveness of
the organization. Using the concepts of multi-criteria decision-making (see Belton
and Stewart 2003; Kirkwood 1997), we can create an approximation for V (x)
representing the first-level decision-maker’s decision process.
Let vi express the decision-maker’s valuation of the contribution yi makes to
overall effectiveness. For example, if x i dollars are allocated to the Air Force for
anti-aircraft defense then yi interceptor squadrons can be purchased and operated
over the time corresponding to the planning period. The first-level decision-maker
values this contribution towards anti-aircraft defense by assigning to the yi a num-
ber, vi (yi ). Let wi represent the relative importance of each measure to the overall
value function, V (x), as seen through the eyes of the first-level decision-maker:
in this case, the relative importance of anti-aircraft defense versus other military
capabilities to overall national defense. Thus we assume the decision-maker bases
allocation decisions on the maximization of V (x) expressed as
These equations, together with the budget constraint, provide enough information
to determine the best allocations and the value for the Lagrange multiplier.
While this formulation captures the information differential between the lev-
els of decision-makers, it is not an easier problem to solve than what was
presented originally. To find the optimal solution requires the calculation of
the N-dimensional system defined by Equation (10.5). Second, the first-level
decision-maker requires complete knowledge of the yi (x i ) and the marginal prod-
ucts, d yi (x ik )/d x i . The decision-maker must be both knowledgeable and under-
standing of this detailed information and be able to use it to determine individual
second-level individual marginal effectiveness. This is counter to our assump-
tion that the first-level decision-maker does not know the operational details of
the second-level decision-maker’s organization. Finally, if Equation (10.5) is not
satisfied, the first-level decision-maker is given little direction on how to re-
allocate funds to satisfy the first-order necessary conditions. A sensible guide
268 K. D. Wall, C. J. LaCivita, A. Richter
might be to allocate more funds to the sub-units that display higher marginal
effectiveness and to allocate less funds to those sub-units with low marginal
effectiveness. This, however, still requires the first-level decision-maker to be in
possession of the detailed knowledge assumed to reside only with the second-level
decision-makers.
The only way to approach this problem is through trial and error in an iter-
ative approach. Given an initial allocation, x0 , satisfying Equation (10.2), the
top-level decision-maker then observes the resulting {yi ; 1 ≤ i ≤ N} and evaluates
Equation (10.4). Using the portion of Equation (10.5) which the top-level decision-
dv
maker knows, wi dyi , he guesses what an improved allocation of x might be. The
i
decision-maker does not know for a fact that his choice will indeed lead to a bet-
ter allocation because he cannot estimate the second portion of Equation (10.5),
(x i∗ ), since only the second-level decision-maker has this information. The new
dyi
dx i
allocation is implemented and the first-level decision-maker observes the new
{yi ; 1 ≤ i ≤ N} and evaluates Equation (10.4). This continues until an allocation is
found for which Equation (10.5) is satisfied or when changes to Equation (10.4)
become sufficiently small.
This entire concept can be explained graphically using a 2-dimensional exam-
ple as in Figure 10.2. In the case of perfect, complete information with a single
decision-maker, the iterative solution process to find the optimal allocation can be
detailed on this graphic. The budget constraint, x 1 + x 2 = B, appears in the fourth
quadrant. The first and third quadrants depict the marginal effectiveness of each x i
assuming an additive form; i.e.,
∂V dvi d yi
Di V = = (x i ). (10.6)
∂ x i d yi d x i
D1V
slope =wfw 1
Quadrant II Quadrant I
vv'0 ·- ------
where
ϕ(λ) = max {wi vi (yi (x i )) − λx i } + λB (10.8)
x
is called the dual function. The dual formulation represents the decision problem
confronting the first-level decision-maker as a minimization. The decision-maker
seeks that value of λ for which the dual function attains its minimum.
The dual function is composed of N separate unconstrained, one-dimensional,
maximization problems that represent the decision problems faced by each of the
second-level decision-makers. We find it convenient to exchange the summation
and maximization so we may define this function in terms of a weighted sum of
individual dual functions, ϕi (λ),
ϕ(λ) = max{wi vi (yi (x i )) − λx i } + λB = ϕi (λ) + λB. (10.9)
xi
this equation is not solved in this form because its solution, x i∗ , satisfying the
first-order condition
dvi d yi ∗
(x ) = λ/wi ; i = 1, 2, . . . , N
d yi d x i i
270 K. D. Wall, C. J. LaCivita, A. Richter
requires knowledge of dvi /d yi , wi and λ. These are known only to the first-level
decision-maker.
The second-level decision-maker knows only yi (x i ) and d yi /d x i . However, we
can define the decision problem of the second-level decision-maker as
max{yi (x i ) − λi x i } (10.11)
xi
given λi where
dvi −1
λi = λ wi = λm −1
i . (10.12)
d yi
∇ϕ(λ∗ ) = 0, (10.13)
where ϕ(λ∗ ) = ϕ(x∗ (λ∗ )). The most striking aspect of the dual formulation is the
fact that this gradient is the budget constraint itself (see Luenberger 1973, 313):
∇ϕ(λ) = x i∗ (λ) − B. (10.14)
Defense resources management 271
In other words, when λ = λ the resulting {x i∗ (λ∗ /wi ); 1 = 1, 2, . . ., N} satisfy the
∗
λk+1 = α k λk .
∗
For each iteration we now replace x i (λ) with x ik (λk ) and λi with λki . Given
k x i k(λ )
k k
slope =wfw 1
Quadrant II Quadrant I
Quadrant Ill
Quadrant IV
or
max [ω E wi vi (yi (x i )) − ωC x i ] .
x
dvi d yi (x i∗ ) 1 ωC
− = 0; 1≤i ≤ N
d yi d x i wi ω E
or
d yi (x i∗ ) ωC
− m −1 = 0; 1≤i ≤ N (10.16)
d xi i
ωE
where
dvi
m i = wi .
d yi
274 K. D. Wall, C. J. LaCivita, A. Richter
This solution is identical to that obtained in the dual formulation if we define
ωC /ω E = λ. Thus, the maximization of overall cost-effectiveness is identi-
cal to the dual function maximization problem given by Equation (10.8):
x i∗ (m −1 ∗ ∗ ∗
i λ ) = x i (λi ). The identical solution method as presented in Section 10.2
can be used.
We find it both reasonable and natural for decision-makers to rationalize deci-
sions by claiming to have found the “correct,” “best,” or “most appropriate”
tradeoff between cost and effectiveness. They are stating that they have found the
most effective solution given the opportunity cost of the resource (as captured by
the Lagrange multipliers). The dual problem iterative solution we have presented
is the basis for the communication between levels in a hierarchical organization.
The first-level decision-maker transmits downward the current opportunity cost
of effectiveness. In return the second-level decision-makers send the first level
their solutions: yi (x ik ) and x ik . The first-level decision-maker checks to
see if the
resource constraint is satisfied. If the constraint is satisfied (i.e., |B − x i | ≤ δ)
then a solution has been found. If not then the first-level decision-maker sends
revised opportunity cost information to the second level and the process repeats.
No marginal information is exchanged. The first-level decision-maker merely
checks the budget requests against the constraint and returns to the second-level
decision-makers new information on the tradeoff between effectiveness and cost.
All evaluations of the second-level decisions and how they integrate with the first-
level decision-maker’s overall tradeoff between effectiveness and cost are done
implicitly.
y1 (x 1 ) = 1 − e−1.9x1
276 K. D. Wall, C. J. LaCivita, A. Richter
and let Program 2 (Coastal Security) be described by the effectiveness production
function
y2 (x 2 ) = 1 − e−0.9x2 .
This information resides internal to each program office and its staff. The director
of the joint planning staff does not know, nor chooses to know, these details. The
director possesses only the “strategic” information—that contained in V (x). The
planning process begins with the director setting a value for λ0 based on his/her
assessment of the given budget, B. If B represents a significant growth over pre-
vious budgets then λ0 will be smaller. This will allow the program managers to
propose more capabilities as the marginal cost of providing effectiveness is low.
Likewise, if B represents a significant reduction from past experience then λ0 will
be larger. This will force the program managers to reduce the amount of capabil-
ities they propose as the marginal cost of effectiveness is high. The director now
uses Equation (10.16) to construct the λi resulting in
where we set yi equal to the previous proposed value given to the director by the
program managers. Initially, in the beginning of the current planning period, these
would be the result of the previous planning period solution.
We now illustrate the planning process derived from the overall cost-
effectiveness paradigm (based on the dual formulation) by simulating the process
using the following steps. Let a1 = 2.95, a2 = 2.1, c1 = 1.9, c2 = 0.9 and λ0 = 1/B.
λki
STEP 3: x ik = −ci−1 n ci
.
STEP 4: B k = + x 1k x 2k .
|B k−B| ≤ δ then STOP. Otherwise
k
STEP 5: Check for budget satisfaction.
If
If B > B + δ then set λ = 1 + β ·
k k+1 B−B
B
λk and k = k + 1 and go to
STEP 2. k
If B k < B − δ then set λk+1 = 1 − β · B B−B λk and k = k + 1 and go to
STEP 2.
2
2
4
4
Iteration
Iteration
8
8
0
0
2
2
4
4
Iteration
Iteration
6
6
Budget = 2.50
8
8
Effectiveness Lambda
0.0 0.2 0.4 0.6 0.8 1.0 0.00 0.05 0.10 0.15 0.20 0.25 0.30
0
0
2
2
4
4
Iteration
Iteration
8
8
0
0
2
2
4
4
Iteration
Iteration
6
6
Budget = 4.50
8
8
Effectiveness Lambda
0.0 0.2 0.4 0.6 0.8 1.0 0.00 0.05 0.10 0.15 0.20 0.25 0.30
0
0
2
2
4
4
Iteration
Iteration
8
8
0
0
2
2
4
4
Iteration
Iteration
6
6
Budget = 6.50
8
8
Defense resources management 281
Table 10.1 Border security alternatives
Alternative x1 y1 (x1 )
1 0.50 0.15
2 0.75 0.40
3 0.81 0.49
4 1.50 0.68
5 1.79 0.75
6 3.50 0.85
Alternative x2 y2 (x2 )
7 0.40 0.11
8 1.00 0.55
9 1.55 0.76
10 2.00 0.83
11 2.80 0.91
12 3.95 0.98
of the combinations as they are presented sequentially (by the program managers).
There are 20 clearly dominated alternatives—alternatives where there exist at least
one other alternative that is either less costly yet equally (or better) effective or
more effective yet equally (or less) costly. We see that for small B there are clearly
defined solutions but the envelop of the set of alternatives, the efficient set, becomes
relatively “flat” for B > 4.0 making the identification of efficient solutions difficult.
We present Figure 10.7 solely for the benefit of the reader to understand the challenge
of each example allocation problem presented below.
As in the previous case we have w1 = 0.6, w2 = 0.4, a1 = 2.95, and a2 = 2.1.
The simulation is composed of the following steps.
STEP 4: B k = x 1k + x 2k .
STEP 5: If |B k − B| ≤ δ then STOP, y1 (x 1k ), y2 (x 2k ) is a satisfactory solution.
Otherwise k
If B k > B + δ then set λk+1 = 1 + β · B−B B
λk and k = k + 1 and go to
STEP 2. k
If B k < B − δ then set λk+1 = 1 − β · B B−B λk and k = k + 1 and go to
STEP 2.
282 K. D. Wall, C. J. LaCivita, A. Richter
1.0
0.9
Possible Alternatives in C–E Space
0.8
0.7
0.6
0.5
0.4
0.3
Budget = 4.00
0.2
0.1
0.0
0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0 4.5 5.0 5.5 6.0 6.5 7.0 7.5 8.0
Iteration
Steps 1, 2, 4, and 5 again represent the director and Step 3 the program man-
agers’ maximization of cost-effectiveness, which is now a direct search over their
respective finite set of solutions. Figures 10.8, 10.9, and 10.10 show the simulation
results for three budgets: B = 2.5, 4.5, and 6.5, respectively. As in the continuous
case, each simulation is presented in four panels: (1) the top left presents λk ; (2)
the top right presents the decisions x 1k (solid line) and x 2k (dashed line); (3) the
lower left presents V (x); and (4) B k = x 1k + x 2k occupies the lower right panel.
All simulations find a satisficing solution. Convergence depends on δ = ηB and
β. We use η = 0.01 and β = 0.45 for B = 2.5 and obtain convergence in three
iterations. We use η = 0.01 and β = 1.10 for B = 4.5 and obtain convergence
in three iterations. When B = 6.5 convergence is obtained in three iterations if
η ≥ 0.02 with β = 1.0 or η = 0.01 with β = 2.0.
The convergence indicated here compares favorably with the existing PPBES
process at the DoD. Undoubtedly, convergence to a satisficing solution within a
three or four level hierarchy would take longer to achieve.
10.6 Summary
The dual formulation of a multi-level public sector resource allocation problem
provides a sound theoretical basis for decision-makers who rationalize their deci-
sions with statements like “. . .the chosen alternative possesses the best balance
between cost and effectiveness.” They are expressing nothing less than the first-
order necessary conditions for a solution of the dual formulation of the problem.
The pivotal variable in such problems is the Lagrange multiplier, or “shadow
price”. This represents the marginal cost of effectiveness.
4
1.0
X2
0.8
3
and
X1
0.6
2
Lambda
0.4
1
0.2
Budget Allocations
0
0.0
0 2 4 6 8 1 3 5 7 9 11
Iteration Iteration
1.0
Budget = 2.50
0.8
6
5
0.6
4
Cost
0.4
3
Effectiveness
2
0.2
1
0
0.0
1 3 5 7 9 11 1 3 5 7 9 11
Iteration Iteration
1.0
X2
0.8
3
and
X1
0.6
2
Lambda
0.4
1
0.2
Budget Allocations
0
0.0
0 2 4 6 8 10 1 3 5 7 9 11
Iteration Iteration
1.0
Budget = 4.50
0.8
6
5
0.6
4
Cost
0.4
3
Effectiveness
2
0.2
1
0
0.0
1 3 5 7 9 11 1 3 5 7 9 11
Iteration Iteration
1.0
X2
0.8
3
and
X1
0.6
2
Lambda
0.4
1
0.2
Budget Allocations
0
0.0
0 2 4 6 8 10 1 3 5 7 9 11
Iteration Iteration
1.0
Budget = 6.50
0.8
6
5
0.6
4
Cost
0.4
3
Effectiveness
2
0.2
1
0
0.0
1 3 5 7 9 11 1 3 5 7 9 11
Iteration Iteration
Acknowledgement
This work benefits from the insights and interpretations of Assistant Professor David
L. Rose who has since passed away. He was an inspiring teacher and brilliant mind
that is fondly remembered, and sorely missed, by all who knew him.
Bibliography
Bazaraa, Mokhtar S. and C.M. Shetty. 1979. Nonlinear Programming: Theory and Algo-
rithms. Toronto, Canada: John Wiley and Sons.
Defense resources management 287
Belton, Valerie and T.J. Stewart. 2003. Multiple Criteria Decision Analysis: An Integrated
Approach. Norwell, MA: Duxbury Press.
Everett III, Hugh. 1963. “Generalized Lagrange Multiplier Method for Solving Problems
of Optimum Allocation of Resources.” Operations Research 11: 399–417.
Fowler, B.W. and P.P. Cason. 1998. “The Cost Exchange Ratio: A New Aggregate Measure
of Cost and Operational Effectiveness.” Journal of the Military Operations Research
Society 3: 57–64.
Hillier, Frederick S. and Gerald J. Lieberman. 1990. Introduction to Operations Research
(5th edition). New York: McGraw-Hill Publishing Company.
Kaplan, Stanely and B. John Garrick. 1981. “On the Quantitative Definition of Risk.” Risk
Analysis 1: 11–27.
Kirkwood, Craig. W. 1997. Strategic Decision Making: Multiobjective Decision Analysis
with Spreadsheets. Belmont, CA: Duxbury Press.
Lasry, A., G.S. Zaric, and M.W. Carter. 2007. “Multi-level Resource Allocation for HIV
Prevention: A Model for Developing Countries.” European Journal of Operational
Research 180: 786–799.
Libby, R. and P.C. Fishburn. 1977. “Behavioral Models of Risk Taking in Business
Decisions: A Survey and Evaluation.” Journal of Accounting Research 15: 272–292.
Luenberger, David G. 1973. Introduction to Linear and Nonlinear Programming. Reading,
MA: Addison-Wesley.
Payne, J.W., D.J. Laughhunn, and R. Crum. 1980. “Translation of Gambles and Aspiration
Level Effects in Risky Choice Behavior.” Management Science 10: 1039–1060.
Petty II, J.W. and D.F. Scott. 1981. “Capital Budgeting Practices in Large American
Firms: A Retrospective Analysis and Update.” In Financing Issues in Corporate Project
Selection, edited by F.G.J. Derkinderen and R.L. Crum. Boston, MA: Martinus Nyhoff.
Quade, E.S. 1989. Analysis for Public Decisions. Englewood Cliffs, NJ: Prentice Hall.
Ruefli, T. 1971. “PPBS – An Analytical Approach.” In Studies in Budgeting, edited by
R.F. Byrne, A. Charnes, W.W. Cooper, O.A. Davis, and D. Gilford. Amsterdam: North-
Holland.
Ruefli, T. 1974. “Analytic Models of Resource Allocation in Hierarchical Multi-Level
Systems.” Socio-Economic Planning Sciences 8: 353–363.
Solow, Daniel 1984. Linear Programming: An Introduction to Finite Improvement Algo-
rithms. New York: Elsevier Science Publishing Company.
Thomas, C. and R.S. Sheldon. 1999. “The ‘Knee of the Curve’ – Useful Clue but
Incomplete Support.” Journal of the Military Operations Research Society 4: 17–29.
Willard, D. 1998. “Cost-Effectiveness, CAIV, and the Knee of the Curve.” Phalanx 31:
24–27.
11.1 Introduction
Games of chance have been popular throughout history. Biblical accounts even
record Roman soldiers gambling for Christ’s robes. In earlier times, chance was
something that occurred in nature, and humans were simply subjected to it as a
ship is to random waves. Until the Renaissance, the future was thought to be a
chance occurrence of completely random events, beyond the control of humans.
With the advent of games of chance the study of risk and uncertainty increas-
ingly mirrored real-life events. Although games of chance were initially played
with great enthusiasm, rarely did anyone consciously determine the odds. It was
not until the mid-1600s that the concept of chance was seriously studied: The
first such endeavor can be credited to Blaise Pascal, a father of the study of
choice, chance, and probability. Fortunately, after many centuries of mathematical
and statistical innovations by visionaries such as Pascal, Bernoulli, Bayes, Gauss,
Laplace, and Fermat, and with the explosion of computing technology, uncertainty
can increasingly be explained with much more elegance through rigorous, method-
ological applications. Today, with the assistance of powerful software packages,
we have the ability to apply these techniques to increase the value of cost–benefit
analysis (CBA), or what might be called “risk–benefit analysis.”
Humans have struggled with risk for centuries, but through trial and error and
through the evolution of human knowledge, have devised ways to describe, quan-
tify, hedge, and even take advantage of risk.1 In the US military, risk analysis, real
options analysis, and portfolio optimization techniques are enablers of a new way
of approaching problems of estimating return on investment (ROI) and estimating
risk-value tradeoffs of strategic real options.2
Several advanced quantitative risk-based concepts are introduced in this chap-
ter, including the application of strategic real options analysis, Monte Carlo risk
simulation, stochastic forecasting, portfolio optimization, and knowledge value
added (KVA). These methodologies rely on standard metrics and existing tech-
niques, such as ROI and discounted cash flow discussed in Chapter 17, and
complement these techniques by pushing the envelope of analytics while not
290 J. C. Mun, T. Housel
replacing them outright. These advanced techniques are used to help make the best
possible decisions, to allocate budgets, forecast outcomes, and create portfolios
with the highest strategic value or ROI (even when conditions surrounding these
decisions are risky or uncertain). These new techniques can be used to identify,
analyze, quantify, value, forecast, and manage (hedge, mitigate, optimize, allo-
cate, and diversify) risks associated with strategically important military options
(see Chapter 14 for example).
Looking at bang for the buck, X (2), Y (1), Z (10), Project Z should
be chosen–with a $1,000 budget, the following can be obtained:
Project X: 20 Project Xs returning $1,000, with $500 risk
Project Y: 4 Project Xs returning $800, with $800 risk
Project Z: 10 Project Xs returning $1,000, with $100 risk
Project X: For each $1 return, $0.5 risk is taken
Project Y: For each $1 return, $1.0 risk is taken
Project Z: For each $1 return, $0.1 risk is taken
Project X: For each $1 of risk taken, $2 return is obtained
Project Y: For each $1 of risk taken, $1 return is obtained
Project Z: For each $1 of risk taken, $10 return is obtained
Conclusion:
Risk is important. Forgoing risks results in making the wrong decision.
and that an unlimited number of projects from each category can be chosen (only
constrained by a budget of US$1,000). With this US$1,000 budget, 20 project Xs
can be chosen, yielding US$1,000 in net returns and US$500 risks, etc.
It is clear from Figure 11.2 that project Z is the best project: That is, for the same
net return (US$1,000), the least amount of risk is undertaken (US$100). Another
way of viewing this selection is that for each US$1 of risk, US$10 in returns are
obtained on average. This example illustrates the concept of bang for the buck or
getting the best value (net monetary benefits) with the least amount of risk.
An even more striking example is provided when there are several different
projects with identical single-point average net benefit or costs of US$10 million
each. Without risk analysis, a decision-maker should, in theory, be indifferent to
choosing any of the projects. With risk analysis, however, a better decision can be
made. For instance, suppose the first project has a 10 percent chance of exceeding
US$10 million; the second, a 15 percent chance; and the third, a 55 percent chance.
Then additional critical information is obtained on the riskiness and uncertainty
of the project or strategy, and a better decision can be made.
292 J. C. Mun, T. Housel
11.3 From traditional risk assessment to
Monte Carlo risk simulation
Military and business leaders have dealt with risk throughout the history of war
and commerce. In most cases, decision-makers viewed risks of a particular project,
acknowledged its existence, and moved on. Little quantification was performed. In
fact, most decision-makers only focus on single-point estimates of a project’s net
benefit or profitability. Figure 11.3 shows an example of a single-point estimate.4
The estimated net revenue of US$30 is simply that: a single-point estimate whose
probability of occurrence is close to zero.5
In the simplest model shown in Figure 11.3, the effects of interdependencies are
ignored, and, in modeling jargon, we have the problem of garbage-in, garbage-out
(GIGO). As an example of interdependencies, the units sold are probably negatively
correlated to the price of the product (since demand curves slope downwards), while
average variable costs may be positively or negatively correlated to units sold (i.e.,
depending on whether there are decreasing or increasing returns to scale in produc-
tion). Ignoring these effects in a single-point net revenue estimate with uncertain
sales will yield grossly misleading results. There are numerous interdependencies
in military options as well—for example, complex issues in logistics, beginning
with the supplier all the way through to the warrior in the field.
In the commercial example illustrated in Figure 11.3, if the unit sales variable
becomes 11 instead of 10, the resulting revenue may not simply be US$35. The
net revenue may actually decrease if there is an increase in variable cost-per-unit,
while the sale price may actually be slightly lower driving the increase in unit
sales. Ignoring these interdependencies will reduce the accuracy of the model.
•
•
•
Unit Sales
Sales Price
Total Reve nue
;ot)
---
~
$100
lnterdependendes
rne::m GIGO
Estimate
Point Estimate
Q)
o.4o
tt 200 0"
m
100 0.20 ~
'<
Based on the tenets of complexity theory, KVA assumes that humans and tech-
nology in organizations add value by taking inputs and changing them (measured
in units of complexity) into outputs through core processes. The amount of change
an asset within a process produces can be used as a measure of its marginal
value-added or benefit. The additional assumptions in KVA include the following:
Phase
Phase Milestone 4
Phase II Milestone 3
Exit
Milestone 2 Exit Stop after
Phase I
Phase III
Stop after
Milestone 1
Phase II
Strategy A Exit
Divide the development Stop after
into multiple milestones, Phase I
Exit
stage-gating the
development and Do nothing Contract
mitigating the risk. Outsourcing
Phase II
Larger Deployment
Exit
POC
Small-Scale Internal development rather
Proof of than off-the-shelf applications
Start Strategy B
Exit
Concept
Apply a quick proof-of-
concept (POC) of the Abandon
technology, replicate and Exit
scale to larger force-wide Do nothing
deployment. If POC fails, Expan
abandon project and
reduce future risks. R&D new technology,
Outsource or obtain off- Buy expand into other
the-shelf platforms if application areas
successful. Purchase
technology Exit
Strategy C
Purchase technology Sell IP and technology,
rather than developing it Exit abandon project
in-house. If application fails,
Do nothing
abandon and sell the
company’s intellectual
property and technology. If
successful, find other
applications for technology.
Project 1 $458.00 $150.76 $1,732.44 1.20 381.67 1.09 0 8.10 2.31 1.20 1.98
Project 2 $1,954.00 $245.00 $859.00 9.80 199.39 1.29 1 1.27 4.83 2.50 1.76
Project 3 $1,599.00 $458.00 $1,845.00 9.70 164.85 1.25 0 9.88 4.75 3.60 2.77
Project 4 $2,251.00 $529.00 $1,645.00 4.50 500.22 1.32 0 8.83 1.61 4.50 2.07
Project 5 $849.00 $564.00 $458.00 10.90 77.89 2.23 0 5.02 6.25 5.50 2.94
Project 6 $758.00 $135.00 $52.00 7.40 102.43 3.60 1 3.64 5.79 9.20 3.26
Project 7 $2,845.00 $311.00 $758.00 19.80 143.69 1.41 1 5.27 6.47 12.50 4.04
Project 8 $1,235.00 $754.00 $115.00 7.50 164.67 7.56 1 9.80 7.16 5.30 3.63
Project 9 $1,945.00 $198.00 $125.00 10.80 180.09 2.58 1 5.68 2.39 6.30 2.16
Project 10 $2,250.00 $785.00 $458.00 8.50 264.71 2.71 1 8.29 4.41 4.50 2.67
Project 11 $549.00 $35.00 $45.00 4.80 114.38 1.78 0 7.52 4.65 4.90 2.75
Project 12 $525.00 $75.00 $105.00 5.90 88.98 1.71 0 5.54 5.09 5.20 2.69
Project 13 $516.00 $451.00 $48.00 2.80 184.29 10.40 0 2.51 2.17 4.60 1.66
Project 14 $499.00 $458.00 $351.00 9.40 53.09 2.30 1 9.41 9.49 9.90 4.85
Project 15 $859.00 $125.00 $421.00 6.50 132.15 1.30 1 6.91 9.62 7.20 4.25
Project 16 $884.00 $458.00 $124.00 3.90 226.67 4.69 1 7.06 9.98 7.50 4.46
Project 17 $956.00 $124.00 $521.00 15.40 62.08 1.24 1 1.25 2.50 8.60 2.07
Project 18 $854.00 $164.00 $512.00 21.00 40.67 1.32 0 3.09 2.90 4.30 1.70
Project 19 $195.00 $45.00 $5.00 1.20 162.50 10.00 0 5.25 1.22 4.10 1.86
Project 20 $210.00 $85.00 $21.00 1.00 210.00 5.05 0 2.01 4.06 5.20 2.50
30.00 55.00
$3,500 $4,000 $4,500 $5,000 $5,500 $6,000 $6,500 $7,000 $7,500 $8,000 $8,500 $3,500 $4,500 $5,500 $6,500 $7,500 $8,500
68.00 $600,000.00
66.00 $550,000.00
RISK ANALYSIS
RISK MODELING
RISK PREDICTION
RISK IDENTIFICATION
Start with a list of projects
Optimization Retirement
13 296,916 9,851,788 6,086,684 3,765,104 0.949% 0.87%
realized label
Loss cost reduction Strategic options value
Value of the
RISK HEDGING
1,154,349 Wait to Invest
3,715,300 1.263
RISK MITIGATION
…the relevant projects next optional step if multiple
are chosen for real …real options analytics are projects exist that require efficient
RISK MANAGEMENT
RISK DIVERSIFICATION
options analysis and the calculated through binomial lattices asset allocation given some …create reports, make
project or portfolio real and closed-form partial-differential budgetary constraints… useful for decisions, and do it all
options are framed… models with simulation… strategic portfolio management… again iteratively over time…
11.8 Conclusion
The leadership of ministries of defense can take advantage of more advanced ana-
lytical procedures to make strategic investment decisions and manage portfolios
of projects using KVA and the real options approach. In the past, due to the lack
of technological maturity, this would have been extremely difficult; hence, busi-
nesses and the government had to resort to intuition, past experience, and relatively
unsophisticated models. Now, with the assistance of new technology and more
mature methodologies, financial leaders have every reason to take the analysis a
step further.
Corporations such as 3M, Airbus, AT&T, Boeing, BP, Chevron, Johnson &
Johnson, Motorola, and many others have already successfully applied these tech-
niques. As emphasized in this chapter (and in the real options application in
Chapter 14) the military has the same opportunity. The relevant software appli-
cations, books, case studies, and public seminars have been created, including
specific case studies developed exclusively for the US Navy.
The only real barrier to implementation is a lack of exposure to the potential
benefits of these new methods. Hopefully, this chapter has revealed the poten-
tial benefits of these analytical techniques and tools that can complement and
enhance CBA techniques currently used by leadership. In order to fully prepare
for twenty-first-century challenges and create highly effective and flexible mil-
itary forces, strategic real options, KVA, and risk analysis can be leveraged by
leadership to improve decision making. Combining real options with KVA can
help ensure maximum strategic flexibility and improve risk-based CBA.
Notes
1 To the people who lived centuries ago, risk was simply the inevitability of chance occur-
rence beyond the realm of human control, albeit many phony soothsayers profited from
their ability to convincingly profess their clairvoyance by simply stating the obvious
or reading the victims’ body language and telling them what they wanted to hear. We
modern-day humans—ignoring for the moment the occasional seers among us and even
with our fancy technological achievements—are still susceptible to risk and uncertainty.
We may be able to predict the orbital paths of planets in our solar system with astound-
ing accuracy or the escape velocity required to shoot a man from the Earth to the Moon
310 J. C. Mun, T. Housel
or drop a smart bomb within a few feet of its target thousands of miles away, but when
it comes to, say, predicting a firm’s revenues the following year, we are at a loss.
2 There are many new US Department of Defense (DoD) requirements for using more
advanced analytical techniques. For instance, the Clinger-Cohen Act of 1996 mandates
the use of portfolio management for all federal agencies. The Government Accountabil-
ity Office’s Assessing Risks and Returns: A Guide for Evaluating Federal Agencies’ IT
Investment Decision-Making (Version 1) (February 1997) requires that IT investments
apply ROI measures. DoD Directive 8115.01 issued October 2005 mandates the use of
performance metrics based on outputs, with ROI analysis required for all current and
planned IT investments. DoD Directive 8115.bb (2006) implements policy and assigns
responsibilities for the management of DoD IT investments as portfolios within the DoD
enterprise—where a portfolio is defined to include outcome performance measures and
an expected ROI. The DoD Risk Management Guidance Defense Acquisition guidebook
requires that alternatives to the traditional cost estimation need to be considered because
legacy cost models tend not to adequately address costs associated with information
systems or the risks associated with them.
3 Risks can be computed many ways, including volatility, standard deviation of lognormal
returns, value at risk, and so forth. For more technical details, see Mun’s Modeling Risk
(2nd edition, 2010).
4 We will demonstrate how KVA combined with the traditional market comparables
valuation method allows for the monetization of benefits (i.e., revenue).
5 On a continuous basis, the probability of occurrence is the area under a curve, e.g., there
is a 90 percent probability revenues will be between US$10 and US$11. The area under
a straight line, however, approaches zero. Therefore, the probability of hitting exactly
US$10.00 is close to 0.00000001 percent.
6 See Chapter 14 for an interesting application of real options analysis in the case where
benefits are not measured in monetary units.
7 Perhaps the most famous early use of Monte Carlo risk simulation was by the Nobel
physicist Enrico Fermi (sometimes referred to as the father of the atomic bomb) in 1930,
when he used a random method to calculate the properties of the newly discovered neu-
tron. Monte Carlo methods were central to the simulations required for the Manhattan
Project: in the 1950s Monte Carlo risk simulation was used at Los Alamos for early
work relating to the development of the hydrogen bomb and became popularized in the
fields of physics and operations research. The Rand Corporation and the US Air Force
were two of the major organizations responsible for funding and disseminating informa-
tion on Monte Carlo methods during this time, and today there is a wide application of
Monte Carlo risk simulation in many different fields—including engineering, physics,
research and development, business, and finance.
8 The outcomes from a Monte Carlo risk simulation include probabilities and various risk
statistics that can be used to make better decisions.
9 The pathways can be valued using partial differential closed-form equations, lattices,
and simulation. Various software packages are available for this task, such as the
author’s The Real Options SLS software by Real Options Valuation, Inc., available at
www.realoptionsvaluation.com [last accessed November 21, 2014].
10 For example, in the US DOTMLPF vs New Program/Service solutions, joint integra-
tion, analysis of material alternatives (AMA), analysis of alternatives (AoA), and spiral
development.
11 Suppose that an optimization model depends on only two decision variables. If each
variable has 10 possible values, trying each combination requires 100 iterations (102
alternatives). If each iteration takes a very short amount of time (e.g., two seconds), then
the entire process could take approximately three minutes of computer time. Instead of
two decision variables, however, consider six; then consider that trying all combina-
tions requires 1,000,000 iterations (106 alternatives). It is easily possible for complete
Real options: a risk-based approach 311
enumeration to take many years to carry out. Therefore, solving complex optimizations
has always been a fantasy until now: With the advent of sophisticated software and
computing power, coupled with smart heuristics and algorithms, such analyses can now
be completed within minutes.
12 There are 2 × 1018 possible permutations for this problem and, if tested by hand, it
would take years to complete. Using Risk Simulator, the problem is solved in about
five seconds, or several minutes if Monte Carlo risk simulation and real options are
incorporated in the analysis.
13 See Mun (2010, Chapters 8 and 9) for details on forecasting and using software to run
time-series, extrapolation, stochastic process, ARIMA, and regression forecasts.
14 See Mun (2010, Chapters 4 and 5) for details on using software to run Monte Carlo
simulations.
15 See Mun (2006) for more technical details on framing and solving real options
problems.
16 See Mun (2010, Chapters 10 and 11) for details on using software to perform portfolio
optimization.
References
Cook, G. R., T. J. Housel, J. Mun, and P. Pavlou. January 2007. “Return on Investment
in Non-revenue Generating Activities: Applying KVA and Real Options to Government
Operations.” Review of Business Research 7(1): 16–19.
DoD. US Department of Defense Directive 8115. October 10, 2005. Information Tech-
nology Portfolio Management. NUMBER 8115.01, ASD(NII)/DoD CIO. Available
at http://dtic.mil/whs/directives/corres/pdf/811501p.pdf [last accessed November 27,
2014].
DoD. US Department of Defense Directive 8115. October 30, 2006. Information Technol-
ogy Portfolio Management Implementation. NUMBER 8115.02, ASD(NII)/DoD CIO.
Available at http://www.dtic.mil/whs/directives/corres/pdf/811502p.pdf [last accessed
November 27, 2014].
GAO. February 1997. Assessing Risks and Returns: A Guide for Evaluating Federal
Agencies’ IT Investment Decision-Making. Version 1. GAO/AIMD-10.1.13. Available
at www.gao.gov/assets/80/76295.pdf [last accessed November 27, 2014].
Housel, T. J., and V. Kanevsky 1995. “Reengineering Business Processes: A Complexity
Theory Approach to Value Added.” INFOR 33: 248–262.
Housel, T. J., and S. K. Nelson. 2005. “Knowledge Valuation Analysis: Applica-
tions for Organizational Intellectual Capital.” Journal of Intellectual Capital 6(4):
544–557.
Housel, T. J., W. Little, and W. Rodgers. October 29–31, 2007. “Estimating the Value
of Non-Profit Organizations Using the Market Comparables Approach.” In Refereed
Proceedings of the Collected Papers of the European Institute for Advanced Studies
in Management (EIASM): 3RD Workshop on Visualising, Measuring and Managing
Intangibles & Intellectual Capital. Ferrara, Italy, pp. 156–162.
Housel, T. J., E. Tarantino, S. Hom, and J. Mun. 2008. “Shipyard Planning Case.” In Making
Cents Out of Knowledge Management, edited by J. Liebowitz, pp 68–90. Lanham, MD:
Scarecrow Press.
Housel, T. J., V. A. Kanevsky, W. Rodgers, and W. Little. May 28, 2010. “The Use of Mod-
ern Portfolio Theory in Non-Profits and Their IT Decisions.” In Refereed Proceedings
of the IC6: World Conference on Intellectual Capital for Communities in the Knowledge
Economy. Paris, France.
312 J. C. Mun, T. Housel
Kendall, T., and T. J. Housel. July 18–21, 2004. “Resolving the ROI on IT Issue (Best Ses-
sion Paper).” In Refereed Proceedings of the 8th World Multi-conference on Systemics,
Cybernetics, and Informatics. Orlando, FL.
Mun, J. 2006. Real Options Analysis: Tools and Techniques for Valuing Strategic Invest-
ments and Decisions (2nd edition). Hoboken, NJ: Wiley Finance.
Mun, J. 2010. Modeling Risk: Applying Monte Carlo Risk Simulation, Strategic Real
Options, Stochastic Forecasting, and Portfolio Optimization (2nd edition). Hoboken, NJ:
Wiley Finance.
OMB (Executive Office of the President; Office of Management and Budget). August
4, 1983, revised 1999. Circular N0. A-76. Washington DC 20503. Available at
http://www.whitehouse.gov/sites/default/files/omb/assets/omb/circulars/a076.pdf [last
accessed September 30, 2014].
Pavlou, P. A., T. J. Housel, W. Rodgers, and E. Jansen. 2005. “Measuring the Return on
Information Technology: A Knowledge-based Approach for Revenue Allocation at the
Process and Firm Level.” Journal of the Association of Information Systems 6: 199–226.
Rabelo, L., and T. J. Housel. May 20, 2008. “Estimating the Knowledge Value-Added
of Information Technology Investments.” In Refereed Proceedings of the Institute of
Industrial Engineers (IIE) Annual Conference. Vancouver, British Columbia, Canada,
pp. 1849–1854.
Rios, C., T. J. Housel, and J. Mun. 2006. “Real Options and KVA in Military Strategy
at the United States Navy.” In Modeling Risk: Applying Monte Carlo Simulation, Real
Options Analysis, Forecasting, and Optimization Techniques, Case Study, pp. 441–452.
Hoboken, NJ: Wiley Finance.
Rodgers, W., and T. J. Housel. 2006. “Improvement of Global Performance Measures
Related to Intangible Assets.” Collected Papers of the 32nd European International
Business Academy (EIBA) Conference (Refereed Proceedings). December 7–9, Fribourg
Switzerland.
12 Extensions of the Greenfield–Persselin
optimal fleet replacement model
Application to the Canadian
Forces CP-140A Arcturus Fleet
David W. Maybury
12.1 Introduction
Today’s Canadian Forces operate increasingly aging military platforms, often
retaining fleets for unprecedented long service lives (DND 2014). The unfamiliar
territory of operating platforms well beyond expected service lifetimes presents
the Canadian Department of National Defence (DND) with a central cost–benefit
analysis (CBA) problem: to replace a fleet of aging vehicles or to continue an
ongoing maintenance regime. As vehicles age, we expect that the overall oper-
ating and maintenance costs (O&M) will eventually reach a level prohibitive to
continued operation, thereby suggesting vehicle replacement as the best alterna-
tive. Aircraft platform replacement decisions arising primarily from aging effects
present a relatively new problem for global militaries. Most replacement decisions
in the past have occurred largely through new capability requirements, thus limit-
ing aging effect data. Only recently have selected platforms yielded sufficient data
to allow for comprehensive studies.
The United States Air Force (USAF) faces substantial challenges in operat-
ing airframes with exceptionally long service lives and, having recognized age
as a factor in O&M costs as early as the 1960s (Johnson 1962), the USAF
started comprehensive fleet lifetime O&M studies in the 1990s. While limited
data in early studies contributed to a confused literature in the 60s and 70s,
from Johnson (1962) to Marks and Hess (1981), large datasets and the applica-
tion of more sophisticated analytical methods available to the USAF in recent
years has enabled investigators to draw the cautious conclusion that age effects
impact O&M costs. In particular, it was demonstrated in Ramsey et al. (1998)
that heavy-maintenance workloads have increased with the chronological age of
the KC-135 tanker aircraft, and an earlier RAND study (Hildebrandt and Sze
1990) estimated that, for every year increment in the age of a USAF mission
design series, O&M costs increase on average by 1.7 perent. Further studies on
age effects using commercial airline data (Ramsey et al. 1998; Dixon 2005), data
from US Navy aircraft (Johnson 1993; Jondrow et al. 2002; Stoll and Davis 1993;
Francies and Shaw 2000) and those of the USAF (Hildebrandt and Sze 1990;
Kiley 2001; Pyles 2003; Keating and Dixon 2003) in the areas of workloads,
material consumption, repairs per flight hour, mean time between failures, and
314 D. W. Maybury
program depot maintenance all show positive growth with age. Investigators warn
(Pyles 2003) that changing accounting practices, budget sluggishness, and rela-
tively fixed maintenance-personnel requirements plague USAF and USN studies,
creating a difficult environment in which to extract age effects from other pres-
sures. Raymond Pyles’s 2003 RAND corporation investigation for the USAF on
the impact of aging airframes on maintenance represents the most comprehen-
sive study to date (Pyles 2003). Using regression methods that address several
issues simultaneously, Pyles discovered a positive relationship between mainte-
nance requirements—in nearly all activities—with airframe age. Not only did
Pyles find that different maintenance tasks realize different age effects for the same
airframe, but he also found that age effects correlate with aircraft complexity and
fly-a-way costs.
In light of growing evidence that suggests age impacts maintenance and hence
O&M costs, DND requires a strategy to determine the optimal replacement time
for a fleet of vehicles after which it becomes increasingly disadvantageous both
economically and operationally to maintain the fleet. As a case study for construct-
ing and implementing a strategy for fleet replacement, we consider the CP-140A
Arcturus maritime surveillance aircraft. Acquired from the Lockheed Aircraft
Corporation by DND in 1993, the CP-140A is the sister aircraft of the CP-140
Aurora, with both airframes based on the P-3 Orion. The Arcturus fleet con-
sists of three vehicles based with 14 Wing at CFB Greenwood and, unlike the
Aurora, the Arcturus does not possess anti-submarine warfare capability. Today,
the Arcturus finds a primary role in pilot training, drug trafficking interdiction,
search-and-rescue, and sovereignty patrols.
As part of the Canada First Defence Strategy announced in 2008, the Gov-
ernment of Canada has undertaken a significant renewal process for Canada’s
military. In late 2008, the Assistant Deputy Minister (Materiel) and the Director
General Aerospace Equipment Program Management sought operations research
input from the Centre for Operations Research and Analysis (CORA) on deci-
sions surrounding the CP-140A Arcturus and the CP-140 Aurora fleets. The
Government of Canada had recently altered the decade-long Aurora Incremental
Modernization Program (AIMP) for CP-140 life extension to help position DND
for the eventual acquisition of a new long-range maritime patrol surveillance air-
craft. The Assistant Deputy Minister (Materiel) wished to better understand the
expected service life of this platform.
Since acquisition, the CP-140A Arcturus fleet has seen annual O&M costs per
vehicle increase dramatically since procurement. Discerning pure age effects from
the upward spiralling annual O&M costs presents a highly non-trivial problem.
Instead of attempting to isolate each age effect from usual inflationary pres-
sures, operational tempo, budget processes, training programs, and random events
(such as the discovery of unanticipated maintenance problems), we apply the
Greenfield–Persselin model (Greenfield and Persselin 2002, 2003) by assuming
that random events occur independently of airframe age, but that the aircraft’s
resiliency diminishes with time. Our position maintains that random events trigger
O&M costs that evolve proportionally—the aircraft becomes more susceptible to
Optimal fleet replacement 315
damage as it ages. By focusing on economic considerations that include tradeoffs
among costs, operational availability, and the effects of uncertainty, we find an
optimal replacement time in a stochastic setting.
In addition to addressing fleet replacement issues, the model we apply also con-
tains forecasting capabilities, yielding insight into future replacement decisions.
Most importantly—and we cannot overstress this point—the model implicitly
attaches an option value to the capacity to delay a decision. Models based solely on
deterministic net present value (NPV) arguments implicitly assume that a military
investment choice is either reversible, in that the military can cancel the invest-
ment if circumstances become unfavourable, or if irreversible, that the military
investment choice occurs as a now or never proposition. Deterministic approaches
oversimplify military fleet replacement decisions by not placing an option value
on the capacity to delay.
In this paper, we model the CP-140A Arcturus O&M costs as a geometric ran-
dom walk with a stochastic discount factor represented by operational availability
(Ao ). We apply the techniques of dynamic programming to calculate an optimal
replacement time by translating the Greenfield–Persselin model into an explicit
first passage time problem. The model we construct represents a prototype for a
fleet replacement strategy using data from the CP-140A Arcturus as a case study.
The methods we use build on the Greenfield–Persselin model and we apply the
model directly to help inform Canadian military leadership on fleet-wide decisions
regarding Canada’s maritime surveillance capabilities.
We organize the paper in four parts. Following the introduction, we present
the Greenfield–Persselin model in Section 12.2. Section 12.3 applies the model to
the CP-140A Arcturus data and provides a mean first passage time approach to the
replacement problem. Finally, in Section 12.4, we present a discussion concerning
military impact and future research.
where s denotes the replacement time. We now imagine that the replacement pro-
cess repeats ad infinitum. Like the simplified forester’s problem, if all else remains
equal, each generation faces the same O&M cost structure. Thus, we can write the
total ownership cost of successively owning a series of fleets as
∞
s
c(s) = p+ dt m(t) exp(−r t) exp(−r si ), (12.2)
i=0 0
where s denotes the replacement time of each generation. The total ownership
cost, Equation (12.2), represents a converging geometric series yielding
s
p+ dt m(t) exp(−r t)
c(s) = 0
. (12.3)
1 − exp(−r s)
Differentiating this expression with respect to the replacement time, s, we find that
the military minimizes total ownership cost by applying the first order condition
which yields
∗
s
0
dt m(t) exp(−r t) + p exp (−r s ∗ )
m(s ∗ ) = r p + r . (12.4)
1 − exp (−r s ∗ )
As an example, if we know that the instantaneous O&M cost rate has the form
m(t) = b exp(αt), then we find that
Optimal fleet replacement 317
and
• the constant p in the numerator represents the part of the total ownership cost
associated with the initial procurement;
• the second term in the numerator represents a NPV contribution to the total
ownership cost arising from O&M activity; and
• the denominator represents a discount factor.
318 D. W. Maybury
Greenfield and Persselin promote the O&M cost function, the NPV, and the
discount factor of Equation (12.5) to a stochastic variable and use dynamic pro-
gramming to calculate the expected replacement time. By calculating the expected
discount factor and the expected NPV at the expected replacement time, Green-
field and Persselin produce a boundary in the O&M costs which signals the
optimal replacement point (Greenfield and Persselin 2002, 2003).
While the Greenfield–Persselin model strictly uses O&M costs as the random
process that triggers replacement, we feel that the strict reliance on monetary
considerations masks the military’s true objective: maximize a military advan-
tage at the best possible cost. By relying only on pure monetary issues, the
Greenfield–Persselin model can call for long holding periods even in cases where
the fleet fails to provide the military with operational availability. A similar prob-
lem occurs in economic modeling. Few people value money for its own sake, but
rather value the consumption possibilities that wealth represents. For this reason,
economic modeling usually resorts to utility functions based on consumption as
opposed to wealth itself. We extend the Greenfield–Persselin model by including
a stochastic discount factor with the O&M cost process. Determining an objective
value function that attempts to measure degrading capability will not only prove
exceedingly difficult to construct, but will also destroy the attractive feature of the
Greenfield–Persselin model in that the fleet is always compared to itself. We pro-
pose using operational availability (Ao ) as the stochastic discount factor. By using
Ao , we penalize the fleet for failing to provide the military with a vehicle deemed
necessary even if the lack of availability reduces costs.
We model the O&M costs per Ao (m u ) as a geometric random walk
dm u = αm u dt + σ m u d W (t), (12.7)
where d W (t) is the Weiner process, α gives the geometric growth rate, and σ sets
the strength of the driving uncertainty. Following the Greenfield–Persselin model,
we can calculate the expected replacement time, the expected discount factor at
replacement, and the expected NPV at replacement using dynamic programming.
Each part can be used to assemble an expected total ownership cost function. The
expected replacement time, a(m ∗u , m u ), satisfies the Bellman relation
da(m ∗u , m u ) = dt + a(m ∗u , m u ) + E da(m ∗u , m u ) , (12.8)
where
2
α 1 α 1 2r
β1 = − − + − + , (12.11)
σ2 2 σ 2 2 σ2
β1 −1 β1
b
1 − mb∗ + p mb∗
β1 β1 r−α u u
m ∗u = (r − α) p + (r − α) β1 ,
β1 − 1 β1 − 1
1 − mb∗
u
(12.12)
We do not include the actual dataset for the CP-140A Arcturus in this report.1
Instead we provide a normalized dataset that sets the initial O&M per Ao at 4.4 in
dimensionless units. We apply a standard likelihood analysis technique to estimate
the model parameters of the geometric random walk in Equation (12.7), yielding
(at two standard deviations)
Notice that we have large uncertainty in the estimated parameters. This reflects
the limited costing data we have CP-140A fleet—we have a small sample size.2 In
Figure 12.1, we see the O&M per Ao sample path (in normalized units) with the
calculated barrier and its associated uncertainty represented by a hatched bar. The
first passage interpretation becomes immediately clear in the presence of real data
(normalized in this case). The uncertainty in the estimated barrier, m ∗u , implies
uncertainty in the expected replacement time. Since the expectation is calculated
conditionally on parameter estimates, the expectation itself becomes uncertain.
To capture the subtleties associated with uncertainty in the parameter estimates,
we treat the problem as a first passage of time to the region delineated by the
upper and lower estimate bounds of m ∗u . We interpret the region between the two
boundaries as a critical region for the decision-maker. Once the process enters the
Optimal fleet replacement 321
critical region
100
O&M costs/Ao
10
1
5 10 15 20 25
Years after procurement
Figure 12.1 CP-140 Arcturus data with the critical region indicated.
critical region, the model calls for the decision-maker to consider a replacement
or an overhaul initiative.
By interpreting entry of the critical region as a flag for initiating a decision
process, we see in Figure 12.1 that the model calls for replacement consideration
of the CP-140A Arcturus at year ten. Using Equation (12.8), the expected time to
enter the critical region from the first year of data is at year 11. The uncertainty
envelope gives the 95 percent confidence band for the future sample path.
The application of the original Greenfield–Persselin model, which does not use
a utility discount, yields Figure 12.2. Without the utility discount, and again using
standard likelihood techiques, we find the parameter estimates
with barrier
The expected time to the critical region jumps to 23 years, representing more than
a 100 percent increase over the model without the stochastic discount factor. We
see that the effect of a stochastic discount factor dramatically changes the predic-
tions of the model. Again, we display the uncertainly envelope at the 95 percent
confidence band.
322 D. W. Maybury
100
10
1
5 10 15 20 25 30
Years after procurement
Figure 12.2 CP-140 Arcturus data with no stochastic discount factor shown with the
critical region indicated.
Given the current level of the O&M per Ao process, the first passage of time
interpretation allows us to calculate the expected sojourn time within the critical
region, and the probability that the process will exit through the upper boundary
before the lower boundary. Both pieces of information provide a clearer picture of
the decision problem to military leadership. The sojourn time calculation can give
decision-makers an estimate of the amount of time remaining before the model
definitively calls for fleet replacement or a major overhaul.
To calculate the expected time of a drift diffusion process within a region, we
can use the backward Fokker-Planck equation (see Gardiner 2003 for details),
which for our model yields the differential equation for the expected time as
d T (m̃ u ) 1 d 2 T (m̃ u )
k1 + k2 = −1, (12.17)
d m̃ u 2 d m̃ 2u
where L̃, and Ũ denote the log-transformed upper and lower boundaries of the crit-
ical region respectively. We find that the expected sojourn time within the critical
region at year 11 for the CP-140A Arcturus in Figure 12.1 is 0.8 years.
As an aside, we can also calculate the probability that the O&M per Ao process
will exit through the upper barrier of the critical region before the lower barrier by
once again invoking the backward Fokker-Planck equation. In this case, we have
where πU (m̃ u ) represents the probability of exit through Ũ before L̃ with the
conditon that
and the condition that πŨ (Ũ ) = 1 and πŨ ( L̃) = 0. The probability that the process
will exit through the upper boundary Ũ before the lower boundary is given by the
solution to Equation (12.19), namely,
−k2 −k2
k1 m̃ u k1 L̃
e −e
πŨ (m̃ u ) = −k2 −k2
, (12.21)
Ũ L̃
e k1 −e k1
−k2 −k2
m̃ L̃
e k1 u −e k1
π L̃ (m̃ u ) = 1 − −k2 −k2
, (12.22)
Ũ L̃
e k1 − e k1
where m̃ u gives the current log-transformed value of the O&M per Ao process. The
probability of exiting through the top of the critical region before the lower bound-
ary can give the decision-maker insight into delaying the procurement initiative
once the process has entered the critical region.
In addition to calculating the sojourn time and probabilities of exit, we can also
calculate the distribution of the hitting time to any level. In particular, since we are
working in the log-transformed space, the probability density function for the first
passage time to level L̃ is given by
# $
( L̃ − m u ) ( L̃ − m u − (α − 1/2σ 2 )t)2
p( L̃, m u ; t) = √ exp − , (12.23)
σ 2πt 3 2σ 2 t
324 D. W. Maybury
0.08
0.07
0.06
Probability density
0.05
0.04
0.03
0.02
0.01
0
0 5 10 15 20 25 30 35
Years after procurement
where α and σ are the growth and volatility parameters of the model. In
Figure 12.3 we show the distribution of the hitting time to the lower barrier in
Figure 12.1 from the first year of data. The hatched region in Figure 12.3 denotes
the 60 percent confidence region around the expectation of 11 years. The hitting
time calculation allows decision-makers to estimate the probability of reaching
a given level which can shed light on the risks of holding the fleet longer than
anticipated.
We should not view the critical region solely in terms of fleet replacement. The
critical region simply denotes the place in O&M per Ao space where the expected
total ownership costs begin to realize an expected marginal cost. We imagine that
several external variables, such as queuing problems or sparing issues, can cre-
ate an adverse growth environment and corrections in these areas might radically
drive the sample path away from the critical region in a non-stochastic manner.
Furthermore, we might find it advantageous to undertake a reset-the-clock over-
haul rather than acquiring a new fleet once the sample path has drifted into the
critical region. Ideally, we see the stochastic model presented in this paper used in
conjunction with other decision tools.
We have treated the O&M cost per Ao as a single stochastic variable in a
drift-diffusion model. The statistical tests we performed on the data gave us no
compelling reason to reject this approach. In reality, the O&M costs and Ao both
evolve stochastically and potentially feed back on one another. Thus, we can
imagine the fleet replacement/overhaul problem as a coupled system of stochastic
differential equations with O&M cost following a geometric random walk, and the
Ao tracking a different process, such as a mean reverting process (e.g. an Ornstein–
Uhlenbeck process with boundaries). We might also advance the model by using a
stochastic interest rate with an appropriate industry specific military deflator (see
Chapter 17). Finally, if military consensus is available, the stochastic discount
factor could be promoted to a nonlinear function of Ao with the inclusion of poten-
tially other relevant performance and economic measures. In particular, since the
fleet will require routine maintenance work, we might imagine a stochastic dis-
count function of Ao that turns on slowly as Ao diminishes. We view this area of
research as an exciting possibility.
The application of the Greenfield–Persselin model and the model extension pre-
sented above rest on the ability to capture the statistical properties of the sample
path in the data. Describing the sample path as a geometric random walk rep-
resents an approximation. If the data supports it, we can use more sophisticated
models, such as jump diffusion (or more general Lévy processes). These modifica-
tions would require a new first passage time calculation along with a re-calculation
of the critical region. Again, this represents exciting opportunities for further work.
However, as the O&M cost per Ao is a price process that evolves over time, a geo-
metric random walk is a good starting place. If on examination of the data, the
sample path cannot be approximated by a geometric random walk, new meth-
ods will be required. Finally, as in all operational research studies, this model is
an approximation of reality that endeavors to provide decision support—in this
326 D. W. Maybury
case a military cost–benefit analysis. The overall replacement/overhaul decision
will require additional input, including examination of replacement alternatives,
budget feasibility, and military obsolescence.
The CP-140A Arcturus fleet represents a special case in that the homogenous
fleet consists of only three aircraft all delivered to the military at the same time.
The application of the model to the CP-140A did not include complications from
multiple aircraft variants, aircraft with different service lives, or aircraft that have
seen life extension modifications. More complicated fleets will require special
treatment of these circumstances to apply the Greenfield–Persselin model more
generally.
Operational tempo, unforeseen engineering and maintenance problems, and
budgeting processes all factor into the evolution of O&M costs of vehicle fleets.
Separating specific variables that directly relate to age from other effects presents
a daunting task. By treating the problem stochastically under the assumption
that fleet resiliency diminishes with time, we can determine the critical time to
consider fleet replacement. Most importantly, the stochastic approach implicitly
attaches an option value to the capacity to delay in making the replacement
decision—as more information becomes available to the decision-makers, bet-
ter choices can be made. Deterministic methods that focus only on marginal
costs or NPV inputs ignore this important feature. The Greenfield–Persselin
model, with a stochastic discount factor, balances economic considerations, uncer-
tainty, and fleet performance in establishing a replacement time. We feel that
the Greenfield–Persselin model class represents an important step in under-
standing optimal fleet replacement times in military life cycle cost–benefit
problems.
Expected age
First, we imagine a pre-determined cut-off cost per utility (perhaps set indepen-
dently by policy) m ∗u , above which DND is not willing to pay. Once we find
m u > m ∗u , we consider fleet replacement or a reset-the-clock overhaul. We imag-
ine that the process starts at initial condition b below m ∗u and, using Bellman’s
recursive equation from the last section, we see that the expected age, a(m ∗u , m u )
Optimal fleet replacement 327
evolves as
Note that Equation (A12.1.1) contains two parts: an additive time increment in
the expected age from the current time increment, and a term that represents
the expected age at replacement resulting from the time increment. After some
manipulation, we can re-express Equation (A12.1.1) as
E(da(m ∗u , m u ))
+ 1 = 0; (A12.1.2)
dt
Expanding the differential using Ito’s Lemma we find
1 2 2 d 2 a(m ∗u , m u ) da(m ∗u , m u )
E(da(m ∗u , m u )) = σ mu + αm u , (A12.1.3)
2 dm 2u dm u
where we have used E(d W ) = 0. Thus, we have the differential equation for the
expected age of replacement as
1 2 2 d 2 a(m ∗u , m u ) da(m ∗u , m u )
σ mu + αm + 1 = 0. (A12.1.4)
2 dm 2u dm u
We should note that, while we obtained Equation (A12.1.4) through dynamic pro-
gramming, we could have obtained the same result from a mean first passage time
approach. If we regard the cutoff, m ∗u , as a barrier, the problem amounts to find-
ing the expected time for m u to diffuse from some initial starting point to the
barrier. The stochastic formalism requires the backward Fokker-Planck equation
(Gardiner 2003) and, in this case, results in Equation (A12.1.4). Thus, we can
approach the replacement problem from either perspective, and we will use either
formalism as appropriate. Since m u = 0 is an absorbing point in the geometric
random walk, the solution to Equation (A12.1.4) reads
Again, we can obtain the same result from calculating the mean first passage time
of a(m ∗u , m) from b through the barrier m ∗u .
Equation (A12.1.6) connects with the deterministic approach since setting σ = 0
recovers the deterministic result. In the deterministic case, we find the exact age
328 D. W. Maybury
at which to replace the vehicle as a function of the pre-determined boundary, m ∗u ,
namely
1 ∗
a(m ∗u , b) = ln m u /b . (A12.1.7)
α
The deterministic case determines m ∗u through the minimization of the total own-
ership cost function. Thus, we see that uncertainty in the system, characterized by
σ , demands that we hold the platform longer than the deterministic result. We see
that the stochastic approach implicitly values the capacity to wait by recognizing
that there is a chance that next year’s O&M cost per utility will decrease. The
term, σ , reflects the size of this chance and thereby increases the waiting time. We
also notice that in order for the Equation (A12.1.6) to make sense, we must have
α − (1/2)σ 2 > 0, otherwise the expectation time to cross the barrier m ∗u becomes
infinite (implying that it will never be advantageous to replace the vehicle).
Rearranging, we find
1 2 2 d 2 ρ(m ∗u , m u ) dρ(m ∗u , m u )
E(dρ(m ∗u , m u )) = σ mu + αm u dt = 0.
2 dm 2u dm u
(A12.1.10)
1 2 2 d 2 ρ(m ∗u , m u ) dρ(m ∗u , m u )
σ mu + αm u − rρ(m ∗u , m u ) = 0. (A12.1.11)
2 dm 2u dm u
ρ(m ∗u , m u ) = Am βu 1 + Bm βu 2 . (A12.1.12)
Optimal fleet replacement 329
Substituting our solution into Equation (A12.1.11) we obtain β1 , β2 as the roots of
the quadratic equation
1 2
σ (β − 1)β + αβ − r = 0, (A12.1.13)
2
namely,
α 1 α 1 2 2r
β1,2 = − − ± − + 2. (A12.1.14)
σ2 2 σ2 2 σ
ρ(m ∗u , 0) = 0, (A12.1.15)
demands that the coefficient associated with negative root vanishes, and the
boundary condition
ρ(m ∗u , m ∗u ) = 1 (A12.1.16)
implies that A = (m ∗u )−β1 . Thus, the full solution for the expected discount factor
at replacement, in light of the boundary conditions, reads
β1
mu
ρ(m u ) = . (A12.1.17)
m ∗u
M(m ∗u , m u ) = m u dt + exp(−r dt) M(m ∗u , m u ) + E(d M(m ∗u , m u )) .
(A12.1.18)
1 2 2 d 2 M(m ∗u , m u ) d M(m ∗u , m u )
σ mu + αm u − r M(m ∗u , m u ) + m u = 0.
2 dm 2u dm u
(A12.1.19)
If we now put the expected net present life-cycle value together with the
expected discount factor, we find the expected ownership cost curve as a function
of the cost:
β1 −1
p + r−α
b
1 − mbu
c(m u ) = β1 . (A12.1.21)
1 − mbu
β1 −1 β1
b
1 − mb∗ + p mb∗
β1 β1 r−α u u
m ∗u = (r − α) p + (r − α) β1 .
β1 − 1 β1 − 1
1− b
m∗u
(A12.1.22)
Note that Equation (A12.1.23) conveys the same information as Equation (12.6)
since the deterministic case links the replacement age and the O&M cost per utility
in a one-to-one manner.
Comparison of Equation (A12.1.22) and Equation (A12.1.23) reveal the essen-
tial difference concerning the value of waiting. The deterministic case balances
the marginal costs and the known savings from delay through a small increment in
time. On the other hand, the stochastic case balances the marginal costs and sav-
ings from delaying until the O&M costs per utility increase by a small increment.
In the stochastic case, the time interval required to increment the O&M costs per
utility by a given amount is random.
Optimal fleet replacement 331
Notes
1 DND did not want the actual cost data included with this paper.
2 This paper elucidates an application of the Greenfield–Persselin model with extension to
a Canadian Forces fleet that was already undergoing a replacement decision process by
DND, and it was for this reason that we chose the CP-140A fleet.
References
Dixon, M. 2005. The Costs of Aging Aircraft: Insights from Commercial Aviation. RAND
Dissertation, Pardee RAND Graduate School.
DND (last access July 2014). Ch-124 Sea King: Intelligence, Surveillance and Recon-
naissance Aircraft. http://www.rcaf-arc.forces.gc.ca/en/aircraft-current/ch-124.page.
Department of National Defence 2, 1.
Francies, P. and G. Shaw 2000. Effect of Aircraft Age on Maintenance Costs. Briefing,
Center For Naval Analyses, Alexandria VA.
Gardiner, C. 2003. Handbook Of Stochastic Methods (3rd edition). Heidelberg:
Springer.
Greenfield, V. and D.M. Persselin. 2002. An Economic Framework For Evaluating Military
Aircraft Replacement. Santa Monica, CA: RAND MR-1489-AF.
——. 2003. “How Old Is Too Old? An Economic Approach To Replacing Military
Aircraft.” Defence and Peace Economics 14(5): 357–368.
Hildebrandt, G. and M. Sze. May 1990. An Estimation Of USAF Aircraft Operating And
Support Cost Relations. Santa Monica, CA: RAND, N-3062-ACQ.
Johnson, J. August 1993. Age Impacts On Operating And Support Costs: Navy Air-
craft Age Analysis Methodology. Patuxent River, MD: Naval Aviation Maintenance
Office.
——. June 1962. Forecasting The Maintenance Manpower Needs Of The Depot System.
Santa Monica, CA: RAND, RM-3163-PR.
Jondrow, J., R. Toast, M. Ye, J. Hall, P. Francis, and B. Measell. 2002. Support
Costs And Aging Aircraft: Implications For Budgeting And Procurement. CNA
Corporation.
Keating, E. and M. Dixon. 2003. Investigating Optimal Replacement Of Aging Air Force
Systems. Santa Monica, CA: RAND MR-1763-AF.
Kiley, G. 2001. The Effects Of Aging On The Costs Of Maintaining Military Equipment.
Washington, DC: Congressional Budget Office.
Marks, K. and R. Hess. 1981. Estimating Aircraft Depot Maintenance Costs. Santa Monica,
CA: RAND R-2731-PA&E.
Nelson, J. 1977. Life Cycle Analysis Of Aircraft Turbine Engines. Santa Monica, CA:
RAND, R-2103-AF.
Pyles, R. 2003. Aging Aircraft: USAF Workload And Material Consumption Life Cycle
Patterns. Santa Monica, CA: RAND MR-1641.
Ramsey, T., C. French, and K. Sperry. 1998. Aircraft Maintenance Trend Analysis. Briefing,
Oklahoma City Air Logistics Center, and Boeing.
Stoll, L. and S. Davis. July 1993. Aircraft Age Impact On Individual Operating And Support
Cost Elements. Patuxent River, MD: Naval Aviation Maintenance Office.
This page intentionally left blank
Part V
Selected applications
This page intentionally left blank
13 Embedding affordability
assessments in military
cost–benefit analysis
Defense modernization in
Bulgaria
Venelin Georgiev
13.1 Introduction
The economic reality of scarce resources combined with the human need for
security requires hard choices. The challenge is to make rational, informed and
effective decisions to meet existing needs with available resources1 in a way that
maximizes security. Defining, motivating, and supporting such decisions with mil-
itary cost–benefit analysis (CBA) is a complex process, the implementation of
which would be impossible without appropriate tools—quantitative and qualita-
tive analytical methods and models for evaluating the effectiveness of the use
of resources to meet the needs of people and society. The limited availability
and competing demands for financial resources on the one hand, and the exis-
tential human need for security on the other, requires careful resource allocation
decisions, and useful tools to justify and support the implementation of those
decisions.
The tools to manage defense resources include various management systems
(requirements generation, financial management, defense acquisition, etc.), and
involves making routine and reasonable decisions. The overall effectiveness of
defense spending depends on the resulting output, namely the security of society,
guaranteed by the defensive capabilities obtained as a product of careful manage-
ment of defense resources and innovative and technologically appropriate defense
investments (see Chapter 3). The efficiency and effectiveness of management deci-
sions depends on objective, scientifically sound and generally accepted military
CBA methods and models such as those described in previous chapters, applied in
practice by competent experts, managers and politicians.
Resource decisions are made within a process that in itself needs to be trans-
parent to decision-makers, e.g., to allow the preservation of a clear audit trail
from national security objectives, through defense objectives, to taxpayers’ money
(Tagarev 2009). Among the various questions related to resources management,
the most important are:
The risk is that there exists an imbalance between the massive investment costs
required to modernize armed forces, and the quality and effectiveness of cur-
rent military CBA methods and models applied to the management of acquisition
projects, especially relating to the evaluation and selection of projects for future
funding.
Another example is provided by the recent experience of a new NATO member,
Bulgaria. The country suffered a significant cut in its 2010 defense budget both in
absolute terms and as a percentage of its gross domestic product (GDP). The new
defense capital budget amounted to only 0.8 percent of the total defense budget
in 2010, compared to an average of 20–22 percent in previous years. This sharp
reduction in the amount of investment funding, among other things, increased
demands on the decision-making process. It created a strong impetus to improve
financial and project management, and in particular to increase the application of
military CBA methods that incorporate careful economic analysis in the selection
and evaluation of force modernization projects.
Bulgaria (BG) faced many of the same problems experienced by the US
described above. For example, within a single planning cycle, cost estimates of
the country’s commitments to NATO defense and security in the form of force
goals increased several times, from the time commitments were originally made to
the actual assessment of costs in the programming and budgeting cycle. Adverse
effects from implementing such a mechanism to plan, evaluate and implement
defense programs and projects can be classified in several ways.
First, attempts to reduce investment costs can lead to the development of incom-
plete and therefore less effective investment projects.3 Here it is sufficient to
mention two examples of false savings. The acquisition of ships for the BG
Navy was made without appropriate energy units to provide power while docked.
Another example involves the purchase of training aircraft for the BG Air Force
that did not include the necessary computer simulator courses required to support
the initial training of pilots. The results of such oversights is that life-cycle costs
of defense investments increase unexpectedly, and/or that it becomes necessary to
launch new investment projects for the acquisition of additional equipment, train-
ing, services or supplies. Ultimately, this leads to lower quality investments that
reduce the desired capability (effectiveness) of the program.
Second, incentives matter. There is strong motivation for program managers to
complete investment projects, even if they prove costly, since cancelling them can
be a career killer. This behavior can bias cost estimates. If optimistic (lower) cost
estimates are systematically used in military CBAs, and future defense funding
constraints are not explicitly incorporated into the analysis, then more investment
projects are started than can realistically be financed by the defense budget. As
a consequence, in the implementation phase financial resources have to be trans-
ferred from one investment project to another and from one major program to
338 V. Georgiev
another, negatively impacting overall capabilities and the performance of defense
forces. Ideally, such inter-program decisions should be conducted ex-ante through
a properly structured CBA and not ex-post under duress.
It is often necessary to reduce the scope, or to suspend or eliminate invest-
ment projects to accommodate budget shortfalls. This was the experience of recent
projects for the acquisition of transport aircraft and helicopters by the BG Air
Force (BGAF), and for automotive equipment by the BG Land Forces. The ulti-
mate result of management decisions based on biased or incomplete analysis is a
delay in providing required defense capabilities, often accompanied by a reduction
in the level and quality of those capabilities.
The issue of reporting the feasibility of investment projects in terms of future fund-
ing realities is becoming more urgent. The 2010 White Paper on Defense and the
Armed Forces of Bulgaria declares the intention of Ministry of Defense (MoD) lead-
ership to develop a long-term investment plan incorporating all force modernization
projects aimed at building and maintaining required defense capabilities. Acknowl-
edging future funding constraints, the plan calls for the MoD to include investment
projects that are “financially feasible in the period until 2020” (MoD 2010).
Ideally, military CBAs would normally require investment plans that ensure
financial feasibility of modernization projects throughout their life cycles, and
thus a longer horizon of investment planning may be necessary. The designers of
these plans should take account of multiple risks, including those related to future
financing opportunities, the depreciation of the defense capital stock over time,
economic viability, the dynamics of innovation and technology, possible changes
in the security environment, and the commitments of the country to international
security (see Chapter 7).
Accepting that innovation is a “new idea that works” (Tzvetkov 2004), analysts
and practitioners need to look for opportunities to utilize military CBA to resolve
analytically challenging tasks in defense resources management, or to approach
existing analytical tasks in more effective ways. The new “real options” approach
introduced in Chapters 11 and 14 in this volume offer an example.4 No matter
whether innovation is of the “push” type, implementing ideas coming from science
and research, or “pulled” from needs identified in practice, the objective is to pro-
vide defense forces that maximize national (and collective) security of individuals
and society.
Implementation of innovative ideas can increase the competitiveness of the
defense sector in the contest for the allocation of public funds among areas such
as health, education, internal order, etc. The economic evaluation of alternatives
(EEoA) presented in Chapter 4 is an example of one such innovative method for
performing CBA in evaluating and selecting investment projects to modernize the
armed forces, under realistic budget scenarios.
One of the major differences between the traditional analysis of alternatives
(AoA) and EEoA is the content of the reported input variables in the analy-
sis (Figure 13.1). While examining operational effectiveness and the life-cycle
costs of defense investments (similar to a conventional AoA), EEoA adds a new
variable—the uncertainty of existing and projected future funding constraints.
Affordability assessments 339
METHOD
(tool)
INPUT OUTPUT
(variables) (results)
Overall Measure
Components of
Level 1 Level 1 Level 1 measures
More specific
Level 2 Level 2 Level 2 Level 2 Level 2 Level 2 Level 2
measures
Overall
Level 1 Level 2
measure
Survivability
fighter
Specifications (MoP)
Source: Guide for Aircrafts’ Tactical and Technical Characteristics, 2000, pp. 66–82.
Source: Guide for Aircrafts’ Tactical and Technical Characteristics, 2000, pp. 66–82.
sufficient time and resources since it can have a major impact on the final results
(see Chapter 8). Analysts should take into account all relevant cost factors and the
total expected effectiveness of each alternative. MoE/MoP results can be reported
as shown in Table 13.2.
As discussed earlier in Chapter 5, analysts use a variety of methods to calcu-
late life-cycle costs. Here, the Government Asset Management Committee (2001)
recommends engineering methods when there is detailed and accurate capital and
operational cost data available; direct cost estimation of particular cost elements;
analogous cost methods when there exists data from similar systems of analogous
size, technology, and operational characteristics; and parametric cost estimation
when historical cost data is available (see Chapter 5).
344 V. Georgiev
13.3.3 Approaches to compare and rank alternatives
The next phase of the military CBA involves comparing and ranking alternatives.
The literature refers to this as “the heart of the analysis.” In the vocabulary of the
EEoA, the multi-role fighter example is an “intra-program” analysis, and the alter-
natives under consideration could be constructed as competing fleets of aircraft
(see Chapter 4).
One of the main challenges in defense management is to achieve a balanced
portfolio of acquisition projects that, among other requirements, are feasible in
terms of budgetary constraints. It is possible to address this challenge through
EEoA (see Chapter 4). This method clearly distinguishes among three key vari-
ables: life cycle costs, operational effectiveness, and the budgets expected to be
available to finance the program/project (funding scenarios). One of the ideas
behind EEoA is to improve coordination between capabilities-based planning,
programmatic defense resources management, and defense acquisition.
An innovative idea of EEoA is to use different budget scenarios to reflect fore-
casts for the future programming period (also see Chapter 7). EEoA prescribes
an explicit optimization approach to make military investment decisions within
realistic funding constraints. This parallels the optimization approach implicit in
PPBS, designed to build an optimal mix of forces, equipment and supplies which
maximizes national security within fiscal constraints.
In general there are six ways to structure an EEoA, as illustrated in
Figure 13.4.5
• In the first five approaches the program or project can be analyzed inde-
pendently of other programs/projects in the defense portfolio. In the sixth
approach, the analysis of costs and effectiveness takes account of competing
(or complementary) programs/projects included in the portfolio.
• The first three approaches assume that alternatives can be constructed by
decision-makers in cooperation with analysts (endogenous alternatives); an
example would be comparing alternatives fleets of multi-role fighters gener-
ated by experts and the programming team (see Chapter 15 for example).
• The last three approaches focus on evaluating previously defined alternatives
(exogenous alternatives) e.g. different companies producing multi-role air-
craft participating in the contest for selection of a supplier of such aircraft for
the needs of the BGAF.
Type of
EEoA
Intra-program
analysis
Build New
Alternatives
Modify Existing
Alternatives
Modify Existing
Alternatives
For example assumptions and constraints for the multi-role fighter could involve
future budgets, the time-span of life-cycle costs and effectiveness, technical and
tactical characteristics, operations and maintenance costs, etc. In discounting
future costs (see Chapter 17) there are several key milestones such as a mis-
sion’s life, base year, lead-time, etc. The mission’s life is the period over which
the system is required. The base year is the year to which costs and benefits for the
alternatives will be discounted. Start year is the year in which initial investments
for the considered alternatives are made. Lead-time is the time from the beginning
of the start year to the beginning of the economic life of an asset for each of the
alternatives. Possible values of these time characteristics for the multi-role fighter
example are shown in Table 13.3.
The next to last phase of a military CBA involves exploring assumptions and
constraints through sensitivity analysis. Sensitivity analysis answers “what-if”
346 V. Georgiev
Table 13.3 Time characteristics assumed for the alternatives
Source: Guide for Aircrafts’ Tactical and Technical Characteristics, 2000, pp. 66–82.
questions and helps to generate confidence in the results. Variations in key param-
eters such as costs, future budgets, or MoE/MoPs can be explored. If a change in
the value of a parameter impacts the ranking of the alternatives, this needs to be
carefully recorded.6
The final phase of any military CBA is to report the results and make recom-
mendations. It is an essential phase—the best alternative needs to be identified and
recommendations regarding its implementation formulated. Recommendations to
decision-makers are extremely important because sometimes the rankings do not
clearly demonstrate which alternatives are truly superior options.
The final results from the application of the AoA approach should be pre-
sented to decision-makers in the most attractive and intuitive manner possi-
ble. The report should include the goals, assumptions and constraints, list of
alternatives, the database for life-cycle cost and effectiveness, appropriate cost-
effectiveness charts, and results of the sensitivity analysis. The report should end
with conclusions and recommendations.
13.4 Conclusion
In conclusion we can summarize by observing that the application of the EEoA
framework offers several significant advantages in military CBA. The multi-
criteria approach to decision-making, which is a basis for the traditional AoA,
focuses on life-cycle costs and operational effectiveness while leaving affordabil-
ity as a secondary consideration. In contrast, EEoA expands the range of variables
in the analysis by explicitly including future funding constraints. In practice,
this can be accomplished through the development and implementation of finan-
cial scenarios for the acquisition project. Given future funding challenges, it is
imperative that all defense acquisition decisions explicitly include affordability.
Notes
1 The amount of financial resources is limited by various factors, such as the size of gross
domestic product, the state of the economy as a whole, the amount of funds allocated
from the state budget in the interest of building and maintaining defense capabilities, the
allocation of earmarked funds for security in areas which include personnel, maintenance,
training, participation in operations, investments, etc.
2 Another consequence is the adoption of investment programs and projects in defense
without having sufficiently comprehensive, detailed and reliable estimates of costs and
Affordability assessments 347
associated resources. In a study conducted in 2009, the Business Executives for National
Security (BENS) concluded that most of the requirements for defense capabilities are
determined in the absence of adequate input as to the suitability of the investment
programs and projects in terms of budgetary constraints.
3 In the language of project management, too often the scope of the investment project
does not match the scope of the defense products, which is the subject of investment.
4 The development and implementation of improved or new methods of military CBA
approaches in the evaluation and selection of investment projects to modernize the armed
forces can be considered a critical part of the innovation process in defense.
5 A detailed discussion of the six approaches to structure an EEoA appears in Chapter 4.
This chapter offers a decision map which significantly eases the practical work of ana-
lysts in choosing the preferred approach to structure an EEoA. Depending on specific
characteristics of the particular acquisition, this decision map facilitates the analytical
process and can serve as a valuable guide for analysts and decision-makers.
6 An appropriate way to find parameters that impact the results is to evaluate their influ-
ence, for example, in terms of percentages of the costs or effectiveness of the alternatives.
From a practical point of view, all factors that are related to spending greater than 20 per-
cent from the total spending for the alternative need to be considered in the sensitivity
analysis.
References
DOA. 2001. Economic Analysis Manual. US Army Cost & Economic Analysis Center.
GAO. March 2009. Michael J. Sullivan, Defense Acquisitions: DoD Must Prioritize
its Weapon System Acquisitions and Balance Them with Available Resources, Testi-
mony, before the Committee on the Budget, House of Representatives, GAO-09-501T,
Washington, DC.
Government Asset Management Committee. 2001. Life Cycle Costing. Sofia: Military
Publishing House.
Guide for Aircrafts’ Tactical and Technical Characteristics. 2000. Sofia: Military Publish-
ing House.
Ministry of Defense. 2010. White Paper on Defence and Bulgarian Armed Forces. Sofia:
Ministry of Defense.
Tagarev, Todor. Spring–Summer 2006. “Introduction to Program-based Defense
Resource Management.” Connections: The Quarterly Journal 5:1 55–69. Avail-
able at www.pfpconsortium.org/file/introduction-to-program-based-defense-resource-
management-by-todor-tagarev [last accessed February 8, 2012].
Tagarev, Todor. 2009. “Introduction to Program-based Force Development.” In Defence
Management: An Introduction, edited by Hari Bucur-Marcu, Philipp Fluri, and Todor
Tagarev. Geneva: DCAF, pp. 75–92.
Tzvetkov, Tzvetan. 2004. Innovations and Investments in Defence. Sofia: Stopanstvo.
14 Real options in military
acquisition
A retrospective case study of the
Javelin anti-tank missile system
Diana I. Angelis, David Ford, and
John Dillard
14.1 Introduction
This chapter offers an application of the real options approach to cost–benefit
analysis (CBA). The Javelin “fire-and-forget” anti-tank missile was originally
designed to replace the US Army’s DRAGON missile system. Introduced in the
late 1970s, the DRAGON had a wired command link to guide its missile to the
target that needed to be optically tracked by a gunner. Besides placing the gunner
at risk, the overall effectiveness of the DRAGON was somewhat compromised
due to limited range, lethality, and reliability problems. It also proved challeng-
ing for gunners to aim the missile and track the target. The goal was to replace
the DRAGON with a lighter weapon with increased range, lethality, and ideally a
“fire-and-forget” tracking and guidance system.
The development of the Javelin, shown in Figure 14.1, required advances in
several immature technologies to achieve the desired benefits. Target locating and
missile guidance subsystems represented particularly troublesome issues. Three
alternative technologies were initially considered. Each of the three technologies
represented realistic options that promised the desired benefits/effectiveness.
Instead of investing in a single technology, the Army decided to award three
“Proof of Principle” (technology demonstration phase) contracts of US$30 million
to each of three competing contractor teams to develop their technologies, fol-
lowed by a “fly-off” missile competition. The Army paid US$90 million in August
1986 for these three options, each with the potential, but not a guarantee, of deliv-
ering the desired benefits. Spreading the US$90 million over three competitors,
the Army in effect acquired the right, but not the obligation, to purchase the most
successful technology at a later date—a textbook example of a “real option.”
Section 14.2 offers a brief review of real options theory introduced in
Chapter 11 of this volume. Section 14.3 follows with a more detailed description
of the three Javelin guidance technology alternatives. A multi-criteria model is
developed in Section 14.4 to measure the benefits or “effectiveness” of the three
Real options: the Javelin anti-tank missile 349
Figure 14.1 The Javelin anti-tank weapon system missile and command launch unit.
Source: US Marine Corps photo 2/10/2010 by Corporal Andres Escatel.
0.3 Lethality
0.7 P(H) × P(K) 5 1.05 4 0.84 7 1.47
0.3 Top attack 6 0.54 7 0.63 9 0.81
0.3 Tactical
advantage
0.4 Weight 9 1.08 5 0.6 3 0.36
0.3 Time to engage 8 0.72 7 0.63 5 0.45
0.2 Time to flight 7 0.42 5 0.3 5 0.3
0.1 Redirect 10 0.3 10 0.3 0 0
capability
0.3 Gunner
safety
0.2 Required training 5 0.3 1 0.06 10 0.6
0.8 Exposure after 2 0.48 8 1.92 10 2.4
launch
0.1 Procurement 1 Ease of 8 0.8 6 0.6 4 0.4
procurement
MoE 5.69 5.88 6.79
(0.2), and then redirect capability (0.1). Gunner safety is measured mostly by
the gunner’s exposure to enemy fire after launch (0.8) and also by the amount
of training required (0.2).
In this example, actual values for each measure (computed or subjectively
assigned) for each of the three competing technology alternatives (LBR; FO;
FLIR) are converted (normalized) on a scale of 0 to 10. For instance, in the case
of gunner safety, the objective is to minimize the amount of time the gunner is
exposed to enemy fire (measured in minutes). With LBR the gunner must stay in
place until the target is hit, leading to a longer exposure time, so LBR receives a
low value of 2. The FO system allows the gunner to hide while guiding the mis-
sile, so he is exposed for a shorter time and, thus, FO receives a better value of 8.
The FLIR system allows the gunner to conceal himself immediately after launch
(“fire-and-forget”) and, thus, is given the maximum value of 10. The other values
shown in Table 14.1 are derived in similar fashion.
The overall MoE achieved by each of the three alternatives under the previous
assumptions is shown at the bottom of Table 14.1. Individual scores shown in
Table 14.1 for each measure are calculated by taking the normalized value for that
measure, multiplying it by the relative weight for that measure, and multiplying
again by the relative weight placed on the higher level objective.5 The overall MoE
for any given alternative is the sum of the individual metric scores.6
The overall MoEs reported in Table 14.1 suggest that FLIR offers the greatest
benefit/effectiveness (6.79), with LBR and FO closely tied for second (offering
354 D. I. Angelis, D. Ford, J. Dillard
overall MoEs of 5.69 and 5.88, respectively). This ranking turns out to be con-
sistent with the Army’s revealed preference for the three guidance technologies:
They preferred FLIR over the other two guidance systems, and they perceived
the FO system as being slightly better than the LBR system (J. Dillard; personal
communication, November 12, 2012).
Program Cost
10
-
7
.&
• + LBR
4 • Fo
3 A FUR
0
$- $50 $100 $150 $200 $250 $300 $350
$Millions
Table 14.3 Marginal analysis of cost and effectiveness for LBR and FO
Table 14.4 Marginal analysis of cost and effectiveness for FO and FLIR
Table 14.5 Probability of development success and expected MoE for Javelin technology
options
LBR FO FLIR
Program Cost
10
6
~ 5 + LBR
::E
• Fa
• --
4
•
3 .A FUR
0
$- $50 $100 $150 $200 $250 $300 $350
$Millions
Table 14.6 Expected cost of Javelin guidance technology alternatives without option to
terminate project
LBR FO FLIR
The values shown in Table 14.6 assume that we do not use a real options
approach. Instead, we will pick one of the technologies based on the cost vs ben-
efit analysis presented in Section 14.5. Whichever technology we choose, we will
have to pay an additional cost to achieve the anticipated effectiveness (MoE) if the
development phase fails.
The real options approach allows the Army to pay for the option to find out if
technology development succeeds before making its final choice. If development
succeeds, the MoE shown in Table 14.1 is achieved and the Army can proceed with
the project if they prefer that option based on the cost vs effectiveness analysis
presented in the previous section. If development fails, the particular development
project can be terminated and there is no further cost.
358 D. I. Angelis, D. Ford, J. Dillard
The value of the option is given by the difference between the expected cost
of the technology development project with no option (from Table 14.6) and the
expected cost of the project with the option to terminate. Values of the options for
the three alternatives appear in Table 14.7.
Differences in values of the options reflect the different levels of uncertainty
associated with each technology. The values shown in Table 14.7 are maximums
in the sense that if we pay any more than the option value we would have been
better off not using an option. If we pay less than the option value, we experience
real cost savings by not expending funds on an unsuccessful technology. Note the
more uncertain the technology (i.e. probability of failure), the greater the value of
an option to terminate the project if technology development fails.
Let us suppose that the Army preferred the LBR technology (based on the cost
vs effectiveness analysis presented in Section 14.5). Then, according to Table 14.7,
they should pay no more than US$92 million for the option to terminate the
project. Since the Army was in effect buying options for all three technologies,
this indicates that the total amount spent on options should not exceed the value
of the option for the preferred technology. Since the maximum value of the LBR
option is US$92 million, if the Army allocated that option value equally across all
the alternatives, they should spend no more than roughly US$30 million for each
option, which is precisely what they did.
Given some technologies are more uncertain than others, a better approach is to
allocate the US$92 million based on the level of uncertainty the Army is trying to
resolve. Using the probability of failure data in Table 14.7 as a notional measure
of risk, 27 percent of the total option value would be allocated to LBR, 33 percent
to FO, and 40 percent to FLIR.
Returning to the previous example, if ex-ante the Army prefers the LBR
technology, then the total cost of the option should not exceed US$92 million. Rec-
ognizing uncertainty (i.e. risk of failure), this means the Army should be willing to
pay roughly US$25 million for the LBR option, US$30 million for the FO option,
and US$37 million for the FLIR option. Doing so allocates the dollars based on
risk while keeping the total cost no more than the maximum option value of the
preferred alternative. Again, since US$92 million is the maximum value of the
LBR FO FLIR
14.8 Conclusions
Several observations can be drawn from the analysis of the Javelin guidance tech-
nology acquisition process. The first is that the benefit of weapon systems or, in
this case, missile guidance systems, is not measured in dollars. This makes using
a traditional option valuation model based on monetary benefits minus costs diffi-
cult (see Chapter 11 for example), if not impossible. Instead, a useful approach is
to turn to principles of multi-criteria decision-making to develop MoEs for each
alternative. The overall MoE can then be joined with cost estimates to illustrate a
trade space of alternatives available to decision-makers.
Secondly, we note that the three proposed guidance systems involved dif-
ferent levels of risk. This information can be used to calculate an “expected
MoE” for each alternative, incorporating uncertainty into the analysis. This prob-
abilistic MoE can be combined with the alternative’s expected cost to present a
risk-adjusted trade space for decision-makers.
Thirdly, we show that a real options approach not only allows uncertainty to be
incorporated explicitly in the analysis, but it also enables calculation of the value
of real options based upon various risks. This leads to different option values for
different alternatives based on technological maturity.
Using this real options approach, we conclude that the Army might have mit-
igated its risks by offering each development team a different amount of money
to develop their proposed technology, based on probabilities of success derived
from technological risk assessments. We note that the final “cost to fix” the FLIR
guidance technology ultimately selected by the Army turned out to be signifi-
cantly higher than the US$30 million originally paid to develop the technology.
This is in line with predictions of the real options model that recognized the FLIR
technology as the riskiest investment.
In conclusion, the real options approach to CBA captures the value of flexibility
in acquisition decisions. Faced with significant uncertainty, implementing a real
options approach can result in more efficient use of scarce resources.
Notes
1 In the classic example of stock purchase options, the exercise decision rule is to sell a
stock if the price rises above a certain price, and the exercise signal is the stock price.
The decision delay is incurred while the option holder waits to see if the stock price
rises above the exercise price.
2 The Advanced Anti-Armor Weapon System – Medium (AAWS-M) project inspired the
Javelin program. The joint Army and Marine Corps operational requirements document
for the Javelin was formally approved-amended in 1986–88. The FLIR approach con-
sisted of a laser beam-riding system, a fiber-optic guided system, and a forward-looking
infra-red system.
Real options: the Javelin anti-tank missile 361
3 The services procuring the Javelin system did not actually use this exact methodology
for the selection of the Javelin guidance technology, but used something similar for a
weighted decision analysis of the three alternatives.
4 This burden is sometimes termed “transaction costs” as opposed to “production costs.”
Note prior discussions in this volume about the interpretation of placing a weight on pro-
duction costs. Ideally, the operational performance of investments should be evaluated
independently of production costs in the spirit of “cost as an independent variable”
(CAIV). The next section discusses a way to combine cost and effectiveness measures.
5 For example, required training has a weight of 0.2 and is a metric that supports gunner
safety, which has a weight of 0.3. LBR received a value of 5 for this metric, so the score
for LBR is 0.3 = (5) × (0.2) × (0.3). The score for FO is 0.060 = (1) × (0.2) × (0.3)
and the score for FLIR is 0.6 = (10) × (0.2) × (0.3). All other scores in Table 14.1 are
calculated in a similar manner.
6 For example, the MoE for LBR is 5.69 = 1.05 + 0.54 + 1.08 + 0.72 + 0.42 + 0.30 +
0.30 + 0.48 + 0.80. Note that the MoE is calculated on a scale of 0 to 10 where an
“ideal” alternative would receive an MoE of 10.
7 As discussed in Chapter 4, and because the MoE combines different measures that are
not necessarily substitutes, it is not possible to use benefit/cost ratios (i.e. the ratio of
MoE to cost) to rank alternatives.
8 Solving the cost-effectiveness problem as defined in this case could be handled by “lev-
eling the playing field” as discussed in Chapter 4, using the fourth or fifth approaches
to structuring an economic evaluation of alternatives (EEoA)—the modified budget or
modified effectiveness approaches—or the sixth (“opportunity cost”) approach.
9 As part of the capability formulation process, technical constraints are deliberately
avoided in requirements documents to allow and encourage a maximum range of
alternative solutions to a particular requirement or capability deficiency.
10 It would be gauged today at approximately Technology Readiness Level 5. DoD
assesses the maturity of critical technologies on a scale from 1 (lowest) to 9 (highest)
as described in the Technology Readiness Assessment Guidance (2011) http://www.acq.
osd.mil/ddre/publications/docs/TRA2011.pdf [last accessed December 4, 2014].
11 The two-partner TI/Martin-Marietta Joint Venture in the full-scale development phase
was also free to maximize competition at the subcontractor level. In their make-or-buy
decision, Texas Instruments elected to make the focal plane array for both of its uses
in the command launch unit and in the missile. The company had made these devices
for other programs, but not in these two distinct configurations (scanning and starting
arrays).
12 Around focal plane array attainment of specified sensitivity and production yield,
system weight, tracker algorithm, and other areas.
13 This constituted a “Nunn-McCurdy breach” of cost and schedule thresholds, which
required Congressional notifications and formal re-baselining taking the better part of
the next year to accomplish. Over that next year, the program sought a new baseline
with many different revised program estimates—climbing from 36 months duration and
US$298 million in cost, to 48 months duration and US$372 million in cost, and finally
to 54 months and US$443 million for the total cost and duration of this phase.
14 The system design has continued to be upgraded, not as blocks of capability, but with
software, warhead, and producibility enhancements.
Bibliography
Amram, M. and N. Kulatilaka. 1999. Real Options: Managing Strategic Investment in an
Uncertain World. Cambridge, MA: Harvard Business School Press.
Brealey, R. and S. Myers. 2000. Principles of Corporate Finance (6th edition). New York:
McGraw Hill.
362 D. I. Angelis, D. Ford, J. Dillard
Buede, D. M. 1986. “Structuring Value Attributes.” Interfaces 16(2): 52–62.
Ceylan, B. K. and D. N. Ford. 2002. “Using Options to Manage Dynamic Uncertainty in
Acquisition Projects.” Acquisition Review Quarterly 9(4): 243–58.
Courtney, H., J. Kirkland, and P. Viguerie. 1997. “Strategy under Uncertainty.” Harvard
Business Review 75(6): 67–79.
Dillard, J. and D. Ford. 2009. “From Amorphous to Defined: Balancing the Risks of Spiral
Development.” Accounting Review (52).
Ford, D. and J. Dillard. 2009. “Modeling the Integration of Open Systems and Evolutionary
Acquisition in DoD Programs.” Acquisition Review Quarterly (51).
Garvin, M. and C. Cheah. 2004. “Valuation Techniques for Infrastructure Decisions.”
Construction Management & Economics 22(5): 373–383.
Garvin, M. and D. Ford. 2012. “Real Options in Infrastructure Projects: Theory, Practice
and Prospects.” Engineering Project Organization Journal 2(1–2): 97–108.
Keeney, R. 1982. “Decision Analysis: An Overview.” Operations Research 30(5): 803–838.
——. 1988. “Structuring Objectives for Problems of Public Interest.” Operations Research
36(3): 396–405.
Lyons, J., D. Long, and R. Chait. 2006. Critical Technology Events in the Development of
the Stinger and Javelin Missile Systems: Project Hindsight Revisited. Washington, DC:
National Defense University Center for Technology and National Security Policy.
Miller, R. and D. Lessard. 2000. The Strategic Management of Large Engineering Projects.
Cambridge, MA: MIT Press.
Triantis, A. 2005. “Realizing the Potential of Real Options: Does Theory Meet Practice?”
Journal of Applied Corporate Finance 17(2): 8–16.
15 An application of military
cost–benefit analysis in a
major defense acquisition
The C-17 transport aircraft
William L. Greer
15.1 Introduction
This case study illustrates the traditional approach to military cost–benefit analysis
(CBA) and acquisition decisions within the Department of Defense (DoD) described
in Chapter 8. It specifically discusses decisions made in 1993–95 to replace C-141
transport aircraft with the C-17 strategic airlifter. It also includes a discussion of
events leading up to the study, the analytical approach taken, the study participants,
critical points in the analyses, decisions reached, and lessons learned.
The C-17 Globemaster II aircraft has been one of the more successful military
transportation assets in history. It provides rapid delivery of cargo or passengers
almost anywhere in the world. Currently it is flown by the US Air Force (USAF)’s
Air Mobility Command (AMC) and by the armed forces of several other coun-
tries, including the United Kingdom’s Royal Air Force, the Royal Australian Air
Force, the Canadian Force’s Air Command, NATO, and the Qatar Emiri Air Force.
The United Arab Emirates recently ordered C-17s, and many other countries are
exploring the possibility of doing so. In addition to its transoceanic strategic-
lift capability, the C-17 can also perform tactical airlift, medical evacuation, and
airdrop missions. It played a significant role in moving goods, vehicles, and per-
sonnel during Operation Enduring Freedom in Afghanistan and Operation Iraqi
Freedom in Iraq, and it has contributed to emergency relief operations around the
world. It is widely considered a success story.
That was not always the case. The program came very close to being cancelled
in the early 1990s. Key decisions made around that time serve as a valuable exam-
ple of the useful role of CBA in major defense investment decisions. This chapter
describes cost-effectiveness analyses (CEAs) that informed the decision to retain
the C-17 program, and serves as a case study in how such assessments can assist
decision-makers. The focus is on the analyses but, just as important, it provides
insights into how these analyses were received and used to make the decision to
invest in the C-17.
The analyses depicted here are drawn from documents (Greer et al. 1993;
Bexfield et al. 2001) prepared by the Institute for Defense Analyses (IDA) for the
DoD. Studies such as these were referred to at that time as cost and operational
effectiveness analyses (COEAs), currently called analyses of alternatives (AoAs).
364 W. L. Greer
These studies include many of the basic considerations of cost and effectiveness
analysis discussed in this book.
15.1.1 Background
In the 1970s, the USAF launched a plan for a new cargo aircraft to carry most
of the largest equipment the Army needed (e.g. tanks, large-tracked and wheeled
vehicles, and Patriot batteries) into small remote airfields—many with runways
shorter than 3,000 feet. Although at the time airlifters existed that could carry large
equipment (the C-5) and land in small airfields (the C-130), no existing airlifters
could perform both these tasks.
In December 1979, the DoD initiated the Cargo-Experimental (C-X) competi-
tion for a new strategic airlift aircraft. The C-X was to be an aircraft that could
deliver a full range of combat equipment over intercontinental distances, operate
from a 3,000-foot runway, possess survivability features, have excellent reliabil-
ity, maintainability, and availability, and have a low life-cycle cost. Early in 1980,
the DoD issued the request for proposals (RFPs) for the new C-X program. Boe-
ing, Lockheed Martin, and McDonnell Douglas responded. After reviewing the
various designs, in August 1981 the DoD selected the aircraft design proposed
by McDonnell Douglas.1 The winning design incorporated many features already
demonstrated on the YC-15, a McDonnell Douglas aircraft that had been devel-
oped and flight tested in the 1970s. This chosen design was later designated the
C-17. Because of the earlier testing of the YC-15, it was viewed as a program with
low development and cost risk.
In addition to carrying large cargo and landing in small airfields, the C-17 was
also designed to back up and turn in a small radius, thereby optimizing the use of
limited ground space. It was also designed to conduct troop and cargo airdrops,
replacing the aging C-141s that previously conducted those missions. It was an
aircraft for all missions.
In 1982, the DoD conducted a large transportation assessment, the Mobility
Study, which established the air mobility requirements at that time, based on per-
ceived wartime needs. To meet these requirements, the USAF determined that 210
C-17s were needed. These requirements were codified in the 1983 Air Mobility
Master Plan. Largely because of budget rather than effectiveness considerations,
the number 210 was subsequently reduced by the Office of the Secretary of
Defense (OSD) to 120 during the 1990 Major Airlift Review. Thus the number
of C-17s under consideration in the early 1990s was 120.
Several complicating issues arose at that time. The earlier assessments of low
development and cost risk proved wrong. Costs for the C-17 continued to rise,
and the several aircraft delivered for testing did not demonstrate the reliability and
performance expected. Static wing loading tests revealed that the wings failed at
lower loads than predicted, forcing even higher costs to strengthen them.
Lockheed Martin, manufacturer of the C-5 and C-141, made an unsolicited pro-
posal in 1991 to upgrade and extend the life of the C-141s at a cost that appeared
lower than that of the troubled C-17 program. An added complication and a renewed
Acquisition of the C-17 aircraft 365
sense of urgency was presented when it was found that C-141s had developed a large
number of microscopic cracks in their wing structures—a condition that temporarily
grounded the entire C-141 fleet until repairs could be made.
This raised a number of questions for DoD: Should it continue with the C-17
program despite escalating costs and lower performance than desired? Should it
cancel the C-17 program and extend the life of C-141s, continuing with wing
repairs until a new C-X design was chosen? Or should it investigate other alterna-
tives? These were the issues confronting Congress and the DoD in 1993 and are
the topics of this case study.
Congress mandated that the DoD conduct an independent assessment of the
C-17 program and report its findings to the US Congress (Senate and House
Armed Services Committees). The legislation further restricted fiscal year (FY)
1994 spending on the C-17 until such a study was completed. After reviewing the
expertise and capabilities in several federally funded research and development
centers (FFRDCs), the OSD selected the IDA to conduct the study. IDA com-
pleted its study in 1994. The findings were used by the DoD and duly reported to
the congressional defense committees.
Table 15.1 summarizes the history of the C-17 program up to the point at which
analyses were conducted to determine whether to go ahead with the program.
With the CEAs in hand, the OSD finally approved the 120-aircraft C-17 pro-
gram in 1995. The road to this decision point was not an easy one. It required
numerous analyses, the main one of which is the cost-effectiveness case study
Year Events
1970s • USAF decided new cargo aircraft, C-X, was needed to carry
heavy and large army equipment into austere airfields
1981 • McDonnell Douglas selected to build C-X, called C-17
1982 • Mobility Study published, calling for more airlift capability
1983 • USAF Airlift Master Plan estimated that 210 C-17s needed to
meet needs identified in mobility study
1990 • Major Airlift Review reduced C-17 requirement from 210 to
120
1991 • Lockheed Martin proposed C-141 service life extension
program (SLEP)
1992 • Joint Staff issued report, concluding that C-17 is more
cost-effective than either new C-5 or C-141 SLEP
• Joint Staff issued Mobility Requirements Study
1993 • Problems with aging C-141s continued
• C-17 exhibited engineering problems and cost increases
• Congress restricted FY94 C-17 spending pending further study
• DoD asked IDA to conduct COEA on C-17 program
1994 • IDA completed C-17 COEA
• COEA sent to Congress after OSD review
• Additional DoD studies were authorized
366 W. L. Greer
summarized in this section. At the end of this section, we also summarize more
recent decisions affecting the C-17 program.
15.1.3 Methodology
The approach taken in this study consisted of the following steps:
15.2 Alternatives
Alternatives refer to the alternative fleets, each made up of a number of different
types of aircraft. This section first discusses the different types of aircraft in the
alternative fleets and then describes the composition of the alternative fleets, i.e.
how many of which types of aircraft are in each. Figure 15.1 provides a schematic
of the aircraft under consideration to indicate the relative sizes and capacities of
these airlifters. Capacity is significant for determining how much each airlifter can
carry, and size proves important for airbases with restricted space or facilities.
15.2.1.1 C-5A/B
At the time of the study in 1993, 76 C-5As and 50 C-5Bs made up the total aircraft
inventories of the active, guard, and reserve fleets.5 Operating costs and utilization
rates are slightly different for the two types, but they carry the same payloads the
same distances. These are the largest airlifters of the USAF. Each C-5 can take off
or land with a maximum payload of nearly 120 tons. At the time the study was
conducted, the average cargo weight carried by the C-5s was 68.9 tons (these esti-
mates are updated frequently by AMC, so more current numbers should be used in
future studies), with a block speed6 of 423 knots. C-5s can carry very heavy or very
large outsize cargo, such as M-1 Abrams tanks, Patriot battery radars, or CH-47
helicopters. They can also carry smaller oversize cargo, such as wheeled vehi-
cles (5-ton and smaller trucks, M-2 Bradley or Stryker vehicles, or high-mobility
multi-purpose wheeled vehicles (HMMWVs)). Of course, the C-5 can also carry
palletized bulk cargo such as ammunition, supplies and food, as well as troops.
15.2.1.2 KC-10A
The KC-10 is a dual-purpose aircraft. It serves as the largest airborne tanker in the
fleet and as a cargo carrier. For purposes of this study, AMC estimated that 23 of
the KC-10s would be dedicated to cargo missions. A single KC-10 can lift 83 tons
of cargo, although it can carry only a fraction of oversize cargo and cannot carry
any outsize cargo. The average payload assumed was 41.7 tons, and the aircraft
moves with a 445-knot block speed.
15.2.2.2 C-Y
The C-Y is a hypothetical new replacement airlifter for the C-141. The C-Y would
have modern off-the-shelf engines and avionics, but would carry the same load as
the C-141. It was used in the analyses as a potential replacement for the C-141 as
each C-141 reached its terminal fuselage life span at 45,000 flying hours, a point
reached before the end of the 25-year period of the analysis. Thus, two acquisition
and operating costs are associated with the C-141 SLEP: the cost of SLEP itself
plus the cost of the replacement airlifter when the C-141 reaches the end of its
operational life.
15.2.2.3 C-17
The C-17 is smaller than the C-5, but larger than the C-141. The C-17 can carry
outsize cargo, but not as much as a single C-5. The C-17 requires no more room at
Acquisition of the C-17 aircraft 371
an airbase than the C-141 does, which is important when airbase space is limited.
It can also operate from shorter runways than can the C-5, affording an opportunity
to use more airfields in-theater and en-route. The C-17 can take off and land with a
maximum payload of 85 tons, although the average payload assumed in the study
was 48.3 tons (again, AMC sources have updated data). The C-17 block speed was
estimated at 423 knots.
15.2.2.4 C-5B+
This is a new aircraft proposed as an alternative to the C-17. It is identical to the
C-5B except that new GE CF6 engines would be used to reduce the high noise
and pollution levels associated with the C-5 TF39 engines. The study assumed the
same performance as the C-5B. Since the current C-5 line has been shut down,
part of its cost involves a restart of the C-5 manufacturing line at Lockheed.
One-aircraft 1 120/102 – – – –
alternatives 2 – 263/225 – – –
3 – – 102/87 – –
Two-aircraft 8 94/80 58/49 – – –
alternatives 9 94/80 – 21/18 – –
with reduced 10 94/80 – – 18/17 –
numbers of 11 94/80 – – – 30/28
C-17s
25 70/60 108/92 – – –
26 70/60 – 42/36 – –
18 70/60 – – 34/32 –
22 70/60 – – – 56/53
23 47/40 160/137 – – –
24 47/40 – 62/53 – –
21 47/40 – – 49/47 –
17 47/40 – – – 83/78
4 26/20 212/181 – – –
5 26/20 – 82/70 – –
6 26/20 – – 66/63 –
7 26/20 – – – 108/103
Two-aircraft 12 – 136/116 49/42 – –
alternatives 13 – 136/116 – 40/38 –
with no C-17s 14 – 136/116 – – 65/62
15 – – 49/42 42/40 –
16 – – 49/42 – 69/66
19 – – – 32/30 83/79
Illustrative 20 47/40 93/79 – 21/20 –
three-aircraft
alternative
Note
a TAI = total aircraft inventory; PAA = primary aircraft authorized.
The entries in Table 15.2 represent alternative additions to the core fleet. Each
alternative also includes the then-current core fleet of C-5s, KC-10s, and Stage III
CRAF forces. The alternatives are sorted into sets of categories: one-aircraft alter-
natives; two-aircraft alternatives with reduced numbers of C-17s; two-aircraft
alternatives with no C-17s; and an illustrative three-aircraft alternative. Alternative 1
has 120 C-17s; all other alternatives have fewer, or no, C-17s.
Note that the one-aircraft alternatives involve only one new type of aircraft
added to the core fleet. As already noted, Alternative 1 contains the full 120 C-17
fleet as programmed at that time by the USAF. Alternatives 2 and 3 proposed
Acquisition of the C-17 aircraft 373
substitution of the C-17s with a SLEPed C-141 fleet or a new-start C-5 fleet,
respectively.
The alternatives that included a reduced number of C-17s proposed a number
of different options including some C-17s in addition to one other aircraft. For
example, Alternative 8 has 94 C-17s and 58 C-141 SLEPs. The smallest number
of C-17s considered in this category was 26 C-17s, the minimum number already
obligated to DoD under contract without incurring termination penalties.
The two-aircraft alternatives with no C-17s include mixes of the C-141 SLEP,
C-5B+, 747, and 767 aircraft, taken two at a time, that provide the same MTM/D
as 120 C-17s.
The study also briefly examined a few three-aircraft alternatives, but it was felt
that DoD would probably not embrace a program that called for three new airlifter
types unless the benefits turned out to be dramatic for a reasonable cost. Although
the results for these alternatives did not turn out to offer dramatic benefits, they
served as a useful example for purposes of comparison.
250
Cargo Required (ktons)
30 days
200 90 days
150
100
50
0
Outsize Oversize Bulk
Cargo Type
Figure 15.2 MRS airlift delivery requirements for MRC-East and MRC-West
combined.
troops to the target country. In this second scenario, CRAF was not activated and
the reserves were not mobilized. A total of 50,000 tons of cargo was required in
four days, in addition to that delivered in the initial brigade airdrop. This case
allowed the analytic team to test the robustness of the alternatives to very different
requirements from those addressed by the MRCs.
The MRC scenarios dominated the analyses conducted. These scenarios
required the full support of the airlifter fleet, including full augmentation by
CRAF. Most of the analyses discussed in this chapter refer to the first set of sce-
narios. It turns out that the LRC scenario added little to the insights obtained from
the two MRC scenarios.
The airlift requirements from the mobility requirements study (MRS) are sum-
marized graphically in Figure 15.2 for two different time periods: 30 days and
90 days. Virtually all outsize and oversize cargo was required to be in-theater
within the first 30 days. There are only minor differences between requirements
for 30-day delivery and for 90-day delivery. To meet continuing resupply require-
ments, the remaining bulk cargo in each case continues to build over time. This
emphasis on rapid delivery of outsize and oversize cargo will become the basis of
measures of merit (or effectiveness) in the study.
140
747
120
Maximum Payload (tons)
C-5
100
C-17 KC-10
80
767
60
C-141
40
20
0
0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000 5500 6000
Range (nmi)
Table 15.4 MOG estimates maximum number of aircraft on ground simultaneously (by
theater and aircraft type)
East 26 26 15
West 24 20 11
Table 15.4 summarizes MOG estimates for the airlifters analyzed at an aggregate
theater-wide level. Note that the largest airlifters consume nearly twice the MOG
space of the smaller ones. For example, a maximum of 26 C-17 aircraft can be on
the ground simultaneously in the East-MRC theater of operation.16
Another important input is the ground time required to load and unload cargo
and refuel aircraft. This reflects the amount of time each aircraft uses its allocation
of MOG. Ground times vary by mission (en-route refueling, on-load, and off-load)
and aircraft type. The C-5s had the longest ground times (3.75 hours on-load;
3.25 hours en route and off-load), especially when compared to 2.25 hours for en
route, on-load, and off-load for the C-17 and C-141. Boeing’s commercial deriva-
tives had the longest on-load times (4–5 hours) but the shortest en route times
(1.5 hours).17
15.3.3.2 MASS/AFM
The MASS/AFM was a large Monte Carlo simulation19 that scheduled and exe-
cuted each airlift mission by tail number. It incorporated a great deal of realism,
such as specific routes used, airbases with MOG limitations, times for loading
and off-loading, and crew duty days. It used the ALM loading results along with
the nature of the TPFDD cargo and PAX at individual airbases to load the cargo.
Only that part of the TPFDD available at a specific base on a specific day (plus
any undelivered leftover cargo from previous days) was available for loading. The
output was tons delivered per day.
15.3.4.1 General
The basic approach was first to establish acquisition schedules and C-141 retire-
ment schedules for all alternatives and to make assumptions regarding the
active/reserve aircraft mix for airlift. The study team used the C-17 schedule
programmed at that time by the USAF, with C-141s retired at such a rate as
378 W. L. Greer
to maintain a constant MTM/D capacity. Other airlift alternatives (C-5B+ and
commercial derivatives) were acquired at rates consistent with their manufactures’
capabilities, with the C-141s also retired at a rate to keep MTM/D constant.21
15.3.4.2 Acquisition
The study used the USAF C-17 acquisition schedule, with a maximum acquisition
of 16 per year. Historical cost data for the C-17 were used to generate a cost-learning
curve to predict future costs for C-17s. Appropriate adjustments to historical trends
were added as needed, such as the cost for extra weight to strengthen the wings
and higher costs associated with more realistic engine estimates for all lots after the
first ones. IDA estimated the full C-17 program acquisition cost at approximately
US$23 billion, about US$3 billion over the Program Office estimate. In the end, the
higher-cost IDA results were used in the report, with a comparison to the Program
Office estimates included as an excursion.
At the time the study was conducted, 26 C-17s were under contract with long-
lead funding. The US government would have incurred a cost penalty for breach of
contract if it failed to buy 26 production C-17s. Thus alternatives that included less
than 26 C-17s were levied an additional termination fee associated with breaking
the contract (US$1.5 billion).
Lockheed Martin Aeronautical Systems provided cost estimates for the alter-
native of restarting the C-5 line as well as for refurbishing the C-141s in a C-141
SLEP. IDA used these and other historical Lockheed data and made adjustments as
required. Examples of adjustments include adding hush kits to dampen the sound
of the engines and adding extra material to the fuselage. The cost to restart the C-5
line was estimated at US$750 million. New C-5 engines were used to reduce noise
levels below those mandated by the Federal Aviation Administration (FAA).
A recently completed Scientific Advisory Board (SAB) report had recom-
mended that the weep hole cracks in then-current C-141s could be repaired and
their life extended to 45,000 flying hours. When the C-141 reached its 45,000-
flying-hour life limit, the SAB determined it had to be replaced, and the cost of a
new C-Y replacement was imposed to include a US$3.5 billion development cost.
Boeing supplied basic cost data for the 747 and 767 commercial derivative air-
craft alternatives. Production rates from 6 to 12 per year for military acquisition
were used, based on Boeing production capabilities and competing markets for
new 747 or 767 aircraft. An additional 20 percent of the acquisition price was
added to account for reinforced floors and rollers and widened side doors, for
an estimate of US$155 million for the militarized 747 and US$88 million for the
militarized 767. Since the relevant production lines were already open, no develop-
mental costs were imposed. Over the course of the study, these same commercial
derivative aircraft also became known as “non-developmental airlift aircraft.”
65-503 1994). Use of the model requires aircraft to be sorted according to whether
they are in the active, associate or reserve fleets. It takes into account the number of
flying hours per year, the cost of the fuel burned, the cost associated with personnel
needed (crew, maintenance), software costs, training costs, the cost of spares, and
the cost of contractor logistics support.
O&S cost estimates from the USAF CORE model for all the aircraft types
are summarized in Table 15.5. The commercial derivative aircraft were assumed
to have contractor logistics support, the cost of which was included in the cost
estimates for alternatives containing commercial aircraft derivatives.
Current trucks
70~c=========~==========~~--------~~-----------,--------~
C-17 :
0 COMMERCIAL DERIVATIVES -----------L-------------------------
65 6, C-141 SLEP i TPFDD
D C-5B+
Outsize cargo delivered in 30 days (kilotons)
60
Equal cost
effectiveness
55 Alternative 1
cost
55
1 26
24
55 18 8
10
5 9
11 25
21 22
55 MRS 3
6 20
17 15
16 4 23
13 12
55 7
2
19
14
55
55 55 55 55 55 55
Figure 15.4 Comparison of cost and effectiveness of alternatives with current trucks.
each symbol identify the alternatives. To assist with visual interpretation, alterna-
tives that contain any C-17s contain a small black dot; those with any C-5B+ have
a square; those with any C-141 SLEP have a triangle; and those with commercial
derivatives such as 747s or 767s have an open circle. For example, Alternative 1
with only a small black dot includes only C-17s (120 according to Table 15.2). In
contrast, alternative 18 contains both C-17s and commercial derivatives (70 C-17s
and 34 747s according to Table 15.2) as indicated by the dot inside the larger open
circle.
For reference, the outsize cargo requirement of the 30-day scenario set by the
MRS is shown as a horizontal dashed line. Any fleet alternatives below the line
would require additional aircraft to satisfy this mission requirement. The vertical
dashed line shows the cost of the 120-C-17 Alternative 1. Any fleet alternatives
to the right have a higher cumulative cost than this fleet made up exclusively
of C-17s. The sloping dashed line indicates a locus of alternatives that have the
same cost-effectiveness as Alternative 1 (i.e. same “bang for the buck” or ratio of
effectiveness to cost).24
This chart clarifies several things. First, the C-141 SLEP alternatives (i.e. the
triangles) appear to be among the worst candidates relative to the C-17 fleet
(Alternative 1), both from a cost and from an effectiveness perspective. To make
them more effective by adding either C-141 aircraft or another aircraft that is
Acquisition of the C-17 aircraft 381
part of the mix, additional costs would be incurred.25 This seems to eliminate a
whole category of alternatives immediately. Second, the fleet alternatives of C-17s
combined with commercial derivatives (open circles with concentric black dots)
appear to be close in cost-effectiveness. This catches our attention and elevates
our consideration of militarized 747s and 767s. In particular, the 747 appears the
better of the two commercial derivatives. Finally, the C-5B+ (squares) alternatives
appear slightly less cost-effective in all their mixes than the C-17 mixes, but run a
close second. They cannot easily be dismissed from this chart.
Based on these initial results, the DAB was impressed with the potential for
using commercial derivatives such as the 747 at lower cost than the C-17 program.
The study team was tasked with examining several excursions and reporting back
as soon as possible. At the same time, C-17 proponents perceived the results as a
threat to their program and reviewed the assumptions of the analyses in detail.
Equal cost
55 effectiveness
50 ~-
-~.,...,-',.,.
45 1
t -------[il-----
26
18 24
-----------------------e·~~::::~t~-~--~--~----~-A------
10 8
21a* 11 5
MRS 21 9 25
40
c6 :
~--- -o 22 3 Q
35 ~~-~--
-----· e - 17
®C
16
15
i
:
4
20
23~
/ (i) 13 I 12
0 7
19
o : 2
30~----------~----------~------~(\
14
~~:--~~~------~--------~
45 50 55 60 65 70
25-year cumulative cost ($B)
Figure 15.5 Comparison of cost and effectiveness of alternatives with new army
FMTV trucks.
60~r=======~=========c~------~-------,--------,
~~~MERCIAL DERIVATIVES;~
j
Q Equal cost
55 O
~ C-141 SLEP effectiveness
C-58+
Outsize cargo delivered in 30 days (kilotons)
50
Alternative 1
cost _,,.; .... "' --
_......... -......... -
.... -,.."""
45
............
..
_1 -------[:] 26
I
E1 E1 9
18 10
@- ,''- ® : 24
MRS
40 ------------------- =: -:(!f ~ ~:::: =------~ - - lG -----------------------
21
15
3
5
----- c®18 10
&
35 ---------- • 23
c
13
---- ®
21 15
.0 2
30 Fleet Composition (TAI)
Legend:
.0
Alternatives
FMTV can load on modified 747
13 A/C 1 2 3 5 9 10 13 15 18 21 23 24 26
C-17 120 – – 26 94 94 – – 70 47 47 47 70
25 C-5B+ – – 94 82 21 – – 49 – – – 62 42
FMTV do not load on modified 747 747 – – – – – 18 40 42 34 49 – – –
C-141 1
20L_________ ________ _ L_ _ _ _ _ _ _ _ __ SLEP
_ _ _–_ _263
_L _ –_ _–_ _– _ _– 136 – – – 160 – –
60
~ ~L_ ~
45 50 55 65 70
In all cases, as the MOG was reduced, the ability of the C-17 to land and take
off on short airfields, to back up, and to take up less space than C-5s and the com-
mercial derivatives made it more favorable. This is illustrated in Figure 15.7 for
five of the alternatives. For clarity, only the most robust and constrained extremes
are shown while the moderate MOG case lies in between.
Changes in the MOG have a powerful influence on the results. In the constrained
case, none of the alternatives meet the MRS requirements, although Alternative 1
with the most C-17s suffers less than the other fleet alternatives with C-17s and
other, more space-intensive airlifters. In terms of systems thinking, this clearly
points to the possibility of investing in adequate infrastructure to improve MOG
as a new alternative that can be integrated with various combinations of aircraft.
Acquisition of the C-17 aircraft 385
50~------~------~------~------~------~------~------~----~
----~------------~---------------
Robust MOG
C-17 UTE equal cost
rate effectiveness
45 reduction 1 26
Outsize cargo delivered in 30 days (kilotons)
Robust MOG
MRS based on
- robust MOG
®-----------
- 15.2 C–17 UTE rate
16 [!] 26
40 -~:::~~~-~~~:~::: __________ c:J 3___________________________________ _
® 16 1
23
35
___ _ --
_,..._.,.,.
_.,.,._ ....
.
...
_,.._ ...
....
Constrained MOG
equal cost _____1 -----
30 C-17 UTE
effectiveness
------ rate
reduction
Constrained MOG
16 26
25
Fleet Composition (TAI)
Alternatives
A/C 1 3 18 23 26
20
C-17
C-5B+
120
–
–
94
70
–
47
–
70
42 D3 Note: C-17 surge UTE rates
23
747 – – 34 – –
C-141
- baseline 15.2 hrs/day
SLEP – – – 160 – - excursion 12.5 hrs/day
15L-----~------~------_L ______ _ L_ _ _ _ _ _~------~------L-----~
52 54 56 58 60 62 64 66 68
25-year cumulative cost ($B)
Figure 15.7 Impact of reduced MOG and reduced C-17 use rates.
It also cautions against assuming that lower cost alternatives always provide the
best payoffs.
Another key assumption in the analysis was questioned by 747 proponents who
objected to the assumption of higher C-17 use rates. The basic analyses assumed
a 12.5-hour-per-day use rate for the 747s and a 15.2-hour-per-day rate for the
C-17s under wartime surge conditions. The use rate attainable by the C-17 was
still unproven when the study was conducted, so some doubted it would exceed
that achievable by commercial aircraft.
Figure 15.7 also illustrates the effect of a smaller use rate for the C-17. The
reduced use rate was assumed to be 12.5 hours/day—the same as the assumed use
rates for the C-5B+ and the militarized commercial derivatives. The smaller use
rate reduced effectiveness and also cost (because of fewer flying hours per year
associated with such a lower use rate). The overall effect was noticeable but was
not nearly as dramatic as that seen in the MOG assumption excursions. While it
was reasonable for the study team to be skeptical about such a large projected use
rate for an unfinished aircraft, deviations from that projection did not dramatically
affect the cost-effectiveness measures for the C-17. Interestingly, subsequent tests
two years after this original study validated the 15.2 hours/day value, rendering the
arguments moot. Study teams, however, can never know these things in advance
and are well served to look carefully at all key assumptions that could influence
the results.
386 W. L. Greer
15.5.2 Assuming airlift capability other than 52 MTM/D
After the first DAB meeting in August 1993, there was a growing sense on the part
of some decision-makers that commercial derivatives in conjunction with some
number of C-17s might be the best solution. When the initial results were briefed
to the DAB, additional airlift capability assumptions were requested. These cases
were developed to determine how few C-17s might be acceptable if that program
was reduced from the 120 aircraft originally assumed in Alternative 1.
IDA was asked to assess how well a number of new alternatives would fare.
The result of one set of analyses is summarized in Figure 15.8. In this case, 10
new alternative fleets are introduced, with MTM/D values ranging from 46 up to
55. An attention to cost was also used to make virtually all the new alternatives
no more costly than Alternative 1. Figure 15.8 illustrates cost and effectiveness
measures for Alternative 1, various mixes of C-17s and commercial derivatives,
and mixes of C-17s and C-5B+s. Since by the time the DAB was expected to
reach its decision it was expected that 40 C-17s would already be delivered or
under construction, many of the new alternatives included 40 C-17s, and none
involves fewer than that number. As Figure 15.8 shows, the set of mixes of C-17s
with commercial derivatives is roughly equal in cost-effectiveness to Alternative
1 with exclusively C-17s. Moreover, for roughly the same budget, the commercial
and C-17 fleet mixes are well above the effectiveness of the C-17 and C-5B+
mixes.28 This chart confirmed the sense derived from earlier charts that a mixed
fleet of C-17s and commercial derivatives might have merit—provided they could
carry the new army trucks.
Some decision-makers felt that a fleet size capable of generating a transportation
capability of 52 MTM/D was too small. After all, the C-17 program had been born
when the requirement was 66 MTM/D. As previous charts have shown, all alterna-
tives fell significantly below the effectiveness demanded by the TPFDD line. Thus
IDA was subsequently asked to examine the possibility of alternatives that generated
higher MTM/D. Figure 15.9 summarizes the results of two different sets of alterna-
tives: those with 52 MTM/D and those with 59 MTM/D, a value with historical roots
when new airlift programs were initiated several years prior to this study period. The
59 MTM/D alternatives are all identified with the prefix “E” on the chart. It is inter-
esting that the same frontier line observed earlier still carries over to these higher
MTM/D cases. Namely, all alternatives fall on or below the equal cost-effectiveness
ratio line that runs through Alternative 1, within statistical uncertainty.29
46 747 – 55 64 74 27 14 effectiveness
MTM/D 52 52 54 55 52 52
1
44
112
,/®_/-; 116
42
.A!> 115
125
--------
40 MRS __...- ---- II
-----------:=~~~~~ ~--------------------~-----------------------
111
124
------ 110
®
Fleet Composition (TAI)
38 C-17/C-5B+ Mixes
Alternatives
---------------- A/C 1 121 122 123 124 125
El C-5B+
MTM/D
–
52
30
46
35
47
42
48
18
50
16
52
122
121
34
32 C-17
COMMERCIAL DERIVATIVES 0
C-58+ D
30L_----~----~------~----_L--~==~====~====~====~
50 52 54 56 58 60 62 64 66
25-year cumulative cost ($B)
Figure 15.10 summarizes the results for LRC-Short. In this case, the effective-
ness measure is the total cargo delivered (not just outsize) in eight days after the
initial first-day airdrop. The 25-year total cost is the same as before. The four-day
TPFDD is used for purposes of comparison, although it has not been achieved
even in eight days by any alternative.
A comparison of the amount of cargo delivered for robust (baseline) MOG con-
ditions and for a 50 percent loss in airfields shows the same general results that
appeared before in the case of the MRC scenarios: alternatives with fleets with C-5s
and 747s suffering a greater reduction in capability than those with C-17s when
MOG is constrained. Those with the C-5B+ seem to suffer the greatest reductions.
65 A/C 1 3 5 9 10 15 18 21 24 26
C-17 120 – 26 94 94 – 70 47 47 70 _..
C-5B+ – 94 82 21 – 49 – – 62 42 Equal cost – _..
Outsize cargo delivered in 30 days (kilotons)
effectiveness
60
747 – – – – 18 42 34 49 – –
... --
----
.... --
Fleet Composition (TAI)
______ ....... ----
Expanded Alternatives (59 MTM/D)
.......
55 A/C E-1 E-2 E-3 E-4 E-7 E-8 E-9 E-10 E-3
.,.•,,.,. .... --
E-1 ... --
C-17 176 120 120 120 94 94 70 47
C-141 SLEP – 124 – – – – – – __ . . . . -- ®E-4 E-7
E-2
50 C-5B+ – – 48 39 69 – –
.... --
–
747 – – – – 57 73 89
E-8
___ .,. .......... "' .... "' @
E-10 / (!) E-9
45 ~[!] 25
18 10
_.. -f
1
24
-"'"' I r:1
Sl ..--"' @ :5 L!J[!] 9
MRS
40 --------~~~---- ~ - ~ --------------------------------------------
,_.--(!}
21 3 :
. . ----"' [!]
15
:I
35 I
I Alternative 1
I
I cost
I
I
1
30L---------~--------~ ---------L--------~----------L---------~--------~
50 55 60 65 70 75 80 85
25-year cumulative cost ($B)
Figure 15.9 Comparison of 59 MTM/D expanded capacity alternatives with the nominal
52 MTM/D alternatives.
In the cost estimates shown earlier, each alternative was developed estimating
the costs of efficient production-lines, independent of annual budgets. A more real-
istic approach would be to constrain all annual expenses over the six-year period
of the Future Years Defense Program (FYDP) to be no greater than those associ-
ated with the C-17 program itself, as expressed in the then-current C-17 Selected
Acquisition Report (SAR). This would force many of the alternatives analyzed to
stretch their acquisition profiles to remain under the spending cap.30 To test how
stretching out programs and the concomitant increases in acquisition costs would
influence the outcome of the analyses, a sensitivity analysis was conducted. The
results for selected fleet alternatives appear in Figure 15.11.
As Figure 15.11 shows, fleet alternatives with mixes of C-17 and commer-
cial derivatives all show increases in total cost of US$1–2 billion. The cost of
Alternative 1 also increases slightly, since the cost team did not feel that the full
120 C-17 program could stay within the cost caps of the C-17 acquisition bud-
get. The C-5B+ fleet alternatives do not increase in cost in this excursion, since
the start-up time for the new C-5 line takes several years and removes it from
competition with other aircraft types in those alternatives. As the figure shows, even
Acquisition of the C-17 aircraft 389
45
Total cargo delivered in 8 days (kilotons)
21
® ®18 1
40 10 ® 5
® [!] Baseline MOG
c
35 15 1
3 ®
30
0
® ®
25 18
® 10
21 One-half MOG
20 excursion
[!] 5
15
c 0
Fleet Composition (TAI)
15 3 Alternatives
10 A/C 1 3 5 10 15 18 21
C-17 C-17 120 – 26 94 – 70 47
5
0 COMMERCIAL DERIVATIVES C-5B+ – 94 82 – 49 – –
fj. C-141 SLEP 747 – – – 18 42 34 49
0 C-58+
0
45 50 55 60 65 70
25-year cumulative cost ($B)
though total lifetime costs do increase if the SAR serves to limit near-term fund-
ing, the C-17/commercial-derivatives fleets still dominate the cost-effectiveness
comparisons.
The relative contribution from the different cost elements can be seen in
Figure 15.12. Figure 15.12 also illustrates the effect of different discounting
assumptions on the cost estimates. Two of the alternatives are compared in
Figure 15.12: Alternative 1 with 120 C-17s, and Alternative 2 with no C-17s but
with 263 C-141 SLEP/C-Ys instead. Figure 15.12 also illustrates the effect of dis-
counting through two examples: no discounting and the baseline OMB-directed
4.5 percent discounting. Note that the large effect of O&S costs that accrue over
the 25-year life-cycle cost period for all alternatives. Discounting emphasizes
near-term costs, which has a bigger effect (reduction) on long-term O&S cost
estimates than on near-term acquisition cost estimates.
65 ~ C-141 SLEP
0 C-58+
Outsize cargo delivered in 30 days (kilotons)
Alternative 1
cost
55
__ ....
_.,.,. ..... ""-
50
- -
................ ........
Equal cost
.... -
........
- -
effectiveness
45 __1 ........ ........
~ . . @-® ® :5
18 10, - " ' I
40 ---------------------~;
MRS 21 ~~::_=---- ~ --~ ---------------------- -
3
_........ _........ [] : Fleet Composition (TAI)
Alternatives
~~- 15 I
..-- I A/C 1 3 5 10 15 18 21
35 --~
:
I
~ ..... .,.,.,...... C-17 120 – 26 94 – 70 47
-- II C-5B+ – 94 82 – 49 – –
I 747 – – – 18 42 34 49
30L-------------~------------~------------~~------------~------------~
70
1
45 50 55 60 65
25-year cumulative cost ($B)
• The cost and performance of the planned C-17 fleet alternative (#1) make it
the preferred military airlifter. It is more resistant to airfield constraints (i.e.
MOG) than the new C-5 and possesses a higher use rate. It is far superior in
both cost and effectiveness to the C-141 SLEP.
• The next most attractive alternatives after the 120 C-17s would be mixed fleets
of C-17s and modified commercial aircraft with specially reinforced floors
and some concession for the height of new FMTV army trucks, such as wider
side doors.
• If new army trucks cannot be loaded on commercial derivatives, then the next
most attractive alternatives to the 120 C-17 program would be mixes of C-17s
and new C-5s.
A mixed fleet of commercial derivatives and some number of C-17s (less than
120) became a serious contender to Alternative 1, the original/baseline 120 C-17
fleet, reducing the total cost of the program. The study did caution that overall
effectiveness could be compromised. The introduction of 747s would provide
fewer aircraft for certain unique military operations such as airdrops, deliveries
to remote and inadequate airfields, low-altitude parachute extractions, and rapid
off-loading of cargo while the airlifter is still moving down the runway (combat
Acquisition of the C-17 aircraft 391
C-130 O&S
C-Y O&S
100 C-141 O&S
C-17 O&S
25-Year Cumulative Cost (FY 93 $B)
C-Y Acquisition
80
C-141 SLEP Acquisition
C-17 Acquisition
60
40
20
0
Alt 1 Alt 2 Alt 1 Alt 2
No Discounting Baseline 4.5% Discounting
off-loading) with engines running. It also provided a fleet less well able to adapt
to limited MOG conditions. Nonetheless, its lower cost recommended it for more
detailed consideration.
reports subsequent decisions made to augment the C-17 fleet from the original
120 aircraft to 223.
The impetus for adding C-17s beyond the 120 initially approved came after
the attacks on September 11, 2001, and the decision to strike back at al-Qaeda
and the Taliban in Afghanistan. Airfield limitations in Afghanistan and Iraq sup-
ported the case for additional C-17s that could more easily operate in austere
environments.
Additional studies have continued to the present time, all addressing the issue
of how many C-17s (and other airlift aircraft) are needed. Table 15.6 summarizes
the conclusions reached in those studies. The programmed number in 2011 is 223
C-17s.
MASS/AFM
The MASS/AFM was a large Monte Carlo simulation31 that scheduled and exe-
cuted each airlift mission by tail number. It incorporated a great deal of realism,
such as specific routes used, airbases with MOG limitations, times for loading and
off-loading, and crew duty days. It used the ALM loading results along with the
nature of the TPFDD cargo and PAX at individual airbases to load the cargo. Only
that part of the TPFDD available at a specific base on a specific day (plus any unde-
livered leftover cargo from previous days) was available for loading. User-selected
priorities for loading out/over/bulk/PAX classes were considered for each airlifter.
The AMC analysts ran the simulations at Scott AFB with inputs from IDA. The
primary outputs for the study were tons delivered per day by commodity type and
class (out/over/bulk/PAX).
As with ALM, MASS/AFM may produce results that would be viewed as opti-
mistic in real operations. For example, the inefficiencies that often exist with high
levels of congestion at en-route bases are not included (although MOG does limit
the number of aircraft being serviced at the base). Again, the decision-makers
and study team need to make reasonable assumptions and focus on the higher
objectives of the study. If an assumption appears to have the potential for making
a large impact on the results, sensitivity analysis may be required.
396 W. L. Greer
Airlift cycle analysis spreadsheet
The ACAS model was a simplified approximation to the analyses produced by
MASS/AFM and was used to conduct the large number of excursions needed in
the study. It used aggregate data and aggregate MOG values, calibrated against
MASS/AFM to give confidence, for these numerous excursions.
Acquisition costs
The study used the Air Force C-17 procurement schedule, with a maximum acqui-
sition of 16 per year. It also used multi-year procurement arrangements to reduce
uncertainty and cost. From the historical cost data for the C-17, IDA derived a cost-
learning curve to predict future costs for C-17s as more manufacturing experience
is attained. This stage was a delicate one, requiring the cooperation of McDonnell
Douglas to supply competition-sensitive information with the understanding that
only aggregate levels of detail would appear in the final report. Publishing a sec-
ond proprietary document for use by the government, while using the aggregate
results in the COEA, attained the desired balance between a need for discussion
of the methodology and a display of the aggregate results and the need to pro-
tect the legitimate interests of the source. Appropriate adjustments to historical
trends were added as needed, such as the cost for extra weight to strengthen the
wings and higher costs associated with more realistic engine estimates for all lots
after the first ones. During the study, the C-17 Program Office disagreed strongly
with the IDA estimates. The basic disagreement arose from different assumptions
about the learning curves. IDA used data for the entire C-17 program, adjusting
the first few low-cost lots for their later prices while the Program Office used
later prices only. IDA estimated the full C-17 program acquisition cost at approx-
imately US$23 billion, about US$3 billion over the Program Office estimate. In
the end, the higher-cost IDA results were used in the report, with a comparison to
the Program Office estimates included as an excursion.
At the time the study was conducted, three test C-17s (not counted as part of
the 120) had been produced, an additional six production aircraft were undergo-
ing operational test and evaluation (OT&E), and a total of 17 production aircraft
had either been delivered or were in various stages of assembly at the Long Beach
McDonnell Douglas facility. Moreover, aircraft through production number 26
were under contract with long-lead funding. Failing to buy 26 production C-17s
would have incurred a cost penalty by the US government for a breach of con-
tract. Thus alternatives with fewer than 26 C-17s were levied an additional cost
associated with breaking the contract (US$1.5 billion).
Lockheed Martin Aeronautical Systems provided cost estimates for restarting
the C-5 line as well as for refurbishing the C-141s in a C-141 SLEP. Again, only
aggregate data appeared in the final report, with company-sensitive data relegated
to the proprietary government-use document. IDA used these data and other his-
torical Lockheed data and made adjustments as required. Examples of adjustments
include adding hush kits to dampen the sound of the engines and extra material
to the fuselage. A recently completed SAB had recommended that the weep hole
Acquisition of the C-17 aircraft 397
cracks in then-current C-141s could be repaired and their life extended to 45,000
flying hours. When the C-141 reached its 45,000-flying-hour life limit, the SAB
determined it had to be replaced, and the cost of a new C-Y replacement was
imposed to include a US$3.5 billion development cost.
For the C-5 restart, a cost to restart the line was imposed on any alternatives
using C-5B+ aircraft. IDA estimated that US$750 million would be needed to
restart the line in Marietta, Georgia. New C-5 engines were used to reduce noise
levels below those mandated by the FAA.
Boeing supplied basic cost data for the 747 and 767 commercial deriva-
tive aircraft. Production rates from six to 12 per year for military acquisition
were used, based on Boeing production capabilities and competing markets for
new 747 or 767 aircraft. An additional 20 percent of the acquisition price was
added to account for reinforced floors and rollers and widened side doors, for
an estimate of US$155 million for the militarized 747 and US$88 million for
the militarized 767. Since these lines are open, no developmental costs were
imposed.
Notes
1 McDonnell Douglas was subsequently merged with the Boeing Company in 1997.
2 Bulk cargo is carried on standard USAF 463L pallets and is transportable by all air-
lifters. A pallet holds bulk cargo measuring no more than 104 × 84 × 96 . Oversize
cargo is larger than the pallet dimensions and consists typically of wheeled vehicles.
Oversize cargo is larger than a single pallet but less than 1,090 × 117 × 105 . All
oversize cargo can fit on a C-141 (or larger aircraft). Outsize cargo is the largest of
these three, fitting only on C-5s and C-17s.
3 The use of a detailed model (i.e. MASS) along with a more aggregate one (i.e. ACAS)
is a common approach that balances study resources and time with precision in results.
Calibration of ACAS to specific MASS results allowed ACAS to be used in rapid sen-
sitivity excursion analyses. Comparisons of ACAS runs with MASS runs for these
same sensitivity excursions showed no loss in precision for using the significantly faster
ACAS.
4 Costs also addressed contract claims that McDonnell Douglas had made as well as
penalty charges associated with terminating the C-17 line short of then-contracted
production levels.
5 In 2010, a total of 111 C-5s were in the USAF inventory. Of these, 49 were C-5Bs.
6 Block speed is the distance flown by an aircraft divided by the time spent once blocks
are removed from the aircraft wheels and it taxies down the runway until it lands and
comes to a complete stop with blocks again in place.
7 Bulk cargo would be no problem in the commercial non-militarized version, since that is
what CRAF carries. The nature of the militarization would be to strengthen the floors to
handle higher pressure points (such as under the wheels of fully loaded five-ton trucks)
and possibly to modify the doors to permit wider cargo or equipment to be loaded from
the side.
8 A considerable number of additional alternatives were added later in the study, not all of
which had a capacity of 52 MTM/D, but these were introduced after the more powerful
comparison displays (to be shown later) were developed. At that point, equal MTM/D
became an interesting but marginally useful criterion. Actual performance in realistic
scenarios became the more important gauge. To start the analyses, IDA first focused on
these 26 equal-MTM/D alternatives.
398 W. L. Greer
9 The TAI entries indicate how many new aircraft are to be procured: The PAA values
are the numbers available to deliver cargo and troops in the scenarios studied and the
numbers used in O&S costing estimates.
10 Today the acronym MRC has been replaced by MCO (Major Combat Operation) and
CC (Conventional Campaign), but the meaning is equivalent. Here, we will use the
MRC notation used in the original 1993 study.
11 MRC-East refers to a theater in Southwest Asia, supported mainly by flights out of
East Coast US bases. Similarly, MRC-West refers to a theater in the Western Pacific
with deployments from the western US airbases. It was assumed that CRAF Stage II
would be authorized at the onset of MRC-East, with Stage III activated when MRC-West
deployments began.
12 Today, LRC has been replaced by IW (Irregular Warfare) or a vignette from the SSSP,
the Steady State Security Posture collection of lower intensity conflicts.
13 This assumption is further supported by the heavy demand for tankers to support fighter
deployments to the MRC theaters as well as the potential for supporting combat aircraft
employment during and following deployment.
14 The use rate is the average number of hours per day that an aircraft flies. Its estimate
depends on the amount of time it spends on the ground which, in turn, involves the
mission capable rate, which is a measure of reliability. It is used in the models to place
an upper limit on the amount of time an aircraft flies.
15 The MOG-value of each airbase is specified by aircraft type and is determined by the
material handling equipment (MHE) available, fuel availability, and the number of ser-
vice positions allotted to airlift (in-theater bases are often mostly occupied by fighter
aircraft). Large aircraft may require more than one service position.
16 It suffices to show the difference in the number of spaces available for small airlifters
as contrasted with larger ones. Detailed databases at the appropriate classification are
maintained by AMC.
17 MOG was usually the most constraining at en-route bases. En-route MOG constraints
may cause longer, less efficient routes to be used.
18 These models have subsequently evolved to more advanced versions, but it is important
to know how the transportation problem was approached in 1993.
19 MASS/AFM has now been replaced at AMC by the Air Mobility Operations Model
(AMOS).
20 The Office of Management and Budget (OMB) changes the discount factor every
year (US OMB 1992). For cost-effectiveness analyses of Federal programs, OMB
mandates the use of discounted dollars.
21 All costs prior to FY 1994 were considered “sunk” and ignored. Only future expendi-
tures were considered.
22 The SSC provided a review of the study prior to presentation to the higher-level DAB.
All offices represented in the DAB had representatives in the SSC.
23 As noted earlier in Figure 15.2, virtually all the required outsize cargo must be delivered
within a 30-day time period. This category of cargo differentiates best among the alterna-
tives because of its criticality to the early stages of battle. Excursions (not shown here) at
20 days showed no new insights, so 30-day deliveries were displayed in the study.
24 As discussed elsewhere in this book, using cost-effectiveness ratios to evaluate alterna-
tives is generally not appropriate and can in fact be very misleading without a budget
constraint or fixed level of effectiveness.
25 These are examples of the modified budget approach and the modified effectiveness
approach for cost-effectiveness comparisons in which neither cost nor effectiveness are
initially constrained to be equal for all alternatives. These are described by Francois
Melese (Melese 2010, 40–45).
26 The reader may wonder how oversize cargo such as trucks can influence the rate of
delivery of outsize cargo. The interaction is complex, involving a competition for space
Acquisition of the C-17 aircraft 399
aboard aircraft that does carry outsize (i.e. C-17s and C-5s) and between outsize cargo
and oversize cargo. If no oversize trucks can be carried by CRAF, those same trucks
begin to crowd out any outsize cargo that C-17s and C-5s would have otherwise carried,
reducing their outsize delivery capability.
27 This is an example of the modified effectiveness (“level the playing field”) approach
described by F. Melese in Chapter 4 of this book.
28 This is an example of the fixed budget approach described by Francois Melese (2010,
33–35).
29 The use of cost-effectiveness ratios can be misleading unless the analyses are con-
structed under the same budget or same effectiveness frameworks (Melese 2010,
17–20). In this case, the use of a common effectiveness framework, i.e. same MTM/D
confers meaning.
30 Note that the economic evaluation of alternatives (EEoA) approaches introduced in
Chapter 4 of this book make a clear distinction between the “life-cycle costs” or “price”
of an alternative, its operational effectiveness (schedule and performance), and the
resources (funding or budget) likely to be available for the overall program.
31 MASS/AFM has now been replaced at AMC by the Air Mobility Operations Model
(AMOS).
References
Baumol, W. J., J. C. Panzar, and R. D. Willig. 1983. “Contestable Markets: An Uprising in
the Theory of Industry Structure: Reply.” American Economic Review 73(3): 492–496.
Bexfield, J. M., T. A. Allen, and W. L. Greer. September 2001. C-17 COEA Case Study
(IDA Document D-2688). Alexandria, VA: Institute for Defense Analyses.
Greer, W. L., J. N. Bexfield, J. R. Nelson, D. A. Arthur, P. B. Buck, W. C. Devers, J. L.
Freeh, B. R. Harmon, D. Y. Lo, H. J. Manetti, A. Salemo, J. A. Schwartz, J. W. Stahl,
V. Suchorebrow, and D. M. Utech. December 1993. Cost and Operational Effectiveness
Analysis of the C-17 Program (IDA Report R-390). Alexandria, VA: Institute for Defense
Analyses.
Greer, W. L. et al. February 2009. (U) Study on Size and Mix of Airlift Force (IDA Paper
P-4425) (Classified study). Alexandria, VA: Institute for Defense Analyses. (Subject is
SECET//NOFORN.)
Joint Staff J4 Logistics. 1992. (U) Mobility Requirements Study (Classified study). Wash-
ington, DC: Joint Staff J4.
Melese, F. January 4, 2010. The Economic Evaluation of Alternatives (EEoA) (NPS-
GSBPP-10-002). Monterey, CA: Naval Postgraduate School.
Office of the Secretary of Defense (OSD). 1995. (U) Mobility Requirements Study Bottom
Up Review Update (Classified study). Washington, DC: OSD.
Office of the Secretary of Defense (OSD). 2010. Quadrennial Defense Review
Report. Washington, DC: OSD. Available at www.defense.gov/qdr/qdr%20as%
20of%2029jan10%201600.pdf [last accessed July 23, 2014].
——. 2005. (U) Mobility Capabilities Study (Classified study). Washington, DC: OSD.
——. 2010. (U) Mobility Capabilities and Requirements Study – 2016 (Classified study).
Washington, DC: OSD.
Secretary of the Air Force. February 4, 1994. US Air Force Cost and Planning Factors (AF
Instruction 65-503). Washington, DC: Secretary of the Air Force.
——. December 2003. Air Mobility Planning Factors (Air Force Pamphlet 10-1403). Scott
Air Force Base, IL: USAF Air Mobility Command. (December 2003 is the most recent
version of AFPAM 10-1403).
400 W. L. Greer
US Congressional Budget Office (CBO). October 2007. Issues Regarding the Current and
Future Use of the Civil Reserve Air Fleet. Washington, DC: US CBO.
US Government Accountability Office (GAO). July 11, 1994. Airlift Requirements:
Commercial Freighters Can Help Meet Requirements at Greatly Reduced Costs
(GAO/NSIAD 94-209). Washington, DC: US GAO. Available at www.gpo.gov/fdsys/
pkg/GAOREPORTS-NSIAD-94-209/content-detail.html [last accessed July 22, 2014].
US Office of Management and Budget (OMB). October 29, 1992. Guidelines and
Discount Rates for Benefit-Cost Analysis of Federal Programs (OMB Circular No.
A-94) (Revised annually). Washington, DC: OMB. Available at www.whitehouse.gov/
omb/circulars_default [last accessed July 22, 2014].
USAF Air Mobility Command (USAF). 1995. (U) Strategic Airlift Force Mix Analysis
(Classified study). Scott Air Force Base, IL: USAF Air Mobility Command.
US Senate Homeland Security and Governmental Affairs; Subcommittee on Federal
Financial Management, Government Information, Federal Services, and International
Security. July 13, 2010. Hearing on the Cost-Effectiveness of Procuring Weapon Sys-
tems in Excess of Requirements. Statement by Dr. William L. Greer, Assistant Director,
System Evaluation Division, Institute for Defense Analyses, Alexandria, VA.
16 Cost-effectiveness analysis of
autonomous aerial platforms
and communication payloads
Randall E. Everly, David C. Limmer,
and Cameron A. MacKenzie
16.1 Introduction
This chapter demonstrates the use of multi-criteria decision-making techniques
(discussed in Chapter 8) to analyze the cost-effectiveness of autonomous aerial
platforms and communication payloads for communication missions in the mili-
tary. We compare the cost-effectiveness of 17 aerial platforms and 9 communica-
tion payloads across three mission scenarios.
Unmanned aerial vehicles (UAVs) can help to supply the information
technology connectivity that is increasingly being demanded by technology
advancements on the battlefield. Autonomous vehicles are well suited for the
communication mission, which is often mundane and tedious. The advent of
lightweight construction materials, high energy-density lithium battery technol-
ogy, and more efficient microprocessors increase UAV capability (Department of
Defense 2013). The larger payload capacities and longer endurance of modern
UAVs combined with communication payloads that are more capable and effi-
cient and weigh less make them more suitable to the communications relay role
than their predecessors. UAVs are also more flexible than more permanent infras-
tructure such as relay and cellular towers, and they can be quickly repositioned to
support warfighters on the move.
Scrutiny on discretionary spending in public budgets, including defense
budgets, will continue to increase in many countries. Future acquisitions will
need to be executed thoughtfully with a clear consideration of the value and cost.
Although other studies (see Ferguson and Harbold 2001; Collier and Kacala 2008)
have analyzed the costs and benefits of UAVs for different missions, no study
has undertaken an extensive cost-effectiveness of the most modern UAVs and
other aerial platforms that can support communication requirements for military
operations in austere environments. This chapter seeks to remedy that deficiency
by performing a cost-effectiveness analysis of aerial platforms and communica-
tion payloads for use as a communication relay in support of distributed military
operations.
We conduct a multi-objective analysis of alternatives to compare the cost-
effectiveness of selected aerial platforms and communication payloads across
three scenarios. This chapter considers 13 different UAVs, 4 alternative aerial
platforms, and 9 communication payloads suitable for the communication relay
402 R. E. Everly, D. C. Limmer, C. A. MacKenzie
mission. UAVs range in size from the hand-launched Raven to the Triton with
its 130-foot wingspan, and communication payloads vary in weight from less
than a pound to over 250 pounds. We follow the approach detailed in Chapter 8
to analyze the cost-effectiveness of these alternatives. An objectives hierarchy
lists the desirable attributes for aerial platforms and communication systems, and
value measures and tradeoff weights lead to a numerical measure of effective-
ness (MoE). The annualized life-cycle costs (LCC) are estimated for the aerial
platforms, and acquisition costs are calculated for the communication payloads.
Selecting the most cost-effective alternative involves consideration of the costs
and MoEs. After selecting the most cost-effective aerial platforms and commu-
nication payloads for a specified scenario, we discuss whether the selected aerial
platforms are compatible with the communication payloads.
Section 16.2 reviews a few applications of cost-effectiveness analysis in defense
and discusses previous studies of UAV effectiveness. We briefly outline in
Section 16.3 the different UAVs, alternative aerial platforms, and communication
payloads. Section 16.4 discusses the cost-effectiveness model. An objectives hier-
archy is presented for the aerial platform and for the communication payload in
Section 16.5. After detailing the methodology for estimating costs for the aerial
platforms in Section 16.6, we analyze the cost-effectiveness of all the alterna-
tives for three different scenarios in Section 16.7, and offer concluding thoughts
in Section 16.8. We rely on subject matter experts to collect and analyze data for
several alternatives, and develop value measures and tradeoff weights based on
our own research, expertise, and experiences. Consequently, the results should
not be viewed as definitive, but the analysis does provide insights into select-
ing the best aerial platform for different communication relay missions. Defense
decision-makers could incorporate their own preferences within this framework to
determine the most cost-effective aerial platform and communication payload for
a given scenario.
M
v( j ) = wi vi (x i ( j )), (16.1)
i=1
406 R. E. Everly, D. C. Limmer, C. A. MacKenzie
where wi is the global tradeoff weight for attribute i , vi (·) is a value function for
attribute i , x i ( j ) is the level of attribute i for alternative j , and M is the total
number of attributes. The range for vi (·) is between 0 and 1, inclusive.
If the attribute is a numerical measure, we assess the value function using
marginal analysis by asking what the incremental or marginal change in value is
when the attribute is increased or decreased. We fit an exponential function to the
assessed values for attribute i using a least-squares approach (see Appendix 16.1
for more details). If the attribute is composed of categorical ratings, such as
“low,” “medium” and “high,” we directly assign values from 0 and 1 for each
category.
Local tradeoff weights are weights assessed for all attributes within a single
objective. We frequently use a swing weight procedure to assess local tradeoff
weights among attributes within an objective (von Winterfeldt and Edwards 1986).
The swing weight procedure assumes that all attributes are at the worst level, and
the decision-maker is asked which attribute he or she desires to move to the best
level. The selected attribute receives a score of 100 and the desirability of “swing-
ing” the other attributes from worst to best are assessed relative to the score of 100
for the most preferred attribute and a score of 0 when all the attributes are at their
worst level. Local tradeoff weights for each attribute are calculated by normalizing
the scores so that these weights sum to one. Multiplying the local weights by the
tradeoff weights for each objective higher up in the objectives Mhierarchy returns
global weights wi for i = 1, . . . , M for each attribute so that i=1 wi = 1.
After the MoE for each alternative is calculated via Equation (16.1) and Equa-
tion (A16.1.1) in the Appendix, we depict the MoE and the cost of each alternative
as a point on a two-dimensional chart. We analyze the cost-effectiveness of each
alternative that belongs to the efficient solution set, where the efficient set is
composed of alternatives that are not dominated by another alternative. Tradeoffs
between the different alternatives are discussed.1
Maximize effectiveness
Maximize effectiveness
30
Data points
Best−fit line
Annual O&S cost ($ millions)
25 Global Hawk
20
Triton
15
10
Reaper
Fire
5 Scout
Gray Eagle
Predator
0
0 50 100 150 200 250
Acquisition cost ($ millions)
Table 16.1 Global tradeoff weights for aerial platform for each scenario
1
Hummingbird Global Hawk
Reaper
Predator
0.8 Gray
Fire Scout Triton
Eagle
PTDS 74K UCAS−D
TIF−25K
0.6 Shadow
Cerberus
MoE
Tower Blackjack
RAID Tower
0.4
Raven
0.2 Wasp
T−Hawk
0
10−2 10−1 100 101 102
Annualized LCC ($ millions)
Table 16.2 Global tradeoff weights for communication payload for each scenario
1
Xiphos 6RU
0
100 101 102 103 104
Acquisition cost ($ thousands)
As can be seen from Figure 16.5, the efficient solution set comprises the Wave
Relay, Wave Relay Quad, Falcon III RF-7800W, Xiphos 1RU, and Xiphos 6RU.
(Figure 16.5 shows the acquisition cost for each radio on a logarithmic scale
to more clearly see the differences in costs.) If the decision-maker places more
importance on receiver sensitivity and traffic type relative to throughput, the Fal-
con III AN/PRC will replace the Falcon III RF-7800W as a member of the efficient
solution set. Regardless, either of the Falcons is only slightly more effective than
the Wave Relay Quad, and the Falcons cost US$18,000 more, and we prefer the
Wave Relay Quad to either of the Falcons. The Xiphos 1RU and 6RU are most
effective with MoEs of 0.79 and 0.93 respectively. Both radios offer high power
output, high throughput, and excellent scalability for multiple users. Their superior
capabilities are worth the additional cost.
Triton
0.8
UCAS−D
Reaper
Global Hawk
0.6
Predator Hummingbird
MoE
Gray Eagle
0.4 Fire Scout
PTDS 74K
0.2
0
10−2 10−1 100 101 102
Annualized LCC ($ millions)
If the UCAS-D’s LCC is closer to that of the Triton, the UCAS-D could be a
cost-effective alternative for the long-range scenario.
The Triton is annually US$23–25 million more expensive than the Reaper and
Predator, but the Triton’s MoE is 0.21 greater than that of the Reaper and 0.30
greater than that of the Predator. This gap in effectiveness is primarily due to
the Triton’s deicing capability. If icing conditions are not a factor, we prefer the
Reaper to the Triton. The Reaper costs US$1.6 million more than the Predator, but
its MoE is 0.09 greater than that of the Predator. All three UAVs will be evaluated
for compatibility with the most cost-effective communication payloads.
0.8
Falcon III AN/PRC WildCat II
0.6
MoE
0.4
0.2
0
100 101 102 103
Acquisition cost ($ thousands)
we are, the Wildcat II will be rated more effective than the Falcon III AN/PRC.
A decision-maker may prefer the lighter Wildcat II that consumes less power than
the Falcon III AN/PRC. The cost difference between the two communication pay-
loads is about US$25,000, which may not be an important discriminator when the
annual cost of the UAV is US$2 million or more.
0.8
Raven Hummingbird
Cerberus Fire Scout
Tower RAID Tower
0.6 PTDS 74K
TIF−25K Shadow
Reaper Global Hawk
MoE
Wasp
Blackjack
Predator UCAS−D
0.4 T−Hawk Gray Eagle
Triton
0.2
0
10−2 10−1 100 101 102
Annualized LCC ($ millions)
performance, and 0.11 for readiness. Within the flexibility objective, we believe
it is important to minimize the weight of the radio, followed by minimizing
power consumption. The attribute for weight has a local tradeoff weight of 0.38,
and power consumption has a local weight of 0.26. Mesh capability and traffic
type each have local weights of 0.18. For the performance objective, increasing
throughput is the most important attribute, to allow the tactical user to transmit
some video. The local weight for throughput is 0.45, and the both of the local
weights for power output and receiver sensitivity are 0.27.
Figure 16.9 displays the cost-effectiveness for the communication payloads.
The Wave Relay radio is the most effective and the least expensive alternative. The
Wave Relay’s MoE is the largest because of the radio’s small weight and relatively
high throughput. If a decision-maker places more importance on power consump-
tion and receiver sensitivity relative to throughput, the Ocelot will become more
effective than the Wave Relay although the former costs almost three times as
much as the Wave Relay. Both communication payloads weigh less than 0.2 lbs.
If a decision-maker is unwilling to trade off as much performance in favor of
flexibility than in our assessment, the WildCat II may be the most effective. The
WildCat II weighs 3.4 lbs, which likely makes it too heavy for the small UAVs.
0.8
Wave Relay
Ocelot WildCat II
0.6 Digital Xiphos 6RU
MoE
0.2
0
100 101 102 103 104
Acquisition cost ($ thousands)
though both communication payloads possess a small form factor, the payload
compartment of the Raven may not be able to support either radio. A previous
study (Menjivar 2012) tested a Raven with a Wave Relay payload. The Wave Relay
radio had to be taped onto the outside of the UAV instead of being secured inside
the payload compartment, which is not ideal for performance of the UAV or the
communication link.
One communication payload that would work with the Raven is AeroViron-
ment’s Digital Data Link. AeroVironment offers the Digital Data Link with the
Raven UAV from the factory. Digital Data Link’s price is positioned between the
Wave Relay and the Ocelot at US$5,000, but it has a slightly lower MoE than either
one at 0.55. However, compatibility between the Digital Data Link and the Raven
is assured, whereas the Ocelot and the Wave Relay units needs further testing to
ensure compatibility.
16.8 Conclusions
This chapter demonstrates the use of traditional multi-criteria decision-making
techniques to analyze the cost-effectiveness of aerial platforms and communi-
cation payloads for communication missions in the military. We compared the
cost-effectiveness of 17 aerial platforms and 9 communication payloads across
three mission scenarios.
The first scenario requires long endurance and high bandwidth capability to
complete a disaster relief mission. The most cost-effective aerial platforms are the
TIF-25K aerostat and the Predator UAV. The TIF-25K combines endurance mea-
sured in weeks and a high useful load with a moderate ceiling. Additionally, the
TIF-25K can utilize a tether to power its payloads from the ground. The Predator
is a very capable UAV platform with a relatively long endurance, long range, and
420 R. E. Everly, D. C. Limmer, C. A. MacKenzie
reasonable cost of ownership. The most cost-effective communication payloads
are the two configurations of the Oceus Networks Xiphos (1RU and 6RU) due to
their high throughput, power output, and excellent scalability. The Xiphos radios
can be relatively heavy (78 lbs to 276 lbs depending on configuration) and power
hungry (855 to 3,275 watts depending on configuration). We select either a TIF-
25K aerostat with a Xiphos 6RU connected to ground power or a Predator with a
Xiphos 1RU configuration.
In the second scenario, a long-range relay is needed to connect users across a
340 NM range, which requires a minimum altitude of 19,102 ft due to radio hori-
zon and UHF capability. The most cost-effective aerial platforms are the Predator,
Reaper, and Triton. Due to its UHF capability, sensitivity, and power output, the
Falcon III AN/PRC is the most cost-effective communication payload. Each of the
three aerial platforms could be the best choice, depending on the decision-maker’s
preference for price (Predator), better altitude and performance (Reaper), or out-
standing performance and deicing capability (Triton). The extra performance and
capability of the Triton comes with a much higher ownership cost than the Predator
or Reaper.
A covert, tactical situation where portability is preferred is the final scenario.
The most cost-effective aerial platform is the Raven, which combines man porta-
bility and adequate range and endurance with the lowest cost of any aerial platform
in this study. The Wave Relay is the most cost-effective communication pay-
load for this scenario because it weighs very little and possesses a relatively
high throughput for its small form factor. However, compatibility between the
Raven and Wave Relay cannot be confirmed at this time. The slightly less effec-
tive and more expensive Digital Data Link is compatible with the Raven and can
be an acceptable solution in the interim until more compatibility testing can be
completed between the Raven and the Wave Relay communication payload.
As with any analysis, our cost-effectiveness analysis relies on a number of
assumptions, which can motivate future research. We assess compatibility between
the most cost-effective aerial platforms and communication payloads within each
scenario based primarily on the manufacturer’s provided specifications. Actual
field testing would provide a proof of concept and help validate the findings of
this chapter. Since complete, transparent, and reliable cost data was not always
available, further research of the fully captured LCCs for UAVs, military towers,
and aerostats will enable a more accurate and detailed analysis. Other missions,
such as ISR, may present an interesting and useful subject for future research,
and other aerial platforms currently under development could be considered.
Finally, other decision-makers may have different objectives constraints, value
functions, and tradeoff weights, which would change the alternatives’ MoEs and
the cost-effectiveness analysis.
This chapter offers a framework for comparing the cost-effectiveness of dissim-
ilar aerial platforms and communication payloads across different mission sets.
Several militaries around the world currently operate UAVs, and their use for com-
munication will likely grow in the future. This research develops value functions
and weighting parameters that change based upon mission requirements and are
Cost-effectiveness of platforms and payloads 421
easily adapted to other mission sets including the traditional ISR mission. Our
analysis demonstrates that the cost-effectiveness of an alternative depends on the
mission requirements, and suggests the military should continue to purchase a
wide array of aerial platforms and communication payloads that can be used for
different missions.
Acknowledgments
We would like to thank Glen Cook and John Gibson of the Naval Postgraduate
School for their advice and feedback during this research and for sharing their
expertise in communication systems. Copies of the spreadsheets used to ana-
lyze the cost-effectiveness of the aerial platforms and communication payloads
can be downloaded and altered to meet a decision-maker’s own preferences from
Cameron MacKenzie’s web page at https://faculty.nps.edu/camacken/ under the
tab “Thesis Supervision.”
Appendix 16.1
Exponential value functions are used for the numerical attributes. The value of
attribute i at a given level x i can be calculated based on whether the decision-
maker prefers “more” or “less” of an attribute:
⎧
⎪ 1 − exp(a[x i − x min ]b )
⎪
⎨ if more is preferred
K
vi (x i ) = (A16.1.1)
⎪
⎩ 1 − exp(a[x max − x i ] )
⎪ b
if less is preferred,
K
Notes
1 See Chapter 4 for an alternative perspective.
2 The rank-sum only requires the decision-maker to rank the objectives rather than deter-
mining the exact tradeoff between multiple objectives. For three objectives, the most
important objective receives an unnormalized weight of 3, the second most important
receives a 2, and the least important receives a 1. Dividing each weight by the sum of
these weights returns normalized weights of 1/2, 1/3, and 1/6 for the first, second, and
third objectives, respectively.
422 R. E. Everly, D. C. Limmer, C. A. MacKenzie
References
AeroVironment. 2013. “Digital Data Link.” Available at https://www.avinc.com/uas/
small_uas/ddl/ [last accessed August 29, 2014].
American Institute of Aeronautics and Astronautics. 2013. “UAV Roundup 2013.”
Aerospace America, July/August. Available at www.aerospaceamerica.org/Documents/
AerospaceAmerica-PDFs-2013/July-August-2013/UAVRoundup2013t-AA-Jul-Aug
2013.pdf [last accessed September 15, 2014].
“Army Deploys 300th RAID Tower, Supporting Forward Base Protection by Per-
sistent Surveillance and Dissemination System PSDS2.” November 23, 2008.
Defense Update, November 23. Available at http://defense-update.com/features/2008/
november/231108_psds2_raid_sensors.html [last accessed August 29, 2014].
Buede, Dennis M., and Terry A. Bresnick. 1992. “Applications of Decision Analysis to the
Military Systems Acquisition Process.” Interfaces 22: 110–125.
Collier, Corey M., and Jeffrey C. Kacala. 2008. “A Cost-Effectiveness Analysis of Tacti-
cal Satellites, High-Altitude Long-Endurance Airships, and High and Medium Altitude
Unmanned Aerial Systems for ISR and Communication Missions.” Master’s thesis,
Naval Postgraduate School.
Department of Defense. 2013. “Unmanned Systems Integrated Roadmap
FY2013-2038.” Reference number 14-S-0553. Available at www.defense.gov/
pubs/DOD-USRM-2013.pdf [last accessed August 29, 2014].
Dobbing, Mary, and Chris Cole. 2014. Israel and the Drone Wars: Examining Israel’s
Production, Use and Proliferation of UAVs. Oxford, UK: Drone Wars UK. Available
at http://dronewarsuk.files.wordpress.com/2014/01/israel-and-the-drone-wars.pdf [last
accessed November 27, 2014].
Ewing Jr., Paul L., William Tarantino, and Gregory S. Parnell. 2006. “Use of Decision Anal-
ysis in the Army Base Realignment and Closure (BRAC) 2005 Military Value Analysis.”
Decision Analysis 3: 33–49.
Ferguson, Charles R., and Douglas A. Harbold. 2001. “High Altitude Long Endurance
(HALE) Platforms for Tactical Wireless Communications and Sensor Use in Military
Operations.” Master’s thesis, Naval Postgraduate School.
Fry, Jereed N., and Steven E. Tutaj. 2010. “A Business Case Analysis for the Vulture
Program.” MBA Professional Report, Naval Postgraduate School, Monterey, CA.
General Atomics Aeronautical. 2012a. “MQ-1C Gray Eagle.” Available at www.
ga-asi.com/products/aircraft/gray_eagle.php [last accessed August 29, 2014].
——. 2012b. “MQ-9 Reaper/Predator B.” Available at www.ga-asi.com/products/
aircraft/predator_b.php [last accessed August 29, 2014].
——. 2013. “MQ-1 Predator.” Available at www.ga-asi.com/products/aircraft/
predator.php [last accessed August 29, 2014].
Harris. 2012a. “AN/PRC-177G(V)1(C).” Available at http://rf.harris.com/capabilities/
tactical-radios-networking/an-prc-117g/default.asp [last accessed August 29, 2014].
——. 2012b. “RF-7800W-OU440.” Available at http://rf.harris.com/capabilities/
tactical-radios-networking/rf-7800w.asp [last accessed August 29, 2014].
Keeney, Ralph L. 1996. Value-Focused Thinking: A Path to Creative Decisionmaking.
Cambridge, MA: Harvard University Press.
Keeney, Ralph L., and Howard Raiffa. 1976. Decisions with Multiple Objectives: Prefer-
ences and Value Tradeoffs. New York: John Wiley & Sons.
Kirkwood, Craig W. 1997. Strategic Decision Making: Multiobjective Decision Analysis
with Spreadsheets. Belmont, CA: Brooks/Cole.
Cost-effectiveness of platforms and payloads 423
Lawler, Paul P. 2010. “Cost Implications of the Broad Area Maritime Surveillance
Unmanned Aircraft System for the Navy Flying Hour Program and Operation and
Maintenance Budget.” Master’s thesis, Naval Postgraduate School.
Menjivar, Jose D. 2012. “Bridging Operational and Strategic Communication Architectures
Integrating Small Unmanned Aircraft Systems as Airborne Tactical Communication
Vertical Nodes.” Master’s thesis, Naval Postgraduate School, Monterey, CA.
Naval Air Systems Command. 2014. “Aircraft and Weapons: Unmanned Combat Air Sys-
tem Demonstration.” Available at www.navair.navy.mil/index.cfm [last accessed May
24, 2014].
Nicholas, Ted G., and Rita Rossi. 2011. U.S. Unmanned Aircraft Systems, 2011. Fountain
Valley, CA: Data Search Associates.
Oceus Networks. 2014. “Oceus Networks Xiphos.” Available at www.oceusnetworks.
com/products/xiphos [last accessed August 29, 2014].
Parsons, Dan. 2013. “Worldwide, Drones Are in High Demand.” National
Defense Magazine, May. Available at www.nationaldefensemagazine.org/archive/
2013/May/Pages/Worldwide,DronesAreinHighDemand.aspx [last accessed September
15, 2014].
Persistent Systems. 2012. “Technology.” Available at www.persistentsystems.com/
pdf/WaveRelay_ WhitePaper_Technology_01.pdf [last accessed August 29, 2014].
——. 2014. “Quad Radio Router.” Available at www.persistentsystems.com/quad/ [last
accessed August 29, 2014].
Raven Aerostar. 2014. “TIF-25K teathered aerostat.” Available at http://raven
aerostar.com/solutions/aerostats/tif-25k [last accessed August 29, 2014].
Thiow Yong Dennis, Lim. 2007. “A Methodological Approach for Conducting a Business
Case Analysis (BCA) of the Global Observer Joint Capability Technology Demonstra-
tion (JCTD).” Master’s thesis, Naval Postgraduate School.
TrellisWare Technologies. 2104a. “TW-130 WildCat II.” Available at https://www.
trellisware.com/wp-content/uploads/TW-130-WildCat-II-Product-Bulletin.pdf [last
accessed March 5, 2014].
——. 2014b. “TW-600 Ocelot.” Available at https://www.trellisware.com/wp-content/
uploads/TW-600_Ocelot_Product_bulletin.pdf[last accessed March 9, 2014].
“US Army Aerostat-Based PTDS Provide IED Warning.” June 8, 2010. Defense Industry
Daily. Available at www.defenseindustrydaily.com/133M-to-Lockheed-Martin-for-US-
Army-Aerostat-based-Warning-System-05835/ [last accessed August 29, 2014].
von Winterfeldt, Detlof, and Ward Edwards. 1986. Decision Analysis and Behavioral
Research. Cambridge, UK: Cambridge University Press.
Yilmaz, Bahadir. 2013. “Cost and Effectiveness Analysis on Unmanned Aerial Vehicle
(UAV) Use at Border Security.” In SPIE Proceedings, edited by Edward M. Cara-
pezza. Paper presented at the Sensors, and Command, Control, Communications, and
Intelligence (C3I) Technologies for Homeland Security and Homeland Defense XII,
Baltimore, Maryland, April 29. doi:10.1117/12.2015914.
17 Time discounting in military
cost–benefit analysis
Jason Hansen and Jonathan Lipow
17.1 Introduction
The proverb “a bird in the hand is worth two in the bush” can be found in John
Ray’s 1670 A Handbook of Proverbs, and remains just as valid today. People place
higher value on what is certain—a bird in hand ready to eat—than they place on
the uncertain—two birds in the bush that may be ready to escape. Since the past
and present are by definition certain, uncertainty is really a characteristic of the
future. Hence, since people value certainty over uncertainty, they also value the
present over the future.
Careful accounting for the difference in value that people place on the present
as opposed to the future is an important requirement of any properly conducted
cost–benefit analysis (CBA). Costs incurred in the present tend to be weighted
more heavily than those that can be deferred to the future. Similarly, monetary
and other benefits received in the present tend to be weighted more heavily that
those that will only be enjoyed at some future date.
The standard method used in military CBA to adjust costs and benefits to reflect
their timing is known as “present value analysis” (PVA) or “time discounting.”
This chapter briefly reviews how to apply PVA to CBA, paying extra attention to
the most problematic aspect—the choice of an appropriate discount rate. The next
section offers a simple example of PVA applied to CBA. Section 17.3 discusses
the challenge of selecting an appropriate discount rate. Risk is explicitly incorpo-
rated into the discussion in Section 17.4. Concluding comments in Section 17.5
reinforce the importance of conducting sensitivity analysis on discount rates and
other key parameters in a military CBA.
FV = PV(1 + r )n , (17.1)
where the future value (FV) is the value n years in the future of a sum of money
whose present value (PV) is deposited in an account today earning interest at an
annual rate of r . Solving for PV, we find that
FV
PV = . (17.2)
(1 + r )n
1 100/ (1 + 10%) =1
90.90
2 100/ (1 + 10%)2 = 82.64
3 100/ (1 + 10%)3 = 75.19
426 J. Hansen, J. Lipow
Table 17.3 Comparative life-cycle costs
Alternatives
Alpha Beta
operation and maintenance costs are reversed: $100 million per year for the first
three years but only $50 million per year for the following three years.
Calculating the six year life-cycle costs (LCC) for each alternative, columns (2)
and (4) in Table 17.3 reveals that Alpha’s LCC price is $875 million, while Beta’s
is only $850 million. But is Beta really the “lowest price” alternative?
Comparing columns (2) and (4) in Table 17.3, we see that Beta costs less to
procure but more to operate initially. Overall, however, helicopter Beta requires a
lot more cash than Alpha over the first three years, and a lot less in the final three
years. One possible interpretation is that if Alpha was selected, there would be
excess funds freed up during those early years that could be deposited in a bank
earning interest. Funds earned in that account could then be used to defray the
higher costs that Alpha would incur in later years. The discounted present value
calculations displayed in columns (3) and (5) of Table 17.3 take these additional
funds into account. This suggests that (assuming a 10 percent interest rate) Alpha,
not Beta, may be the lowest price option.
r = α + ηλ, (17.3)
where α is the pure rate of time discount, λ is the rate at which per capita real con-
sumption is expected to grow over the long term, and η is the marginal elasticity
of utility with respect to real consumption. Each variable is interpreted carefully
below.
We can think of α as the portion of r that discounts future utility or well-
being (Heal 2007). It captures the reality that people are impatient, and “like
small children. . . hate waiting for anything,” as well as capturing the degree to
which people fear unexpected death or similar, equally unforeseen as undesirable,
events (Hansen and Lipow 2013). The larger α, the greater the desire to enjoy
benefits earlier rather than later.
We can think of ηλ as the portion of r that discounts future consumption bene-
fits (Heal 2007). The value of η captures the marginal utility of consumption, and
is also a measure of risk aversion. A value of η = 0 implies that a person is risk
neutral so that their marginal utility of consumption is constant over time. Hansen
and Lipow (2013) point out that:
where E is the expectations operator and γ = 1/(α + 1). In (17.4) the natural log
operator represents a common method in economics to represent an individual’s
utility, or well-being. The first operator on the right hand side of the equation
Time discounting in war and peace 429
Table 17.4 Parameter definitions
computes utility in period one based on the certain sum of money the citizen
retains. The second piece of the equation does the same thing for period two,
except that now the amount the citizen retains is uncertain so the expectations
operator must be applied. The sum across both periods estimates the citizen’s
utility.
To operationalize this simple model, consider the parameters in Table 17.4.
In Table 17.5, we assume values for operating costs of both alternative propul-
sion systems, g2 , in both states of nature that can occur in period two (war and
peace). We then plug values from Tables 17.4 and 17.5 into Equation (17.4) to
obtain the expected utility of the two alternatives reported in Table 17.5.
From Table 17.5, diesel has the lowest expected cost in period two. Since
cost and utility in the first period are identical for nuclear and diesel, a conven-
tional CBA focused exclusively on minimizing expected costs of propulsion would
recommend diesel-electric propulsion, regardless of the discount rate applied.
However, nuclear propulsion is the choice that maximizes the expected utility,
or well-being, of the representative citizen.
As the example above demonstrates, failure to take risk into account can eas-
ily lead to situations where CBA rejects the superior project or program while
endorsing the inferior. The most popular methodology for correcting this problem
is to modify the discount rate, r , by adding an additional “risk premium.” Breeden
(1979) shows the value of the risk premium (RP) that maximizes expected utility
is given by
1
RPx = − × cov(u c , X), (17.5)
E(u c )
where R is a proxy for wealth or consumption (often, simply the return on a stock
index such as the S&P 500), r is Ramsay’s discount rate (often proxied by rates
on government bonds), and βx is the covariance of X and R.
Consistent with this approach, the US government’s guidelines for CBA (US
OMB 1992, 12) acknowledges that “the absolute variability of a risky outcome can
be much less significant than its correlation with other significant determinants of
social welfare, such as real national income.” Surprisingly, however, Circular A-94
does not then go on to mandate use of the formula given in (17.5).
A possible explanation is that the formula works fairly well when the covariance
of X and R can be estimated using historical data, but is much more difficult to
apply when trying to evaluate new investments with little or no historical data. In
such circumstances, a popular graduate finance textbook explains that since “you
cannot hope to estimate the relative risk of assets with any precision. . . examine
the project from a variety of angles and look for clues to its riskiness” (Brealey
and Myers 1991, p. 200).
17.5 Conclusion
For the government analyst, “looking for clues” can be quite a challenge. For
most military CBAs, “clues” may not even exist.5 So what can be done to account
for risk in a military CBA? While far from perfect—indeed far from adequate—
Hansen and Lipow’s (2013) proposed approach has the virtue of being relatively
easy to implement in the United States. The analyst first identifies the discount
rate mandated by OMB Circular A-94 (US OMB 1992). This value for r , which
is based on government bond prices, is given by Appendix C of A-94 and is
updated every year. The analyst then makes a qualitative judgment as to whether
each flow of cost or benefit associated with an investment project is positively or
negatively correlated with per capita consumption, and documents this judgment
by writing a paragraph that explains his or her reasoning. The analyst then adds
1 percentage point to the mandated value of r used to discount benefits and costs
that are positively correlated with consumption, while subtracting 1 percentage
point from the value of r used to discount flows that are negatively correlated with
consumption.
While such an approach is likely superior to current practice, the resultant esti-
mates of the appropriate discount rates will still not precisely reflect the true social
Time discounting in war and peace 431
rate of discount. Barring a methodological breakthrough, the accounting for time
and risk in discounting are likely to remain an Achilles’ heel of military CBA.
This reinforces the value of sensitivity analysis, including Monte Carlo simulation
and other modeling efforts discussed in earlier chapters, to generate a thorough
understanding of the sensitivity of military CBA to key parameters.
Notes
1 The formula to use in PVA is
FV
PV = .
(1 + r )n
Here PV is the present value of a future value FV in n years, discounted into present
terms at the rate of discount r .
2 For simplicity of exposition, assume the helicopters are fully depreciated at the end of
this six year life-cycle and have no scrap value.
3 Evans (2005), for example, estimates η for 20 OECD countries and derives estimates that
range from 1.08 to 1.82, with values ranging from 1.15 to 1.45 for the United States.
4 Naval pedants would be correct to point out that, formally, a diesel-electric submarine
does not use steam propulsion, and hence can’t “steam.”
5 What exactly is the covariance of the S&P and the costs associated with building and
operating an aircraft carrier? That is not an easy question. What is the covariance of
the S&P and the benefits (or flows of effectiveness) that we expect to receive from that
aircraft carrier? That isn’t a question at all—it is a riddle worthy of the Sphinx.
References
Bazelon, C., and Smetters, K. 1999. “Discounting Inside the Washington DC Beltway.” The
Journal of Economic Perspectives, 13(4): 213–228.
Breeden, D. T. 1979. “An Intertemporal Asset Pricing Model with Stochastic Consumption
and Investment Opportunities.” Journal of Financial Economics, 7(3): 265–296.
Brealey, R. A., and Myers, S. C. 1991. Principles of Corporate Finance (4th edition).
New York: McGraw-Hill.
Evans, D. J. 2005. “The Elasticity of Marginal Utility of Consumption: Estimates for 20
OECD Countries.” Fiscal Studies, 26(2): 197–224.
Hansen, J., and Lipow, J. 2013. “Accounting for Systematic Risk in Benefit–Cost Analysis:
A Practical Approach.” Journal of Benefit-Cost Analysis, 4(3): 361–373.
Heal, G. 2007. “Discounting: A Review of the Basic Economics.” The University of
Chicago Law Review, 74(1): 59–77.
Ramsey, F. P. 1928. “A Mathematical Theory of Saving.” The Economic Journal, 38(152):
543–559.
Varian, H. R. 1992. Microeconomic Analysis (3rd edition). New York: Norton.
US Office of Management and Budget. 1992. Guidelines and Discount Rates for Benefit–
Cost Analysis of Federal Programs, OMB Circular No. A-94 Revised.
Weitzman, M. L. 2001. “Gamma Discounting.” American Economic Review, 91(1):
260–271.
Index