NextGen Air Transportation System
FAA's Metrics Can Be Used to Report on Status of Individual Programs, but Not of Overall NextGen Implementation or Outcomes
Gao ID: GAO-10-629 July 27, 2010
To prepare for forecasted air traffic growth, the Federal Aviation Administration (FAA), in partnership with other federal agencies and the aviation industry, is planning and implementing the Next Generation Air Transportation System (NextGen), a new satellite-based air traffic management system that will replace the current radar-based system and is expected to enhance the safety and capacity of the air transport system. GAO was asked to review FAA's metrics for (1) tracking the status of NextGen programs and the implementation of NextGen capabilities, the reliability of those metrics, and any limitations or gaps and (2) measuring the performance and outcomes of NextGen capabilities that are implemented and any limitations. GAO analyzed FAA program progress reports and associated metrics for monitoring. GAO also reviewed agency performance and accountability reports and discussed internal performance reporting methods with FAA officials.
FAA has metrics that allow it to monitor the progress of its programs for acquiring software and hardware. These metrics include Earned Value Management (EVM) measurements that show how well a program is meeting its planned cost and schedule targets for system development. Previous GAO reports have identified issues with FAA's implementation of EVM, which continue to affect the accuracy and reliability of some of FAA's program status reports. For example, for one acquisition program, FAA implemented EVM metrics only for the contractor's performance and not for the government's. As a result, the EVM data did not pick up delays that occurred after the contractor delivered the system and the EVM system did not provide early warnings of delays and potential cost overruns. In addition, GAO's previous work has shown that FAA is not able to report on how slippage in one program's schedule or budget will ultimately affect the implementation of other NextGen acquisition programs or operational capabilities whose progress depends on the completion of the first program. GAO has made recommendations to address these issues, which FAA and the Department of Transportation have begun to implement. FAA has also designated specific positions within the NextGen Integration and Implementation Office-known as solution set coordinators-to monitor and track progress toward implementing a portfolio of operational improvements into the national airspace system. However, the role of the coordinators and the process for resolving any disputes across FAA lines of business have not been clearly defined or delineated and it is uncertain whether the processes in place in this portfolio management structure will strengthen oversight and create a greater likelihood that required activities are completed on time. FAA has broad goals for NextGen as a whole, such as increasing capacity and reducing noise and emissions, but has not yet developed specific goals and outcome-based performance metrics to track the impact of and benefits realized from the entire NextGen endeavor. The agency has multiple efforts underway to develop such metrics: FAA's Air Traffic Organization (ATO), which manages the air traffic control system, has started to compile and review a set of metrics for measuring outcomes and performance associated with NextGen improvements. These metrics are likely to measure such things as the extent to which improvements increase throughput at airports, reduce emissions, and reduce flight times, but they are in the early stages of development. Recently, FAA also committed to developing performance metrics with industry, but it has no timeline or action plan for completing this effort. Separately, the Joint Planning and Development Office (JPDO), which is responsible for the long-term planning for NextGen and partnering with other federal agencies, has been working to develop a list of potential metrics, which range from fuel consumed per distance flown to curb-to-curb travel time. Without specific goals and metrics for the performance of NextGen as a whole, together with a timeline and action plan for implementation, it is not clear whether NextGen technologies, systems, and capabilities will achieve desired outcomes and be completed within the planned time frames. The FAA Administrator should clarify dispute resolution processes within FAA's portfolio management structure, and develop a timeline and action plan to agree with stakeholders on a list of specific goals and outcome-based performance metrics for NextGen. DOT agreed to consider GAO's recommendations and provided technical comments that GAO incorporated as appropriate.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Gerald L. Dillingham
Team:
Government Accountability Office: Physical Infrastructure
Phone:
(202) 512-4803
GAO-10-629, NextGen Air Transportation System: FAA's Metrics Can Be Used to Report on Status of Individual Programs, but Not of Overall NextGen Implementation or Outcomes
This is the accessible text file for GAO report number GAO-10-629
entitled 'NextGen Air Transportation System: FAA's Metrics Can Be Used
to Report on Status of Individual Programs, but Not of Overall NextGen
Implementation or Outcomes' which was released on July 27, 2010.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as
part of a longer term project to improve GAO products' accessibility.
Every attempt has been made to maintain the structural and data
integrity of the original printed product. Accessibility features,
such as text descriptions of tables, consecutively numbered footnotes
placed at the end of the file, and the text of agency comment letters,
are provided but may not exactly duplicate the presentation or format
of the printed version. The portable document format (PDF) file is an
exact electronic replica of the printed version. We welcome your
feedback. Please E-mail your comments regarding the contents or
accessibility features of this document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to Congressional Committees:
United States Government Accountability Office:
GAO:
July 2010:
NextGen Air Transportation System:
FAA's Metrics Can Be Used to Report on Status of Individual Programs,
but Not of Overall NextGen Implementation or Outcomes:
GAO-10-629:
GAO Highlights:
Highlights of GAO-10-629, a report to congressional committees.
Why GAO Did This Study:
To prepare for forecasted air traffic growth, the Federal Aviation
Administration (FAA), in partnership with other federal agencies and
the aviation industry, is planning and implementing the Next
Generation Air Transportation System (NextGen), a new satellite-based
air traffic management system that will replace the current radar-
based system and is expected to enhance the safety and capacity of the
air transport system.
GAO was asked to review FAA‘s metrics for (1) tracking the status of
NextGen programs and the implementation of NextGen capabilities, the
reliability of those metrics, and any limitations or gaps and (2)
measuring the performance and outcomes of NextGen capabilities that
are implemented and any limitations. GAO analyzed FAA program progress
reports and associated metrics for monitoring. GAO also reviewed
agency performance and accountability reports and discussed internal
performance reporting methods with FAA officials.
What GAO Found:
FAA has metrics that allow it to monitor the progress of its programs
for acquiring software and hardware. These metrics include Earned
Value Management (EVM) measurements that show how well a program is
meeting its planned cost and schedule targets for system development.
Previous GAO reports have identified issues with FAA‘s implementation
of EVM, which continue to affect the accuracy and reliability of some
of FAA‘s program status reports. For example, for one acquisition
program, FAA implemented EVM metrics only for the contractor‘s
performance and not for the government‘s. As a result, the EVM data
did not pick up delays that occurred after the contractor delivered
the system and the EVM system did not provide early warnings of delays
and potential cost overruns. In addition, GAO‘s previous work has
shown that FAA is not able to report on how slippage in one program‘s
schedule or budget will ultimately affect the implementation of other
NextGen acquisition programs or operational capabilities whose
progress depends on the completion of the first program. GAO has made
recommendations to address these issues, which FAA and the Department
of Transportation have begun to implement. FAA has also designated
specific positions within the NextGen Integration and Implementation
Office–known as solution set coordinators–to monitor and track
progress toward implementing a portfolio of operational improvements
into the national airspace system. However, the role of the
coordinators and the process for resolving any disputes across FAA
lines of business have not been clearly defined or delineated and it
is uncertain whether the processes in place in this portfolio
management structure will strengthen oversight and create a greater
likelihood that required activities are completed on time.
FAA has broad goals for NextGen as a whole, such as increasing
capacity and reducing noise and emissions, but has not yet developed
specific goals and outcome-based performance metrics to track the
impact of and benefits realized from the entire NextGen endeavor. The
agency has multiple efforts underway to develop such metrics: FAA‘s
Air Traffic Organization (ATO), which manages the air traffic control
system, has started to compile and review a set of metrics for
measuring outcomes and performance associated with NextGen
improvements. These metrics are likely to measure such things as the
extent to which improvements increase throughput at airports, reduce
emissions, and reduce flight times, but they are in the early stages
of development. Recently, FAA also committed to developing performance
metrics with industry, but it has no timeline or action plan for
completing this effort. Separately, the Joint Planning and Development
Office (JPDO), which is responsible for the long-term planning for
NextGen and partnering with other federal agencies, has been working
to develop a list of potential metrics, which range from fuel consumed
per distance flown to curb-to-curb travel time. Without specific goals
and metrics for the performance of NextGen as a whole, together with a
timeline and action plan for implementation, it is not clear whether
NextGen technologies, systems, and capabilities will achieve desired
outcomes and be completed within the planned time frames.
What GAO Recommends:
The FAA Administrator should clarify dispute resolution processes
within FAA‘s portfolio management structure, and develop a timeline
and action plan to agree with stakeholders on a list of specific goals
and outcome-based performance metrics for NextGen. DOT agreed to
consider GAO‘s recommendations and provided technical comments that
GAO incorporated as appropriate.
View [hyperlink, http://www.gao.gov/products/GAO-10-629] or key
components. For more information, contact Gerald Dillingham, Ph.D., at
(202) 512-2834 or dillinghamg@gao.gov.
[End of section]
Contents:
Letter:
Background:
FAA Has Metrics to Report on Program Status but Does Not Have Metrics
to Measure Overall Implementation of NextGen Capabilities:
Metrics Have Yet to Be Developed to Measure the Performance of NextGen
Improvements in Relation to Specific NextGen Goals, but Some
Performance Metrics Are Available for Specific Programs:
Conclusions:
Recommendations for Executive Action:
Agency Comments:
Appendix I: Objectives, Scope, and Methodology:
Appendix II: Further Information on the ERAM Program:
Appendix III: GAO Contact and Staff Acknowledgments:
Tables:
Table 1: Status of NextGen's Six Transformational Programs:
Table 2: Description of FAA's NextGen Solution Sets:
Table 3: Selection of JPDO's Proposed Performance Metrics:
Table 4: Selection of FAA Proposed Performance Metrics:
Table 5: Alignment of NextGen Activities with Existing Flight Plan
Metrics:
Table 6: ERAM Status for Achieving Key Program Milestones:
Figures:
Figure 1: NextGen Governmental Organizational Structure:
Figure 2: Program Assessment Report for ERAM as of March 2010:
Figure 3: Program Assessment Report for ADS-B as of March 2010:
Figure 4: Implementation Status of Prior Selected GAO Recommendations
to FAA and DOT as of June 2010:
Figure 5: Current Phase of ERAM Testing at FAA's En Route Centers as
of May 2010:
[End of section]
United States Government Accountability Office:
Washington, DC 20548:
July 27, 2010:
Congressional Committees:
Air traffic is growing, and with it, congestion and flight delays,
which can cause significant economic losses. The Federal Aviation
Administration (FAA) predicts that, by 2025, the number of passengers
will increase 57 percent--from about 700 million to about 1.1 billion
per year--and the number of flights from about 80,000 to more than
95,000 every 24 hours. Today's air transportation system will be
strained to meet these air traffic demands, especially on some routes
to and from major cities and hubs, but improvements to the national
airspace system can mitigate the anticipated increase in flight delays
and any resulting decrease in economic productivity. Accordingly, FAA
and other federal agencies have worked in partnership to develop a
plan for the Next Generation Air Transportation System (NextGen).
[Footnote 1] NextGen involves every aspect of air transportation, from
arrival at the airport to departure from the destination airport.
NextGen requires the acquisition of new integrated systems (software
and hardware), flight procedures, aircraft performance capabilities,
and supporting infrastructure to transform the current air
transportation system into one that uses satellite-based surveillance
and navigation and network-centric operations.[Footnote 2] These
acquisition programs and their associated improvements are intended to
increase the efficiency and capacity of the air transportation system
while maintaining its safety so that it can accommodate anticipated
future growth. The initial planning for NextGen, starting with Vision
100[Footnote 3] in 2003, focused on implementing improvements through
2025. More recently, FAA has emphasized improvements that can be
implemented in the near term and midterm, defined as between 2012 and
2018.[Footnote 4] At the same time, stakeholders and Members of
Congress have expressed concerns about the pace of FAA's
implementation of NextGen, citing the schedule delays that plagued
FAA's previous air traffic control modernization efforts and led GAO
to place air traffic modernization on its High-Risk List between 1996
through 2008. Additionally, given increasing demands for a more
effective, transparent, and accountable federal government, it is
important that federal agencies establish meaningful goals for
improving performance, monitor progress in achieving their goals, and
use information about performance to make decisions that can improve
results. Metrics are important for demonstrating progress toward
achieving goals and providing information on which to base
organizational and management decisions.
In light of the scale and complexity of NextGen implementation and
concerns about past modernization efforts and the pace of
implementation, you asked us to review the metrics and process FAA
uses to monitor the status of NextGen implementation. To do so, we
examined (1) FAA's metrics for tracking the status of NextGen
acquisition programs and the implementation of NextGen capabilities,
the reliability of these metrics and the data underlying them, and any
limitations or gaps in FAA's efforts to track the status of NextGen
implementation; and (2) how FAA currently measures the performance of
NextGen programs and capabilities, FAA's progress in developing a full
suite of metrics to measure the outcomes and performance of NextGen
capabilities once implemented, and any limitations or gaps in FAA's
approach to developing these metrics.
To determine what programmatic metrics FAA has available for
monitoring NextGen programs and programs critical to NextGen
implementation, we reviewed reports used to justify programs prior to
investment, progress reports submitted by program managers, and
reports based on the FAA database that houses program information. We
also interviewed FAA acquisition and finance officials and selected
NextGen program managers to understand how program managers develop
and report their metrics to internal and external stakeholders and to
gain an understanding of the database that houses this information. To
learn how FAA plans to monitor and measure progress toward
implementing NextGen operational capabilities--beyond the status of
acquisition programs--we reviewed documents that outline FAA's
solution set organization and management approach, and interviewed
officials involved in coordinating and managing solution sets.
[Footnote 5] To determine the reliability of these programmatic
metrics and to analyze the extent of any gaps or limitations, we
reviewed past GAO reports on FAA's acquisition process and the
reliability of the data FAA uses to develop its metrics, as well as
the implementation status of prior recommendations. We reviewed
program and process reviews from FAA's acquisition offices to identify
key areas of FAA's internal oversight focus and key findings reached
in such reviews about FAA's acquisition procedures and policies. To
determine FAA's progress in developing metrics for measuring the
outcomes of NextGen improvements, we first reviewed how FAA currently
reports on its performance, both internally and externally, and how
information on the performance of specific NextGen improvements is
incorporated into those metrics. We reviewed FAA's performance and
accountability reports and discussed internal performance reporting
methods with relevant FAA officials. Specifically, we reviewed FAA's
Flight Plan, Performance and Accountability Report, NextGen
Implementation Plans published in 2009 and 2010, Enterprise
Architecture, and reports to the Office of Management and Budget (OMB)
(known as "Exhibit 300" reports).[Footnote 6] To understand FAA's
approach and progress toward developing a suite of NextGen metrics, we
interviewed FAA officials with responsibilities for NextGen planning
and implementation, particularly officials within the Air Traffic
Organization (ATO) and the Joint Planning and Development Office
(JPDO) responsible for modeling NextGen benefits and developing
NextGen performance metrics. To evaluate metrics that FAA is
considering, we compared proposed metrics with key attributes of
successful performance metrics that we identified in past GAO
work.[Footnote 7] We also interviewed several key stakeholders for
NextGen, including representatives from airlines, equipment
manufacturers, federal partner agencies, and the air traffic
controllers union to get their views on the metrics they deem most
appropriate to measure the performance of NextGen.
In this report, we discuss two types of metrics: programmatic and
performance. Programmatic metrics are used to track the progress of
programs or capabilities, and include such things as time, cost, and
schedule. For instance, some programs use earned value management
(EVM), a technique for showing how well a program is meeting cost and
schedule milestones.[Footnote 8] In contrast, performance metrics
measure the impact or results of a program or activity once it is
implemented relative to desired outcomes or goals, such as reductions
in delays or fuel consumption and increased throughput at an airport.
[Footnote 9] Effective performance metrics require baselining, or
determining the current status of whatever is being measured, so that
targets can then be set. These metrics will, if developed well,
measure how well something is progressing toward its intended target.
We conducted this work from June 2009 through July 2010 in accordance
with generally accepted government auditing standards. Those standards
require that we plan and perform the work to obtain sufficient,
appropriate evidence to provide a reasonable basis for our findings
and conclusions based on our audit objectives. We believe that the
evidence obtained provides a reasonable basis for our findings and
conclusions based on our audit objectives. Appendix I contains more
detail information on our objectives, scope, and methodology.
Background:
FAA is currently tracking over 30 acquisitions related to
modernization and improvement of the national airspace. While not all
of these acquisition programs are considered to be "NextGen" programs,
several are instrumental to NextGen implementation. Examples of
instrumental programs include the En Route Automation Modernization
(ERAM); the Airport Surface Detection Equipment-Model X (ASDE-X),
which increases runway safety and airport efficiency by putting in
place tools to improve operations, surveillance, and data sharing on
the airport surface (e.g., runways and taxiways); and the Wide Area
Augmentation System (WAAS), which provides aircraft more accurate
position information for more direct flight paths and precision
approaches to airports. In particular, ERAM, a new system
architecture, will replace the current En Route computer system and
its backup and is considered to be the backbone that will support
NextGen. ERAM is meant to provide all of today's functionality and add
new capabilities needed to support the transformation to NextGen.
Besides these three instrumental programs, FAA has identified six
major acquisition programs that it considers to be transformational
NextGen programs, as follows:
* Automatic Dependent Surveillance Broadcast (ADS-B) is a satellite-
based information broadcasting system that is designed, along with GPS-
based navigation technologies, to enable more precise control of
aircraft during en route flight, approach, and descent.
* System-Wide Information Management (SWIM) is the information
management architecture for the national airspace system, acting as
its "World Wide Web." SWIM will manage surveillance, weather, and
flight data, as well as aeronautical and system status information,
and will provide the information securely to users.
* NextGen Data Communications (Data Comm) is intended to provide a
digital communications link for two-way exchanges between controllers
and flight crews for air traffic control clearances, instructions,
advisories, flight crew requests, and reports.
* NextGen Network Enabled Weather (NNEW) is planned to serve as the
core of the NextGen weather support services and provide a common
weather picture across the national airspace system.
* National Airspace Voice Switch (NVS) is to replace existing switches
and provide the foundation for all air-to-ground and ground-to-ground
voice communications in the future air traffic control environment.
* Collaborative Air Traffic Management Technologies (CATMT)
encompasses the development of systems to distribute and manage
aeronautical information, manage airspace reservations, and manage
flight information from preflight to postflight analysis.
Acquisition programs are overseen by program offices within the Air
Traffic Organization (ATO) and headed by program managers who are
responsible for gathering and reporting programmatic data to FAA's
acquisition tracking database, known as the Simplified Program
Information Reporting and Evaluation (SPIRE) tool, which FAA uses to
track and report the progress of all approved acquisitions toward its
schedule and cost performance targets. Detailed cost, schedule, and
EVM metrics are developed, and reporting begins after the Joint
Resources Council (JRC) approves a program for funding.[Footnote 10]
As table 1 shows, three NextGen transformational programs, ADS-B,
SWIM, and CATM, have received final investment approval and are being
reported with EVM and associated metrics, while the three remaining
programs, DataComm, Voice Switch, and Weather, have not yet received
such approval. For the time being, their progress is being tracked
against schedule milestones.
Table 1: Status of NextGen's Six Transformational Programs:
Program: ADS-B;
Status of latest investment decision: Final investment;
Date of decision: June 2006;
Next major milestone: Under deployment at several sites (at and near
airports); full deployment of ground-based transceivers expected in FY
2013.
Program: SWIM;
Status of latest investment decision: Final investment;
Date of decision: June 2009;
Next major milestone: Deployment of capabilities expected to start in
FY 2010.
Program: Collaborative Air Traffic Management Technologies (CATMT);
Status of latest investment decision: Final investment;
Date of decision: Sept 2008 (Work package 2) Jan. 2010 (Work package
3);
Next major milestone: Integration of weather data in 2011.
Program: DataComm;
Status of latest investment decision: Final investment decision
Segment 1;
Date of decision: September 2011;
Next major milestone: Deployment schedule not baselined until final
investment decision, expected in FY 2016.
Program: NAS Voice Switch (NVS);
Status of latest investment decision: Final investment decision
Segment 1;
Date of decision: August 2012;
Next major milestone: Market survey of potential contractors scheduled
for FY 2010.
Program: NextGen Weather;
Status of latest investment decision: Initial and final investment
decision have yet to be scheduled;
Date of decision: FY 2012-2015;
Next major milestone: [TBD].
Source: FAA.
[End of table]
ATO's NextGen and Operations Planning Office is leading near-term (now
through 2015) and midterm (2015 through 2018) NextGen planning and
implementation efforts.[Footnote 11] These efforts are guided by the
NextGen Implementation Plan, which identifies the NextGen capabilities
that are to be implemented between 2012 and 2018.[Footnote 12] NextGen
capabilities are defined in portfolios of related operational
improvements called solution sets, which together, will bring about
the midterm system. FAA currently is managing seven solution sets,
described in table 2.
Table 2: Description of FAA's NextGen Solution Sets:
Solution set: Initiate Trajectory-Based Operations;
Description of solution set: This will lead to a shift from current
"clearance-based" to "trajectory-based" air traffic control and will
enable aircraft to fly negotiated flight paths that take both
controller and pilot preferences and optimal airspace system
performance into consideration.
Solution set: Increase Arrivals and Departures at High-Density;
Description of solution set: Airports will improve arrival and
departure capacity for airports with heavily used airspace.
Solution set: Increase Flexibility in the Terminal Environment;
Description of solution set: This will increase access and help manage
the separation of aircraft in and around airports and allow for
improved management of aircraft on the airport surface, as well as
improved access to runways in low visibility.
Solution set: Improve Collaborative Air Traffic Management;
Description of solution set: This will support a more flexible air
traffic system capable of adjustments to routings or altitude to match
airspace and airport capacity, and accommodate controller and pilot
preferences to the maximum extent possible.
Solution set: Reduce Weather Impact;
Description of solution set: This will support integration of a broad
range of weather information into air traffic control decision-making.
Solution set: Improve Safety, Security and Environmental Performance;
Description of solution set: This will deploy an automated system to
identify airborne security threats and communicate that information to
the appropriate agency.
Solution set: Transform Facilities;
Description of solution set: This will support planning for future
NextGen facilities.[A]
Source: FAA.
[A] This definition is limited to activities funded for fiscal year
2010.
[End of table]
Each of the solution sets includes or will include numerous
acquisition programs and a variety of other activities that will be
carried out across offices within ATO, such as the Office of
Operations, and several other lines of business across FAA, such as
the Office of Aviation Safety, the Office of Airports, and others. For
example, implementing the solution set Increase Flexibility in the
Terminal Environment requires that ADS-B, DataComm, SWIM, ERAM, NNEW
and other programs be implemented; flight procedures be developed by
the Flight Procedure Standards Branch; and safety analyses be
conducted by ATO, and requirements and standards be developed by the
Flight Technology and Procedures Division, among numerous other
actions.
FAA has created a new position - solution set coordinator - to
coordinate and manage the implementation of each solution set across
the agency. While solution set coordinators manage the day-to-day
implementation of solution sets, the NextGen Management Board, which
includes the heads of ATO and the key agency lines of business,
oversees NextGen implementation efforts within FAA and has the
authority to force timely resolution of emerging NextGen
implementation issues.[Footnote 13] The Board's role is to measure the
progress of deployments and of key activities that support decision-
making; ensure essential resources are available, including
reprioritizing resources as necessary; issue policies and guidance;
and identify officials--like program managers--within organizations
who will be accountable for delivering system changes.
JPDO is responsible for the long-term planning and development for
NextGen and as such, is involved in modeling the costs, benefits, and
risks associated with alternative scenarios of NextGen implementation
over the long term. Originally chartered in Vision 100[Footnote 14] to
plan and coordinate the transition to NextGen, JPDO began to focus on
planning for NextGen beyond 2018 after organizational changes were
made in May 2008.[Footnote 15] JPDO has recently undergone a
leadership change, has been repositioned within FAA's organization,
and now reports directly to the FAA Deputy Administrator-who is the
FAA executive in charge of NextGen. JPDO is also responsible for
ensuring and fostering interagency coordination and collaboration and
is closely tied to the Senior Policy Committee-the governing body for
NextGen chaired by the Secretary of Transportation and made up of
cabinet-level officials from the partner agencies. Figure 1 shows the
current governmental organization surrounding NextGen activities.
Figure 1: NextGen Governmental Organizational Structure:
[Refer to PDF for image: illustration of Organizational Structure]
Top level:
Secretary of Transportation:
* FAA Administrator;
* FAA Deputy Administrator;
* NextGen Management Board.
Second level:
* Senior Policy Committee (Secretary of Transportation, chairman):
reports to Secretary of Transportation;
* JPDO board: reports to Senior Policy Committee;
* NextGen liaison to Secretary of Transportation and JPDO Director:
connected to each of the top level entities and JPDO Board;
* Joint Planning and Development Office (JPDO): connected to NextGen
liaison.
Third level:
Partner agencies: connected to all second level entities:
* Department of Commerce (National Oceanic and Atmospheric
Administration);
* Department of Defense;
* Department of Homeland Security;
* National Aeronautics and Space Administration;
* White House Office of Science and Technology Policy.
Fourth level, connected to NextGen Management Board: ATO offices with
NextGen responsibilities:
Chief Operating Officer, Air Traffic Organization:
* Financial Services;
* NextGen and Operations Planning;
* Safety;
* Strategy and Performance;
* Operations:
- En Route and Oceanic;
- System Operations;
- Terminal;
- Technical Operations.
Fourth level, connected to NextGen Management Board: Other lines of
business with NextGen responsibilities:
* Airports;
* Aviation Safety;
* Financial Services/Chief Financial Officer;
* Information Services/Chief Information Officer;
* International;
* Policy, Planning, and Environment;
* Regions and Center Operations.
Sources: FAA and JPDO.
[End of figure]
The participation of industry and other stakeholders is critically
important to the success of NextGen's implementation. Numerous venues
exist for stakeholders to participate, although our prior work has
shown that industry stakeholders, the air traffic controller and
aviation safety specialist unions, and the partner agencies have
participated unevenly for a variety of reasons.[Footnote 16] The
NextGen Institute, a JPDO mechanism designed to involve private-sector
expertise, tools, and facilities in the development and implementation
of NextGen, has been leaderless since March 2010 when the head of the
NextGen Institute resigned, and a replacement has yet to be named.
Last year, FAA requested that RTCA create a NextGen Midterm
Implementation Task Force to reach consensus within the aviation
community on the operational improvements that can be implemented by
2018 and would be most beneficial to users.[Footnote 17] The Task
Force focused on maximizing benefits in the near term and paid
particular attention to aligning its recommendations with how aircraft
operators decide to invest in aircraft equipment. On September 9,
2009, the Task Force issued its final report, which contained a list
of recommendations for FAA. FAA agreed with the Task Force
recommendations and worked with the Task Force to incorporate and
address their recommendations in FAA's plans. In January 2010, FAA
released its response to the task force, which outlines FAA's specific
responses, including an action plan detailing when certain tasks will
be completed. A key recommendation of the Task Force was for FAA to
work with industry to develop performance metrics to show the progress
and benefits of NextGen.
FAA Has Metrics to Report on Program Status but Does Not Have Metrics
to Measure Overall Implementation of NextGen Capabilities:
FAA Uses Programmatic Metrics to Provide Updates on Program Status,
but Additional Information and Context Could Help Observers and
Overseers Understand Problems:
FAA's SPIRE tool can organize information provided by program managers
into various reports that give high-level indications of a program's
status. For example, figures 2 and 3 show the program assessment
reports for ERAM and ADS-B, respectively. These reports use a color-
coded chart to summarize program managers' assessments of performance
for cost and schedule indicators, with green signifying that the
program is on target, yellow that there are potential issues with
meeting targets, and red that there is significant risk that the
target will not be met. These reports also contain space for the
program managers to clarify the status of the program, implementation
concerns, and information on any other issues deemed necessary to
highlight, such as problems, issues, and corrective actions for
ensuring that milestones and costs are maintained within the
established targets.
Figure 2: Program Assessment Report for ERAM as of June 2010:
[Refer to PDF for image: illustration of report]
Key:
Metric indicates no issues: Green;
Metric indicates some issues: Yellow;
Metric indicates serious issues: Red.
Program Phase: Deployment;
June 2010:
Program assessment:
Financial: Metric indicates no issues;
Schedule: Metric indicates serious issues;
Technical: Metric indicates some issues;
Resources: Metric indicates no issues;
External interest: Metric indicates some issues;
Program manager: Metric indicates some issues.
Supporting indicators:
Financial: Cost Performance Index (CPI);
Measure of cost performance: 1.02.
Financial: To Complete Performance Index (TCPI);
Measure of cost performance: TCPI = 0.85.
Financial: Cost Variance At Completion (CVAC);
Measure of cost performance: Metric indicates no issues.
Financial: Obligation rate;
Measure of cost performance: [Symbol].
Schedule: Schedule Performance Index (SPI);
Measure of schedule performance: 1.00.
Schedule: Schedule Variance At Completion (SVAC);
Measure of schedule performance: Metric indicates serious issues.
Schedule: APB standard level 1 milestones;
Measure of schedule performance: Metric indicates serious issues.
Schedule: Fiscal year schedule goals;
Measure of schedule performance: Metric indicates serious issues.
Technical: Requirements stability;
Measure of cost performance: Metric indicates no issues.
Technical: System defects;
Measure of cost performance: Metric indicates some issues.
Technical: Test results;
Measure of cost performance: Metric indicates some issues.
Technical: Deployment;
Measure of cost performance: Metric indicates some issues.
Technical: Cumulative mitigated risk impact;
Measure of cost performance: Metric indicates serious issues.
Technical: Performance Variance At Completion (PVAC);
Measure of cost performance: Metric indicates no issues.
Resources: Funding;
Measure of appropriate funding: Metric indicates no issues.
External interest: Under review by DOT, IG, GAO, or other;
External review activity: Metric indicates some issues.
Program manager comments:
The approved 2003 schedule for the ERAM program was to reach full
operational status at the first operational site, the Salt Lake City
(ZLC) Air Route Traffic Control Center by the end of December 2009.
However, continuing resolution of PRs needed for the key site to
transition to continuous operations continues to delay this milestone.
These PRs could not have been discovered in the Technical Center
Laboratory environment and only became evident when the system was
tested operationally at the key sites. The Salt Lake City keysite has
run on ERAM operationally three separate times for periods of seven
days from late January through March 2010. During these periods there
were no delays, separation violations, safety impacts, or catastrophic
failures of ERAM. However, there were issues noted which required work-
arounds in operational procedures and required additional staffing to
conduct operations. While this was workable for short durations of up
to a week, these issues precluded sustained continuous operations due
to the impact on staffing.
The Western Service Center has developed a list (the NAS Operations
List or NOL) containing PRs that are needed in order to sustain
continuous operations at the ZLC key site. These PRs are being fixed
and packaged in three incremental software releases. The initial two
releases have been delivered to the key sites and both Salt Lake City
and Seattle have run for 4 hours on the midshift on the first software
release with no new critical issues identified. The key sites are
expected to conduct operational runs on the 2nd release in early
August and the last software release is planned to be delivered during
August as well. Once enough operational experience is gained on these
releases the schedule to an In-Service Decision (ISD), including
conduct of independent operational test and evaluation, and the
resultant waterfall deployment schedule will be determined. The key
sites are expected to come up on a continuous operational basis on
ERAM by November. The current projection for an In-Service Decision is
the 2nd quarter of FY 2011.
The program office is currently assessing the cost impact to the
program baseline as a result of the schedule delay in achieving ISD
and the waterfall deployment at all ARTCCs.
Source: FAA.
[End of figure]
Figure 3: Program Assessment Report for ADS-B as of June 2010:
[Refer to PDF for image: illustration of report]
Key:
Metric indicates no issues: Green;
Metric indicates some issues: Yellow.
Program Phase: Production;
June 2010:
Program assessment:
Financial: Metric indicates no issues;
Schedule: Metric indicates no issues;
Technical: Metric indicates no issues;
Resources: Metric indicates no issues;
External interest: Metric indicates some issues;
Program manager: Metric indicates no issues.
Supporting indicators:
Financial: Cost Performance Index (CPI);
Measure of cost performance: 1.04.
Financial: To Complete Performance Index (TCPI);
Measure of cost performance: TCPI = 0.97.
Financial: Cost Variance At Completion (CVAC);
Measure of cost performance: Metric indicates no issues.
Financial: Obligation rate;
Measure of cost performance: [Symbol].
Schedule: Schedule Performance Index (SPI);
Measure of schedule performance: 0.98.
Schedule: Schedule Variance At Completion (SVAC);
Measure of schedule performance: Metric indicates no issues.
Schedule: APB standard level 1 milestones;
Measure of schedule performance: Metric indicates some issues.
Schedule: Fiscal year schedule goals;
Measure of schedule performance: Metric indicates no issues.
Technical: Requirements stability;
Measure of cost performance: Metric indicates no issues.
Technical: System defects;
Measure of cost performance: N/A.
Technical: Test results;
Measure of cost performance: Metric indicates no issues.
Technical: Deployment;
Measure of cost performance: Metric indicates no issues.
Technical: Cumulative mitigated risk impact;
Measure of cost performance: Metric indicates no issues.
Technical: Performance Variance At Completion (PVAC);
Measure of cost performance: Metric indicates no issues.
Resources: Funding;
Measure of appropriate funding: Metric indicates no issues.
External interest: Under review by DOT, IG, GAO, or other;
External review activity: Metric indicates some issues.
Program manager comments:
External Interest is yellow because the Program has on-going OIG and
GAO audits.
Juneau IOC was achieved on 04/28/2010. Philadelphia IOC was achieved
on March 28. Louisville IOC was achieved on November 19, 2009. Gulf of
Mexico IOC was achieved on December 17, 2009.
The Final rule was published in the Federal Register on 5/28/10.
ISD due in September and is on track. ISR checklist is 61% complete.
Some automation issues are being worked.
Source: FAA.
[End of figure]
As figures 2 and 3 show, FAA data and reports can indicate potential
problems, and the program manager's comment box can provide additional
information on their nature. For example, the report on ERAM in figure
2 shows schedule performance in red, specifically, the program is
behind in meeting key milestones and technical performance in yellow,
including system defects that need to be dealt with and could affect
the cumulative risk and the level 1 milestones associated with the
program. The program manager comment box provides more detailed
information on the nature of the defects and problems associated with
the program-in this case, interruptions in flight data processor
software. These reports also indicate how well the program is
performing from an EVM perspective; that is, they show how well the
program is meeting its planned cost and schedule targets for system
development. In these reports, the information on the left side of the
chart indicates performance relative to EVM. "CPI" values of greater
than 1 on the chart indicate that the program is performing well, in
that for every dollar spent, more than a dollar of value is received,
whereas a "CPI" less than 1 would indicate less than a dollar of value
is received for every dollar spent. Similarly, a "SPI" value of 1.00
indicates that work is being accomplished at the planned rate, while a
"SPI" value of less than 1 indicates that work is behind schedule to
some degree.
While we did not specifically review the validity of EVM metrics for
these programs as part of this engagement, we have previously reported
issues with FAA's implementation of EVM. Our review of FAA's reports
and interviews with program managers and other officials indicates
that these issues continue to affect the accuracy and reliability of
some of FAA's program status reports. For example, in 2008, we
reported that FAA did not apply sound earned value techniques to the
full ERAM program baseline.[Footnote 18] In particular, FAA had
rigorous EVM processes to govern contractor deliverables for ERAM, but
it did not have the same processes in place for government work. As a
result, FAA could not ensure that the earned value data reported for
the total program were reliable. The consequences of this risk can be
seen in figure 2. While the figure shows that milestones for the ERAM
program as a whole are being met, the EVM contract-level data do not
reliably reflect these ongoing schedule issues affecting the program
at this time. In another recent GAO engagement, we found that because
EVM is not applied at the full program level for ERAM and only at the
contractor level, it is unclear in FAA's reporting whether delays will
affect the program's overall costs.[Footnote 19] The ERAM program
manager told us that the overall program cost is likely to be in
excess of what was originally planned due to the ongoing software
defects and schedule slippages the program is currently experiencing,
but FAA's program assessment report for ERAM does not indicate this
issue. If the EVM processes had been implemented appropriately, the
EVM data could likely have provided an early warning of the problems
the program is currently experiencing and enabled managers to take
timely and aggressive action to mitigate them. (For a further
discussion on ERAM and a more complete description of issues involved
in its implementation, see appendix II). For other programs like ADS-
B, EVM data will be more accurate because EVM metrics have been
established for the full program.
In other previous work, we found that FAA's measurement and reporting
of its acquisition performance could mask budget increases and
schedule delays that could negatively affect the transition to
NextGen.[Footnote 20] Consequently, budget increases and delays in one
program that could slow the implementation of NextGen capabilities may
not be apparent to Congress or aviation stakeholders. Our review of
FAA's reports and interviews with program managers and other officials
indicates that these problems persist. For example, according to the
ERAM program manager, the implementation schedule for ERAM will affect
the implementation schedule for ADS-B if the implementation of ERAM
extends beyond April 2011. Generally, the individual program offices
understand their programs' dependence on ERAM's implementation, but
FAA has not developed a full listing of how ERAM schedule slippages
could affect or put other programs' implementation schedules at risk
or delay the implementation of capabilities or improvements.[Footnote
21] We recommended that FAA improve the usefulness of ATO's
acquisition performance reporting by including information in the
agency's Performance and Accountability Report or elsewhere on the
potential effect of any budget or schedule slippages on the overall
transition to NextGen. This recommendation remains open, as FAA has
not definitively indicated how it will track slippages that will
affect other dependent NextGen programs. Currently, FAA manages its
acquisitions using its Acquisition Management System (AMS), which
establishes policy and guidance for life-cycle acquisition management;
however, AMS was not designed for managing NextGen programs in an
integrated way. To assist in managing its NextGen portfolios, FAA is
employing a solution set management approach, discussed in the next
section of this report, which is designed to monitor all the
activities of a particular operational improvement to ensure
integration is on track. As this approach is more fully implemented,
it will likely clarify the impact of slippages in one program's
schedule on the implementation status of other NextGen programs and
operational capabilities.
In addition to the issues described above, we have made several
recommendations to FAA and DOT on acquisition performance measurement
and reporting systems, and FAA has made many improvements in response,
as shown in figure 4.
Figure 4: Implementation Status of Prior Selected GAO Recommendations
to FAA and DOT as of June 2010:
[Refer to PDF for image: illustrated table]
GAO Report: GAO-08-42 (Dec. 2007):
Recommendation (summarized): Improve the objectivity, reliability, and
inclusion of core programs in ATO‘s acquisition performance measures;
Agency responsible: FAA;
Status of recommendation: Closed (implemented).
GAO Report: GAO-08-42 (Dec. 2007):
Recommendation (summarized): Improve the clarity of ATO‘s annual
acquisition performance measurement process by disclosing in its
Performance and Accountability Reports that the measurement for on-
budget performance covers 8 months and is measured against the most
recently approved budget baselines;
Agency responsible: FAA;
Status of recommendation: Closed (implemented).
GAO Report: GAO-08-42 (Dec. 2007):
Recommendation (summarized): Include information on any mitigation
plans ATO has developed to lessen the effects of program slippages on
the implementation of NextGen systems;
Agency responsible: FAA;
Status of recommendation: Open but in process.
GAO Report: GAO-08-756 (July 2008):
Recommendation (summarized): Modify acquisition policies governing EVM
to require the use of a product-oriented standard work breakdown
structure;
Agency responsible: FAA;
Status of recommendation: Closed (implemented).
GAO Report: GAO-08-756 (July 2008):
Recommendation (summarized): Modify acquisition policies governing EVM
to enforce existing EVM training requirements and expand these
requirements to include senior executives responsible for investment
oversight and program staff responsible for program oversight;
Agency responsible: FAA;
Status of recommendation: Open but in process.
GAO Report: GAO-08-756 (July 2008):
Recommendation (summarized): Modify acquisition policies governing EVM
to define acceptable reasons for rebaselining and when seeking to
rebaseline a program, require (1) a root cause analysis to determine
why significant cost and schedule variances occurred, and (2)
mitigation plans to address the root cause;
Agency responsible: FAA;
Status of recommendation: Closed (implemented).
GAO Report: GAO-08-756 (July 2008):
Recommendation (summarized): Improve FAA‘s oversight processes by
including an evaluation of contractors‘ performance data as part of FAA‘
s program assessment criteria;
Agency responsible: FAA;
Status of recommendation: Closed (implemented).
GAO Report: GAO-08-756 (July 2008):
Recommendation (summarized): Modify acquisition policies governing EVM
to enforce existing EVM training requirements and expand these
requirements to include senior executives responsible for investment
oversight and program staff responsible for program oversight;
Agency responsible: FAA;
Status of recommendation: Closed (implemented).
GAO Report: GAO-10-2 (Dec. 2009):
Recommendation (summarized): Modify policies governing EVM to ensure
that they address the weaknesses that we identified. Direct managers
of key system acquisition programs to implement the EVM practices;
Agency responsible: DOT;
Status of recommendation: Open.
GAO Report: GAO-10-2 (Dec. 2009):
Recommendation (summarized): Direct managers of key system acquisition
programs to take action to reverse current negative performance
trends, as shown in the earned value data, to mitigate the potential
cost and schedule overruns;
Agency responsible: DOT;
Status of recommendation: Open.
Source: GAO review of DOT and FAA responses to past recommendations.
[End of figure]
NextGen Solution Set Approach Encompasses Program Metric Data and
Other Initiatives and Processes but Has Yet to Be Fully Developed:
The NextGen solution set organization and structure hold promise for
monitoring NextGen implementation, but are still under development,
and questions about appropriate roles have yet to be resolved. The
solution set management team will be responsible for monitoring all
aspects of NextGen implementation by tracking schedule and budget
data, as well as changes in policies and processes affecting such
things as certifications, standards, and staffing levels. Within the
management team, solution set coordinators will be in charge of
collecting and monitoring the status of all aspects of operational
improvements and supporting activities within their solution sets. The
coordinator's area of responsibility is vast and shifting because each
solution set encompasses numerous capital acquisitions, programs,
projects, and processes handled by various FAA offices. As of April
20, 2010, FAA had filled four permanent coordinator positions, one
position was filled through a temporary assignment, and one position
was vacant. In addition, the position of solution set manager, who
oversees all the solution set coordinators, was being filled on a
temporary basis. Filling key positions with qualified personnel is an
ongoing challenge for FAA, as we have previously reported.[Footnote 22]
To support its monitoring of solution set activities, FAA is
developing two key tools, the portfolio management tool and project-
level agreements. The portfolio management tool is a database for
tracking and monitoring key milestones and the status of funding that
has been obligated and committed for individual budget line items.
According, to FAA officials currently, about 50 percent of the
programs and projects that receive funding are loaded into the
portfolio management tool; none of these are linked to specific
NextGen operational improvements. This process has been slowed, in
part, as FAA continues to staff the office. Once the remaining
programs and projects are loaded into the portfolio management tool
and linked to operational improvements, the solution set coordinators
can monitor and report progress at a portfolio level. According to FAA
officials, efforts to load the necessary information into the tool are
ongoing and are expected to be completed by the first quarter of
fiscal year 2011. Project-level agreements are annual agreements
between the NextGen Integration and Implementation Office and the
performing service organization (e.g., the Air Traffic Organization-
Terminal for Flexible Terminals and Airports program, or FAA's Office
of Environment and Energy for the Advanced Noise/Emissions Reduction
program) for monitoring and reporting milestones and obligations to
FAA management, OMB, and other stakeholders. In total, there are 95
project-level agreements for fiscal years 2009 and 2010, including 49
for fiscal year 2009 and 46 for fiscal year 2010 that were signed as
of July 15, 2010. It is unclear whether the remaining nine agreements
will be signed by the end of fiscal year 2010, as planned, because,
according to FAA officials, work was slowed during the first part of
fiscal year 2010 by issues associated with operating under a
continuing resolution.
Our analysis of the solution set structure raises questions about
whether sufficient processes are in place that will strengthen
oversight and create a greater likelihood that actions required by
various lines of business to produce operational improvements are
implemented in a timely fashion across the agency. The project-level
agreement outlines the key responsibilities of the performing
organization, such as reporting information in the portfolio
management tool and managing obligations and milestones. If a dispute
arises or the performing organization does not perform its functions
in a timely manner, the agreement provides for an informal resolution
between the signers to the agreement (i.e., the solution set
coordinator and the performing organization). According to FAA
officials, if a dispute could not be resolved, it would be brought to
the NextGen Management Board, which has authority to resolve any
issues, such as ensuring that appropriate resources are available or
issuing policy and guidance to force timely resolution. However, FAA
has no written policy for resolving this type of dispute beyond what
is described in the project-level agreement. Given that the solution
set model is relatively new, there is little experience to draw upon
to understand the impact of tasks not being completed on time or
funding not being spent properly.
Metrics Have Yet to Be Developed to Measure the Performance of NextGen
Improvements in Relation to Specific NextGen Goals, but Some
Performance Metrics Are Available for Specific Programs:
FAA Is Considering a Number of NextGen Performance Metrics, but Little
Progress Has Been Made:
FAA has broad goals for NextGen, such as enhancing safety, reducing
aviation's environmental impact, and increasing operations and
efficiency, but specific goals for NextGen as a whole have yet to be
determined and FAA has not agreed on a set of overall performance
metrics that it can use to measure progress. In order to measure
outcomes and performance as implementation progresses, the Senior
Policy Committee-which is the interagency governing body for NextGen-
will need to identify milestones or performance goals for NextGen as a
whole across federal partner agencies. Relative to the broad goals
outlined for NextGen, FAA will then need to identify a set of metrics
and begin collecting baseline performance information against which to
measure the effects of its NextGen activities.
GAO has identified criteria for sound performance management for
federal agencies that may assist FAA as it continues to develop
specific NextGen performance goals and metrics. According to previous
GAO work, agencies that have been successful in measuring performance
had performance measures that demonstrate results, are limited to the
vital few, cover multiple priorities, and provide useful information
for decision making.[Footnote 23] Furthermore, GAO work cited specific
attributes[Footnote 24] that are key to successful performance
measures:
* Linkage-Measure is aligned with division and agencywide goals and
mission and clearly communicated throughout the organization.
* Clarity-Measure is clearly stated and the name and definition are
consistent with the methodology used to calculate it.
* Measurable target-Measure has a numerical goal.
* Objectivity-Measure is reasonably free from significant bias or
manipulation.
* Reliability-Measure produces the same result under similar
conditions.
* Core program activities-Measures cover the activities that an entity
is expected to perform to support the intent of the program.
* Limited overlap-Measure should provide new information beyond that
provided by other measures.
* Balance-Balance exists when a suite of measures ensures that an
organization's various priorities are covered.
* Governmentwide priorities-Each measure should cover a priority such
as quality, timeliness, and cost of service.
Having performance metrics with these attributes will help FAA
management and stakeholders, such as Congress, make decisions about
how to fund and monitor the progress of NextGen.
While there are currently no agreed-upon NextGen performance goals or
metrics available, JPDO and ATO are working to develop such
performance tools. First, JPDO has developed a list of potential
performance metrics for measuring progress toward the goals of federal
partner agencies-not just FAA, as shown in table 3.
Table 3: Selection of JPDO's Proposed Performance Metrics:
Performance area: Expand capacity;
Metrics: Throughput and average delay.
Metrics: Difference in delay between good and bad weather delays.
Metrics: Cancellations and consequent passenger delay time.
Metrics: Curb-to-curb travel time.
Performance area: Safety;
Metrics: Percentage of proposed improvements evaluated for safety.
Performance area: Environment;
Metrics: Fuel consumed per unit of distance flown.
Performance area: National defense;
Metrics: Reduced flight time from flexible use of special use airspace.
Performance area: Security;
Metrics: Time passengers spend in airport security.
Source: JPDO.
[End of table]
Second, ATO recently stated in the 2010 NextGen Implementation Plan
that it would begin to consider what NextGen performance metrics are
feasible for both FAA and industry. ATO officials told us that FAA is
forming a team of staff from FAA and MITRE[Footnote 25] to develop
metrics as part of the agency's response to a recommendation from the
RTCA Task Force. However, this effort has only recently begun, and no
timeline or action plan has yet been established. Under the direction
of the NextGen Management Board, this group will be charged with
identifying performance metrics in consultation with industry.
According to an FAA official, one of the group's first tasks will be
to review an extensive list of several hundred potential metrics that
FAA has considered in the past and to recommend those metrics that the
group considers the most appropriate for use. Table 4 shows an initial
list of FAA's proposed metrics. This list will be revised and changed
as the work group gets underway. Additionally, FAA has modeling
efforts under way to estimate the impacts of NextGen technologies on
safety, environmental operations and total delay reduction.
Specifically, FAA estimates that, in aggregate, planned NextGen
technologies--including performance based procedures and planned
runway improvements will reduce delays by about 21 percent by 2019 as
measured against doing nothing at all and will save 1.4 billion
gallons of fuel from air traffic operations. These modeling efforts
are somewhat preliminary and still under development and these
estimates of benefits are not currently performance targets for
planned NextGen improvements.
Table 4: Selection of FAA Proposed Performance Metrics:
Performance area: Access and equity;
Performance indicators:
Fleet penetration;
Number of airports with enhanced NextGen capabilities;
Number of centers with enhanced NextGen capabilities.
Performance area: Capacity;
Performance indicators:
Peak-hour throughput at airports;
Peak-hour airspace throughput;
Difference between throughput and demand.
Performance area: Cost-effectiveness;
Performance indicators:
Number of air traffic controllers per operation;
Number of air traffic controllers per flight hour;
Air traffic controller cost per operation;
Air traffic controller cost per flight hour;
Air traffic controller cost per flight mile;
Number of air traffic controllers per facility or facility type.
Performance area: Efficiency;
Performance indicators:
Delay;
Excess fuel consumed;
Excess distance flown;
Distance flown at suboptimal altitude;
Peak-hour average taxi-time.
Performance area: Environment;
Performance indicators:
Aircraft emissions below 3,000 feet;
Full flight emissions;
Terminal noise contour area;
Noise population exposure;
Temperature change, premature mortality, and noise exposure.
Performance area: Flexibility;
Performance indicators:
Percent of flight trajectory flown at optimal parameters;
Percent of user requests granted.
Performance area: Predictability;
Performance indicators:
Delay variance;
Variance of excess fuel consumed;
Variance of excess distance flown;
Variance of distance flown at suboptimal altitude;
Variance of peak-hour average taxi-time;
Variance of demand;
Variance of peak-hour throughput.
Performance area: Safety;
Performance indicators:
Number of accidents per operation;
Number of losses of separation per operation;
Number of pilot deviations per operation;
Number of air traffic management induced accidents;
Runway incursions.
Source: FAA.
Note: The metrics in this table were extracted from FAA's draft
NextGen performance assessment plan. FAA acknowledged that some of
these metrics could not be quantified empirically and were more
appropriate for cost-benefit analysis.
[End of table]
While this list of metrics is preliminary and will be changed as the
work group proceeds, we identified some areas in the course of our
work that do not appear on this list. For example, one industry
stakeholder suggested that average gate-to-gate times for city pairs
would be a useful metric, as several NextGen improvements are meant to
shorten both the time spent flying between two cities, as well as the
time spent taxiing and waiting at the airport. Furthermore, we
recently recommended that FAA develop airport-specific on-time
performance targets to better prioritize its actions and demonstrate
their benefits.[Footnote 26] The Senate FAA reauthorization bill
[Footnote 27] also proposes metrics that are not included here, such
as flown versus filed flight times for key city pairs.
Lastly, developing and agreeing on the right set of goals and metrics
is difficult because many aspects of performance and actions that will
influence outcomes are not exclusively under FAA's control. For
example, to assess its progress in achieving benefits associated with
implementing performance-based navigation procedures, FAA currently
measures the number of procedures it develops annually. While FAA may
be moving away from this approach, stakeholders argue that, by not
measuring outcomes associated with those procedures, such as improved
runway utilization and reduced travel times, FAA has not developed
procedures that have the most significant benefit.[Footnote 28]
However, achieving the benefits of the new procedures requires actions
both outside and within FAA. Outside FAA, airlines must train their
pilots and crews to use the procedures and equip their aircraft for
flight using the procedures. Similarly, JPDO, as the federal agency
coordinator for NextGen, shares responsibility for developing and
agreeing on several proposed metrics whose outcomes will be affected
by the actions of multiple agencies and stakeholders. For example,
JPDO's proposed metric, the outcome of "time passengers spend in
airport security," will be influenced not only by security procedures
from the Department of Homeland Security, but also by airport
configurations and airline scheduling patterns, among other things. In
these cases, developing and agreeing on metrics will require
collaboration with partner agencies, airports, industry, and a variety
of other stakeholders, and will also require commitments from other
parties to take responsibility for various outcomes and aspects of
performance.
FAA Reports Some Performance Metrics for Existing Modernization and
NextGen Programs to OMB, but These Metrics Are Not Always Outcome
Oriented:
FAA reports quarterly to OMB on metrics identified in Exhibit 300
reports for major acquisitions approved through the Joint Resources
Council, but these metrics are not always outcome based or focused on
the performance of the system; therefore, they do not always clearly
indicate progress toward performance goals.[Footnote 29] For example,
one metric for ERAM is related to achieving the capability to utilize
64 ground radar sensors as compared with 24 under the current system.
Performance measures should clearly represent or be related to the
performance they are designed to assess.[Footnote 30] In this case,
the metric measures an output from the ERAM system-utilization of
radar sensors-but does not measure any outcome of having the
capability to utilize more radar sensors - such as improved adherence
to aircraft separation standards and a resulting increase in capacity
or reduction in congestion. Such an outcome would indicate progress
toward an FAA strategic goal to increase capacity. Other metrics FAA
reports are focused on outcomes and will show progress toward goals.
For example, another ERAM metric-to be measured in 2012-is for 10
percent fewer flight delays to be attributable to ERAM as compared
with the average annual number of flight delays attributable to its
predecessor system between 2000 and 2008. This metric will allow FAA
to measure progress toward FAA's goal to improve on-time arrivals.
Hence, the metric is clearly related to the performance that it is
designed to assess and it identifies a baseline from which to measure
progress.
Performance metrics for NextGen programs are also identified in
Exhibit 300[Footnote 31] reports to OMB and include a similar mix of
outcome and output measures. For example, the report on the ADS-B
program identifies a number of clear and specific outcome-based
performance metrics, such as reducing passenger delay hours by 28
percent in the Gulf of Mexico, or maintaining a 13 percent reduction
in the accident rate in Alaska. However, for other aspects of the ADS-
B program, metrics are related to outputs with no corresponding link
to outcomes or goals. For example, one metric is to maintain the
service that transmits weather information via ADS-B, but there are no
metrics associated with the outcome or benefit of having such services
available, the quality of those services, or how the availability of
those services furthers progress toward any of FAA's stated goals,
such as reducing the impact of weather on delays. Without links to
outcomes and goals, metrics will not help to measure progress toward
those goals and the agency may not emphasize the quality of the
services it provides or the resulting benefits to users. In this case,
additional metrics, such as the rate at which aircraft operators
subscribe to ADS-B services or the rates of satisfaction reported by
users of the information, would provide FAA and observers with more
information to indicate the performance of the program and the
benefits derived from public expenditures.
Information on NextGen Outcomes and Performance Is Limited in FAA
Performance Reporting:
In addition to reporting to OMB, FAA uses its annual Flight Plan and
Performance and Accountability Report to report its performance and
activities. However, these documents discuss only a few NextGen
capabilities and programs that are expected to have an effect on
existing agencywide metrics and do not include any performance
information specific to ongoing NextGen capabilities that are being
implemented.[Footnote 32] For example, for one of FAA's metrics-
decreasing the commercial air carrier fatality rate-the Flight Plan
reports that the deployment of ADS-B will help drive the commercial
fatality rate down.[Footnote 33] As discussed in the previous section,
one of the performance metrics associated with ADS-B is to maintain a
13 percent accident reduction rate in Alaska. However, the performance
and accountability report does not currently indicate whether or how
the ongoing deployment of ADS-B has affected the accident rate in
Alaska. Such information would help stakeholders understand the
progress of ADS-B on the performance metric. Outcome goals should be
included in the annual performance plan whenever possible and annual
performance plans should identify performance goals that cover all of
the program activities in an agency's budget.[Footnote 34]
In other cases, NextGen improvements are meant to enhance performance
in certain areas, but the Flight Plan and Performance and
Accountability Report do not mention those planned improvements. For
example, some air traffic modernization and NextGen activities - such
as implementing performance-based navigation procedures[Footnote 35] -
are meant to increase aviation fuel efficiency, and this is one of the
performance metrics in FAA's Flight Plan, yet these reports include no
discussion of the activities underway that are intended to affect this
metric. Table 5 identifies the current metrics described in the Flight
Plan for the goals of increasing safety and capacity; indicates
whether NextGen activities are included in those metrics; and shows
our analysis of whether NextGen activities are captured by the
performance reports.
Table 5: Alignment of NextGen Activities with Existing Flight Plan
Metrics:
Flight Plan goal: Increase safety:
Flight Plan performance metric: Commercial air carrier fatality rate;
NextGen activities captured (per FAA reports): [Check];
Areas NextGen will affect: [Check].
Flight Plan performance metric: General aviation fatal accident;
NextGen activities captured (per FAA reports): [Check];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Alaska accidents;
NextGen activities captured (per FAA reports): [Check];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Runway incursions;
NextGen activities captured (per FAA reports): [Empty];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Commercial space launches;
NextGen activities captured (per FAA reports): [Empty];
Areas NextGen will affect: [Empty].
Flight Plan performance metric: Operational errors;
NextGen activities captured (per FAA reports): [Check];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Safety management system;
NextGen activities captured (per FAA reports): [Empty];
Areas NextGen will affect: [Check].
Flight Plan goal: Increase capacity:
Flight Plan performance metric: Average daily airport capacity (35 OEP
airports);
NextGen activities captured (per FAA reports): [Check];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Average daily capacity (7 metro areas);
NextGen activities captured (per FAA reports): [Check];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Annual service volume;
NextGen activities captured (per FAA reports): [Empty];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Adjusted operational availability;
NextGen activities captured (per FAA reports): [Empty];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Noise exposure;
NextGen activities captured (per FAA reports): [Empty];
Areas NextGen will affect: [Check].
Flight Plan performance metric: Aviation fuel efficiency;
NextGen activities captured (per FAA reports): [Empty];
Areas NextGen will affect: [Check].
Flight Plan performance metric: NAS on-time arrivals;
NextGen activities captured (per FAA reports): [Check];
Areas NextGen will affect: [Check].
Source: GAO analysis of FAA Flight Plan and NextGen Implementation
Plan.
[End of table]
Further clarification and consistency in reporting the outcomes and
performance of new technologies and capabilities as they are deployed,
and how those activities will further affect progress toward
agencywide goals, would provide users with additional context to
discern the impact of ongoing air traffic modernization and NextGen
activities. Recently, FAA has begun an initiative that aims to align
NextGen activities and performance with FAA's Flight Plan and expects
to deliver a report by early 2011.
Conclusions:
NextGen is an undertaking of significant breadth and complexity and
touches several federal agencies, nearly every office within FAA, and
nearly every existing system and piece of infrastructure currently
operating in the national airspace system. As a result, determining
the status and performance of the effort as a whole is, therefore, a
broad, complex undertaking, requiring multiple reports and pieces of
information from multiple parties. While we currently have several
open recommendations related to improving FAA's use of EVM and its
acquisition management system, FAA's current reporting mechanisms can
give overseers and interested parties certain information that can
indicate potential problems with the cost and pace of individual
programs' implementation. However, these mechanisms are insufficient
to report on the status of NextGen portfolios or how delays and cost
overruns in one acquisition can impact implementation of other
programs or capabilities.
FAA's portfolio approach to implementation is designed to help the
agency assess and convey the implementation status of interrelated
capabilities and operational improvements. However, because
implementation of solution sets requires action across various lines
of business with separate budgets within FAA, it is important to
ensure that processes are in place that will strengthen oversight and
create a greater likelihood that required activities are completed on
time. While the NextGen Management Board is ultimately responsible for
resolving disputes, there are no written policies and procedures to
guide its resolution of disputes between the parties to project-level
agreements. A lack of clear dispute resolution procedures raises
questions about how quickly and effectively any such disputes will be
resolved.
Finally, while several of FAA's efforts to develop, agree on, and
implement a suite of performance metrics are relatively recent or
still in progress, action is needed that will provide stakeholders
with a clear vision of what is required by each one. Without a
timeline and action plan that stakeholders have agreed on, it remains
to be seen if these actions will enable FAA to provide stakeholders,
interested parties, Congress, and the American people with a clear
picture of where implementation stands at any given time, and whether
the technologies, capabilities, and operational improvements that are
being implemented are resulting in positive outcomes and improved
performance for operators and passengers.
Recommendations for Executive Action:
To ensure that FAA can effectively manage NextGen solution sets, we
recommend that the Secretary of Transportation direct the FAA
Administrator to develop written policies and procedures for dispute
resolution across different FAA lines of business and outline the
appropriate roles of the solution set managers, program managers, and
the NextGen Management Board in managing these portfolios of
improvements.
To ensure that the outcomes and performance expected from NextGen
improvements are understood and can be monitored, we recommend that
the Secretary of Transportation direct the FAA Administrator to
develop a timeline and action plan to work with industry and federal
partner agencies to develop an agreed-upon list of outcome-based
performance metrics and goals for NextGen broadly and for specific
NextGen portfolios, programs, and capabilities. The Administrator
should then share this list with the appropriate congressional
oversight committees. Furthermore, the Administrator should establish
a clear timeline to align NextGen performance metrics with FAA's
agencywide goals and performance plans.
Agency Comments:
The Department of Transportation provided comments on a draft of this
report via e-mail. In those comments, the department agreed to
consider the report's recommendations. The department also provided
technical comments, which we have incorporated in this report as
appropriate.
We are sending copies of this report to interested congressional
committees, the Secretary of Transportation, the Administrator of the
Federal Aviation Administration, and other parties. In addition, the
report will be available at no charge on the GAO Web site at
[hyperlink, http://www.gao.gov].
If you or your staff have any questions about this report, please
contact me at (202) 512-2834 or dillinghamg@gao.gov. Contact points
for our Offices of Congressional Relations and Public Affairs may be
found on the last page of this report. GAO staff who made major
contributions to this report are listed in appendix III.
Signed by:
Gerald Dillingham, Ph.D.
Director, Physical Infrastructure Issues:
List of Committees:
The Honorable Jerry Costello:
Chairman:
The Honorable Thomas Petri:
Ranking Member:
Subcommittee on Aviation:
Committee on Transportation and Infrastructure:
House of Representatives:
The Honorable John D. Rockefeller:
Chairman:
The Honorable Kay Bailey Hutchinson:
Ranking Member:
Committee on Commerce, Science, and Transportation:
United States Senate:
The Honorable Bart Gordon:
Chairman:
The Honorable Ralph Hall:
Ranking Member:
Committee on Science and Technology:
House of Representatives:
[End of section]
Appendix I: Objectives, Scope, and Methodology:
In response to a congressional request, we examined the Federal
Aviation Administration's (FAA) ability to monitor the implementation
of the Next Generation Air Transportation System (NextGen) portfolio
of air traffic control systems and programs and whether they will
deliver the desired benefit to the national airspace system.
Specifically, we reviewed (1) FAA's metrics for tracking the status of
NextGen acquisition programs and the implementation of NextGen
capabilities, the reliability of these metrics and the data underlying
them, and any limitations or gaps in FAA's efforts to track the status
of NextGen implementation; and (2) how FAA currently measures the
performance of NextGen programs and capabilities, FAA's progress in
developing a full suite of metrics to measure the outcomes and
performance of NextGen capabilities once implemented, and any
limitations or gaps in FAA's approach to developing these metrics.
To determine FAA's metrics for tracking the status of NextGen
acquisition programs and acquisition programs critical to NextGen
implementation (i.e., ERAM), and the implementation of NextGen
capabilities, we analyzed the process for individual program managers
to report their metrics to internal and external stakeholders and
gained an understanding of the database that houses the information.
We reviewed program and process reviews from FAA's acquisition offices
to identify key areas of oversight focus and key findings that have
been reached in such reviews regarding acquisition procedures and
policies. To ensure acquisition program information is reliable, we
drew on past work in which we undertook detailed reviews of the status
of FAA acquisition programs, and we obtained updated information as
necessary from FAA by reviewing documents and interviewing agency
officials. Based on these past reviews, we determined that FAA's
acquisition data were sufficiently reliable for the purposes of our
report. To determine any limitations in FAA's effort to track the
status of NextGen implementation, we obtained and analyzed recent
metric reports for various FAA acquisitions to determine what
information is readily available for FAA management and stakeholders
outside FAA to monitor NextGen programs. We interviewed FAA officials,
including acquisition, finance, and program managers. We then reviewed
this information in light of our past recommendations and findings to
determine the extent to which a program's implementation status can be
discerned from the available information, how well that information
allows reviewers to understand whether any issues may result in delays
to a program, and whether those delays will affect the implementation
of other programs or operational improvements. We also interviewed
ERAM program office officials and representatives of the National Air
Traffic Controllers Association (NATCA) in Washington, D.C.; Salt Lake
City; and Seattle to obtain information related to discrepancies in
program reports associated with ERAM. We did not conduct an individual
or in-depth review of the effectiveness of the specific programs
selected for performance reporting. We also did not identify a
comprehensive list of programs that were excluded from acquisition
performance reporting. This was beyond the scope and intent of this
study.
To gain an understanding of how FAA currently measures the performance
of NextGen programs and capabilities, we reviewed documents that
outline FAA's solution set organization and management approach, and
interviewed officials involved in coordinating and managing solutions
sets. We also reviewed our past reports on FAA's acquisition metrics
and the status of prior recommendations. To determine FAA's progress
in developing metrics to measure the outcomes and performance of
NextGen capabilities once implemented, we first reviewed how FAA
currently reports on its performance, both internally and externally,
and how information on the performance of specific NextGen
capabilities is incorporated into those metrics. We reviewed FAA's
performance and accountability reports and discussed internal
performance reporting methods with relevant FAA officials.
Specifically, we reviewed FAA's Flight Plan, Performance and
Accountability Report, 2009 and 2010 NextGen Implementation Plans,
Enterprise Architecture, and reports to the Office of Management and
Budget (OMB) (known as Exhibit 300 reports).[Footnote 36] To
understand FAA's approach and progress toward developing a suite of
NextGen metrics, we interviewed FAA officials with responsibilities
for NextGen planning and implementation, particularly officials within
the Air Traffic Organization (ATO) and the Joint Planning and
Development Office (JPDO) responsible for modeling NextGen benefits
and developing NextGen performance metrics. To evaluate the metrics
that FAA is considering for potential gaps and limitations, we
compared proposed metrics with key attributes of successful
performance metrics that we identified in past GAO work.[Footnote 37]
Metrics should cover key program activities and represent program and
agency goals and priorities to help identify those activities that
contribute to the goals and priorities. To the greatest extent
possible, metrics should be objective, that is reasonably free of bias
or manipulation that would distort an accurate assessment of
performance, and clearly defined such that they can be understood by
stakeholders both internally and externally. When appropriate, metrics
should be measurable and quantifiable, including having annual
targets, to facilitate future assessments of whether goals or
objectives were achieved. We also interviewed several key stakeholders
for NextGen, including representatives from airlines, equipment
manufacturers, federal partner agencies, and NATCA to get their views
on the metrics they deem most appropriate to measure the performance
of NextGen.
We conducted this performance audit from June 2009 through July 2010
in accordance with generally accepted government auditing standards.
Those standards require that we plan and perform the work to obtain
sufficient, appropriate evidence to provide a reasonable basis for our
findings and conclusions based on our audit objectives. We believe
that the evidence obtained provides a reasonable basis for our
findings and conclusions based on our audit objectives.
[End of section]
Appendix II: Further Information on the ERAM Program:
The En Route Automation Modernization system (ERAM) replaces the
existing en route air traffic control automation system at FAA's Air
Route Traffic Control Centers (ARTCC). ERAM will replace the hardware
and software in the en route Host Computer System and its backup
system, the Direct Access Radar Channel, as well as associated
interfaces, communications, and support infrastructure at 20 en route
centers across the country. This effort is critical because ERAM is
expected to upgrade hardware and software for facilities that control
high-altitude air traffic. ERAM will modernize the En Route
infrastructure to provide a supportable open standards-based system
that will be the basis for future capabilities and enhancements ERAM
will provide existing functionality and new capabilities needed to
support NextGen.
New Information Indicates That Software Issues Will Delay ERAM's
Deployment:
FAA's ability to keep ERAM on schedule remains uncertain because
problems with software at key test sites could delay work at other
sites and on other systems. Recently, FAA halted testing and is
revising its implementation schedule to reflect the effects of these
software problems. Two key sites, Salt Lake and Seattle en route
centers, although achieving initial operating capability (IOC)
[Footnote 38] in June 2009 and September 2009 respectively, under
FAA's master schedule, failed to have an operational ready decision
(ORD)[Footnote 39] by December 2009. As these sites conducted ERAM
testing on live air traffic during the past year, usually late at
night when air traffic volume was low, FAA and controllers found both
critical and non-critical software issues that prompted the center, at
times, to revert to the HOST system. Specifically, instructions to a
controller to hand off control of an aircraft in one sector to a
controller in an adjacent sector failed, and flight data were lost or
reassigned to another flight. Although some progress has been made to
correct these problems, some of these issues remain. FAA continued
working with its contractor, Lockheed Martin, to correct many software
issues, but further testing on live air traffic continues to produce
critical safety errors. As a result, in March 2010, FAA decided, with
the support of the air traffic controllers' union, to halt all ERAM
testing on live traffic and to revise the deployment schedule. Such
revisions will affect numerous sites across the country (see table 6
for the original schedule).
Table 6: ERAM Status for Achieving Key Program Milestones:
Site: Salt Lake;
Government acceptance (GA): 5/3/2008;
Initial operational capability (IOC): 6/18/2009;
Operation readiness decision (ORD): 10/30/2009.
Site: Seattle;
Government acceptance (GA): 5/8/2008;
Initial operational capability (IOC): 9/21/2009;
Operation readiness decision (ORD): 10/30/2009.
Site: Minneapolis;
Government acceptance (GA): 7/24/2008;
Initial operational capability (IOC): 11/6/2009;
Operation readiness decision (ORD): 12/6/2009.
Site: Denver;
Government acceptance (GA): 5/20/2008;
Initial operational capability (IOC): 10/25/2009;
Operation readiness decision (ORD): 11/30/2009.
Site: Albuquerque;
Government acceptance (GA): 6/19/2008;
Initial operational capability (IOC): 2/11/2010;
Operation readiness decision (ORD): 3/13/2010.
Site: Dallas-Ft. Worth;
Government acceptance (GA): 2/17/2009;
Initial operational capability (IOC): 2/14/2010;
Operation readiness decision (ORD): 3/16/2010.
Site: Chicago;
Government acceptance (GA): 4/15/2009;
Initial operational capability (IOC): 2/27/2010;
Operation readiness decision (ORD): 3/29/2010.
Site: Houston;
Government acceptance (GA): 4/23/2009;
Initial operational capability (IOC): 3/3/2010;
Operation readiness decision (ORD): 4/2/2010.
Site: Oakland;
Government acceptance (GA): 1/8/2009;
Initial operational capability (IOC): 3/13/2010;
Operation readiness decision (ORD): 4/12/2010.
Site: Cleveland;
Government acceptance (GA): 11/6/2008;
Initial operational capability (IOC): 3/20/2010;
Operation readiness decision (ORD): 4/19/2010.
Site: Kansas City;
Government acceptance (GA): 6/19/2008;
Initial operational capability (IOC): 4/9/2010;
Operation readiness decision (ORD): 5/9/2010.
Site: Indianapolis;
Government acceptance (GA): 10/28/2008;
Initial operational capability (IOC): 4/25/2010;
Operation readiness decision (ORD): 5/25/2010.
Site: New York;
Government acceptance (GA): 5/18/2009;
Initial operational capability (IOC): 5/7/2010;
Operation readiness decision (ORD): 6/6/2010.
Site: Memphis;
Government acceptance (GA): 9/18/2008;
Initial operational capability (IOC): 5/21/2010;
Operation readiness decision (ORD): 6/20/2010.
Site: Los Angeles;
Government acceptance (GA): 2/19/2009;
Initial operational capability (IOC): 6/4/2010;
Operation readiness decision (ORD): 7/9/2010.
Site: Washington, D.C.;
Government acceptance (GA): 1/28/2009;
Initial operational capability (IOC): 6/19/2010;
Operation readiness decision (ORD): 7/19/2010.
Site: Boston;
Government acceptance (GA): 3/19/2009;
Initial operational capability (IOC): 7/9/2010;
Operation readiness decision (ORD): 8/8/2010.
Site: Atlanta;
Government acceptance (GA): 8/10/2009;
Initial operational capability (IOC): 7/31/2010;
Operation readiness decision (ORD): 8/30/2010.
Site: Jacksonville;
Government acceptance (GA): 10/23/2008;
Initial operational capability (IOC): 8/18/2010;
Operation readiness decision (ORD): 9/17/2010.
Site: Miami;
Government acceptance (GA): 5/20/2009;
Initial operational capability (IOC): 9/24/2010;
Operation readiness decision (ORD): 10/24/2010.
Source: FAA.
[End of table]
Testing at Key Sites Yielded Software Issues:
In testing and evaluating ERAM at two key sites, the Seattle and Salt
Lake en route centers, FAA has encountered both anticipated and
unanticipated software issues.[Footnote 40] Before this testing began
last year at these sites, FAA formally accepted the system from the
contractor--a contractual milestone known as government acceptance
(GA)--which indicates in this case, that equipment has performed to
specification at all 20 sites.[Footnote 41] After GA, FAA designated
the Salt Lake and Seattle sites as key sites for initial live testing
on traffic in order to reach the next milestone, IOC, which FAA
defines as the separation of two aircraft by the ERAM system for as
little as 1 minute. (Figure 5 shows the en route centers progress in
reaching GA and IOC.) However, during the testing process, controllers
discovered a variety of software issues when ERAM was turned on. When
FAA encountered a software issue that required a change, it assigned a
code to the issue to indicate its severity and tracked the
contractor's efforts to correct the issue in a database. During the
life of the program, approximately 15,000 software issues have been
identified. According to FAA officials, problems with the software are
expected and common during the testing phase, especially for a program
as large as ERAM (1.3 million lines of code) and time to correct them
is built into the schedule.
Figure 5: Current Phase of ERAM Testing at FAA's En Route Centers as
of May 2010:
[Refer to PDF for image: illustrated U.S. map]
Initial operating capability (3 Air Route Traffic Control Centers):
Minneapolis;
Salt Lake;
Seattle.
Government acceptance (20 Air Route Traffic Control Centers):
Albuquerque;
Atlanta;
Boston;
Chicago;
Cleveland;
Denver;
Fort Worth;
Houston;
Indianapolis;
Jacksonville;
Kansas City;
Los Angeles;
Memphis;
Miami;
Minneapolis;
New York;
Oakland;
Salt Lake;
Seattle;
Washington.
Sources: FAA and Map Resources (map).
[End of figure]
While some testing at FAA's Technical Center preceded testing at the
two key sites, the Tech Center could not fully test the system because
each of the 20 en route centers has unique airspace or operational
issues that controllers have resolved over the years. Specifically,
the Tech Center ERAM test room has only 12 ERAM positions compared
with the 40 or more controller stations in an enroute center. As a
result the Tech Center could test only limited scenarios. According to
FAA officials, the testing accomplished what was expected, but the
Tech Center environment was not robust enough to capture all issues.
The more extensive testing that has since been conducted in the field
with multiple operational facilities and systems has identified many
issues, both expected and, as field personnel have become involved,
unexpected and these unexpected issues have slowed the scheduled work.
NATCA officials stated that the original ERAM schedule was aggressive
and did not account for such issues and is thus the principal factor
driving the delay.
FAA Halted ERAM Testing in Response to Mounting Software Issues It
Must Address at the Key Sites:
FAA has halted ERAM testing because so many software issues have to be
addressed at the key sites. FAA anticipated the potential for software
issues and initially scheduled approximately 6 to 9 months between IOC
and ORD to fix critical software issues. From the beginning, FAA has
tracked more than 15,000 issues; currently, however, there are
approximately 1,400 unresolved software issues, ranging from not
tracking targets correctly to unwarranted safety alerts for altitude
risk. Out of the 1,400 issues, about 200 are critical and will need to
be resolved before testing on live traffic can resume, according to
FAA and NATCA.
The union was not initially involved in ERAM development but its
recent inclusion may resolve some issues. NATCA officials stated the
many of problems arising with ERAM could have been corrected earlier
if NATCA had been involved. ERAM was designed during a period when
controllers did not participate in efforts to design and test new
systems. Because active users of the system from different locations
could not provide insight early on, issues that could have been
addressed early in the design phase were not addressed. To ensure that
controllers will be involved as efforts go forward, unlike during the
design of ERAM requirements, FAA and NATCA recently entered into a
memorandum of understanding (MOU)[Footnote 42] that is designed to
bring in controllers for testing and evaluation of ERAM to alleviate
some of the same types of problems that arose earlier because they
were not involved. Under this agreement, NATCA will have ERAM
technical, evaluation and training representatives as well as a team
of 16 controllers (including 12 from en route and 4 from terminal
facilities) who will be detailed to test and validate software fixes
with contractor engineers at the FAA Tech Center.
FAA Indicates Revised Schedule May Not Delay Other Programs, but
Software Changes Will Slightly Increase Costs:
FAA acknowledges ERAM is unlikely to be operational at all 20 sites as
originally planned due the unexpectedly large number of software
issues. According to FAA, which has not released a revised schedule,
it is working to fix the 200 or so critical software issues identified
before ERAM testing on live traffic will resume at the key sites. FAA
expects this testing to resume by early fall 2010 with the remaining
sites reaching IOC soon thereafter.
ERAM is a key platform for NextGen programs and keeping it on schedule
is critical to maintaining the schedules for many NextGen programs,
most notably ADS-B. FAA officials stated that the revised schedule for
ERAM is not likely to delay the deployment of ADS-B, which is the
first scheduled NextGen program to come online over the next couple of
years. However, if additional delays occur and push the completion of
testing beyond April 2011, ADS-B's deployment may be delayed.
Specifically, Houston is the first en route center slated to receive
ERAM version 3 in April 2011, which will support ADS-B demonstrations
in the Gulf of Mexico, and reaching IOC on schedule is a critical step
to ensure that ADS-B's deployment schedule is not delayed. ERAM
officials stated they are coordinating with the ADS-B office and that
it has been notified of the potential schedule slippage.
Because FAA has assumed ownership of the system from the contractor,
it is responsible for additional costs associated with any software
changes. Because of the large number of unanticipated software issues,
FAA will have to pay the contractor more than the $2.1 billion
originally budgeted to complete the program. FAA is working to revise
its test schedule and backlog of outstanding software issues and plans
to provide a revised cost and schedule estimate.
[End of section]
Appendix III: GAO Contact and Staff Acknowledgments:
GAO Contact:
Gerald L. Dillingham, Ph.D. (202) 512-2834 or dillinghamg@gao.gov:
Staff Acknowledgments:
In addition to the contact named above, individuals making key
contributions to this report include Andrew Von Ah (Assistant
Director), Kevin Egan, Elizabeth Eisenstadt, Brandon Haller, Rich
Hung, Bert Japikse, Dominic Nadarski, and Josh Ormond.
[End of section]
Footnotes:
[1] NextGen was designed as an interagency effort in order to leverage
various agencies' expertise and funding to advance NextGen while
avoiding duplication. In addition to FAA, federal partner agencies
include the Departments of Commerce (particularly its National Oceanic
and Atmospheric Administration), Defense, Homeland Security, and
Transportation; the National Aeronautics and Space Administration; and
the White House Office of Science and Technology Policy.
[2] GAO, Next Generation Air Transportation System: Challenges with
Partner Agency and FAA Coordination Continue, and Efforts to Integrate
Near, Mid-, and Long-term Activities Are Ongoing, [hyperlink,
http://www.gao.gov/products/GAO-10-649T] (Washington, D.C.: Apr. 21,
2010).
[3] Vision 100--Century of Aviation Reauthorization Act (Pub. L. No.
108-176, 117 Stat. 2490 (2003)).
[4] FAA requested that RTCA--a private, not-for-profit corporation
that develops consensus-based recommendations on communications,
navigation, surveillance, and air traffic management system issues-
create a NextGen Midterm Implementation Task Force (the Task Force),
composed of industry stakeholders, to reach consensus within the
aviation community on the operational improvements that can be
implemented between now and 2018. The Task Force provided
recommendations to FAA in September 2009 and FAA responded to all of
these recommendations in its 2010 NextGen Implementation Plan. FAA is
continuing to work with industry through RTCA to address the
recommendations as implementation continues.
[5] The solution set organization is located in the NextGen
Integration and Implementation Office and provides a portfolio
framework to manage the successful implementation of both immediate
improvements and the large-scale integration of NextGen capabilities.
[6] FAA's Flight Plan is a 5-year strategic plan that outlines agency
goals and metrics. FAA's NextGen Implementation Plan is an annual
workplan that defines the midterm operational capabilities the agency
plans to deliver between now and 2018. FAA's Enterprise Architecture
provides the structure to relate organizational mission, vision, and
goals to business processes and the technical or information
technology infrastructure required to execute them. An Exhibit 300-
also called a Capital Asset Plan and Business Case-is used to justify
resource requests for major investments and is intended to enable an
agency to demonstrate to its own management, as well as to OMB, that a
major project is well planned.
[7] GAO, Tax Administration: IRS Needs to Further Refine Its Tax
Filing Season Performance Measures, [hyperlink,
http://www.gao.gov/products/GAO-03-143] (Washington, D.C.: Nov. 22,
2002).
[8] Earned value management compares the actual work performed at
certain stages of a job to its actual costs--rather than comparing
budgeted and actual costs, the traditional management approach to
assessing progress. By measuring the value of the work that has been
completed at certain stages in a job, earned value management can
alert program managers, contractors, and administrators to potential
cost growth and schedule delays before they occur and to problems that
need correcting before they worsen.
[9] Performance metrics ensure that project managers are accountable
in meeting expected performance goals and that projects are aligned
with an agency's strategic goals.
[10] Within FAA, the Joint Resources Council is an executive body
consisting of associate and assistant administrators, acquisition
executives, the Chief Financial Officer, the Chief Information
Officer, and legal counsel. The council makes agency-level decisions,
including those that determine whether an acquisition meets a mission
need and should proceed. The council also approves changes to a
program's baseline, budget submissions, and the national airspace
system's architecture baseline.
[11] ATO is responsible for operating, maintaining, and modernizing
the nation's current air traffic control system.
[12] The 2010 NextGen Implementation Plan also identifies how FAA
plans to respond to recommendations from the NextGen Midterm
Implementation Task Force (the Task Force). FAA requested that RTCA,
(see footnote 4) create the Task Force to reach consensus within the
aviation community on the operational improvements that can be
implemented between now and 2018.
[13] The NextGen Management Board is chaired by the FAA Deputy
Administrator and consists of ATO's Chief Operating Officer and direct
reports (including all Senior Vice Presidents and Vice Presidents, as
well as the Director of the NextGen Integration and Implementation
Office), the Associate Administrator for Aviation Safety, the Deputy
Associate Administrator for Airports, the Assistant Administrator for
Regions and Center Operations, the Assistant Administrator for
Financial Services/Chief Financial Officer, the Assistant Associate
Administrator for Policy, Planning and Environment, the Director of
JPDO, Department of Defense Liaison, and a representative from MITRE.
Also on the board are members of other key stakeholder groups,
including representatives from the air traffic controller and aviation
safety specialist unions. According to FAA, the Board is currently
undergoing restructuring.
[14] Vision 100--Century of Aviation Reauthorization Act (Pub. L. No.
108-176, 117 Stat. 2490 (2003)).
[15] 2010 NextGen Implementation Plan, March 2010, and GAO, Next
Generation Air Transportation System: Status of Systems Acquisition
and the Transition to the Next Generation Air Transportation System,
[hyperlink, http://www.gao.gov/products/GAO-08-1078] (Washington,
D.C.: Sept. 11, 2008).
[16] [hyperlink, http://www.gao.gov/products/GAO-08-1078] and
[hyperlink, http://www.gao.gov/products/GAO-10-649T].
[17] Organized in 1935 and once called the Radio Technical Commission
for Aeronautics, RTCA is today known just by its acronym, RTCA. RTCA
is a private, not-for-profit corporation that develops consensus-based
performance standards for air traffic control systems. RTCA serves as
a federal advisory committee, and its recommendations are the basis
for a number of FAA's policy, program, and regulatory decisions.
[18] GAO, Air Traffic Control: FAA Uses Earned Value Techniques to
Help Manage Information Technology Acquisitions, but Needs to Clarify
Policy and Strengthen Oversight, [hyperlink,
http://www.gao.gov/products/GAO-08-756] (Washington, D.C.: July 18,
2008).
[19] GAO, Information Technology: Agencies Need to Improve the
Implementation and Use of Earned Value Techniques to Help Manage Major
System Acquisitions, [hyperlink, http://www.gao.gov/products/GAO-10-2]
(Washington, D.C.: Oct. 8, 2009).
[20] GAO, Air Traffic Control: FAA Reports Progress in System
Acquisitions, but Changes in Performance Measurement Could Improve
Usefulness of Information, [hyperlink,
http://www.gao.gov/products/GAO-08-42] (Washington, D.C.: Dec. 18,
2007).
[21] FAA's Enterprise Architecture for the national airspace system
shows the interdependencies and capabilities that may be affected by
various programs, but this document cannot indicate specific
scheduling milestones that might be affected.
[22] GAO, Federal Aviation Administration: Human Capital System
Incorporates Many Leading Practices, but Improving Employees'
Satisfaction with Their Workplace Remains a Challenge, [hyperlink,
http://www.gao.gov/products/GAO-10-89] (Washington, D.C.: Oct. 28,
2009).
[23] [hyperlink, http://www.gao.gov/products/GAO-03-143].
[24] All attributes are not equal and failure to have a particular
attribute does not necessarily indicate that there is a weakness in
that area or that the measure is not useful; rather, it may indicate
an opportunity for further refinement.
[25] MITRE is a not-for-profit organization chartered to work in the
public interest. It manages four Federally Funded Research and
Development Centers, including one for FAA. MITRE has its own
independent research and development program that explores new
technologies and new uses of technologies to solve problems in the
near term and in the future.
[26] GAO, National Airspace System: Setting On-Time Performance
Targets at Congested Airports Could Help Focus FAA's Actions,
[hyperlink, http://www.gao.gov/products/GAO-10-542] (Washington, D.C.:
May 26, 2010).
[27] S.1451, 111th Cong, § 317.
[28] For example, these procedures decrease flight miles, which reduce
an aircraft's fuel burn and carbon emissions.
[29] OMB transmits some of this information into its "IT Dashboard."
See GAO, Information Technology: OMB's Dashboard Has Increased
Transparency, but Data Accuracy Improvements Needed, [hyperlink,
http://www.gao.gov/products/GAO-10-701] (Washington, D.C.: July 17,
2010).
[30] GAO, The Results Act: An Evaluator's Guide to Assessing Agency
Annual Performance Plans, [hyperlink,
http://www.gao.gov/products/GAO/GGD-10.1.20] (Washington, D.C.: Apr.
1998).
[31] OMB 300, also called the Capital Asset Plan and Business Case, is
a document that agencies submit to OMB to justify resource requests
for major information technology investments.
[32] The Flight Plan outlines the agency's four goals (Increase
Safety, Increase Capacity, Organizational Excellence, and
International Leadership) along with numerous performance metrics, and
the Performance and Accountability Report shows the results.
[33] Commercial aviation does not include general aviation.
[34] [hyperlink, http://www.gao.gov/products/GAO/GGD-10.1.20].
[35] Performance-based navigation includes such things as Area
Navigation (RNAV), which enables aircraft to fly on any path within
coverage of ground-or space-based navigation aids, permitting more
access and flexibility for point-to-point operations; and Required
Navigation Performance (RNP), which, like RNAV, enables aircraft to
fly on any path within coverage of ground-or space-based navigation
aids, but also includes an onboard performance monitoring capability.
RNP also enables closer en route spacing without intervention by air
traffic control and permits more precise and consistent arrivals and
departures.
[36] FAA's Flight Plan is a 5-year strategic plan that outlines agency
goals and metrics. FAA's NextGen Implementation Plan is an annual
workplan that defines the midterm operational capabilities the agency
plans to deliver between now and 2018. FAA's Enterprise Architecture
is the structure to relate organizational mission, vision, and goals
to business processes and the technical or IT infrastructure required
to execute them. An exhibit 300-also called a Capital Asset Plan and
Business Case-is used to justify resource requests for major
investments and is intended to enable an agency to demonstrate to its
own management, as well as to OMB, that a major project is well
planned.
[37] [hyperlink, http://www.gao.gov/products/GAO/GGD-10.1.20].
[38] IOC is the declaration by site personnel that the system is ready
for conditional operational use in the national airspace system.
[39] ORD signifies the end of conditional use, at which time,
switchover to the new product is complete.
[40] The Minneapolis Center became a third key site in 2010.
[41] Prior to field testing at the key centers testing was done at
FAA's Technical Center.
[42] The MOU was signed December 2009.
[End of section]
GAO's Mission:
The Government Accountability Office, the audit, evaluation and
investigative arm of Congress, exists to support Congress in meeting
its constitutional responsibilities and to help improve the performance
and accountability of the federal government for the American people.
GAO examines the use of public funds; evaluates federal programs and
policies; and provides analyses, recommendations, and other assistance
to help Congress make informed oversight, policy, and funding
decisions. GAO's commitment to good government is reflected in its core
values of accountability, integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each
weekday, GAO posts newly released reports, testimony, and
correspondence on its Web site. To have GAO e-mail you a list of newly
posted products every afternoon, go to [hyperlink, http://www.gao.gov]
and select "E-mail Updates."
Order by Phone:
The price of each GAO publication reflects GAO‘s actual cost of
production and distribution and depends on the number of pages in the
publication and whether the publication is printed in color or black and
white. Pricing and ordering information is posted on GAO‘s Web site,
[hyperlink, http://www.gao.gov/ordering.htm].
Place orders by calling (202) 512-6000, toll free (866) 801-7077, or
TDD (202) 512-2537.
Orders may be paid for using American Express, Discover Card,
MasterCard, Visa, check, or money order. Call for additional
information.
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]:
E-mail: fraudnet@gao.gov:
Automated answering system: (800) 424-5454 or (202) 512-7470:
Congressional Relations:
Ralph Dawn, Managing Director, dawnr@gao.gov:
(202) 512-4400:
U.S. Government Accountability Office:
441 G Street NW, Room 7125:
Washington, D.C. 20548:
Public Affairs:
Chuck Young, Managing Director, youngc1@gao.gov:
(202) 512-4800:
U.S. Government Accountability Office:
441 G Street NW, Room 7149:
Washington, D.C. 20548: