Aviation Safety
FAA Needs to Strengthen the Management of Its Designee Programs
Gao ID: GAO-05-40 October 8, 2004
The safety of the flying public and the reliability of the nation's aircraft depend, in part, on the Federal Aviation Administration's (FAA) regulation and certification of the aviation industry. FAA delegates the vast majority of its safety certification activities to about 13,600 private persons and organizations, known as "designees," which are currently grouped into 18 different programs. Among other tasks, designees perform physical examinations to ensure that pilots are medically fit to fly and examine the airworthiness of aircraft. GAO reviewed (1) the strengths of FAA's designee programs, (2) the weaknesses of those programs and factors contributing to those weaknesses, and (3) potential improvements to the programs.
The key strength of FAA's designee programs is their ability to leverage agency resources. Allowing technically qualified individuals and organizations to perform 90 percent of certification activities enables FAA to better concentrate its limited staff resources on the most safety-critical functions, such as certifying new and complex aircraft designs. For the aviation industry, designee programs enable individuals and companies to obtain required FAA certifications--such as approvals of aircraft designs--in a timely manner, thus reducing delays and costs to industry that might result from scheduling direct reviews by FAA. For example, officials from Boeing told us that using designees has added significantly to the company's ability to improve daily operations by decreasing certification time. Inconsistent FAA oversight and application of program policies are key weaknesses of the designee programs. FAA headquarters has evaluated only 6 of the 18 designee programs over the last 7 years. FAA conducted the evaluations on an ad hoc basis and lacks requirements or criteria for periodically evaluating these programs. FAA uses these evaluations to determine whether designee programs are complying with agency policies. In addition, FAA field offices do not always oversee designee activities according to agency policy. For example, a recent FAA study found that inspectors were not reviewing designated pilot examiners' work on an annual basis as policy requires. Potential reasons for inconsistent oversight include (1) incomplete databases that FAA uses to manage its oversight of designees, (2) workload demands for FAA staff that limit the time spent on designee oversight, and (3) the lack of adequate training for FAA staff who oversee designees. While we did not find a direct link between inconsistent oversight of these programs and specific safety problems, the lack of consistent oversight limits FAA's assurance that designees perform their work according to federal standards. Opportunities exist for FAA to improve (1) program oversight to ensure consistent compliance with existing policies by FAA staff and (2) the completeness of databases used in designee oversight. For example, FAA could evaluate more of its field offices and designees--efforts modeled partly on the assessments conducted by some FAA regional offices--to ascertain the extent to which policies are being followed.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-05-40, Aviation Safety: FAA Needs to Strengthen the Management of Its Designee Programs
This is the accessible text file for GAO report number GAO-05-40
entitled 'Aviation Safety: FAA Needs to Strengthen the Management of
Its Designee Programs' which was released on November 16, 2004.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to the Ranking Democratic Member, Subcommittee on Aviation,
Committee on Transportation and Infrastructure, House of
Representatives:
October 2004:
AVIATION SAFETY:
FAA Needs to Strengthen the Management of Its Designee Programs:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-05-40]:
GAO Highlights:
Highlights of GAO-05-40, a report to Ranking Democratic Member,
Subcommittee on Aviation, House Committee on Transportation and
Infrastructure:
Why GAO Did This Study:
The safety of the flying public and the reliability of the nation‘s
aircraft depend, in part, on the Federal Aviation Administration‘s
(FAA) regulation and certification of the aviation industry. FAA
delegates the vast majority of its safety certification activities to
about 13,600 private persons and organizations, known as ’designees,“
which are currently grouped into 18 different programs. Among other
tasks, designees perform physical examinations to ensure that pilots
are medically fit to fly and examine the airworthiness of aircraft.
GAO reviewed (1) the strengths of FAA‘s designee programs, (2) the
weaknesses of those programs and factors contributing to those
weaknesses, and (3) potential improvements to the programs.
What GAO Found:
The key strength of FAA‘s designee programs is their ability to
leverage agency resources. Allowing technically qualified individuals
and organizations to perform 90 percent of certification activities
enables FAA to better concentrate its limited staff resources on the
most safety-critical functions, such as certifying new and complex
aircraft designs. For the aviation industry, designee programs enable
individuals and companies to obtain required FAA certifications”such as
approvals of aircraft designs”in a timely manner, thus reducing delays
and costs to industry that might result from scheduling direct reviews
by FAA. For example, officials from Boeing told us that using designees
has added significantly to the company‘s ability to improve daily
operations by decreasing certification time.
Inconsistent FAA oversight and application of program policies are key
weaknesses of the designee programs. FAA headquarters has evaluated
only 6 of the 18 designee programs over the last 7 years. FAA conducted
the evaluations on an ad hoc basis and lacks requirements or criteria
for periodically evaluating these programs. FAA uses these evaluations
to determine whether designee programs are complying with agency
policies. In addition, FAA field offices do not always oversee designee
activities according to agency policy. For example, a recent FAA study
found that inspectors were not reviewing designated pilot examiners‘
work on an annual basis as policy requires. Potential reasons for
inconsistent oversight include (1) incomplete databases that FAA uses
to manage its oversight of designees, (2) workload demands for FAA
staff that limit the time spent on designee oversight, and (3) the lack
of adequate training for FAA staff who oversee designees. While we did
not find a direct link between inconsistent oversight of these programs
and specific safety problems, the lack of consistent oversight limits
FAA‘s assurance that designees perform their work according to federal
standards.
Opportunities exist for FAA to improve (1) program oversight to ensure
consistent compliance with existing policies by FAA staff and (2) the
completeness of databases used in designee oversight. For example, FAA
could evaluate more of its field offices and designees”efforts modeled
partly on the assessments conducted by some FAA regional offices”to
ascertain the extent to which policies are being followed.
Aircraft Undergoing Certification at Organizational Designee Facility:
[See PDF for image]
[End of figure]
What GAO Recommends:
GAO recommends that FAA: (1) establish a program to evaluate all
designee programs, giving priority to those programs that have not been
evaluated, (2) develop mechanisms to improve compliance with existing
designee oversight policies, and (3) upgrade its databases to provide
complete and consistent information on all designee programs and the
extent to which oversight is occurring.
FAA officials generally agreed with our recommendations, but expressed
concerns about our use of an expert panel to identify weaknesses in the
programs.
www.gao.gov/cgi-bin/getrpt?GAO-05-40.
To view the full product, including the scope and methodology, click on
the link above. For more information, contact JayEtta Z. Hecker, (202)
512-2834, heckerj@gao.gov.
[End of section]
Contents:
Letter:
Results in Brief:
Background:
Designee Programs Leverage FAA Resources and Provide Industry with
Timely Certification Reviews:
FAA's Lack of Consistent Oversight of Designee Programs Is Affected by
Incomplete Data, Workload Demands, and Lack of Training:
FAA Has Potential Opportunities to Improve Designee Programs:
Conclusions:
Recommendations for Executive Action:
Agency Comments:
Appendixes:
Appendix I: Objectives, Scope, and Methodology:
Appendix II: Experts Participating on GAO's Panel:
Appendix III: Roles and Responsibilities of Designees:
Appendix IV: Survey Instrument and Results:
Appendix V: GAO Contacts and Staff Acknowledgments:
GAO Contacts:
Staff Acknowledgments:
Bibliography:
Tables:
Table 1: Comparison of Designee Programs Administered by Three FAA
Offices:
Table 2: Experts' Ranking of Top Strengths of the Designee Programs:
Table 3: Experts' Ranking of Top 5 Oversight Weaknesses:
Table 4: Experts' Ranking of Top Ways to Improve FAA's Designee
Programs:
Table 5: Organizations Interviewed by GAO During Site Visits:
Table 6: The Number of Panelists Participating in Each Phase and
Response Rate:
Table 7: Experts' Responses to GAO's Survey:
Figures:
Figure 1: FAA Offices That Manage the Different Designee Programs and
Numbers of Designees (as of May 2004):
Figure 2: Designees Support FAA Throughout the United States:
Abbreviations:
AME: Aviation Medical Examiner:
DAS: Designated Alteration Station:
DER: Designated Engineering Representative:
DOT: Department of Transportation:
DPE: Designated Pilot Examiner:
FAA: Federal Aviation Administration:
GAO: Government Accountability Office:
NVIS: National Vital Information Subsystem:
ODA: organization designation authorization:
PTRS: Program Tracking and Reporting Subsystem:
Letter October 8, 2004:
The Honorable Peter A. DeFazio:
Ranking Democratic Member:
Subcommittee on Aviation:
Committee on Transportation and Infrastructure:
House of Representatives:
Dear Mr. DeFazio:
The safety of the flying public and reliability of the nation's
aircraft depends, in part, on the Federal Aviation Administration's
(FAA) regulation and certification of the aviation industry. Although
FAA staff perform many activities crucial to maintaining the safety of
air transportation, since the 1920s, FAA has depended on
congressionally authorized designee programs to help the agency ensure
that the aviation industry meets certain safety standards. FAA's
designee programs authorize about 13,400 private individuals and about
180 organizations nationwide, known as "designees," to act as
representatives of the agency to conduct many safety certification
activities, such as administering flight tests to pilots, inspecting
repair work by maintenance facilities, conducting medical examinations
of pilots, and approving designs for aircraft parts. These designees
are currently grouped into 18 different programs and are overseen by
three FAA offices--Flight Standards Service, Aerospace Medicine, and
Aircraft Certification Service--all of which are under the Office of
the Associate Administrator for Regulation and Certification. Given the
vastness of the U.S. aviation industry, designees enable FAA to carry
out thousands of certification functions each year. FAA staff[Footnote
1] are responsible for overseeing the work of individual designees and
ensuring that organizational designees (also referred to as
"delegations")--companies such as repair stations that have been
delegated the authority to perform inspections of aircraft that have
undergone major repairs--have systems in place, including staff and
procedures, to perform the delegated functions. Organizational
designees are responsible for overseeing their employees who perform
the delegated functions. Based, in part, on congressional direction,
FAA plans to change its designee programs within the next several years
so that the agency can rely more on organizational rather than
individual designees.
In response to your request, this report addresses the following
questions: (1) What are the strengths of FAA's designee programs? (2)
What are the weaknesses of the programs and the factors that contribute
to those weaknesses? and (3) What can be done to address the identified
weaknesses or otherwise improve the programs?
To address these questions, we obtained and analyzed information from a
variety of sources. We identified 62 aviation experts with knowledge
and expertise in FAA's designee programs, who participated on a Web-
based panel that provided the group's views on the strengths and
weaknesses of the designee programs and ways to improve the programs.
An initial list of experts was identified through referrals by FAA
officials, the National Air Traffic Controllers Association, the
Professional Airway System Specialists, and the Aerospace Repair
Station Association and through citations in the literature on
aviation. We then asked these initially identified experts for
additional experts. We continued this process until we had about 10 to
20 experts in each of four categories: (1) designees, (2) FAA
inspectors and engineers, (3) independent experts and university
academics, and (4) private sector and aviation industry associations.
We obtained the experts' views by employing an iterative and controlled
feedback process for obtaining individual views and then allowing each
participant to respond to the entire panels' comments.
In addition, we obtained and analyzed information from FAA databases
that maintain records on designees for fiscal years 1998 through 2003.
We assessed the reliability of the databases and found the data
sufficiently reliable for the types of analyses that we conducted for
this report--including nationwide analyses of the number of designees
by program, the geographical location of designees, and the number of
designees per FAA staff responsible for designee oversight. However, we
found that specific data needed for oversight were not found in some
databases, as we discuss later in this report. We also conducted semi-
structured interviews with FAA officials, representatives of FAA
inspectors and engineers who oversee designees, and designees in
Seattle, Atlanta, Los Angeles, and Oklahoma City to obtain information
on FAA's oversight of designees. We also interviewed officials from
Transport Canada (the Canadian civil aviation authority) to obtain
descriptive information on their designee programs. In addition, we
reviewed past studies of FAA's use of designees by us, the Department
of Transportation's (DOT) Office of Inspector General, and others. We
conducted our work from April 2003 through October 2004 in accordance
with generally accepted government auditing standards. Additional
information on our methodology and the experts who participated on our
panel are found in appendixes I and II.
Results in Brief:
Designees perform more than 90 percent of FAA's certification
activities, thus greatly leveraging the agency's resources. By
permitting nearly 13,400 technically qualified individuals and about
180 organizations to perform thousands of certification tasks each
year, FAA officials believe that the designee programs allow the agency
to concentrate on what it considers to be the most critical safety
areas, a view shared by our panel of experts. For example, while
designees conduct routine certification functions, such as approvals of
aircraft technologies that the agency and designees have had previous
experience with, FAA focuses on new and complex aircraft designs or
design changes. In addition, the use of designees expands FAA's access
to technical expertise within the aviation community. For the aviation
industry, the designee programs enable individuals and organizations to
obtain required FAA certifications--such as approvals of the design,
production, and airworthiness of aircraft--in a timely manner, thus
reducing delays and costs to the industry that might result from
scheduling direct reviews by FAA staff. For example, officials from an
aircraft manufacturer told us that the use of designees has added
significantly to the company's ability to enhance and improve daily
operations by decreasing certification delivery time and increasing the
flexibility and utilization of company resources. In addition,
designees are convenient to the aviation industry due to their wide
dispersal throughout the United States.
FAA's inconsistent monitoring of its designee programs and oversight of
its designees are key weaknesses of the programs. For example, while
FAA has evaluated 6 of its 18 designee programs over the last 7 years
and has plans to evaluate 2 more, it has no plans to evaluate the
remaining programs because of limited resources, according to a program
official. FAA conducted these evaluations on an ad hoc basis usually at
the request of FAA headquarters directors or regional office managers.
The agency does not have requirements or criteria for periodically
evaluating these programs. FAA uses these evaluations to determine
whether designee programs are being carried out in compliance with
agency policies. However, FAA has not implemented some recommendations
from these evaluations. For example, a 2000 evaluation of designated
alteration stations recommended that FAA establish a process to
periodically assess the effectiveness and applicability of existing
oversight policies concerning designated alteration stations and
consider feedback from FAA field offices and designees as part of that
process. The agency has not implemented this recommendation. In
addition, we found that FAA field offices do not consistently implement
agency policies on monitoring, selecting, and terminating designees.
For example, inspectors in one region were not reviewing designated
pilot examiners work on an annual basis and conducting oversight as
required by agency policy. The primary goal of FAA's standards and
policies, and its oversight of designees, is the safety of U.S.
aviation. While we did not find systematic safety problems associated
with FAA's oversight of designees, the agency's inconsistent oversight
limits its assurance that the designees' work is performed uniformly in
accordance with those standards and policies. Finally, we identified
several factors that may have hindered FAA's ability to systematically
monitor the designee programs and consistently apply designee oversight
policies. First, FAA's oversight is hampered, in part, by the limited
usefulness of some agency databases that are designed to capture
information on designees. While all the databases have descriptive
information on designees, such as their types of designations and
status (i.e., active and/or terminated), the databases lack complete
and consistent information on designees' performance and do not provide
a comprehensive picture of whether FAA staff are carrying out their
responsibilities to oversee designees. Second, the workload demands on
FAA staff may limit the time they spend on designee oversight. Finally,
FAA does not require refresher training for all staff who oversee
designees, thereby increasing the risk that some staff do not retain
the information, skills, and competencies required to perform their
oversight responsibilities.
Opportunities exist for FAA to address these weaknesses by improving
(1) oversight of the designee programs to ensure consistent compliance
with existing policies by FAA inspectors, engineers, and flight
surgeons and (2) the accuracy and comprehensiveness of computerized
information on designees so that the databases can be more useful tools
for designee oversight. Those opportunities were identified by experts
on our panel and our review of practices within FAA and procedures
adopted by other countries in administering their programs. For
example, FAA could more consistently conduct internal evaluations of
its field offices and designee programs--evaluations modeled in part on
the assessments performed by some regional and program offices--to
ascertain the extent to which its policies and procedures are being
followed. FAA's internal review of designated pilot examiners in one
regional office could provide a model for evaluations that could be
performed by other FAA regions and for other designee programs. The
review, which was based on (1) a comprehensive statistical analysis of
designee activity in the region, (2) a survey of pilots who were tested
by those designees, and (3) audits of designee files and surveillance
reports by FAA inspectors, provided a reasonable method to assess
program outcomes, identify the root causes of the lack of compliance
with agency policy, and develop corrective action plans to address the
root causes. Accurate, comprehensive data on FAA oversight and designee
activities are integral to monitoring and evaluating the programs. The
database used by FAA's Office of Aerospace Medicine to monitor the
activities and performance of aviation medical examiners provides
information and uses that could serve as a model for the other offices-
-Flight Standards Service and Aircraft Certification Service--that lack
comprehensive databases on designee activities. Although this database
was designed to simplify the processing of airmen medical certification
information, Aerospace Medicine uses it to extract information on the
status of aviation medical examiners and monitor their activity levels.
Careful consideration of such opportunities are important both because
of the central importance that the designee programs hold for FAA as
well as the agency's plans to expand the use of organizational
designees, which will further transform FAA's role to that of
monitoring the performance of organizations rather than overseeing the
individuals who perform the certification activities. Transport Canada,
which expanded its use of organizational designees in the late 1980s,
identified the establishment of standardized oversight practices and
frequent audits of Canadian designees as important components of its
programs.
To improve management control of the designee programs, and thus
increase assurance that designees meet FAA's performance standards, we
recommend that the Secretary of Transportation direct the FAA
Administrator to establish a program to evaluate all designee programs,
giving priority to those programs that have not been evaluated, and
develop mechanisms to more consistently monitor and improve compliance
with existing designee oversight policies, including identifying and
sharing best practices among FAA programs and field offices. We also
recommend that FAA strengthen the effectiveness of its designee
databases by improving the consistency and completeness of information
on designees activities and performance and FAA oversight. FAA
officials generally agreed with these recommendations. However, the
agency expressed concerns about our methodology for obtaining expert
opinions of the designee programs. Further information is provided in
the "Agency Comments" section of this report.
Background:
FAA has relied on designee programs since the 1920s to help the agency
meet its responsibility for ensuring that the aviation industry meets
FAA's safety standards.[Footnote 2] The programs authorize private
persons and organizations, known as individual and organizational
designees, respectively, to act on behalf of the agency to perform many
activities to ensure the safety of air transportation. Of the nearly
13,600 designees nationwide, approximately 13,400 are individual
designees and about 180 are organizational designees, as of May 2004.
These designees are grouped into 18 different programs and are overseen
by three FAA offices--Flight Standards Service, Aerospace Medicine, and
Aircraft Certification Service--all of which are under the Office of
the Associate Administrator for Regulation and Certification. Figure 1
shows the 18 different designee programs, the number of designees, and
the FAA offices that manage them.
Figure 1: FAA Offices That Manage the Different Designee Programs and
Numbers of Designees (as of May 2004):
[See PDF for image]
[End of figure]
Designees perform a large percentage of certification activities on
behalf of FAA, such as determining whether aircraft designs,
manufacturing, and maintenance meet specific safety standards and
certifying the competency of persons that operate aircraft. FAA policy
calls for the agency to delegate activities by evaluating the risk
involved with such delegation; assessing whether the aviation industry
has the experience to perform designated tasks; and delegating
activities with defined standards, processes, and oversight procedures.
FAA policy also states that some tasks are not delegated. For example,
FAA does not permit designees to make rules, conduct surveillance or
enforcement activities against aircraft manufacturers and airlines, or
issue and modify aircraft type and production certificates.
Individual and organizational designees' roles and responsibilities
vary according to program. For example, individual designees, such as
engineering designees, evaluate whether aircraft designs meet FAA
safety standards, designated mechanic examiners administer practical
tests to mechanic applicants, designated pilot examiners administer
practical tests to pilot applicants, and aviation medical examiners
certify that pilots are medically fit to operate aircraft. Most
individual designees can charge service fees to applicants. Most
organizational designees perform similar activities as individual
designees, but the organization holds the designation rather than the
employees who work for them.[Footnote 3] The organization is
responsible for managing, overseeing, and training its employees who
perform the delegated functions. Organizational designees must develop
procedures manuals that describe how the organizations will comply with
FAA requirements and describe their internal evaluation processes,
including internal auditing procedures. An example of an organizational
designee is a designated alteration station, which is a company that
can issue supplemental type certificates, which are required for
aircraft that have been modified from their original design. Further
information on the roles and responsibilities of the various types of
designees are presented in appendix III.
FAA policy calls for selecting and appointing designees based on
several factors, including designees' experience and qualifications,
FAA field or program offices' ability to oversee designees, and the
need for particular types of designees. Although the selection and
appointment policies and procedures differ somewhat for different
designee types, these policies generally call for specific and thorough
technical reviews of the designee applicants' qualifications, including
verifying the applicants' work experience, testing the applicants'
knowledge and skills, and examining on-the-job performance. According
to FAA policy, FAA officials or flight surgeons evaluate the
applicants' experience and qualifications and determine whether to
appoint or deny the applicant's request for designation.
FAA's field and program offices are responsible for supervising,
monitoring, and tracking designees' activities to ensure that designees
are performing their authorized functions in accordance with the
appropriate regulations, policies, and procedures. FAA policy states
that its inspectors, engineers, and flight surgeons should ensure the
integrity of the designee programs by evaluating designee performance,
interacting with designees on a regular basis, and evaluating technical
data prepared by designees. For instance, FAA inspectors are expected
to oversee designated pilot examiners by verifying their attendance at
required training seminars and meetings, ensuring that they have
developed and implemented a plan of action for the practical tests they
conduct on pilot applicants, observing annually at least one practical
test administered to a pilot applicant, and verifying that the designee
has sufficient work activity to justify continuance of the designation.
By comparison, FAA inspectors and engineers are expected to oversee
organizational designees by ensuring that the organizations' procedures
manuals comply with FAA policies on approving the design, production,
and airworthiness of aircraft and assessing the technical capabilities
of the organization. In addition, FAA officials are expected to provide
guidance and oversight of organizational designees by participating in
many aspects of major approvals. For instance, FAA officials provide
guidance and oversight for projects involving new aircraft design
concepts and technology.
Most designees' appointments are effective for 1 year, with the
exception of individual and organizational designated airworthiness
representatives, who are appointed for up to 5 years and all other
types of organizational designees, whose appointments do not expire.
FAA can terminate designees for various reasons, including insufficient
work activity, unacceptable performance, lapse of qualifications, and
lack of FAA need or ability to manage them. Designees can generally
appeal FAA's decision to terminate them, except when the decision to
terminate has been based on FAA's lack of resources to manage them.
Table 1 compares aspects of designee oversight, including how designees
are selected and terminated, among the three FAA program offices with
designee responsibilities.
Table 1: Comparison of Designee Programs Administered by Three FAA
Offices:
Program areas: Designee selection;
Office of Aircraft Certification Service: Local FAA panel reviews
designee applicants' qualifications and makes appointment;
Office of Flight Standards Service: National selection board (National
Examiner Board) reviews designee applicants' qualifications and creates
a list of qualified candidates; Field office managers make appointment
from the list of qualified candidates;
Office of Aerospace Medicine: FAA regional flight surgeons review the
qualifications of designee applicants and make appointments.
Program areas: Designee oversight;
Office of Aircraft Certification Service: FAA inspectors or engineers
are required to annually witness the performance of designees; FAA is
required to conduct a technical evaluation and an Aircraft
Certification Systems Evaluation Program[A] evaluation of delegated
organizations every 2 years; Organizational designees are required to
perform and document self-evaluation activities;
Office of Flight Standards Service: FAA inspectors are required to
conduct annual surveillance of most designees; Organizational designees
are required to perform and document self-evaluation activities;
Office of Aerospace Medicine: FAA regional flight surgeons are not
required to conduct site visits of designees, but are required to
assess designee performance in order to renew authorizations.
Program areas: Database used to monitor designees;
Office of Aircraft Certification Service: Designee Information Network;
Office of Flight Standards Service: Program Tracking and Reporting
Subsystem and National Vital Information Subsystem;
Office of Aerospace Medicine: Airmen Medical Certification Information
Subsystem.
Program areas: Training for designees and FAA staff who oversee
designees;
Office of Aircraft Certification Service: Designees are required to
attend initial indoctrination and refresher training every 2 years; FAA
staff are required to attend initial training in areas of
specialization and take the Delegation Management Course. Refresher
training is not required for staff; Organizational designees are
responsible for training authorized representatives who perform
delegated functions.[B];
Office of Flight Standards Service: Designees are required to attend
initial indoctrination and refresher training every 2 years; FAA staff
are required to attend initial training in areas of specialization. A
specific training course on designee oversight has not been developed.
Refresher training is not required for staff; Organizational designees
are responsible for training authorized representatives who perform
delegated functions.[B];
Office of Aerospace Medicine: Designees and FAA staff are required to
attend initial indoctrination and refresher training every 3 years.
Program areas: Termination of designees;
Office of Aircraft Certification Service: Field office managers
terminate designees;
Office of Flight Standards Service: Field office managers terminate
designees;
Office of Aerospace Medicine: Regional flight surgeons terminate
designees.
Source: GAO analysis of FAA information.
[A] Aircraft Certification Systems Evaluation Program evaluations were
designed to determine if FAA-delegated facilities are complying with
the requirements of applicable federal regulations and the procedures
established to meet those requirements.
[B] Training covers such areas as functions delegated to the
authorization, the organization's processes and procedures, and FAA
policy and guidance material.
[End of table]
FAA has proposed expanding the number of organizational designees and
reducing the number of individual designees by creating an organization
designation authorization (ODA) program. The ODA program would allow
FAA to expand and standardize the approval functions of organizational
designees and expand eligibility for organizational designees,
including organizations not eligible under current FAA rules.
Organizational designees under the current programs would be phased out
during the first 3 years of implementing the new program, and the
organizational designees would be expected to reapply for an ODA. FAA
issued a Notice of Proposed Rulemaking for the ODA program in January
2004. While FAA has received many comments in opposition to the
proposed program including several that raise concerns that the
proposed program would provide less specific and less technical
oversight by FAA and would, over time, reduce the safety of the flying
public, FAA has also received comments that the proposed program would
improve the effectiveness of the agency's oversight of designees.
In addition, FAA has been mandated to develop and implement a certified
design organization program, which would affect some designees
currently responsible for approving the design and production of
aircraft, and aircraft parts and equipment.[Footnote 4] Under this
program, certain organizational designees that design and produce
aircraft parts and equipment would no longer be designees, rather they
would conduct their approval functions under a newly created FAA
certificate. As a certificate holder, the certified design
organizations would be subject to more formal processes when FAA grants
or revokes the certificate. FAA would develop those processes as part
of its requirement to develop a plan to implement a certified design
organization program by 2007.
Designee Programs Leverage FAA Resources and Provide Industry with
Timely Certification Reviews:
Designees perform more than 90 percent of FAA's certification
activities, thus greatly leveraging the agency's resources and enabling
staff to concentrate on other areas of aviation safety, according to
our panel of experts, FAA and industry officials, and FAA staff who
oversee designees. The approximately 13,600 designees augment FAA's
workforce of about 4,100 inspection staff who are responsible for
ensuring industry's adherence to FAA regulations. According to FAA
officials, designees are crucial to the certification process by
conducting routine activities, thereby allowing the agency to target
its direct involvement to the most critical certification functions.
For example, designated airworthiness representatives and designated
manufacturing inspection representatives routinely support company
efforts to perform design enhancements by conducting design conformity
inspections in accordance with established procedures, while FAA's
Aircraft Certification Service focuses on new and complex aircraft
designs or design changes. This information is consistent with the
strengths of the FAA's designee programs identified by our expert
panel. Table 2 shows the top five strengths identified by our expert
panel. There was considerable agreement among the experts on these
strengths. All were identified as a "great" or "very great" strength of
the designee programs by most of the panelists. No more than 2 of the
62 participating experts felt that these strengths had "no" importance
toward accomplishing FAA's safety responsibilities. (See app. IV for
additional strengths identified by our expert panel.)
Table 2: Experts' Ranking of Top Strengths of the Designee Programs:
Ranking: 1;
Strength: Use of designees expands available FAA resources.
Ranking: 2;
Strength: Use of designees allows for more timely approvals than by not
using designees.
Ranking: 3;
Strength: Use of designees expands available technical expertise and
specialization.
Ranking: 4;
Strength: Designees provide greater scheduling flexibility and access
to the public.
Ranking: 5;
Strength: Use of designees enables FAA staff to concentrate on other
areas of aviation safety.
Source: GAO analysis of expert panel information.
Note: Rankings based on responses from 62 experts and the frequency of
responses indicating a "great" or "very great" strength.
[End of table]
According to all of the private industry experts on our panel and many
of the other panelists, the use of designees allows the aviation
industry and others to obtain more timely approvals and issuance of
aircraft certifications than would be possible if FAA staff were solely
responsible for those tasks. The designee programs provide more timely
service to the aviation industry, while assuring the airworthiness of
aeronautical products by utilizing aviation industry expertise to
perform many certification activities under the oversight of FAA,
according to agency officials. In addition, the designee programs
provide the industry with greater scheduling flexibility and access to
aviation safety-related services, such as access to aircraft and pilot
certification services. For example, Boeing officials told us that the
use of designees has added significantly to the company's ability to
enhance and improve daily operations by providing consistent
certification processes, decreasing certification delivery time, and
increasing the flexibility and utilization of Boeing resources, which
could reduce costs. Many experts on our panel also concurred that the
designee programs are convenient to the aviation industry, as aviation
organizations are able to control their production deadlines and not
depend on FAA's schedule for certification and approval. Figure 2 shows
the geographic distribution of designees and their wide dispersal
throughout the United States.
Figure 2: Designees Support FAA Throughout the United States:
[See PDF for image]
[End of figure]
Additionally, the use of designees expands FAA's access to technical
expertise within the aviation community, as many designees are industry
experts. Forty-six of the 62 experts on our panel thought this was a
"great" or "very great" strength of the designee programs, including
all of the experts from the aviation industry. For example, designated
engineering representatives review thousands of calculations, tests,
and data involved in aircraft designs, on behalf of the agency to
ensure compliance with FAA regulations. Other designees, such as
designated manufacturing inspection representatives and designated
airworthiness representatives, are technical experts in the "production
conformity"[Footnote 5] or inspection of certain aircraft products or
parts and issue certificates or approvals for engines, propellers, and
other aircraft parts. Still other designees are aviation medical
examiners--physicians who have been delegated the authority to perform
physical examinations to determine if applicants are qualified to
receive airman medical certificates[Footnote 6] and student pilot
certificates.
FAA's Lack of Consistent Oversight of Designee Programs Is Affected by
Incomplete Data, Workload Demands, and Lack of Training:
Our work shows that inconsistent oversight is a key weakness of the
designee programs. Oversight occurs at two levels: at FAA headquarters,
which is responsible for monitoring the practices of its field offices,
and at FAA field offices that are directly overseeing designees. First,
while FAA has evaluated 6 of its 18 designee programs since 1997 and
plans to evaluate 2 more programs, it has no plans to evaluate the
remaining programs because of limited resources. Moreover, the agency
has not implemented some key recommendations from these evaluations.
Second, FAA field offices do not always oversee designee activities
according to FAA policy, nor do the field offices apply consistent
criteria for selecting and terminating designees. The primary goal of
FAA's standards and policies, and its oversight of designees, is the
safety of U.S. aviation. While we did not find systematic safety
problems associated with FAA's oversight of designees, the agency's
inconsistent oversight limits its assurance that the designees' work is
performed uniformly in accordance with those standards and policies.
FAA's ability to systematically evaluate the designee programs and
consistently apply its designee oversight policies may be impeded by
three conditions: (1) incomplete data on FAA's oversight of designee
activities, (2) workload demands placed on FAA staff who oversee
designees, and (3) the lack of adequate training for FAA staff who
perform oversight duties.
FAA Provides Inconsistent Monitoring of Its Field Offices:
To monitor the effectiveness of its designee programs and determine
whether field offices are following FAA policy in their oversight of
designees, FAA has evaluated only 6 of its 18 designee programs over
the last 7 years. These evaluations encompass about 35 percent of FAA's
designees. Moreover, these evaluations vary in quality and
comprehensiveness. While FAA has plans to evaluate two additional
designee programs over the next several years, it does not plan to
evaluate the other 10 designee programs because of limited resources,
according to a program official. FAA conducts evaluations of its
designee programs on an ad hoc basis, usually at the request of FAA
headquarters directors or regional office managers and uses these
evaluations to determine whether the programs are being implemented in
accordance with agency policies. The agency does not have requirements
or criteria for periodically evaluating these programs and identifying
the root causes for field offices and staff not consistently following
FAA policies. According to FAA officials, the agency is developing
quality management standards that will be used to evaluate field
offices, including their oversight of designee programs. Both Flight
Standards and Aircraft Certification Services plan to obtain approval
of their quality management standards in 2006, but have no timeframe
for conducting additional evaluations. While Aerospace Medicine has not
evaluated its designee program, it uses regular management meetings
with all the regional flight surgeons to monitor field oversight
activities.
For the 11 designee programs within Flight Standards Service, the
office has evaluated the designated pilot examiner program in some
field offices and has plans to evaluate oversight practices for aircrew
program designees in 2005 and designated mechanic examiners by 2006.
However, the office has no current plans to review the oversight
practices for the additional eight types of designees because of
limited resources, according to a program official. In 2000, FAA's
Flight Standards Service created a Quality Assurance Team to undertake
standardized evaluations of its field offices to determine how they are
conducting business, identify deficient areas, and make improvements as
needed.[Footnote 7] As of July 2004, the Quality Assurance Team had
evaluated the oversight of designated pilot examiners at 60 out of 104
Flight Standards field offices to determine whether each office is
following FAA policies and standards. The team plans to assess the
designated pilot examiner oversight practices of the remaining field
offices in 2005. Among the completed evaluations, Flight Standards has
identified program weaknesses, such as computerized data records that
lack information on required surveillance of designees. The evaluation
process calls for reporting any identified deficiencies to the
appropriate offices and regions for corrective action. However, the
evaluations by the Quality Assurance Team do not identify the root
causes or reasons for field offices and staff not consistently
following FAA policies and standards. According to program officials,
root causes of the problems are not identified because that is not the
purpose of the audits.
In addition, in 2000, Flight Standards' Southwest Region reviewed the
designated pilot examiner program in its nine field offices. While the
review did not find any pilots who had been inappropriately
certificated, it did find that inspectors were not reviewing pilot
examiners' work on an annual basis and conducting oversight as required
by FAA policy.[Footnote 8] The review by the Southwest Region was more
comprehensive than the reviews undertaken by the Quality Assurance
Team. Both the region and the Quality Assurance Team audited data on
designees that were maintained in office files and in a computerized
database for compliance with agency policy. However, unlike the Quality
Assurance Team, the Southwest Region also gathered and analyzed
information on designee activity and surveyed newly certificated pilots
and conducted a 2-day conference with designated pilot examiners from
the region. This more rigorous evaluation allowed the region to assess
the outcomes of this designee program, identify root causes of the lack
of compliance with agency policy, and develop corrective action plans,
including increased training for inspectors, to address the root
causes. Flight Standards has not applied this more comprehensive
evaluation to its other eight regions or other designee programs to see
if similar problems exist and to take any needed corrective action.
By comparison, from 1997 through 2000, FAA's Aircraft Certification
Service assessed five[Footnote 9] of its six designee programs and took
action to identify and correct the root causes of some identified
weaknesses.[Footnote 10] For example, in 2000,[Footnote 11] the office
assessed one designee program--designated alteration stations--in the
aftermath of the fatal crash of Swissair Flight 111 in 1998, which
killed 229 passengers and crewmembers. The Transportation Safety Board
of Canada, which investigated the crash, suspected that an
entertainment system, the installation of which had been approved by an
FAA designee, may have been one factor contributing to a deadly
electrical fire on board the aircraft.[Footnote 12] The Board concluded
that FAA's designee program did not ensure that the designated
alteration station employed personnel with sufficient aircraft-
specific knowledge to appropriately assess the integration of the
entertainment system's power supply with aircraft power. In response to
the Canadian report, in 1999, FAA investigated its oversight of the
designated alteration station involved in the crash and concluded that
FAA's oversight of the designee that installed the entertainment
systems was in accordance with FAA policy.[Footnote 13] However, the
report went on to note that aspects of FAA's policy for overseeing
designated alteration stations lacked clarity and needed revision. To
address this problem, the report recommended a nationwide study of
FAA's oversight of designated alteration stations. This subsequent
study, conducted in 2000, found general oversight weaknesses, including
the lack of a national standard policy on management and oversight of
designated alteration stations and a general lack of FAA supervision of
these designees. To address the root cause of the problems identified,
the 2000 study recommended revisions to FAA's order concerning
oversight of designated alteration stations, which were made and issued
in August 2002. The 2000 review further recommended that the office
establish a process to periodically assess the effectiveness and
applicability of existing policies concerning designated alteration
stations and consider feedback from FAA field offices and designees.
The Aircraft Certification Service has not implemented this
recommendation to directly assess the policies in place, but continues
to rely on informal feedback from FAA field offices and industry.
In addition, FAA has not fully implemented its 2002 policy to conduct
technical evaluations of 49 organizational designees, located primarily
in the Aircraft Certification Service.[Footnote 14] Technical
evaluations allow the agency to determine whether the products and data
produced by the organizations are technically acceptable and comply
with FAA policies. According to FAA officials, the agency had conducted
10 technical evaluations as of June 2004. FAA is allowing
organizational designees time to perform approvals under their new
procedures before performing the technical evaluations, according to
the agency. In the meantime, according to FAA officials, these
organizational designees are being evaluated under the current Aircraft
Certification Systems Evaluation Program, which require an evaluation
every 2 years.
Field Offices Provide Inconsistent Oversight of Designees:
Concerns about the consistency and adequacy of designee oversight that
FAA field offices provide have been raised in previous
reports,[Footnote 15] including FAA's evaluations of various designee
programs, which we discussed earlier in this report; by individuals we
interviewed during site visits; and by our expert panel. Table 3 shows
the top five oversight weaknesses identified by our experts. The top-
ranked weakness--inconsistent oversight by FAA offices--was identified
as a "great" or "very great" weakness by 36 of the 62 experts. No more
than 6 of the 62 experts felt that these top five factors posed "no
weakness" and between 5 and 13 other experts--believed these factors
presented "little" weakness. (See app. IV for additional weaknesses
identified by our expert panel.)
Table 3: Experts' Ranking of Top 5 Oversight Weaknesses:
Ranking: 1;
Weakness: FAA offices level of oversight and interpretation of rules
are inconsistent.
Ranking: 2;
Weakness: Inactive, unqualified, or poor performing designees are not
identified and removed expeditiously.
Ranking: 3;
Weakness: It is difficult to terminate poor performing designees.
Ranking: 4;
Weakness: Inadequate surveillance and oversight of designees.
Ranking: 5;
Weakness: FAA has not made oversight of designees a high enough
priority.
Source: GAO analysis of expert panel information.
Note: Rankings based on responses from 62 experts and the frequency of
responses indicating a "great" or "very great" weakness.
[End of table]
Designees and industry officials that we spoke with indicated that
FAA's level of oversight and interpretation of rules are inconsistent
among regions and among offices within a region. For example, several
designees whom we spoke to provided the example of one Aircraft
Certification field office that was stricter in its application of FAA
standards than other offices--i.e., the stricter office would not
approve submittals for supplemental type certificates that would be
approved by other FAA offices. As a result, applicants tend to "shop
around" to find those offices that will provide expedited approvals,
according to these designees. Another designee and an aviation parts
manufacturer told us that FAA field offices required different
paperwork and interpreted FAA rules differently for the same work. For
example, a manufacturer of fortified cockpit doors found that field
offices in Los Angeles and Seattle interpreted regulations differently
and required different paperwork to process the same type of approval.
Designated mechanic examiners that we spoke with provided similar
examples of inconsistencies among field offices. They cited instances
in which one field office would reject applications that another field
office would approve. Further, an industry representative that we spoke
with provided examples of inconsistencies among FAA offices concerning
whether approval in the form of a supplemental type certificate is
needed--with some offices requiring a supplemental type certificate and
other offices considering the same type of manufacturing or maintenance
work minor and requiring no approval. A designated engineering
representative noted that different FAA staff required different levels
of detail in the standard FAA form that engineering designees submit to
show their completed work. Another industry representative noted the
lack of standardized requirements for data submittals from certain
types of designees, such as designated engineering representatives. A
standardized checklist would help various FAA field offices to
consistently interpret regulations, according to the industry
representative. According to FAA officials, in certain cases, there are
reasons for inconsistent application of rules. For example, in the case
of cockpit doors, the projects typically varied across offices
depending on data submitted by previous applicants and the capability
of the applicant. In order to reduce unnecessary administrative burdens
on applicants, FAA's policy specifies that once an applicant had
demonstrated that a design change met FAA requirements, subsequent
applicants for a similar alteration may not be required to conduct all
the same tests required of the previous applicant, according to FAA
officials. Agency officials further stated that checklists are created
for each project and that standardized checklists cannot be used
because each project is unique. This was disputed by FAA staff that we
spoke with, some of whom had created standardized checklists to use for
all the designees that they oversaw.
We also found that, in some cases, the ability of FAA field offices to
oversee designees is affected by designees working outside of their
normal locality and the amount of written details about that work that
is provided to FAA. FAA policy allows designees to work outside of
their assigned geographic area but, in certain circumstances, requires
designees to notify the local FAA office.[Footnote 16] This situation
can occur, for example, when specialized engineering expertise is
needed by an aviation parts manufacturer and the closest designee with
that expertise is located in a remote FAA region; in which case, the
company may request the services of a designee from outside the region.
We spoke with one designated engineering representative based in
Atlanta who regularly worked outside his geographic area. In 2001, 7 of
12 projects that he approved as a designee were outside the Atlanta
area; in 2002, 20 of 33 projects were outside the area, and in 2003, 4
of 28 projects were outside the area. When he works out of his
geographic area, he normally contacts the field office where he is
conducting his work only after the work is completed and submits the
required paperwork to his FAA office in Atlanta upon completion of a
project. He and other designated engineering representatives told us
that they are likely to include minimal details in the forms submitted
to FAA because that information can be requested under the Freedom of
Information Act. An FAA engineer also told us that designated
engineering representatives may be reluctant to include details on how
they certify aviation products. Since FAA inspectors have little
opportunity to witness the work being performed by designees that work
outside their area, inspectors rely heavily on paperwork reviews. When
the paperwork provides insufficient details about the designees'
activities, FAA staff spend additional time requesting the needed
information from designees, according to an FAA engineer.
In addition, Flight Standards Service staff told us that more direction
and clarity was needed concerning the amount of surveillance that
inspectors should be conducting over designees. Policy guidance
describes how inspectors are to conduct surveillance of designees, and
the service develops a national workplan each year that determines the
number of inspections of designees that inspectors and engineers will
perform. Several FAA field office managers that we spoke with believed
that the oversight called for in the national workplan does not allow
them to target oversight to those designees that need more or less
surveillance. In addition, according to several FAA inspectors that we
spoke with, it is difficult during their site visits of designees to
identify those who are improperly certifying applicants or conducting
inappropriate activities, such as approving parts beyond their
authorization. The inspectors told us that they usually find out about
improper designee activities by noticing mistakes on the forms
submitted by designees and receiving complaints from designees'
clients. A designee that we spoke with further explained that, because
FAA visits are arranged in advance, designees have time to make sure
things are done correctly during the visit. Flight surgeons, by
comparison, are not required to conduct site visits of designees. Due
to limited number of staff and resources available to conduct site
visits, flight surgeons primarily conduct those visits only after
problems have been identified by others, such as complaints by clients.
We also found that field offices did not consistently follow
established policy for selecting designees. While we did not find
evidence that unqualified designees were selected, this situation may
result in not selecting the best qualified candidates. Nineteen of the
62 experts on our panel believed that FAA does not consistently follow
its own designee selection criteria[Footnote 17]--which are based on
designee candidates experience and qualifications, FAA field offices'
ability to oversee designees, and the need for particular types of
designees[Footnote 18]--but rather appoints designees based on personal
associations. Moreover, 9 of the 17 FAA inspectors and engineers on our
panel rated the practice of awarding delegation status based on
personal associations with FAA management as a "great" or "very great"
weakness of the designee programs. FAA policy requires multiple parties
to review applicant's qualifications and reach consensus on appointment
decisions, but we found that field offices sometimes add their own
criteria. For example, Flight Standards Service has established a
National Examiners Board to review all designee applications and
prepare a list of qualified candidates from which field office managers
must select designees. The board was established to provide an
objective, standardized process and to move away from the previous ad
hoc practices of appointing designees that were often based on
selecting personal acquaintances. However, we found that this process
does not always work as intended. For example, in a Flight Standards
field office that we visited, an applicant for designed airworthiness
representative is required to have a letter of recommendation from the
manager of the field office. According to an inspector at that field
office, this practice has resulted in screening out otherwise qualified
individuals. According to FAA officials, personal associations is an
important factor in selecting and appointing designees. They consider
personal knowledge and experience with the applicant an important
consideration in the selection process, without which it is difficult
to know whether applicants have the necessary qualifications and
abilities.
In addition, FAA's internal evaluations confirm our work that FAA
offices provide inconsistent oversight and interpretation of rules
concerning designees, which limits the assurance that the agency has
that the designees are performing certification work properly. For
example, as mentioned previously in this report, in 2000, an FAA
evaluation of designee pilot examiner oversight in one region found
that inspectors were not conducting oversight as required by agency
policy.[Footnote 19] That review further found that up to 30 percent of
the designated pilot examiners in the region were not conducting
complete practical tests of pilot certificate applicants and not
consistently holding pilot applicants to the standards of the practical
test.[Footnote 20] In addition, an FAA-industry study found that
project approvals by certain designated engineering representatives,
which do not require FAA review, combined with the lack of designee and
FAA technical expertise in certain specialized areas, have resulted in
designs that were deficient or not in compliance with FAA regulations.
[Footnote 21]
We also found that FAA offices do not always identify and remove
inactive or poor performing designees expeditiously, which may be due
to reluctance on the part of managers, engineers, and inspectors to
take disciplinary action. FAA policy calls for providing counseling,
remedial training, or limiting or terminating designees' authority for
insufficient work activity and poor performance. For example, since
1998, Aircraft Certification Service has terminated approximately 770
designees for such reasons as insufficient activity, lapse in
qualifications, or lack of care.[Footnote 22] However, a 2002 study
conducted jointly by FAA and industry found that it was the perception
of some FAA field staff who oversee designees that terminating
designees is difficult because of fear of litigation. According to the
report, this perception had resulted in little, if any, disciplinary
action being taken against designees when it may be warranted.[Footnote
23]
Our interviews with FAA field office managers and staff confirmed that
they are reluctant to take disciplinary action against designees. For
example, managers in the Seattle and Oklahoma City field offices and
inspectors and engineers in the Atlanta and Los Angeles field offices
told us that rather than take disciplinary action against poor
performing designees, they wait and terminate the designee during the
renewal process, as long as designees have not committed any criminal
acts. According to these officials, FAA field offices prefer to not
renew poor performing designees rather than terminate them because FAA
management wants to avoid legal appeals that designees can make if the
agency decides to terminate them for poor performances. According to
FAA field inspectors that we spoke with, it is difficult for them to
terminate poor performing designees--such as those who continue to omit
information in their documented work despite training and counseling--
because the process is lengthy and time-consuming. According to one FAA
engineer, when she tried to remove a designated engineering
representative for making incorrect approvals, she was required by FAA
policy to first notify the designee of FAA's intent to terminate the
designation, and then to document the specific reasons for the
recommended removal. The process took 2 to 3 years, according to the
inspector. After designees are removed, they are allowed up to two
appeals, which can further lengthen the removal process. FAA officials
acknowledged that misunderstandings of the removal process among
inspector staff will continue without the development of specific
guidance and training on the designee termination process. Our analysis
of data from the Aircraft Certification Service found that the office
terminated 15 designees because of "lack of care or judgment" and
terminated 121 by not renewing their designations over the last 5
years.
In addition, FAA field and program office managers have some discretion
over terminating poor performing and inactive designees, but because
FAA's criteria for terminating designees is not specifically defined,
each field and program office determines when poor performance or lack
of activity constitutes grounds for termination. According to a manager
in FAA's Civil Aerospace Medical Institute, one region may terminate an
aviation medical examiner who is consistently more than 30 days late in
transmitting medical certification data, while another region may
terminate an aviation medical examiner who is consistently more than 60
days late. An FAA engineer told us that designees in the Aircraft
Certification Service are seldom terminated because of low activity
level. Of the approximately 770 designees for Aircraft Certification
Service that were terminated since 1998, according to information we
analyzed in FAA's Designee Information Network database, about 230 (30
percent) were terminated for inactivity. In addition, a manager for a
Flight Standards field office told us that the criteria the office uses
for terminating poor performing designees include whether the
termination will result in a loss of income for individual designees.
This criterion is not included in FAA policy nor considered by other
field or program officials with whom we spoke.
Consistent application of oversight policies is important to ensure
that designees follow FAA policies and that they remain free from
pressures from employers or clients that may lead them to bypass those
policies. For example, in 1999, FAA found that designated mechanics'
examiners in the Orlando, Florida, area had not adhered to FAA's
standards and had fraudulently indicated that hundreds of mechanic
applicants had passed the certification examination. This resulted in
FAA retesting many of the mechanics. In addition, some designated
engineering representatives are salaried employees of the manufacturers
whose products they are approving on behalf of FAA. In one case, a
designee told us that another designated engineering representative was
an executive officer of the company whose products he was approving,
creating an apparent conflict of interest. The designee also told us
that designees are under pressure by their employers to certify
products. He stated that designated manufacturing inspection
representatives are sometimes pressured by their employers to approve
aviation products for export, under the threat of being fired.
According to FAA officials, agency policy discourages the appointment
of designated engineering representatives who are executives within a
company where the primary job duties are schedule-driven and devoted to
the output of the company's whole saleable products. Other designees,
such as designated pilot examiners, are employed by flight schools and
test pilot applicants for those schools. Since those designees depend
upon the flight school for employment and referral of applicants, there
could be an incentive for the designated pilot examiner to compromise
the integrity of pilot tests. Such situations present the potential
risk that designees may be pressured by employers to bypass FAA
requirements in order to meet schedules or attract additional students.
FAA officials acknowledge that an inherent conflict of interest exists
in the designee programs, but did not view it as a weakness because
designees can be held liable for deficiencies in their work. However,
concerns were expressed to us by several FAA field managers and
inspectors, that smaller organizations, such as repair shops, may be
willing to risk liability and bypass agency requirements.
Poor Data, FAA Staff Workload, and Insufficient Training for FAA Staff
May Contribute to Oversight Weaknesses:
FAA's oversight of its field offices and designees is hampered by the
lack of comprehensive information in some of the agency's databases
that are used to capture information on designees,[Footnote 24] the
workload demands facing FAA staff who oversee designees, and
insufficient training for FAA staff on designee oversight.
Designee Databases:
The databases for the offices of Flight Standards Service and Aircraft
Certification Service were not designed to capture information
concerning oversight performed by the managing offices and do not
provide a comprehensive picture of FAA engineers' and inspectors'
oversight activities or the activity levels of designees. For example,
FAA policies require FAA inspectors in the Aircraft Certification
Service who oversee manufacturing designees to update the designee
management database, Designee Information Network, every time they
oversee or monitor a designee's performance. However, no data field is
provided to capture information on these oversight visits. A field for
comments is available for FAA staff to indicate when a designee
performance evaluation was conducted, but our review of the data files
for 1998 through 2003, found that this information was not consistently
noted. Moreover, FAA policy does not require its engineers to document
their oversight of engineering designees in the database. Thus, FAA
cannot readily ascertain how often staff in the Aircraft Certification
Service monitored and evaluated designees, other than the minimum
levels required to renew the designees' authority.[Footnote 25]
According to officials in that office, information on how often staff
review designee performance is recorded in designees' paper case files,
which are maintained at the field or program offices. In addition, the
Designee Information Network does not contain information on the number
and type of approvals that the individual designees are conducting. As
a result, FAA lacks a single, comprehensive data source that could be
used to facilitate designee oversight by providing FAA a means to
prioritize oversight activities and engineer workload. According to FAA
officials, the fact that all oversight information is not captured in a
single database does not directly affect the agency's ability to
effectively oversee designees.
Two other databases--the Program Tracking and Reporting Subsystem
(PTRS) and National Vital Information Subsystem (NVIS)--used by Flight
Standards Service inspectors to monitor designees also do not
completely track designees' activity level. According to FAA officials,
PTRS was designed to track activities by FAA inspectors, such as noting
when FAA inspectors conducted surveillances of designees, while NVIS
was developed to track basic profile information on designees, such as
their names, addresses, types of certification, designated
authorizations, and status. PTRS can be used to track the activity
levels of designees; however, it requires the FAA inspector to input
the data each time they receive a certification package from a
designee, but past reviews have found problems with incomplete
information in the database. For example, in 2003, the Quality
Assurance Team mentioned earlier found that required information on
designee oversight was missing from the two databases or incorrect. The
team noted records that would indicate the type of surveillance
conducted (such as an observation of a complete or partial test) were
missing from PTRS and records in NVIS that lacked renewal dates and
contained inaccurate information on designee training and
authorizations. By comparison, Aerospace Medicine has one database--
Airmen Medical Certification Information Subsystem--to track
information on aviation medical examiners, including information on the
number of medical certificates issued by each medical examiner and
demographic, training, and oversight information for each designee. Our
review of that database found reasonably complete information; we did
not check the accuracy of the information.
FAA Staff Workload:
FAA's oversight of the designee programs may also be weak, in part,
because of the workload demands facing agency staff who oversee
designees. In addition, the amount of time that FAA staff spend on
other aviation safety activities, such as monitoring air carrier
operations, affects the amount of time spent on designee oversight. FAA
policy recognizes that each designee oversight scenario is unique and
allows variations in determining the extent of oversight needed to meet
minimum annual requirements. FAA policy also states that the ability to
provide adequate oversight depends on balancing the level of FAA
staffing to the agency's workload and the number of designees. FAA
policy, however, does not specify an acceptable workload for meeting
this criterion. For example, each managing office must periodically
verify adequate FAA staffing numbers based on the type and amount of
the work performed by staff who oversee designees. FAA policy provides
no further guidance for determining adequate numbers for proper
oversight. FAA officials stated that the level of specificity in the
guidance is adequate for determining staff workload with designees and
that it would be difficult to determine an exact staffing ratio because
of factors such as the size of facilities, the experience of designees,
and the complexity of projects. However, the lack of clear policy
guidance and staffing standards results in wide variation in the ratio
of designees to FAA staff among offices and programs and makes it
difficult for the agency to measure and account for its staff
resources. For example, our review of FAA data showed that, on average,
the ratio of designees to FAA staff is about 6 to 1 in the Aircraft
Certification Service, about 5 to 1 in Flight Standards, and about 440
to 1 in Aerospace Medicine. The ratios for individual FAA staff ranged
from 1 designee to 1 FAA staff in several Aircraft Certification
offices to about 870 designees to 1 FAA staff in Aerospace Medicine.
Information we gathered from site visits at three of FAA's nine regions
also showed a wide range of workload ratios. For example, information
we gathered at Flight Standard's Northwest Mountain Region showed
ratios among field offices ranging from 1 designee to 1 inspector to
100 designees to 1 inspector. Variations in the ratios of designees to
FAA staff are due to the type of designee and the complexity of their
work, according to FAA officials. However, several engineers in the
Aircraft Certification Service with whom we spoke expressed concerns
that a designee to staff ratio higher than 10 to 1 limits the time they
have to adequately monitor the work performed by designees. One
Aircraft Certification engineer told us that while he was currently
responsible for overseeing 10 designated engineering representatives,
in the past, he had been responsible for between 30 and 60 designees,
which was too many to adequately oversee. Flight Standards Service
officials acknowledged that staffing standards need to be established.
The National Academy of Sciences is currently evaluating the staffing
standards for the office of Regulation and Certification, which
encompasses Flight Standards, Aircraft Certification, and Aerospace
Medicine, and expects to complete the study in 2005.
Past reports by us and others pointed out that escalating workloads
and/or high turnover rates for FAA staff continue to diminish FAA's
ability to oversee designees. For example, over 10 years ago, we
reported that, in response to a dramatically escalating workload, FAA
had delegated aircraft certification duties to designees without
defining a clear role for its staff to ensure that they were
effectively involved in the certification process.[Footnote 26] Since
then, FAA has issued comprehensive policies governing the selection,
appointment, and oversight of individual and organizational designees.
We also pointed out high turnover rates (107 percent over the previous
10 years) for FAA engineers who oversee designees. In addition,
internal FAA documents from 2000 cited the disparity between the
agency's Aircraft Certification Service' workload and its staffing
levels, noting that staff resources have not kept pace with increasing
workload. To update the information in our 1993 report, a 2002 study
prepared for FAA confirmed that the two FAA field offices--Seattle and
Los Angeles--responsible for the majority of commercial transport
airplane oversight still had high turnover rates (115 percent over an
8-year period) and that over 50 percent of the engineers in those
offices had less than 5 years of FAA experience.[Footnote 27] The
report further noted that the consistently high turnover rate and
associated low experience levels were indicators of the limited time
available for FAA engineers to acquire the necessary experience and to
understand the increasingly complex systems and human factors
associated with modern aircraft, which are among the skills needed to
oversee the work of designees. FAA noted that the annual turnover rate
of engineers at the Seattle and Los Angeles field offices had declined
in recent years, indicating that from fiscal years 1999 through 2004,
the average annual rates were 3 percent and 4 percent, respectively for
the two offices.[Footnote 28]
In addition, designees told us that FAA staff who oversee designated
engineering representatives change frequently. A designee monitored by
a Seattle field office told us that he estimated that every 3 years he
reported to a different FAA staff person. Another designee told us that
in the last 5 years, he had reported to six different FAA staff. As a
result of frequent changes in FAA staff, the designees felt frustrated
in the amount of time that it took to establish a good working
relationship with each new FAA staff person. We found a similar
situation in an Atlanta field office, where an FAA engineer explained
that high turnover of engineers in the office made it difficult to
oversee the activities of designated engineering representatives. The
difficulty arises, according to the FAA engineer, because designees
typically submit forms at the end of each quarter to document their
activities, which FAA engineers then review. When a designee's FAA
advisor changes during a quarter, the only information that the new
advisor has concerning the designee's work is the information contained
in the form, because the new advisor does not have information
concerning discussions between the prior FAA staff person and the
designee. Furthermore, as we mentioned earlier in this report, both an
FAA engineer and designated engineering representatives told us that
designated engineering representatives are reluctant to include details
on how they certified a product, fearing that the information could be
requested and made public under the Freedom of Information Act.
FAA provided us with information on how the size of its workforce has
changed over time in comparison with the number of designees they
oversee for some designee programs. For example, based on FAA's
staffing information, the number of designees overseen by engineers and
inspectors in the Aircraft Certification Service decreased slightly
from 6.7 to 1 in fiscal year 1999 to 6.5 to 1 in fiscal year 2003.
However, FAA could not provide similar information for Flight Standards
Service or Aerospace Medicine to determine how the agency's workforce
has changed over time in comparison to designees. Some members of our
expert panel commented that the number of FAA staff who oversee
designees has not increased at the same rate the aviation industry has
grown. Experts also stated that FAA staff do not have time to provide
adequate oversight of designees for whom they are responsible for
overseeing. Additionally, FAA inspectors and engineers that we spoke
with commented that as FAA's dependence on designees continues to
increase, their ability to conduct oversight--consisting of designee
supervision, monitoring, and tracking, as required by FAA policy--will
continue to decrease. According to some FAA engineers that we spoke
with, dramatic increases in their workload has resulted in their
ability to review only a minimal percentage of work conducted by
designees.
The situation in Aerospace Medicine provides another example of
workload issues potentially hampering oversight. Between July 2002 and
June 2003, the nine regional flight surgeons in Aerospace Medicine each
headed a team of about three or four FAA staff and monitored over 4,900
designated medical examiners, who conducted more than 420,000 medical
examinations. Given high workload demands on the flight surgeons and
their staff, in many cases, they are not able to perform site
inspections to ensure that designee offices and facilities meet FAA
standards, according to Aerospace Medicine officials. These officials
also noted that site visits would help FAA ensure that designees are in
compliance with FAA's facility and equipment requirements, such as
verifying that the designees have access to acceptable facilities to
perform physical examinations, meet minimum vision and hearing test
equipment standards, and have access to approved diagnostic
instruments. According to regional flight surgeons, due to the limited
number of staff and resources available to conduct site visits, they
primarily conduct those visits only after problems arise due to
unprofessional behavior or unethical practices on the part of the
designated examiners. Such questionable designee practices are brought
to the attention of regional flight surgeons by the Civil Aerospace
Medical Institute, FAA field staff, and through complaints by the
designees' clients. According to FAA officials, limited resources also
hinder the flight surgeons' ability to identify unprofessional or
unethical designated medical examiners.
Inspectors in Flight Standards also told us of workload demands
affecting designee oversight. For instance, one FAA inspector provided
an example of a designated pilot examiner who conducted approximately
400 practical tests in 1 year. FAA policy calls for inspectors to
conduct one annual inspection of each designated pilot examiner and to
carry out additional surveillance of pilot examiners who perform more
than 50 practical tests per quarter. Because of high workload, the
inspector was only able to conduct one annual inspection of the
designee with high activity and was not able to conduct the required
additional surveillance.
The ability of FAA staff to oversee designees is also affected by the
amount of time that they spend on a wide variety of other aviation
safety activities and the priorities that are given to the various
activities. For example, FAA officials from Flight Standards Service
commented that inspectors are also responsible for other activities
such as taking enforcement actions, evaluating air carrier operations,
monitoring general aviation activities, and conducting accident
investigations. Several FAA engineers that we spoke with said that
their first work priority was to conduct accident investigations and
draft airworthiness directives; their second priority was to draft
policy and regulations; and their third priority was designee
oversight. FAA staff that we interviewed estimated that they spend
about 5 to 15 percent of their time overseeing designees, depending
largely on the number of designees for whom they are responsible.
According to one estimate by an FAA engineer who is responsible for
overseeing 25 designees in the Aircraft Certification Service,
approximately 10 percent of his time--or about 4 hours per week--is
devoted to designee oversight. Inspectors and engineers also pointed
out that poor-performing designees can significantly increase their
workload as they require greater surveillance and more frequent
interactions.
Training for FAA Staff:
FAA's oversight of the designee programs may also be weak, in part,
because of insufficient training for staff who oversee designees.
Twenty-one of the 62 experts on our panel cited a lack of training in
designee oversight for FAA inspectors and engineers as a "great" or
"very great" weakness of the designee programs. Six out of 15 FAA
inspectors or engineers on our expert panel considered this situation
to be a "great" or "very great" weakness. (Six experts felt the lack of
training was not a weakness, and 6 other experts felt it posed little
weakness.) Flight Standards Service officials acknowledged that
additional oversight training would be helpful to address training
weaknesses.
FAA's Aircraft Certification Service and Aerospace Medicine have
established initial training requirements for newly hired staff, which
include courses on designee oversight. For example, the Aircraft
Certification Service requires staff to take the Delegation Management
Job Functions Course, which focuses on overseeing designees and is
designed to teach the skills necessary to select, supervise, and
terminate designees. FAA's Aerospace Medicine requires regional flight
surgeons to take initial training on policies and regulations
pertaining to designees. Aerospace Medicine staff who assist flight
surgeons do not receive initial training concerning designees, but
periodically attend training at the Civil Aerospace Medical Institute
in Oklahoma City or are informed of relevant policy changes through
teleconferences, according to officials in the office. By comparison,
Flight Standards Service does not provide initial training to its
inspectors on designee oversight. Instead, this office requires new
inspectors to attend initial training in their areas of specialization.
Flight Standards is currently evaluating the Delegation Management
Course used by Aircraft Certification to determine if the course meets
inspectors' needs for overseeing designees, according to several
officials in Flight Standards.
Once inspectors and engineers in Flight Standards and Aircraft
Certification services have fulfilled their initial training
requirements, they are encouraged, but not required, to participate in
refresher training. In contrast, FAA requires designees to receive
formal refresher training every 2 or 3 years. By not requiring its
oversight staff to take refresher training, FAA cannot maintain
reasonable assurance that its inspectors and engineers stay current on
changes to policies and procedures. In fact, one FAA manager told us
that, in his office, FAA engineers who oversee designees needed
additional training, especially in the area of managing designees. In
addition, several experts on our panel stated that, given the disparity
in training requirements, it would be possible that designees could
gain a better knowledge of FAA's policies and procedures than the FAA
staff who oversee them. FAA officials stated that inspectors and
engineers receive training through workshops, video training sessions,
and FAA academy training. However, they do not receive refresher
training, which is required for designees. This is in contrast to
regional flight surgeons, who are required to attend refresher training
every 3 years, which is the same training required for designees.
Additionally, previous recommendations for improving inspector
training have not been implemented. For example, as mentioned
previously in this report, in 2000, FAA found that inspectors in field
offices in the Southwest Region were not reviewing designated pilot
examiners' work on an annual basis and conducting oversight as
required. The report recommended that the Southwest Region conduct
standardized initial and refresher training for FAA inspectors,
supervisors, and managers on the agency's oversight policies and
procedures pertaining to designated pilot examiners. In response to the
recommendation, the region implemented a training course that included
briefings at each field office to raise the awareness of FAA inspectors
concerning the importance of designee oversight, to explain current
policy, and to offer techniques for effective oversight. The region
also used the briefings as the basis of curriculum for new training
courses for FAA inspectors and has recommended that such courses be
made available for all Flight Standards Service inspectors nationwide.
According to agency officials, FAA plans to implement a national policy
based on this recommendation in October 2004 and expects the policy to
be implemented by 2005.
FAA Has Potential Opportunities to Improve Designee Programs:
Experts on our panel, best practices within FAA, and practices adopted
by other countries in administering their respective designee programs,
including experiences in implementing organizational delegation
systems, suggest that there are potential opportunities for FAA to
improve (1) program oversight to ensure consistent compliance with
existing policies by FAA inspectors, engineers, and flight surgeons and
(2) the accuracy and comprehensiveness of computerized information on
designees so that the databases can be more useful tools for designee
oversight. Given the central importance that the designee programs hold
for FAA and future agency plans to expand the use of organizational
designees with the creation of the ODA program, FAA has incentives to
carefully consider such opportunities.
Several Opportunities Identified to Improve Oversight of Designee
Programs:
Our work indicated that additional opportunities exist to improve FAA's
oversight of its designee programs to ensure consistent compliance with
existing policies by FAA inspectors, engineers, and flight surgeons.
For example, our expert panel offered a number of suggestions to
improve the designee programs that address some of the weaknesses we
identified, including improvements in selecting and terminating
designees and ensuring that FAA staff who oversee designees are
knowledgeable about FAA policy. In addition, many experts agreed that
it was important for FAA to hold designees accountable for their
findings.[Footnote 29] For example, one expert pointed out that the
designated engineering representative and organizational designee
programs should be overhauled so that the designees are responsible and
accountable for certifications and that FAA needed to put in place a
process to monitor that additional responsibility. An FAA official told
us that accountability is a central part of their designee programs,
since failure to perform delegated functions in accordance with agency
standards and expectations will result in removal of the delegation. In
addition, all of the experts on our panel indicated that it was
important for FAA to conduct audits of existing designee programs to
determine if field offices are providing adequate oversight.[Footnote
30] As we mentioned previously in this report, FAA has audited only 6
of its 18 designee programs. Table 4 lists the top ranked actions in
terms of importance and feasibility identified by the experts; these
actions were identified as "high" or "highest" in importance and
feasibility for implementation by most of our experts. Appendix IV
provides a complete list of suggestions made by our expert panel.
Table 4: Experts' Ranking of Top Ways to Improve FAA's Designee
Programs:
Ranking: 1;
Suggested improvement: Hold designees accountable for their findings.
Ranking: 2;
Suggested improvement: Ensure that FAA employees who oversee designees
are knowledgeable about the regulations, policies, and processes
applicable to the designees' particular specialization.
Ranking: 3;
Suggested improvement: Select designees according to their
qualifications and experience rather than on personal associations
with FAA managers.
Ranking: 4;
Suggested improvement: Clearly define and consistently follow the
criteria for selecting designees.
Ranking: 5;
Suggested improvement: Increase penalties (including the ability to
terminate their status as designees) for individual and organizational
designees found to violate standards or who do not exercise proper
judgment.
Source: GAO analysis of expert panel information.
Note: Rankings based on responses from 62 experts and the frequency of
responses indicating a "high" or "highest" importance to implement.
[End of table]
Consistent evaluation and monitoring of designee activities is crucial
to hold designees accountable for their findings, and some FAA offices
have best practices that may be broadly applicable across the designee
programs. For example, as we discussed earlier in this report, FAA's
internal review of pilot examiners in the Southwest Region was
implemented to determine whether the designees in the region were
conducting valid practical tests of general aviation pilot applicants
and to determine the quality of FAA oversight provided by field offices
in the region. Findings from the internal review were based on a
comprehensive statistical analysis of pilot examiners' activities in
the region, a survey of newly certified private pilots in the region,
audits of pilot examiner files, surveillance reports from FAA
inspectors, and interviews with field office managers and staff. The
review provided a reasonable method to assess program outcomes,
identify the root causes of the lack of compliance with agency policy,
and develop corrective action plans to address the root causes. FAA's
Organization Effectiveness Branch Manager commented that the
methodology for the internal review was reliable, and suggested that
the review was informative for developing regional policy. The Branch
Manager also commented that in order to address FAA national policy, a
national survey would be necessary. Flight Standards Service has not
expanded its use of this methodology to other regions or to other
designee programs.
Canada's practice of systematically evaluating and/or monitoring its
designee programs provides additional examples of opportunities for
improving FAA's oversight of its organizational designee programs and
its plans to implement ODA. Transport Canada oversees both individual
and organizational designees (which are called "delegates"), and
focuses on aircraft design and design modifications.[Footnote 31]
Transport Canada oversees delegates using regional offices and
headquarters staff, similar to FAA. FAA, however, oversees a much
larger number of designees. For example, Canada has approximately 760
aviation medical examiners and 80,000 pilots, while the United States
has about 5,000 aviation medical examiners and about 630,000 pilots.
Transport Canada has implemented a policy to provide a consistent and
standard approach for conducting safety oversight of its organizational
delegates, which includes conducting audits of delegated organizations
on a cycle ranging from 6 to 36 months--an initial audit within 6
months of certification and comprehensive follow-up audits on a
recurring basis. They have also established a centralized
standardization office to ensure that field offices are consistently
interpreting rules and procedures. The centralized office evaluates and
approves technical submissions from applicants and delegated
organizations to determine compliance with regulations. The office is
also responsible for the development, coordination, and implementation
of a national audit plan in auditing delegated organizations. By
comparison, FAA policy calls for conducting annual inspections, and
procedural audits and technical evaluations every 2 years. Annual
inspections focus on a review of the system that the delegated
organization has in place to perform the delegated functions and a
review of the activities conducted by individuals. As mentioned
previously, FAA conducted 10 technical evaluations (out of 49) as of
June 2004. According to FAA, it has established centralized offices
responsible for standardization of policies. However, our work has
shown that FAA field offices do not implement policies in a standard
manner, as discussed earlier in this report.
Transport Canada's experiences in developing an organizational
delegation system in the late-1980s also provide relevant lessons for
FAA as it begins developing the ODA program. According to the Chief of
Delegation and Quality Divisions in Transport Canada, an inconsistent
level of oversight was a major challenge that Transport Canada faced as
it implemented its organizational delegation system. To address this
challenge, the agency established a centralized standardization office
to ensure that field offices are consistently interpreting rules and
procedures. Based on this experience, the Transport Canada official
told us that FAA needs to plan for the inconsistencies that will arise
during the implementation of the ODA program. The larger size of the
U.S. designee programs increases the likelihood that the level of
oversight will be inconsistent, according to the Transport Canada
official. Moreover, the official also commented that, in hindsight,
they should have developed and conducted an audit of organizational
delegates early in the implementation process. The Canadian official
told us that Transport Canada did not conduct audits early on because
staff were preoccupied with reviewing and approving organizations'
procedures manuals. Transport Canada's quality assurance review later
determined that they were not doing audits of organizational delegates
on time, nor conducting audit follow-ups, which contributed to
inconsistent oversight.
Database Monitoring Performance of Aviation Medical Examiners Could Be
Model for Other Designee Databases:
Accurate, comprehensive information on designee activities is an
important prerequisite for designee oversight and is integral to
monitoring and evaluating the programs. The Airmen Medical
Certification Information Subsystem--a database used by FAA's Office of
Aerospace Medicine to monitor the performance of aviation medical
examiners--provides a model for the other designee programs. Although
this database was designed to simplify the processing of airmen medical
certification information, Aerospace Medicine also uses it as a tool to
oversee aviation medical examiner designees and monitor their activity
levels. For instance, regional FAA flight surgeons use information from
the database to determine if they need to more closely monitor aviation
medical examiners with high activity levels or to determine how long it
takes to transmit medical information to FAA. Each flight surgeon is
periodically provided performance data for their designees that include
the number of medical certificates issued by each designee, the number
of errors found in those certificates, and the number of accidents and
incidents involving pilots that received medical certificates from
designated medical examiners, according to an Aerospace Medicine
official. Additionally, according to FAA officials, regional flight
surgeons also use data from the database and link it with the Airmen
Registry to determine the region where FAA needs additional examiners.
Applying this model to Flight Standards Service and Aircraft
Certification Service would provide those offices and inspectors and
engineers with more detailed performance information on designees and
provide a foundation for more consistent oversight of the numerous
designee programs.
FAA officials agreed that improvements were needed to these databases,
but expressed a concern that it would cost $50 million to make
upgrades, which may have implications for other safety programs that
would then receive less funding. Such concerns might be addressed by
looking for ways to share the costs of the designee programs with the
aviation industry, similar to other federal agencies that charge user
fees to process applications for approvals or licenses. For instance,
the Federal Drug Administration charges pharmaceutical companies
application fees to recover the cost of the agency's review of new
drugs.[Footnote 32] As another example, U.S. Customs and Border
Protection charges fees to brokers--private individuals and companies
that are licensed and regulated by the agency to aid importers and
exporters in moving merchandise through Customs. Brokers pay Customs a
$100 permit application fee and a $125 annual user fee. FAA does not
charge designees an initial application fee or a renewal fee, which
could help recover the cost of processing these applications, because
it has been prohibited in law from promulgating new user fees since
1997.[Footnote 33] Moreover, designees charge companies and the general
public fees to have a product certified or to perform a pilot practical
test. Some designees earn up to $60,000 or more a year and have made
designated activities their sole source of income. FAA inspectors,
engineers, and flight surgeons, on the other hand, provide the same
service free as a function of their government employment. In prior
reports, we have stated our belief that, to the extent possible,
commercial users of the aviation system should pay their share of the
costs that they impose on the system.[Footnote 34] Charging fees to
designees to offset the cost to FAA to administer the designee programs
is an analogous situation.
Conclusions:
Designees perform a valuable function for FAA and the aviation
industry, enabling FAA to leverage its staff resources and industry to
obtain FAA-issued certificates in a timely manner. By using designees,
however, FAA places great trust in the integrity and honesty of
designees to adhere to the same requirements, instructions, and
procedures as FAA staff do; therefore, periodic validation and
consistent oversight by FAA staff is necessary to ensure that such
trust is well placed. To date, FAA has not ensured that the oversight
process for its many designee programs is implemented consistently by
different field offices. While we did not find systematic safety
problems associated with FAA's oversight of designees, the agency's
inconsistent oversight limits its assurance that the designees' work is
performed in accordance with the agency's standards and policies. We
found examples of weaknesses in FAA's designee programs--such as
inspectors with too great a workload to conduct required surveillance
of designees--that underscore the need for FAA to ensure that its staff
are consistently following agency policy concerning designee oversight
and to validate those policies and their application by periodic
evaluations. However, FAA has evaluated only 6 of its 18 designee
programs to date. Our study indicated that reasons for FAA's
inconsistent oversight may include limitations on designee data that
FAA maintains, along with heavy workload, and potentially inadequate
training for FAA staff overseeing designees. FAA lacks a comprehensive
information system to effectively monitor and oversee the thousands of
activities performed by designees. Without such information, FAA
management cannot readily determine whether its field staff is
overseeing designees according to policy nor whether designees are
performing according to FAA's standards. Heavy workload for FAA staff
responsible for overseeing designees might preclude thorough
assessment--or any assessment--of some designees' performance.
Finally, by not requiring refresher training for FAA staff, the agency
increases the risk that staff do not retain the information, skills,
and competencies required to perform their oversight responsibilities.
Potential opportunities exist for FAA to address these weaknesses and
provide more consistent oversight of the designee programs by expanding
the use of existing agency practices, such as the Office of Aerospace
Medicine's practice of maintaining information on aviation medical
examiners performance and activity levels and using that information in
conjunction with designee oversight. Charging application and renewal
fees to designees to help offset the cost of administering these
programs would be in line with practices by other agencies and prior
GAO reports on cost-sharing with the aviation industry. However, FAA is
prohibited, by law, from imposing new user fees unless they are
specifically authorized by law.
It is especially important for FAA to consider ways to improve the
oversight of its designee programs as the agency moves forward with the
organization designation authorization program, which would expand the
number and types of organizational designees and further transform
FAA's role to that of monitoring the performance of others. Moreover,
concerns have been raised that under the proposed program FAA would
provide less specific and less technical oversight of the new
organizational designees than under the current program. Expanding the
use of good oversight practices already used within FAA for some
designee programs and examining lessons that may be learned from
Canada's oversight of organizational designees and efforts suggested by
our expert panel, would increase FAA's assurance that its designees are
meeting FAA safety standards and that any future changes to the
designee programs maintain those standards.
Recommendations for Executive Action:
To improve management control of the designee programs, and thus
increase assurance that designees meet FAA's performance standards, GAO
recommends that the Secretary of Transportation direct the FAA
Administrator to take the following three actions:
1. Establish a program to evaluate all designee programs, placing a
priority on those 12 programs that have not been evaluated. At a
minimum, the evaluations should examine field office compliance with
existing policies, identify root causes of noncompliance with those
policies, and establish and monitor corrective action plans.
2. Develop mechanisms to improve the compliance of FAA program and
field offices with existing policies concerning designee oversight. The
mechanisms should include additional training for staff who directly
oversee designees. As part of this effort, FAA should identify best
oversight practices that can be shared by all FAA program and field
offices and lessons learned from the program evaluations and
incorporate, as appropriate, suggestions from our expert panel.
3. Enhance the effectiveness of FAA designee oversight tools, such as
databases, by improving the consistency and completeness of information
on designees' activities and performance and FAA oversight. To the
extent necessary, FAA should examine charging fees to designees to help
pay for the costs of such efforts. If FAA identifies a need for such
fees, the agency should request the Congress to authorize them.
Agency Comments:
We provided a draft of this report to DOT for review and comment. FAA's
Deputy Associate Administrator for Regulation and Certification and
other DOT officials provided oral comments. DOT generally agreed with
our recommendations and acknowledged that automating the data
concerning oversight of designees and enhancing training for FAA
employees who oversee designees are useful steps to enhance the
programs. The department also provided clarifying comments and
technical corrections, which we incorporated as appropriate. In
addition, the department noted that designee programs have been a
cornerstone of aviation safety for 50 years. The constantly improving
level of safety in the U.S. aviation system is due, in no small
measure, to the professional performance of the thousands of designees
who evaluate aircraft designs, assess pilot capability, or conduct the
myriad of other reviews designees perform, according to DOT. DOT also
pointed out that statistics and data show that every day of the year,
the pilots and aircraft that pass through these designee systems fly
safely from departure to destination.
However, DOT officials expressed concern about the use of the Delphi
method in our review of 18 different programs with nearly 14,000
designees. First, they emphasized that, at best, the Delphi method
provides a means to consolidate and prioritize expert opinion, but even
under the best of circumstances, the results are opinion, not
necessarily factual data. The use of Delphi was further complicated in
this particular case, according to DOT, by the span of knowledge that
would be necessary to be considered an "expert" on designees when the
scope of expertise runs from aviation medicine, to aircraft engineering
and production methods, to parachute rigging. They stated that no
individual could be considered an expert in all the programs, and the
solicitation of opinions from the panel of experts would reflect the
specific experience of each individual--but could not be considered a
general statement of the strengths or weaknesses of all the programs.
By consolidating the responses from individuals with expertise from
these diverse fields, the officials questioned whether the results
could be useful for guiding decisions to improve any of the individual
designee programs. Further, the DOT officials cautioned that the Delphi
results should be carefully qualified in the final report, along with
explicit statements about the limitations on the use of the
information.
We disagree with DOT's characterization of our use of the Delphi
method; furthermore, we believe we used this methodology (which is
described in detail in app. I) appropriately. In particular, we used a
"modified" version of the Delphi method in order to compensate for some
the limitations inherent in the Delphi method as well as to adapt the
method to the specific needs of this engagement. For example, we
created a Web-based panel that allowed us to include many more experts
than had we convened a live panel. In addition, the Web-based panel
allowed us to keep the experts' identities anonymous, minimizing the
biasing effects often associated with live group discussions. We also
carefully selected the experts starting with a list provided by FAA,
and took into consideration that not all of the panelists would possess
expertise in many of the designee programs. To help adjust for that
fact, during the first round of questions, we asked experts to indicate
if their responses referred only to specific designee programs and, in
a few cases, experts indicated such. During the second round, the
experts were given the choice of responding to each question that they
did not know or had no opinion. In short, while DOT criticizes the
responses from the experts as "opinions," we believe the responses are
more appropriately characterized as carefully considered judgments of
systematically selected experts. Lastly, as described below, the report
only focuses on issues that were identified by the panel and other
sources.
Second, in reviewing a draft of this report, DOT officials expressed
concern about the way the Delphi results had been presented. They
emphasized, for example, that while the draft mentioned the number of
respondents who considered a factor a "great" or "very great" weakness,
the draft should also state the number who considered a factor "no"
weakness or of "little" weakness. Presenting what DOT considers both
ends of the response spectrum in the body of the report would allow a
full understanding of the results, according to the department. We
agreed that the number of experts responding "no" and "little" should
also be presented whenever the responses to individual questions were
mentioned in the report, and we revised the report accordingly.
Finally, DOT officials emphasized the need to consider, what they
called, the "totality" of the questions and responses in order to
evaluate any inconsistencies among responses. For example, they said
that while our report uses the responses from a single question to
indicate concern regarding the selection process for designees, the
responses from other questions could be interpreted to conclude that
there was little concern about the competency of the designees that
were selected or the quality of their work. Taken together, these
officials felt that these responses in total present a different
perspective on the outcome of the designee selection process than the
first question alone. DOT officials stated that our highlighting the
responses to one question and not balancing it with the results of
others, presents an incomplete picture of the panel's overall findings
and could mislead those who read the report but do not look at the
details in appendix IV. We disagree with DOT's characterization of our
analysis. First, we considered all responses from the expert panel and
provided them in their entirety in the appendix. Furthermore, for the
body of the report, we only focus on issues that were identified by
multiple sources. For example, the report highlights the issue of
selecting designees based on personal association because it was
identified by other sources during our field work and our review of
prior evaluations of the designee programs. Other issues raised by some
of the panel experts concerning the selection process were not
identified by other work we conducted and, therefore, not highlighted
in the report.
As agreed with your office, unless you publicly announce the contents
of this report earlier, we plan no further distribution until 21 days
from the report date. At that time, we will send copies of this report
to interested congressional committees, the Secretary of
Transportation, and the Administrator, FAA. We will also make copies
available to others upon request. In addition, the report will be
available at no charge on the GAO Web site at
[Hyperlink, http://www.gao.gov].
Please call me at (202) 512-2834 if you or your staff have any
questions concerning this report. Major contributors to this report are
listed in appendix V.
Sincerely yours,
Signed by:
JayEtta Z. Hecker:
Director, Physical Infrastructure Issues:
[End of section]
Appendixes:
Appendix I: Objectives, Scope, and Methodology:
This report addresses the following research questions: (1) What are
the strengths of FAA's designee programs? (2) What are the weaknesses
of the programs and the factors that contribute to those weaknesses?
and (3) What can be done to address the identified weaknesses or
otherwise improve the programs?
To address these questions, we used a variety of methods and sources of
information. We obtained and analyzed data for fiscal years 1998
through 2003 from four Federal Aviation Administration (FAA)
databases[Footnote 35] that maintain records on designees. We assessed
the reliability of the databases by (1) performing electronic testing
of required data elements; (2) reviewing existing information about the
data and the system that produced them; and (3) interviewing agency
officials knowledgeable about the data to learn how the information
system was structured, controlled, and used. We determined that the
data were sufficiently reliable for our purposes of describing the
number of designees by program, identifying the geographical location
of designees, and calculating the number of designees per FAA staff
responsible for designee oversight. However, we found that specific
data needed for oversight were not found in some databases, as we
discuss in this report.
In addition, we reviewed FAA program guidance concerning designee
management to obtain an understanding of designee roles and
responsibilities. We did not verify how FAA delegates authorized
functions and what certification activities were delegated. We also
reviewed FAA's Notice of Proposed Rulemaking on the organization
designation authorization program and public comments on the proposed
rule, conducted computer literature searches to obtain information on
other countries' designee programs, and interviewed officials from the
Canadian civil aviation authority. In addition, we reviewed past
studies, by us and others, of FAA's designee programs. (See the
bibliography at the end of this report.) We identified recommendations
that had been made to improve the programs and determined whether those
recommendations had been acted upon by the agency. Information obtained
from the reports and the databases was not equally comprehensive and
available for all types of designees.
We obtained information and data on FAA's designee programs on visits
to four locations--Los Angeles, Seattle, Atlanta, and Oklahoma City.
We selected the locations based on (1) number of designees in the
region; (2) activity-level of designees; (3) ratio of inspectors,
engineers, or flight surgeons to designees; and (4) location of both
Aircraft Certification directorate offices and Flight Standards Service
regional offices. Additionally, these offices were selected because of
the following: (1) the Seattle office has the largest number of
aircraft certification designees, (2) the Atlanta office has the
largest number of flight standards designees along with the most
certification activity, and (3) the Oklahoma City office manages some
designee data and is the location of FAA's training institute. We
interviewed individual FAA inspectors and engineers who oversee
designees at the offices we visited as well as officials from the
National Air Traffic Controllers Association and Professional Airway
System Specialists--unions that represent FAA inspectors and engineers.
We also interviewed designees in Los Angeles, Seattle, and Atlanta. The
cities and organizations where we conducted our work are shown in table
5.
Table 5: Organizations Interviewed by GAO During Site Visits:
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification: Transport
Airplane Directorate.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification: Manufacturing
Inspection Office.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification: Manufacturing
Inspection District Office.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification: Manufacturing
Inspection Satellite Office.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification: Boeing
Certificate Management Office.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aerospace Medicine.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Northwest
Mountain Region.
Location: Seattle, WA, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Northwest
Mountain Region: Seattle Flight Standards District Office.
Location: Seattle, WA, area;
Type of entity: Organizational designated airworthiness representative;
Organization: The Boeing Company.
Location: Seattle, WA, area;
Type of entity: Designated airworthiness representative;
Organization: Pacific Propellers.
Location: Seattle, WA, area;
Type of entity: Designated alteration station;
Organization: Goodrich Aviation Technical Services, Inc.
Location: Seattle, WA, area;
Type of entity: Special Federal Aviation Regulations No. 36, repair
station;
Organization: Alaska Airlines.
Location: Atlanta, GA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification.
Location: Atlanta, GA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification: Small Airplane
Directorate.
Location: Atlanta, GA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aerospace Medicine.
Location: Atlanta, GA, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Southern
Region.
Location: Atlanta, GA, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Southern
Region;
* Flight Standards Regional Office.
Location: Atlanta, GA, area;
Type of entity: Designated engineering representative;
Organization: Garrett Aviation.
Location: Atlanta, GA, area;
Type of entity: Designated engineering representative;
Organization: Propulsion Consultants Inc.
Location: Atlanta, GA, area;
Type of entity: Designated engineering representative;
Organization: Delta Airlines.
Location: Los Angeles, CA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification.
Location: Los Angeles, CA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aircraft Certification: Manufacturing
Inspection District Office.
Location: Los Angeles, CA, area;
Type of entity: Federal government;
Organization: FAA's Office of Aerospace Medicine.
Location: Los Angeles, CA, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Western
Pacific Region.
Location: Los Angeles, CA, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Western
Pacific Region: Los Angeles Flight Standards District Office.
Location: Los Angeles, CA, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Western
Pacific Region: Riverside Flight Standards District Office.
Location: Los Angeles, CA, area;
Type of entity: Designated pilot examiner;
Organization: Aviation Services.
Location: Los Angeles, CA, area;
Type of entity: Designated airworthiness representative/Designated
engineering representative;
Organization: CDO Associates.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Aviation Data Systems.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Designee Standardization Branch.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Delegation and Continued Airworthiness Programs
Branch.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Medical Systems Branch.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Aerospace Medical Education Division.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Civil Aerospace Medical Institute.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Aerospace Human Factors Research Division.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Southwest
Region.
Location: Oklahoma City, OK, area;
Type of entity: Federal government;
Organization: FAA's Office of Flight Standards Service, Southwest
Region: Oklahoma City Flight Standards District Office.
Source: GAO.
[End of table]
In addition, we convened a Web-based panel of experts selected for
their knowledge and expertise in the area of FAA's designee programs.
An initial list of experts was identified through referrals by FAA
officials, the National Air Traffic Controllers Association, the
Professional Airway System Specialists, and the Aerospace Repair
Station Association and through citations in the literature on
aviation. We then asked these initially identified experts for
additional experts. We continued this process until we had about 10 to
20 experts in each of four categories: (1) designees, (2) FAA
inspectors and engineers, (3) independent experts and university
academics, and (4) private sector and aviation industry associations.
(See app. II for the list of participating experts.)
To structure and gather expert opinion from the panel, we employed a
modified version of the Delphi method.[Footnote 36] To obtain opinions
from the large, diverse group of experts, we incorporated an iterative
and controlled feedback process--an important feature of the Delphi
method. We did not encourage experts to arrive at a consensus nor make
forecasts. During this process, we obtained opinions from the experts
using questionnaires administered over the Internet. The experts'
identities were kept anonymous during this step of the process. The
anonymity of this approach helped minimize potential biasing effects
often associated with live group discussions. Biasing effects of live
expert discussion sessions may include the dominance of individuals and
group pressure for conformity.[Footnote 37] The dominance bias would
tend to limit the input of less dominant individuals, and the group
pressure bias would tend to suppress true opinion, particularly on more
controversial issues. These concerns were particularly important given
the need for a broad range of expertise from individuals with varying
backgrounds and perspectives. Also, by creating a Web-based panel we
were able to include many more experts than we could have if we had
convened a live panel.
In the first phase of the expert panel, which ran from October 2 to 31,
2003, we asked the panelists to respond to three open-ended questions:
(1) What, if any, are the three most significant strengths of the FAA
designee programs? (2) What, if any, are the three most significant
weaknesses of the FAA designee programs? And (3) What, if any, are your
suggestions for addressing the weaknesses of or otherwise improving the
FAA designee programs? We further asked them to indicate those
responses that referred only to specific types of designees. The three
questions, based on our study objectives, were pre-tested to ensure
that the questionnaire was clear and unambiguous, did not place undue
burden on individuals completing it, and was independent and unbiased.
We made relevant changes before we deployed the first questionnaire to
all participants on the Internet.
We performed a content analysis of the responses to the open-ended
questions in order to compile a list of all the strengths, weaknesses,
and improvements mentioned by the experts. We contacted the experts,
when necessary, if responses were unclear. About 25 percent of the
coded responses were reviewed by an independent coder to ensure that
the initial coding decisions were consistent and valid. To maintain
standards of methodological integrity, any disagreements in coding
between the coder and reviewer were discussed until consensus was
reached.
The content coded results from the phase I questionnaire consisted of a
list of distinct and specific strengths, weaknesses, and suggested
improvements, which were used to construct the phase II questionnaire.
The phase II questionnaire also served as a feedback mechanism to the
panelists about what other experts thought were important strengths and
weaknesses. The phase II questionnaire was also pre-tested, revised,
and then administered on the Internet from January 5 to March 30, 2004.
In phase II, the panelists rated the strengths, weaknesses, and
suggested improvements on various relevant dimensions using a five-
category scale (e.g., "no weakness" to "very great weakness," or
"definitely infeasible" to "definitely feasible"). In analyzing the
responses to the phase II questionnaire, we calculated the frequency of
responses to identify the strongest levels of opinions on each item
regarding the strength, weakness, or attractiveness (based on
importance and feasibility) of suggested improvements. We ranked the
results based on the number of responses at the top two categories
(e.g., the number of "great weakness" and "very great weakness"
responses) that were rated as the more frequently identified responses.
Initially, 78 experts agreed to participate in the panel. Fifty-eight
panelists actually completed the phase I questionnaire, resulting in a
response rate of 74 percent. There was some attrition during the
subsequent phase. Of the 76 experts who agreed to participate in phase
II, 62 actually completed the questionnaire (including some who did not
participate in phase I). This resulted in a 82 percent response rate
for phase II (see table 6).
Table 6: The Number of Panelists Participating in Each Phase and
Response Rate:
Phase: I;
Experts who agreed to participate: 78;
Experts responding to questionnaire: 58;
Response rate in percentile: 74%.
Phase: II;
Experts who agreed to participate: 76;
Experts responding to questionnaire: 62;
Response rate in percentile: 82%.
Source: GAO.
[End of table]
We conducted our work between April 2003 and October 2004 in accordance
with the generally accepted government auditing standards.
[End of section]
Appendix II: Experts Participating on GAO's Panel:
Independent expert or university affiliation:
Roger Bacchieri, Chair,
Air Traffic Management Division,
Daniel Webster College.
Independent expert or university affiliation:
Patricia Backer, Chair,
Department of Aviation and Technology,
San Jose State University.
Independent expert or university affiliation:
William Caldwell, Chair,
Department of Aviation,
Central Missouri State University.
Independent expert or university affiliation:
Thomas J. Connolly, Associate Dean,
College of Aviation,
Embry-Riddle Aeronautical University.
Independent expert or university affiliation:
Bart J. Crotty,
Aviation Safety/Security Consultant,
former FAA Airworthiness Inspector,
former FAA Designated Airworthiness Representative.
Independent expert or university affiliation:
Alfred Dickinson, Director,
Aviation Safety Program,
University of Southern California.
Independent expert or university affiliation:
Carey L. Freeman, Chair,
Aviation Department,
Hampton University.
Independent expert or university affiliation:
Jim Frisbee,
Aviation Consultant,
former Director of Quality Assurance,
Northwest Airlines.
Independent expert or university affiliation:
Larry Gross,
Associate Professor of Aviation Technology,
Purdue University.
Independent expert or university affiliation:
Gary Kitely, Executive Director,
Council on Aviation Accreditation.
Independent expert or university affiliation:
Nick Lacey,
Aviation Consultant,
Mortem Beyer and Agnew,
former Director of FAA's Flight Standards Service.
Independent expert or university affiliation:
Doug Latia,
Associate Professor,
Aviation Technology Department,
Purdue University.
Independent expert or university affiliation:
Fred Leonelli,
former manager of FAA's Aircraft Maintenance Division.
Independent expert or university affiliation:
Kent Lovelace, Chair,
Department of Aerospace,
University of North Dakota.
Independent expert or university affiliation:
Jacqueline B. Sanders,
Assistant to the Provost,
Mercer County Community College.
Independent expert or university affiliation:
Glynn Dale Sistrunk, Chair,
Department of Professional Aviation,
Louisiana Tech University.
Aviation industry:
Mark Arcelle,
Senior Manager of Fleet Engineering,
FedEx Express.
Aviation industry:
Melissa Bailey,
Vice President of Air Traffic Regulation and Certification,
Aircraft Owners and Pilots Association.
Aviation industry:
Tony Broderick,
Aviation Safety Consultant,
former FAA Associate Administrator.
Aviation industry:
Eric Byer,
Manager of Government Industry Affairs,
National Air Transportation Association.
Aviation industry:
Aubrey Carter,
General Manager of Enabling Technology,
Delta Airlines.
Aviation industry:
Elias Cotti,
Director of Technical Operations,
National Business Aviation Association.
Aviation industry:
Brian Finnegan, President,
Professional Aviation Maintenance Association.
Aviation industry:
John Frisbee,
Manager of Quality Assurance,
Champion Airline.
Aviation industry:
Rick Hoy, Manager,
Regulatory Compliance,
Delta Airline.
Aviation industry:
Sarah Macleod, Executive Director,
Aeronautical Repair Station Association.
Aviation industry:
Doug MacNair, Vice President,
Government Relations,
Experimental Aircraft Association.
Aviation industry:
Nick Mateo, Senior Director,
Technical Services,
Continental Airlines.
Aviation industry:
Thomas McSweeny,
Director of International Safety and Regulatory Affairs, Boeing,
former FAA Associate Administrator of Regulation and Certification,
former FAA Director of Aircraft Certification Service.
Aviation industry:
Rick Oehme, Vice President,
Quality and Engineering,
America West Airlines.
Aviation industry:
Richard Peri, Vice President,
Government and Industry Affairs,
Aircraft Electronic Association.
Aviation industry:
Robert Robeson, Jr., Vice President of Civil Aviation,
Aerospace Industries Association.
Aviation industry:
Stan Sorscher, Labor Representative,
Society of Professional Engineering Employees in Aerospace.
Aviation industry:
Ronald Swanda, Vice President of Operations,
General Aviation Manufacturers Association.
Aviation industry:
Mark Szkirpan, Senior Specialist of Regulatory Affairs,
American Airlines.
Designees:
David Bryman, D.O.,
Senior Aviation Medical Examiner.
Designees:
Thomas W. Carroll,
Designated Airworthiness Representative,
former FAA Supervisory Aviation Safety Inspector.
Designees:
Harold Coralnick, M.D.,
Senior Aviation Medical Examiner.
Designees:
Dominick P. DaCosta,
Designated Airworthiness Representative,
Designated Engineering Representative,
Chief Executive Officer of DERS Group Inc.
Designees:
Joseph Kilpatrick,
Designated Engineering Representative.
Designees:
Osvaldo Lopez,
Designated Engineering Representative.
Designees:
Joe Norris,
Designated Airworthiness Representative.
Designees:
David Orfant,
CDO Associates,
Designated Airworthiness Representative (Manufacturing and
Maintenance),
Designated Engineering Representative.
Designees:
Thomas C. Willis,
Designated Airworthiness Representative (Maintenance).
Source: GAO.
Note: In addition, to the experts listed above, 7 inspectors from FAA's
Flight Standards Service, 10 engineers from FAA's Aircraft
Certification Service, and 1 other designee participated on the panel.
[End of table]
[End of section]
Appendix III: Roles and Responsibilities of Designees:
Designee type: Individual designees: Aviation Medical Examiners;
Responsibilities: Authorized to accept applications and perform
physical examinations necessary to determine qualification for the
issuance of airman medical certificates and combined medical/student
pilot certificates. Designees can issue, defer, or deny the
certificates, as appropriate.
Designee type: Individual designees: Designated Engineering
Representatives;
Responsibilities: Authorized to examine and approve certain
engineering technical data for their employer. Designees can either be
employed by a company or act as free agents.
Designee type: Individual designees: Designated Manufacturing
Inspection Representatives;
Responsibilities: Authorized to perform conformity inspections[A] and
issue airworthiness certificates and approvals for products and parts
produced by FAA-approved production approval holders.[B] Production
approval holders or its authorized suppliers employ this type of
designee.
Designee type: Individual designees: Training Center Evaluators;
Responsibilities: Authorized to accept applications and conduct
practical tests leading to the issuance of pilot and flight instructor
certificates.
Designee type: Individual designees: Designated Pilot Examiners;
Responsibilities: Authorized to accept applications for flight tests,
conduct those tests, and issue temporary pilot certificates to
qualified applicants.
Designee type: Individual designees: Aircrew Program Designees;
Responsibilities: Authorized to perform airman certification in one
type of aircraft for an operator's pilots who have been trained under
the operator's FAA-approved training program.
Designee type: Individual designees: Designated Airworthiness
Representatives (maintenance);
Responsibilities: Authorized to perform certain inspections, including
issuing recurrent airworthiness certificates and approvals for
maintenance conducted by repair stations and air carriers.
Designee type: Individual designees: Designated Airworthiness
Representatives (manufacturing);
Responsibilities: Authorized to perform conformity inspections, issue
airworthiness certificates and approval for products and parts produced
by FAA-approved production approval holders. Designees are independent
individuals, but may be employed by the production approval holder.
Designee type: Individual designees: Designated Mechanic Examiners;
Responsibilities: Authorized to accept applications for and conduct
oral and practical tests for issuing mechanic certificates.
Designee type: Individual designees: Designated Parachute Rigger
Examiners;
Responsibilities: Authorized to accept applications for, and conduct,
oral and practical tests for issuing parachute rigger certificates.
Designee type: Individual designees: Designated Aircraft Dispatcher
Examiners;
Responsibilities: Authorized to accept applications for, and conduct,
written and practical tests necessary for issuing aircraft dispatcher
certificates and, at the discretion of a local Flight Standards
inspector, issue temporary aircraft dispatcher certificates to
qualified applicants.
Designee type: Individual designees: Designated Flight Engineer
Examiners;
Responsibilities: Authorized to perform airman certification for an
operator's flight engineer candidates who have been trained under the
operator's FAA-approved training program.
Designee type: Individual designees: Computer Testing Designee;
Responsibilities: Authorized to administer computerized airman
knowledge tests through computer test sites located throughout the
United States and authorized foreign locations.
Designee type: Organizational designees: Organizational Designated
Airworthiness Representatives (maintenance);
Responsibilities: Organizations that (1) hold repair station
certificates with appropriate ratings or air carrier operating
certificates with FAA-approved Continuous Airworthiness Maintenance
programs and (2) are authorized to issue recurrent airworthiness
certificates and export airworthiness approvals for certain products.
Designee type: Organizational designees: Organizational Designated
Airworthiness Representatives (manufacturing);
Responsibilities: Organizations that hold FAA production approvals and
are authorized to issue airworthiness certificates and approvals and
make conformity determinations.
Designee type: Organizational designees: Designated Alteration
Stations;
Responsibilities: Companies that hold a current domestic repair
station certificate and are manufacturers of a product for which they
have alteration authority. The designees are authorized to issue
supplemental type certificates, perform prototype conformity
inspections,[C] and issue experimental airworthiness certificates for
the purpose of flight-testing and the standard airworthiness
certificate after the supplemental type certificate has been issued.
Designee type: Organizational designees: Special Federal Aviation
Regulations No. 36, Repair Stations;
Responsibilities: Companies that are authorized to generate engineering
technical data that are acceptable to the FAA. These data can be used
only by the specific designee for major repairs.
Designee type: Organizational designees: Delegation Option
Authorizations;
Responsibilities: Companies that are authorized to obtain type
certificates, approve type design changes, conduct conformity
inspections, and issue airworthiness certificates and approvals.
Source: GAO analysis of FAA documents.
[A] Conformity inspection is an assessment necessary to determine that
aviation products and related parts conform to an approved design and
can be operated safely.
[B] Production approval holders are aircraft manufacturers that hold a
type or production certificate and can produce modification or
replacement parts.
[C] Prototype conformity inspection is an examination to verify an
applicant's compliance with federal regulations and determine that
prototype products and related parts conform to proposed design
drawings and specifications.
[End of table]
[End of section]
Appendix IV: Survey Instrument and Results:
This appendix presents the results from the expert panel on the
identified strengths, weaknesses, and what can be done to address the
program weaknesses or otherwise improve the designee programs. Included
here are the questions and the ranking of responses developed based on
the frequency of responses to questions that were completed by members
of the panel selected for this study (referred to as "phase I" and
"phase II"). We administered the questionnaires for phases I and II
over the Internet.
As discussed in appendix I, in phase I of the expert panel, we asked
the panelists to respond to open-ended questions about the identified
strengths, weaknesses, and the potential of other alternatives to
improve FAA's designee programs. We performed a content analysis on the
responses to the open-ended questions in order to develop close-ended
questions for phase II of the expert panel. The purpose of the second
phase was to provide the panelists with the opportunity to consider the
other panelists' responses to the first phase and to respond in a
structured, quantifiable way. Phase II consisted of 64 closed-ended
questions on the categorized responses to phase I. Sixty-two of the 76
experts completed the phase II survey (about 82 percent response rate).
Table 7 summarizes the results from phase II, ranked based on the
number of responses at the top two points on the categories (e.g., the
number of "great" and "very great" responses) that were rated as the
more frequently identified responses.
Table 7: Experts' Responses to GAO's Survey:
Strengths of FAA's designee programs:
1; How important, if at all, is each of the following strengths of
FAA's designee programs toward accomplishing FAA's safety
responsibilities?
Strengths: a; Use of designees expands available FAA resources;
No: 0;
Some: 3;
Moderate: 9;
Great: 16;
Very great: 33;
Don‘t know/No opinion: 1;
No response: 0.
Strengths: b; Use of designees allows for more timely approvals than by
not using designees;
No: 0;
Some: 3;
Moderate: 10;
Great: 15;
Very great: 33;
Don‘t know/No opinion: 1;
No response: 0.
Strengths: c; Use of designees expands available technical expertise;
No: 2;
Some: 2;
Moderate: 11;
Great: 19;
Very great: 27;
Don‘t know/No opinion: 1;
No response: 0.
Strengths: d; Use of designees enables FAA staff to concentrate on
other areas of aviation safety;
No: 2;
Some: 5;
Moderate: 13;
Great: 19;
Very great: 20;
Don‘t know/No opinion: 2;
No response: 1.
Strengths: e; Designees provide greater scheduling flexibility and
access to the public;
No: 1;
Some: 7;
Moderate: 12;
Great: 12;
Very great: 27;
Don‘t know/No opinion: 2;
No response: 1.
Strengths: f; Use of designees allows for greater geographic coverage;
No: 3;
Some: 7;
Moderate: 12;
Great: 15;
Very great: 23;
Don‘t know/No opinion: 1;
No response: 1.
Strengths: g; Designees also perform liaison role improving relations
between FAA and aviation community;
No: 5;
Some: 12;
Moderate: 16;
Great: 19;
Very great: 7;
Don‘t know/No opinion: 3;
No response: 0.
Strengths: h; Designees help educate FAA engineers and inspectors;
No: 10;
Some: 7;
Moderate: 15;
Great: 17;
Very great: 8;
Don‘t know/No opinion: 4;
No response: 1.
Strengths: i; Designees provide consistent certification because they
receive recurrent training;
No: 7;
Some: 14;
Moderate: 16;
Great: 13;
Very great: 7;
Don‘t know/No opinion: 4;
No response: 1.
Strengths: j; Designees provide a pool of resources from which to draw
when filling positions at FAA;
No: 11;
Some: 15;
Moderate: 16;
Great: 10;
Very great: 4;
Don‘t know/No opinion: 5;
No response: 1.
Weaknesses of FAA's designee programs:
2; How much of a weakness is each of the following factors related to
the workload of FAA inspectors and aircraft certification engineers who
oversee designees?
Factors: a; Numbers of FAA inspectors and engineers not increasing
commensurate with industry growth;
No: 1;
Some: 5;
Moderate: 17;
Great: 12;
Very great: 20;
Don‘t know/No opinion: 7;
No response: 0.
Factors: b; Backlog of work submitted by designees awaiting approval/
concurrence by FAA;
No: 2;
Some: 6;
Moderate: 18;
Great: 16;
Very great: 11;
Don‘t know/No opinion: 8;
No response: 1.
Factors: c; FAA inspectors and engineers do not have enough time to
provide adequate oversight of designees for whom they are responsible;
No: 5;
Some: 4;
Moderate: 16;
Great: 17;
Very great: 10;
Don‘t know/No opinion: 8;
No response: 2.
Factors: d; Insufficient number of FAA inspectors/engineers compared
with designees to provide adequate oversight;
No: 3;
Some: 8;
Moderate: 23;
Great: 13;
Very great: 8;
Don‘t know/No opinion: 7;
No response: 0.
Factors: e; Applies only to Designated Pilot Examiners (DPE): High
turnover rate of FAA inspectors responsible for overseeing DPEs;
No: 2;
Some: 2;
Moderate: 7;
Great: 5;
Very great: 2;
Don‘t know/No opinion: 25;
No response: 19.
3; How much of a weakness is each of the following factors related to
the designee selection process?
Factors: a; Local FAA offices appoint designees based on personal
associations rather than qualifications and experiences;
No: 7;
Some: 15;
Moderate: 12;
Great: 10;
Very great: 11;
Don‘t know/No opinion: 7;
No response: 0.
Factors: b; Shortage of designees in some geographic areas and in
certain specializations;
No: 4;
Some: 6;
Moderate: 23;
Great: 15;
Very great: 5;
Don‘t know/No opinion: 9;
No response: 0.
Factors: c; FAA limits the number of designees;
No: 8;
Some: 7;
Moderate: 18;
Great: 13;
Very great: 7;
Don‘t know/No opinion: 8;
No response: 1.
Factors: d; FAA does not follow its own selection criteria;
No: 4;
Some: 17;
Moderate: 13;
Great: 11;
Very great: 8;
Don‘t know/No opinion: 8;
No response: 1.
Factors: e; The selection process lacks sufficient rigor to ensure
that designees are competent and will perform high quality work;
No: 3;
Some: 22;
Moderate: 14;
Great: 9;
Very great: 9;
Don‘t know/No opinion: 4;
No response: 1.
Factors: f; Variation in the qualifications of designees;
No: 3;
Some: 19;
Moderate: 19;
Great: 13;
Very great: 4;
Don‘t know/No opinion: 3;
No response: 1.
Factors: g; The application process for becoming a designee takes a
long time;
No: 8;
Some: 18;
Moderate: 15;
Great: 10;
Very great: 3;
Don‘t know/No opinion: 8;
No response: 0.
Factors: h; The selection process is not well defined;
No: 8;
Some: 24;
Moderate: 13;
Great: 8;
Very great: 4;
Don‘t know/No opinion: 4;
No response: 1.
4; How much of a weakness is each of the following factors related to
designee activities?
Factors: a; Applicants for certification shop for "easy" designees;
No: 0;
Some: 11;
Moderate: 14;
Great: 11;
Very great: 18;
Don‘t know/No opinion: 6;
No response: 2.
Factors: b; Employer pressure of financial incentives may lead to
conflicts of interest;
No: 10;
Some: 13;
Moderate: 15;
Great: 8;
Very great: 12;
Don‘t know/No opinion: 4;
No response: 0.
Factors: c;
Some applicants for certification are unfamiliar with FAA requirements
and designee's authority limits;
No: 7;
Some: 12;
Moderate: 21;
Great: 12;
Very great: 6;
Don‘t know/No opinion: 1;
No response: 3.
Factors: d; Designees' fees are inconsistent and unregulated;
No: 8;
Some: 15;
Moderate: 12;
Great: 9;
Very great: 8;
Don‘t know/No opinion: 7;
No response: 3.
Factors: e; Designees perform beyond their delegated authority;
No: 7;
Some: 21;
Moderate: 10;
Great: 10;
Very great: 6;
Don‘t know/No opinion: 6;
No response: 2.
Factors: f; Designees provide inconsistent service;
No: 6;
Some: 14;
Moderate: 25;
Great: 9;
Very great: 6;
Don‘t know/No opinion: 2;
No response: 0.
Factors: g; Designees perform more activities in less time than
standards would seem to require;
No: 8;
Some: 19;
Moderate: 11;
Great: 10;
Very great: 4;
Don‘t know/No opinion: 9;
No response: 1.
Factors: h; Designees are not current on regulations and orders;
No: 11;
Some: 19;
Moderate: 18;
Great: 9;
Very great: 4;
Don‘t know/No opinion: 1;
No response: 0.
Factors: i; Designees are constrained geographically;
No: 9;
Some: 20;
Moderate: 15;
Great: 10;
Very great: 3;
Don‘t know/No opinion: 3;
No response: 2.
Factors: j; Companies with organizational designations appoint
inexperienced engineers to make approvals and do not train them in the
certification process;
No: 8;
Some: 12;
Moderate: 8;
Great: 5;
Very great: 7;
Don‘t know/No opinion: 19;
No response: 3.
Factors: k; Erroneous certification by designees;
No: 7;
Some: 18;
Moderate: 15;
Great: 5;
Very great: 7;
Don‘t know/No opinion: 7;
No response: 3.
Factors: l; Designees do not understand their full authority;
No: 8;
Some: 19;
Moderate: 18;
Great: 5;
Very great: 6;
Don‘t know/No opinion: 4;
No response: 2.
Factors: m; The current scope of organizational delegation is narrow;
No: 9;
Some: 18;
Moderate: 10;
Great: 8;
Very great: 3;
Don‘t know/No opinion: 12;
No response: 2.
Factors: n; Designees perform outside of their jurisdiction without the
knowledge and authorization of local FAA offices;
No: 9;
Some: 17;
Moderate: 12;
Great: 5;
Very great: 4;
Don‘t know/No opinion: 15;
No response: 0.
Factors: o; Limitations on the approval authority of designees;
No: 10;
Some: 23;
Moderate: 16;
Great: 5;
Very great: 2;
Don‘t know/No opinion: 4;
No response: 2.
Factors: p; Applies only to Designated Mechanic Examiners: Inflexible
procedures for testing candidates for A&P certificates;
No: 4;
Some: 3;
Moderate: 6;
Great: 2;
Very great: 2;
Don‘t know/No opinion: 25;
No response: 20.
5; How much of a weakness is each of the following factors related to
FAA oversight?
Factors: a; FAA offices' level of oversight and interpretation of rules
are inconsistent;
No: 2;
Some: 5;
Moderate: 17;
Great: 16;
Very great: 20;
Don‘t know/No opinion: 0;
No response: 2.
Factors: b; Inactive, unqualified, or poor performing designees are
not identified or removed expeditiously;
No: 1;
Some: 8;
Moderate: 22;
Great: 15;
Very great: 12;
Don‘t know/No opinion: 3;
No response: 1.
Factors: c; It is difficult to terminate poor performing designees;
No: 2;
Some: 9;
Moderate: 16;
Great: 6;
Very great: 17;
Don‘t know/No opinion: 12;
No response: 0.
Factors: d; Inadequate surveillance and oversight of designees;
No: 6;
Some: 13;
Moderate: 15;
Great: 8;
Very great: 14;
Don‘t know/No opinion: 4;
No response: 2.
Factors: e; FAA has not made oversight of designees a high enough
priority;
No: 4;
Some: 9;
Moderate: 18;
Great: 12;
Very great: 8;
Don‘t know/No opinion: 8;
No response: 3.
Factors: f; Multitude of bulletins, advisory circulars, and other
documents from FAA have resulted in conflicting information and
procedures;
No: 1;
Some: 11;
Moderate: 23;
Great: 10;
Very great: 10;
Don‘t know/No opinion: 2;
No response: 5.
Factors: g; FAA management does not agree with engineers' or
inspectors' judgment about disciplining or removing poor performing
designees;
No: 4;
Some: 11;
Moderate: 7;
Great: 9;
Very great: 10;
Don‘t know/No opinion: 18;
No response: 3.
Factors: h; Oversight process is burdensome for FAA staff;
No: 6;
Some: 11;
Moderate: 17;
Great: 8;
Very great: 10;
Don‘t know/No opinion: 9;
No response: 1.
Factors: i; Designees are not held accountable for their findings;
No: 10;
Some: 19;
Moderate: 8;
Great: 7;
Very great: 11;
Don‘t know/No opinion: 5;
No response: 2.
Factors: j; FAA does not terminate poorly performing organizational
designees because that would put an entire company out of business;
No: 4;
Some: 13;
Moderate: 10;
Great: 9;
Very great: 8;
Don‘t know/No opinion: 17;
No response: 1.
Factors: k; FAA does not have adequate authority to impose penalties on
certain types of designees;
No: 9;
Some: 14;
Moderate: 11;
Great: 8;
Very great: 8;
Don‘t know/No opinion: 10;
No response: 2.
Factors: l; Lack of FAA process to evaluate the designee programs;
No: 5;
Some: 11;
Moderate: 19;
Great: 10;
Very great: 5;
Don‘t know/No opinion: 7;
No response: 5.
Factors: m; Lack of independent review of data. Designees perform the
analysis of the data that they then approve. The data are not reviewed
by a different person;
No: 6;
Some: 15;
Moderate: 10;
Great: 9;
Very great: 5;
Don‘t know/No opinion: 12;
No response: 5.
Factors: n; FAA engineers duplicate efforts of designees;
No: 4;
Some: 20;
Moderate: 8;
Great: 4;
Very great: 6;
Don‘t know/No opinion: 17;
No response: 3.
Factors: o; FAA engineers are reluctant to delegate routine activities
to designees;
No: 6;
Some: 11;
Moderate: 10;
Great: 6;
Very great: 4;
Don‘t know/No opinion: 21;
No response: 4.
Factors: p; FAA management pressures FAA engineers to give designees'
findings less scrutiny than standards require;
No: 4;
Some: 14;
Moderate: 3;
Great: 1;
Very great: 9;
Don‘t know/No opinion: 25;
No response: 6.
Factors: q; FAA inspectors and engineers lack the level of professional
experience necessary to oversee designees;
No: 9;
Some: 19;
Moderate: 12;
Great: 7;
Very great: 2;
Don‘t know/No opinion: 9;
No response: 4.
Factors: r; The designee programs lack formal methods of appeal when
designees' privileges are revoked;
No: 14;
Some: 13;
Moderate: 12;
Great: 5;
Very great: 2;
Don‘t know/No opinion: 12;
No response: 4.
Factors: s; Designees as well as the FAA inspectors/engineers who
oversee them have little or no familiarity with the products upon which
findings are being made;
No: 6;
Some: 23;
Moderate: 12;
Great: 2;
Very great: 4;
Don‘t know/No opinion: 10;
No response: 5.
Factors: t; FAA field office staffs do not have complete knowledge of
designees within their jurisdictions;
No: 5;
Some: 13;
Moderate: 23;
Great: 3;
Very great: 3;
Don‘t know/No opinion: 10;
No response: 5.
Factors: u; Applies only to Aviation Medical Examiners (AME): Error
letters are inaccurate indicators of an AME's performance;
No: 1;
Some: 0;
Moderate: 3;
Great: 1;
Very great: 1;
Don‘t know/No opinion: 31;
No response: 25.
6; How much of a weakness is each of the following factors related to
training for designees, FAA inspectors, and FAA engineers?
Factors: a; FAA engineers and inspectors do not receive adequate
training in designee oversight;
No: 6;
Some: 6;
Moderate: 12;
Great: 14;
Very great: 7;
Don‘t know/No opinion: 12;
No response: 5.
Factors: b; Lack of adequate and accessible designee training;
No: 5;
Some: 15;
Moderate: 19;
Great: 10;
Very great: 4;
Don‘t know/No opinion: 5;
No response: 4.
Factors: c; Designees are technically well versed in the area in which
they are authorized but poorly educated in the relevant regulations;
No: 8;
Some: 22;
Moderate: 13;
Great: 7;
Very great: 5;
Don‘t know/No opinion: 3;
No response: 4.
Factors: d; Seminar instructors for designee training are not current
or knowledgeable in the subject matter;
No: 14;
Some: 12;
Moderate: 14;
Great: 6;
Very great: 4;
Don‘t know/No opinion: 8;
No response: 4.
Factors: e; Training disparity between FAA engineers and designees
results in designees being more current on new orders, advisories, and
policies;
No: 12;
Some: 13;
Moderate: 10;
Great: 4;
Very great: 5;
Don‘t know/No opinion: 13;
No response: 5.
Overall Weaknesses of FAA‘s Designee Programs:
7; How much weakness, overall, is there in each of the following main
areas of FAA's designee programs?
Weakness: a; FAA oversight;
No: 2;
Some: 10;
Moderate: 18;
Great: 14;
Very great: 12;
Don‘t know/No opinion: 5;
No response: 1.
Weakness: b; Workload of FAA inspectors and aircraft certification
engineers who oversee designees;
No: 4;
Some: 5;
Moderate: 16;
Great: 15;
Very great: 11;
Don‘t know/No opinion: 10;
No response: 1.
Weakness: c; Training for FAA inspectors and FAA engineers who oversee
designees;
No: 3;
Some: 5;
Moderate: 18;
Great: 12;
Very great: 10;
Don‘t know/No opinion: 11;
No response: 3.
Weakness: d; Designee selection process;
No: 4;
Some: 18;
Moderate: 15;
Great: 10;
Very great: 7;
Don‘t know/No opinion: 6;
No response: 2.
Weakness: e; Training for designees;
No: 7;
Some: 13;
Moderate: 17;
Great: 11;
Very great: 6;
Don‘t know/No opinion: 4;
No response: 4.
Weakness: f; Designee activities;
No: 6;
Some: 20;
Moderate: 12;
Great: 12;
Very great: 4;
Don‘t know/No opinion: 6;
No response: 2.
Addressing Weaknesses with or Otherwise Improving FAA Inspector/
engineer Workload:
8; Increase the number of engineers/inspectors so that FAA staff have
more time available for oversight of designees:
Question: a; How important is it to implement this improvement?
No: 4;
Some: 10;
Moderate: 14;
Great: 17;
Very great: 12;
Don‘t know/No opinion: 4;
No response: 1.
Question: b; How feasible is it to implement this improvement?
No: 2;
Some: 8;
Moderate: 21;
Great: 12;
Very great: 11;
Don‘t know/No opinion: 7;
No response: 1.
9; Increase the priority given to the oversight of designees within
FAA:
Question: a; How important is it to implement this improvement?
No: 2;
Some: 8;
Moderate: 17;
Great: 24;
Very great: 4;
Don‘t know/No opinion: 5;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 6;
Moderate: 13;
Great: 21;
Very great: 14;
Don‘t know/No opinion: 6;
No response: 2.
10; Establish specific ratio for FAA engineers/inspectors to designees:
Question: a; How important is it to implement this improvement?
No: 5;
Some: 15;
Moderate: 11;
Great: 15;
Very great: 6;
Don‘t know/No opinion: 8;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 6;
Moderate: 16;
Great: 13;
Very great: 16;
Don‘t know/No opinion: 8;
No response: 3.
Addressing Weaknesses with or Otherwise Improving the Designee
Selection Process:
11; Select designees according to their qualifications and experience
rather than on personal associations with FAA managers:
Question: a; How important is it to implement this improvement?
No: 0;
Some: 6;
Moderate: 7;
Great: 19;
Very great: 28;
Don‘t know/No opinion: 1;
No response: 1.
Question: b; How feasible is it to implement this improvement?
No: 1;
Some: 2;
Moderate: 6;
Great: 10;
Very great: 40;
Don‘t know/No opinion: 2;
No response: 1.
12; Clearly define and consistently follow the criteria for selecting
designees:
Improving FAA's designee programs:
Question: a; How important is it to implement this improvement?
No: 0;
Some: 5;
Moderate: 7;
Great: 23;
Very great: 24;
Don‘t know/No opinion: 1;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 1;
Some: 1;
Moderate: 7;
Great: 24;
Very great: 24;
Don‘t know/No opinion: 3;
No response: 2.
13; Establish a review process for determining demand for designees by
type, specialty, activity level, and geographic location:
Question: a; How important is it to implement this improvement?
No: 5;
Some: 9;
Moderate: 20;
Great: 16;
Very great: 8;
Don‘t know/No opinion: 2;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 3;
Some: 1;
Moderate: 11;
Great: 28;
Very great: 11;
Don‘t know/No opinion: 4;
No response: 4.
14; Streamline procedures for the appointment of designees:
Question: a; How important is it to implement this improvement?
No: 4;
Some: 15;
Moderate: 18;
Great: 14;
Very great: 7;
Don‘t know/No opinion: 3;
No response: 1.
Question: b; How feasible is it to implement this improvement?
No: 1;
Some: 1;
Moderate: 15;
Great: 25;
Very great: 16;
Don‘t know/No opinion: 3;
No response: 1.
15; Centralize the designee selection process:
Question: a; How important is it to implement this improvement?
No: 16;
Some: 6;
Moderate: 14;
Great: 14;
Very great: 9;
Don‘t know/No opinion: 2;
No response: 1.
Question: b; How feasible is it to implement this improvement?
No: 7;
Some: 9;
Moderate: 13;
Great: 14;
Very great: 13;
Don‘t know/No opinion: 5;
No response: 1.
Addressing Weaknesses with or Otherwise Improving Designee Activities:
16; Improve FAA communication with designees, including communications
on regulations and orders and complicated certification situations:
Question: a; How important is it to implement this improvement?
No: 0;
Some: 8;
Moderate: 8;
Great: 24;
Very great: 18;
Don‘t know/No opinion: 3;
No response: 1.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 0;
Moderate: 5;
Great: 28;
Very great: 23;
Don‘t know/No opinion: 4;
No response: 2.
17; Clarify designations, including authority and limits:
Question: a; How important is it to implement this improvement?
No: 1;
Some: 8;
Moderate: 8;
Great: 25;
Very great: 14;
Don‘t know/No opinion: 4;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 1;
Moderate: 5;
Great: 24;
Very great: 25;
Don‘t know/No opinion: 5;
No response: 2.
18; Make company or organizational designees part of a different group
within the company than the group seeking the certification:
Question: a; How important is it to implement this improvement?
No: 6;
Some: 6;
Moderate: 13;
Great: 14;
Very great: 14;
Don‘t know/No opinion: 8;
No response: 1.
Question: b; How feasible is it to implement this improvement?
No: 3;
Some: 8;
Moderate: 11;
Great: 17;
Very great: 13;
Don‘t know/No opinion: 9;
No response: 1.
19; Determine if there are additional safety-critical areas that should
be beyond the scope of designees' authority:
Question: a; How important is it to implement this improvement?
No: 5;
Some: 8;
Moderate: 12;
Great: 16;
Very great: 12;
Don‘t know/No opinion: 8;
No response: 1.
Improving FAA's designee programs:
Question: b; How feasible is it to implement this improvement?
No: 2;
Some: 4;
Moderate: 11;
Great: 21;
Very great: 15;
Don‘t know/No opinion: 8;
No response: 1.
20; Provide individual designees with identification cards listing
their delegated authorizations that could be requested by and displayed
to customers:
Question: a; How important is it to implement this improvement?
No: 6;
Some: 7;
Moderate: 17;
Great: 11;
Very great: 16;
Don‘t know/No opinion: 3;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 1;
Some: 0;
Moderate: 9;
Great: 16;
Very great: 30;
Don‘t know/No opinion: 5;
No response: 1.
21; Increase FAA participation in complex approvals conducted by a
Designated Alteration Station (DAS):
Question: a; How important is it to implement this improvement?
No: 1;
Some: 5;
Moderate: 9;
Great: 14;
Very great: 13;
Don‘t know/No opinion: ; 17;
No response: 3.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 0;
Moderate: 5;
Great: 22;
Very great: 14;
Don‘t know/No opinion: 18;
No response: 3.
22; Implement FAA's Organization Designation Authorization proposal
and provide training for FAA employees on how to oversee a delegated
organization:
Question: a; How important is it to implement this improvement?
No: 8;
Some: 5;
Moderate: 9;
Great: 14;
Very great: 12;
Don‘t know/No opinion: 11;
No response: 3.
Question: b; How feasible is it to implement this improvement?
No: 1;
Some: 3;
Moderate: 8;
Great: 21;
Very great: 13;
Don‘t know/No opinion: 13;
No response: 3.
23; Require designees performing work outside of their geographic
boundaries to notify their home FAA office and the FAA office where the
work is being performed:
Question: a; How important is it to implement this improvement?
No: 8;
Some: 11;
Moderate: 13;
Great: 16;
Very great: 8;
Don‘t know/No opinion: 4;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 2;
Some: 2;
Moderate: 4;
Great: 18;
Very great: 27;
Don‘t know/No opinion: 5;
No response: 4.
24; Implement legislative proposal to establish "certified design
organizations" (also called "design organization certificates"):
Question: a; How important is it to implement this improvement?
No: 14;
Some: 5;
Moderate: 9;
Great: 8;
Very great: 7;
Don‘t know/No opinion: 16;
No response: 3.
Question: b; How feasible is it to implement this improvement?
No: 2;
Some: 6;
Moderate: 15;
Great: 10;
Very great: 9;
Don‘t know/No opinion: 17;
No response: 3.
25; Develop a fee structure of what designees may charge:
Question: a; How important is it to implement this improvement?
No: 23;
Some: 16;
Moderate: 5;
Great: 8;
Very great: 7;
Don‘t know/No opinion: 1;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 9;
Some: 12;
Moderate: 10;
Great: 9;
Very great: 13;
Don‘t know/No opinion: 5;
No response: 4.
26; Provide designees with broader authority:
Question: a; How important is it to implement this improvement?
No: 13;
Some: 16;
Moderate: 13;
Great: 11;
Very great: 3;
Don‘t know/No opinion: 4;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 3;
Some: 2;
Moderate: 17;
Great: 24;
Very great: 9;
Don‘t know/No opinion: 4;
No response: 3.
27; Make public the fees charged by designees.
Question: a; How important is it to implement this improvement?
No: 20;
Some: 12;
Moderate: 14;
Great: 5;
Very great: 7;
Don‘t know/No opinion: 2;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 9;
Some: 4;
Moderate: 17;
Great: 11;
Very great: 15;
Don‘t know/No opinion: 4;
No response: 2.
28; Establish a standard for limiting the number of certifications that
a designee can perform in a given period of time:
Question: a; How important is it to implement this improvement?
No: 23;
Some: 14;
Moderate: 10;
Great: 9;
Very great: 2;
Don‘t know/No opinion: 3;
No response: 1.
Question: b; How feasible is it to implement this improvement?
No: 8;
Some: 11;
Moderate: 15;
Great: 13;
Very great: 8;
Don‘t know/No opinion: 6;
No response: 1.
29; Assign designees to applicants instead of allowing applicants to
choose designees:
Question: a; How important is it to implement this improvement?
No: 24;
Some: 13;
Moderate: 11;
Great: 4;
Very great: 4;
Don‘t know/No opinion: 4;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 12;
Some: 10;
Moderate: 12;
Great: 9;
Very great: 9;
Don‘t know/No opinion: 8;
No response: 2.
Addressing Weaknesses with or Otherwise Improving FAA Oversight:
30; Hold designees accountable for their findings:
Question: a; How important is it to implement this improvement?
No: 1;
Some: 0;
Moderate: 4;
Great: 22;
Very great: 31;
Don‘t know/No opinion: 0;
No response: 4.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 3;
Moderate: 6;
Great: 16;
Very great: 30;
Don‘t know/No opinion: 2;
No response: 5.
31; Ensure that FAA employees who oversee designees are knowledgeable
about the regulations, policies, and processes applicable to the
designee's particular specialization:
Question: a; How important is it to implement this improvement?
No: 0;
Some: 0;
Moderate: 6;
Great: 17;
Very great: 32;
Don‘t know/No opinion: 1;
No response: 6.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 0;
Moderate: 6;
Great: 19;
Very great: 30;
Don‘t know/No opinion: 2;
No response: 5.
32; Increase penalties (including the ability to terminate their
status as designees) for individual and organizational designees found
to violate standards or who do not exercise proper judgment:
Question: a; How important is it to implement this improvement?
No: 4;
Some: 3;
Moderate: 8;
Great: 15;
Very great: 28;
Don‘t know/No opinion: 2;
No response: 2.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 2;
Moderate: 4;
Great: 27;
Very great: 24;
Don‘t know/No opinion: 3;
No response: 2.
33; Establish strict criteria and process for identifying and removing
designees that are underperforming, unqualified, or inactive:
Question: a; How important is it to implement this improvement?
No: 0;
Some: 7;
Moderate: 11;
Great: 20;
Very great: 22;
Don‘t know/No opinion: 2;
No response: 0.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 1;
Moderate: 11;
Great: 22;
Very great: 25;
Don‘t know/No opinion: 2;
No response: 1.
34; Improve coordination among the regional offices and headquarters
to standardize designee oversight:
Question: a; How important is it to implement this improvement?
No: 1;
Some: 3;
Moderate: 15;
Great: 20;
Very great: 18;
Don‘t know/No opinion: 2;
No response: 3.
Question: b; How feasible is it to implement this improvement?
No: 1;
Some: 4;
Moderate: 9;
Great: 24;
Very great: 18;
Don‘t know/No opinion: 3;
No response: 3.
35; Obtain feedback from users, designees, and other stakeholders
regarding the certification process and quality of oversight:
Question: a; How important is it to implement this improvement?
No: 2;
Some: 5;
Moderate: 14;
Great: 18;
Very great: 19;
Don‘t know/No opinion: 1;
No response: 3.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 4;
Moderate: 6;
Great: 28;
Very great: 17;
Don‘t know/No opinion: 3;
No response: 4.
36; Conduct audits to determine if designees have been given adequate
oversight:
Question: a; How important is it to implement this improvement?
No: 0;
Some: 7;
Moderate: 15;
Great: 26;
Very great: 9;
Don‘t know/No opinion: 2;
No response: 3.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 3;
Moderate: 9;
Great: 27;
Very great: 16;
Don‘t know/No opinion: 4;
No response: 3.
37; Improve FAA's public relations with those in the aviation community
who use designees by providing timely, knowledgeable responses to
public inquiries:
Question: a; How important is it to implement this improvement?
No: 3;
Some: 2;
Moderate: 16;
Great: 22;
Very great: 12;
Don‘t know/No opinion: 3;
No response: 4.
Question: b; How feasible is it to implement this improvement?
No: 0;
Some: 1;
Moderate: 8;
Great: 29;
Very great: 15;
Don‘t know/No opinion: 5;
No response: 4.
38; Establish a "whistleblower" program that would grant protection to
FAA employees who identify problems with the designee programs:
Question: a; How important is it to implement this improvement?
No: 10;
Some: 4;
Moderate: 9;
Great: 17;
Very great: 15;
Don‘t know/No opinion: 3;
No response: 4.
Question: b; How feasible is it to implement this improvement?
No: 4;
Some: 9;
Moderate: 8;
Great: 16;
Very great: 18;
Don‘t know/No opinion: 3;
No response: 4.
39; Develop competency testing and performance standards for designees:
Question: a; How important is it to implement this improvement?
No: 3;
Some: 9;
Moderate: 14;
Great: 18;
Very great: 13;
Don‘t know/No opinion: 1;
No response: 4.
Question: b; How feasible is it to implement this improvement?
No: 3;
Some: 6;
Moderate: 11;
Great: 21;
Very great: 13;
Don‘t know/No opinion: 3;
No response: 5.
40; Increase the support by FAA management of engineers' and
inspectors' judgment about disciplining poor performing designees:
Question: a; How important is it to implement this improvement?
No: 2;
Some: 8;
Moderate: 9;
Great: 18;
Very great: 12;
Don‘t know/No opinion: 7;
No response: 6.
Question: b; How feasible is it to implement this improvement?
No: 2;
Some: 4;
Moderate: 15;
Great: 17;
Very great: 11;
Don‘t know/No opinion: 6;
No response: 7.
41; Develop a formal process of appeal for designees facing discipline
or termination:
Question: a; How important is it to implement this improvement?
No: 1;
Some: 8;
Moderate: 14;
Great: 19;
Very great: 9;
Don‘t know/No opinion: 4;
No response: 7.
Question: b; How feasible is it to implement this improvement?
No: 1;
Some: 1;
Moderate: 10;
Great: 20;
Very great: 18;
Don‘t know/No opinion: 6;
No response: 6.
42; Increase requirements for oversight and surveillance to be
conducted by FAA inspectors and engineers:
Question: a; How important is it to implement this improvement?
No: 8;
Some: 6;
Moderate: 11;
Great: 17;
Very great: 11;
Don‘t know/No opinion: 4;
No response: 5.
Question: b; How feasible is it to implement this improvement?
No: 4;
Some: 4;
Moderate: 18;
Great: 14;
Very great: 12;
Don‘t know/No opinion: 6;
No response: 4.
43; Choose FAA aircraft certification offices with oversight
responsibility based on their knowledge of the product involved rather
than the geographic location of the designee:
Question: a; How important is it to implement this improvement?
No: 5;
Some: 6;
Moderate: 11;
Great: 18;
Very great: 8;
Don‘t know/No opinion: 7;
No response: 7.
Question: b; How feasible is it to implement this improvement?
No: 2;
Some: 4;
Moderate: 15;
Great: 13;
Very great: 11;
Don‘t know/No opinion: 10;
No response: 7.
44; Renew designees based on performance standards, rather than
allowing renewal to be automatic:
Question: a; How important is it to implement this improvement?
No: 4;
Some: 11;
Moderate: 13;
Great: 12;
Very great: 14;
Don‘t know/No opinion: 3;
No response: 5.
Question: b; How feasible is it to implement this improvement?
No: 2;
Some: 3;
Moderate: 11;
Great: 20;
Very great: 18;
Don‘t know/No opinion: 4;
No response: 4.
45; Make FAA engineers responsible for understanding and approving the
results of designee actions rather than checking only the paperwork
associated with those actions:
Question: a; How important is it to implement this improvement?
No: 8;
Some: 9;
Moderate: 9;
Great: 12;
Very great: 12;
Don‘t know/No opinion: 6;
No response: 6.
Question: b; How feasible is it to implement this improvement?
No: 6;
Some: 9;
Moderate: 10;
Great: 15;
Very great: 9;
Don‘t know/No opinion: 7;
No response: 6.
46; Reduce the administrative (paperwork) burden of designee oversight:
Question: a; How important is it to implement this improvement?
No: 3;
Some: 8;
Moderate: 18;
Great: 19;
Very great: 5;
Don‘t know/No opinion: 4;
No response: 5.
Question: b; How feasible is it to implement this improvement?
No: 3;
Some: 5;
Moderate: 18;
Great: 18;
Very great: 8;
Don‘t know/No opinion: 6;
No response: 4.
47; Establish a panel of senior FAA inspectors/engineers to review
allegations of impropriety by designees. Provide the panel with the
authority to improve penalties.
Question: How important is it to implement this improvement?
No: 6;
Some: 12;
Moderate: 14;
Great: 11;
Very great: 12;
Don‘t know/No opinion: 5;
No response: 2.
Question: How feasible is it to implement this improvement?
No: 5;
Some: 7;
Moderate: 16;
Great: 14;
Very great: 12;
Don‘t know/No opinion: 6;
No response: 2.
48; Develop an automated system to allow designees to complete and
submit documents electronically only when they are done correctly.
Question: How important is it to implement this improvement?
No: 5;
Some: 11;
Moderate: 14;
Great: 16;
Very great: 7;
Don‘t know/No opinion: 4;
No response: 5.
Question: How feasible is it to implement this improvement?
No: 1;
Some: 5;
Moderate: 14;
Great: 19;
Very great: 12;
Don‘t know/No opinion: 6;
No response: 5.
49; Develop specific statements or checklists that identify the steps
in the certification process and the extent of the designee's
authority.
Question: How important is it to implement this improvement?
No: 4;
Some: 5;
Moderate: 24;
Great: 16;
Very great: 4;
Don‘t know/No opinion: 4;
No response: 5.
Question: How feasible is it to implement this improvement?
No: 1;
Some: 1;
Moderate: 17;
Great: 21;
Very great: 11;
Don‘t know/No opinion: 7;
No response: 4.
50; Eliminate geographic boundaries imposed on aircraft certification
designees:
Question: How important is it to implement this improvement?
No: 10;
Some: 7;
Moderate: 9;
Great: 9;
Very great: 10;
Don‘t know/No opinion: 12;
No response: 5.
Question: How feasible is it to implement this improvement?
No: 2;
Some: 3;
Moderate: 10;
Great: 14;
Very great: 12;
Don‘t know/No opinion: 15;
No response: 6.
51; Have FAA inspectors and engineers who oversee designees report to
a central FAA focal point who is independent of their supervisors:
Question: How important is it to implement this improvement?
No: 11;
Some: 5;
Moderate: 17;
Great: 13;
Very great: 3;
Don‘t know/No opinion: 8;
No response: 5.
Question: How feasible is it to implement this improvement?
No: 5;
Some: 8;
Moderate: 19;
Great: 11;
Very great: 6;
Don‘t know/No opinion: 7;
No response: 6.
52; Prohibit designees from approving any documents that they have
produced:
Question: How important is it to implement this improvement?
No: 12;
Some: 13;
Moderate: 9;
Great: 6;
Very great: 9;
Don‘t know/No opinion: 8;
No response: 5.
Question: How feasible is it to implement this improvement?
No: 8;
Some: 7;
Moderate: 11;
Great: 10;
Very great: 13;
Don‘t know/No opinion: 8;
No response: 5.
53; Applies only to Designated Engineering Representatives (DER): Make
the selection and oversight process for company DERs the same as for
consultant DERs.
Question: How important is it to implement this improvement?
No: 5;
Some: 4;
Moderate: 12;
Great: 8;
Very great: 7;
Don‘t know/No opinion: 16;
No response: 10.
Question: How feasible is it to implement this improvement?
No: 0;
Some: 2;
Moderate: 8;
Great: 14;
Very great: 11;
Don‘t know/No opinion: 17;
No response: 10.
54; Limit the ability of designees to contest their removal:
Question: How important is it to implement this improvement?
No: 20;
Some: 15;
Moderate: 9;
Great: 7;
Very great: 4;
Don‘t know/No opinion: 6;
No response: 1.
Question: How feasible is it to implement this improvement?
No: 7;
Some: 7;
Moderate: 18;
Great: 12;
Very great: 9;
Don‘t know/No opinion: 7;
No response: 2.
Addressing Weaknesses with or Otherwise Improving Training:
55; Improve availability of training for FAA inspectors and engineers
to advance technical competence related to oversight of designees:
Question: How important is it to implement this improvement?
No: 2;
Some: 4;
Moderate: 8;
Great: 20;
Very great: 22;
Don‘t know/No opinion: 3;
No response: 3.
Question: How feasible is it to implement this improvement?
No: 1;
Some: 1;
Moderate: 7;
Great: 30;
Very great: 16;
Don‘t know/No opinion: 3;
No response: 4.
56; Ensure standard training of designees within specific specialties
to improve consistency of their work:
Question: How important is it to implement this improvement?
No: 1;
Some: 3;
Moderate: 14;
Great: 20;
Very great: 20;
Don‘t know/No opinion: 0;
No response: 4.
Question: How feasible is it to implement this improvement?
No: 1;
Some: 3;
Moderate: 9;
Great: 24;
Very great: 20;
Don‘t know/No opinion: 1;
No response: 4.
57; Require consistent training for all designees with the same skill
designation to improve the consistency among designees:
Question: How important is it to implement this improvement?
No: 1;
Some: 6;
Moderate: 13;
Great: 27;
Very great: 13;
Don‘t know/No opinion: 0;
No response: 2.
Question: How feasible is it to implement this improvement?
No: 0;
Some: 5;
Moderate: 6;
Great: 31;
Very great: 16;
Don‘t know/No opinion: 1;
No response: 3.
58; Increase number of subject matter workshops for designees, with
instruction provided by industry experts, FAA specialists, engineers,
and designees:
Question: How important is it to implement this improvement?
No: 1;
Some: 5;
Moderate: 11;
Great: 23;
Very great: 17;
Don‘t know/No opinion: 1;
No response: 4.
Question: How feasible is it to implement this improvement?
No: 0;
Some: 1;
Moderate: 13;
Great: 22;
Very great: 20;
Don‘t know/No opinion: 2;
No response: 4.
59; Require FAA inspectors and engineers to receive recurrent training
related to the oversight of designees:
Question: How important is it to implement this improvement?
No: 2;
Some: 3;
Moderate: 14;
Great: 22;
Very great: 14;
Don‘t know/No opinion: 3;
No response: 4.
Question: How feasible is it to implement this improvement?
No: 1;
Some: 2;
Moderate: 10;
Great: 27;
Very great: 14;
Don‘t know/No opinion: 4;
No response: 4.
60; Require additional training for designees in regulations that apply
to their work:
Question: How important is it to implement this improvement?
No: 1;
Some: 10;
Moderate: 14;
Great: 22;
Very great: 9;
Don‘t know/No opinion: 2;
No response: 4.
Question: How feasible is it to implement this improvement?
No: 0;
Some: 0;
Moderate: 10;
Great: 32;
Very great: 15;
Don‘t know/No opinion: 2;
No response: 3.
61; Improve and expand designee training, including routine skills
testing:
Question: How important is it to implement this improvement?
No: 2;
Some: 11;
Moderate: 11;
Great: 16;
Very great: 15;
Don‘t know/No opinion: 2;
No response: 5.
Question: How feasible is it to implement this improvement?
No: 2;
Some: 4;
Moderate: 13;
Great: 24;
Very great: 12;
Don‘t know/No opinion: 3;
No response: 4.
62; Have experienced designees mentor designee candidates:
Question: How important is it to implement this improvement?
No: 3;
Some: 6;
Moderate: 19;
Great: 23;
Very great: 6;
Don‘t know/No opinion: 1;
No response: 4.
Question: How feasible is it to implement this improvement?
No: 2;
Some: 3;
Moderate: 18;
Great: 18;
Very great: 15;
Don‘t know/No opinion: 1;
No response: 5.
63; Make the training and standardization seminar for designees an
annual requirement:
Question: How important is it to implement this improvement?
No: 9;
Some: 5;
Moderate: 15;
Great: 13;
Very great: 15;
Don‘t know/No opinion: 1;
No response: 4.
Question: How feasible is it to implement this improvement?
No: 4;
Some: 3;
Moderate: 11;
Great: 23;
Very great: 16;
Don‘t know/No opinion: 1;
No response: 4.
64; Applies only to Designated Alteration Station (DAS): Require
additional training for FAA inspectors and engineers in areas such as
designee selection and oversight, regulations that pertain to the
activities of designees, and the recognition of a management structure
that provides appropriate direction and support for DAS operations:
Question: How important is it to implement this improvement?
No: 1;
Some: 2;
Moderate: 6;
Great: 10;
Very great: 13;
Don‘t know/No opinion: 18;
No response: 12.
Question: How feasible is it to implement this improvement?
No: 0;
Some: 2;
Moderate: 6;
Great: 11;
Very great: 14;
Don‘t know/No opinion: 18;
No response: 11.
Source: GAO analysis of expert panel information.
[End of table]
[End of section]
Appendix V: GAO Contacts and Staff Acknowledgments:
GAO Contacts:
JayEtta Z. Hecker (202) 512-2834:
Gerald Dillingham (202) 512-2834:
Teresa Spisak (202) 512-3952:
Staff Acknowledgments:
In addition to the above individuals, Howard Cott, Colin Fallon, Isidro
Gomez, Curtis Groves, Brandon Haller, David Hooper, Jennifer Kim, Rosa
Leung, Elizabeth A. Marchak, and Larry Thomas made key contributions to
this report.
[End of section]
Bibliography:
Booz-Allen & Hamilton, Challenge 2000 Recommendations for Future Safety
Regulation: Shifting Roles and Responsibilities Between FAA and
Industry (Prepared for Federal Aviation Administration, Office of
Policy, Planning, and International Aviation) (McLean, VA: Apr. 19,
1996).
Department of Transportation, Office of Inspector General, Report on
the FAA's Designated Pilot Examiner Program, E5-FA-4-007 (Washington,
D.C.: Feb. 25, 1994).
Department of Transportation, Office of Inspector General, Pilot
Examiner Program, R2-FA-7-001 (Washington, D.C.: Oct. 22, 1996).
Federal Aviation Administration, Designated Engineering Representative
Oversight Team Report (Washington, D.C.: Oct. 11, 1994).
Federal Aviation Administration, Aircraft Certification Service: DER
Oversight Evaluation (Washington, D.C.: Sept. 11, 1997).
Federal Aviation Administration, Designated Alteration Station System
Assessment Final Report (Washington, D.C.: Sept. 21, 2000).
Federal Aviation Administration, Commercial Airplane Certification
Process Study: An Evaluation of Selected Aircraft Certification,
Operations, and Maintenance Processes (Washington, D.C.: March 2002).
GAO, Aviation Safety: FAA Generally Agrees With But is Slow in
Implementing Safety Recommendations,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-96-193]
(Washington, D.C.: September 1996).
GAO, Aircraft Certification: New FAA Approach Needed to Meet Challenges
of Advanced Technology,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-93-155]
(Washington, D.C.: September 1993).
National Research Council, Improving the Continued Airworthiness of
Civil Aircraft: A Strategy for the FAA's Aircraft Certification Service
(Washington, D.C: 1998).
RTCA, Final Report of RTCA Task Force 4 Certification (Washington,
D.C.: Feb. 26, 1999).
(540056):
FOOTNOTES
[1] Those staff are safety inspectors in Flight Standards Service,
engineers in Aircraft Certification Service, and flight surgeons in
Aerospace Medicine.
[2] Title 49, U.S.C. 447702(d) provides FAA's legislative authority to
use designees and Title 14, C.F.R., Part 183, sets out the types of
designations FAA may issue and the process for selecting designees.
[3] Such employees, who actually perform the delegated activities, are
referred to as "authorized representatives."
[4] Public Law 108-176, Vision 100 - Century of Aviation
Reauthorization Act, requires FAA to develop a plan for implementing a
certified design organization program by 2007.
[5] Production conformity is an inspection necessary to determine that
aviation products and related parts conform to an approved design and
can be operated safely.
[6] A pilot must have both a pilot certificate and a medical
certificate in order to fly an aircraft, with the exception of glider
and balloon pilots, who are not required to have a medical certificate.
The pilot certificate never expires. The medical certificate must be
updated every 6 months to 3 years, depending on the type of pilot
certificate (e.g., airline transport pilots must have their medical
certificate updated more frequently than private pilots).
[7] The Quality Assurance Team was established as a result of a 1999
recommendation by the International Civil Aviation Organization that
Flight Standards Service conduct standardized evaluations of its field
offices.
[8] Federal Aviation Administration, Southwest Region General Aviation
Pilot Examiner Review Final Report (Fort Worth, TX: Sept. 1, 2000).
[9] See Federal Aviation Administration, Designated Alteration Station
System Assessment Final Report (Sept. 21, 2000); Aircraft Certification
Service Evaluation of the Airworthiness Designee Management Program
(Dec. 1998); and Aircraft Certification Service DER Oversight
Evaluation (Sept. 11, 1997).
[10] The office has not assessed its smallest designee program--the
delegation option authorization program, which has six designated
organizations.
[11] Federal Aviation Administration, Designated Alteration Station
System Assessment Final Report (Sept. 21, 2000).
[12] Transportation Safety Board of Canada, Aviation Investigation
Report, In-Flight Fire Leading to Collision with Water, Swissair
Transport Limited McDonnell Douglas MD-11 GH-IWF, Peggy's Cove, Nova
Scotia 5 nm SW, 2 September 1998, report number A98H0003 (no date).
[13] Federal Aviation Administration, Special Certification Review Team
Report on: Santa Barbara Aerospace STC ST00236LA-D Swissair Model MD-11
Airplane In-flight Entertainment System (June 14, 1999).
[14] These include 31 designated alteration stations, 12 Special
Federal Aviation Regulations No. 36 (repair stations), and 6 delegation
option authorizations.
[15] See bibliography at the end of this report.
[16] When designated engineering representatives conduct work related
to field approvals outside of their assigned geographic areas, they are
not required to contact the field office where they are conducting that
work. On the other hand, when their work is related to issuing type
certificates or supplemental type certificates outside their assigned
geographic area, they are required to contact the FAA field office
where they are conducting that work.
[17] Nineteen experts indicated this factor was a "great" or "very
great" weakness of the designee programs; 4 experts felt this factor
was not a weakness; 17 experts felt that this posed "little" weakness.
[18] Each type of designee has unique qualification requirements, which
are defined in FAA Order 8100.8 Chapter 4.
[19] Federal Aviation Administration, Southwest Region General Aviation
Pilot Examiner Review Final Report (Fort Worth, TX: Sept. 1, 2000).
[20] Practical test standards are areas of operating aircraft, such as
flight procedures or flight maneuvers, in which pilot applicants must
demonstrate their knowledge and skills before receiving pilot
certificates. FAA developed these standards for FAA inspectors and
designated pilot examiners to use when conducting practical tests to
pilot applicants.
[21] Federal Aviation Administration, Commercial Airplane
Certification Process Study: An Evaluation of Selected Aircraft
Certification, Operations, and Maintenance Processes (Washington,
D.C.: March 2002).
[22] During that time period, 2,850 additional designees were
terminated for reasons not associated with disciplinary action, such as
change of employment, retirement, or the request of the designee.
[23] See footnote 21.
[24] The four databases are the National Vital Information Subsystem
and Program Tracking and Reporting Subsystem used by Flight Standards
Service, the Designee Information Network used by Aircraft
Certification Service, and the Airmen Medical Certification Information
Subsystem used by Aerospace Medicine.
[25] Most designees' appointments are effective for 1 year, with the
exception of individual and organizational designated airworthiness
representatives, who are appointed for up to 5 years, and all other
types of organizational designees, which are appointed indefinitely.
According to FAA policy, the minimum level of oversight requires FAA
engineers and inspectors to review designees' files for project
activity in order to renew the designees' authority.
[26] GAO, Aircraft Certification: New FAA Approach Needed to Meet
Challenges of Advanced Technologies, GAO/RCED-93-155 (Washington,
D.C.: Sept. 16, 1993).
[27] See footnote 21.
[28] The turnover rates reported in the two studies were cumulative
over the time period, while FAA provided information on an annual
basis. The turnover rates from FAA, therefore, are not comparable to
the rates from the two studies.
[29] Only one expert indicated that greater accountability of designees
was not necessary.
[30] In response to this question, two experts had no opinion and three
experts declined to answer.
[31] Established in the 1980s, Canada's two types of organizational
delegates, Design Approval Organizations and Airworthiness Engineering
Organizations, are authorized to evaluate and approve technical data to
determine compliance with safety requirements.
[32] The user fee program was established by the Prescription Drug User
Fee Act of 1992.
[33] P.L. 105-66 (October 27, 1997).
[34] GAO, Transportation Financing: Challenges in Meeting Long-Term
Funding Needs for FAA, Amtrak, and the Nation's Highways, GAO/T-RCED-
97-151 (Washington, D.C.: May 7, 1997); GAO, Airport and Airway Trust
Fund: Issues Raised by Proposal to Replace the Airline Ticket Tax, GAO/
RCED-97-23 (Washington, D.C.: Dec. 9, 1996).
[35] The four databases are the (1) National Vital Information
Subsystem, (2) Program Tracking and Reporting Subsystem, (3) Designee
Information Network, and (4) Airmen Medical Certification Information
Subsystem.
[36] For examples of recent use of this methodology see, GAO, Drinking
Water: Experts' Views on How Future Federal Funding Can Best Be Spent
to Improve Security, GAO-04-29 (Washington, D.C.: Oct. 31, 2003);
International Trade: Experts' Advice for Small Businesses Seeking
Foreign Patents, GAO-03-910 (Washington, D.C.: June 26, 2003); Economic
Models of Cattle Prices: How USDA Can Act to Improve Models to Explain
Cattle Prices, GAO-02-246 (Washington, D.C.: Mar. 15, 2002);
Environmental Protection: Federal Incentives Could Help Promote Land
Use That Protects Air and Water Quality, GAO-02-12 (Washington, D.C.:
Oct. 31, 2001).
[37] James P. Wright, "Delphi-Systematic Opinion Gathering," The GAO
Review (Spring 1972): 20-27.
GAO's Mission:
The Government Accountability Office, the investigative arm of
Congress, exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office
441 G Street NW, Room LM
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm
E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director,
NelliganJ@gao.gov
(202) 512-4800
U.S. Government Accountability Office,
441 G Street NW, Room 7149
Washington, D.C. 20548: