Information Security
Agencies Need to Implement Consistent Processes In Authorizing Systems for Operation
Gao ID: GAO-04-376 June 28, 2004
The Office of Management and Budget (OMB) requires agencies to certify the security controls of their information systems and to formally authorize and accept the risk associated with their operation (a process known as accreditation). These processes support requirements of the Federal Information Security Management Act of 2002 (FISMA). Further, OMB requires agencies to report the number of systems authorized following certification and accreditation as one of the key FISMA performance measures. In response to the Congressional request, GAO (1) identified existing governmentwide requirements and guidelines for certifying and accrediting information systems, (2) determined the extent to which agencies have reported their systems as certified and accredited, and (3) assessed whether their processes provide consistent, comparable results and adequate information for authorizing officials.
The National Institute of Standards and Technology (NIST) and other agencies, including the Department of Defense, have provided guidance for the certification and accreditation of federal information systems. This guidance includes new guidelines just issued by NIST, which emphasize a model of continuous monitoring, as well as compliance with FISMA-required standards for minimum-security controls. Many agencies report that they have begun to use the new guidance in their certification and accreditation processes. The reported percentage of systems certified and accredited for operation as of the first half of 2004 was 63 percent for 24 major federal agencies. However, the picture is not uniform across the government, with 7 of the agencies reporting greater than 90 percent of their systems certified and accredited but 6 reporting fewer than half. GAO's analyses also highlighted instances in which agencies do not consistently report FISMA performance measurement data, as well as other factors that lessen the usefulness of these data, such as the limited assurance of data reliability and quality. All the agencies GAO surveyed reported that their certification and accreditation processes met criteria consistent with those identified in federal guidance, such as a current risk assessment and security control evaluation. However, our review of documentation for the certification and accreditation of 32 selected systems at four of these agencies showed that these criteria were not always met--results similar to those found by agency inspectors general. Further, three of these four agencies did not have routine quality review processes to determine whether such criteria are met--processes that could help agency accrediting officials receive consistent information on which to base their decisions. Several agencies cited obstacles in implementing their certification and accreditation processes, including resource and staffing limitations. Some agencies have taken actions to improve their processes, such as redefining system boundaries to better manage systems.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-04-376, Information Security: Agencies Need to Implement Consistent Processes In Authorizing Systems for Operation
This is the accessible text file for GAO report number GAO-04-376
entitled 'Information Security: Agencies Need to Implement Consistent
Processes In Authorizing Systems for Operation' which was released on
July 28, 2004.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to Congressional Requesters:
United States General Acctounting Office
GAO:
June 2004:
INFORMATION SECURITY:
Agencies Need to Implement Consistent Processes in Authorizing Systems
for Operation:
GAO-04-376:
GAO Highlights:
Highlights of GAO-04-376, a report to the Chairman, House Committee on
Government Reform and the Chairman of its Subcommittee on Technology,
Information Policy, Intergovernmental Relations and the Census:
Why GAO Did This Study:
The Office of Management and Budget (OMB) requires agencies to certify
the security controls of their information systems and to formally
authorize and accept the risk associated with their operation (a
process known as accreditation). These processes support requirements
of the Federal Information Security Management Act of 2002 (FISMA).
Further, OMB requires agencies to report the number of systems
authorized following certification and accreditation as one of the key
FISMA performance measures.
In response to the committee and subcommittee request, GAO (1)
identified existing governmentwide requirements and guidelines for
certifying and accrediting information systems, (2) determined the
extent to which agencies have reported their systems as certified and
accredited, and (3) assessed whether their processes provide
consistent, comparable results and adequate information for
authorizing officials.
What GAO Found:
The National Institute of Standards and Technology (NIST) and other
agencies, including the Department of Defense, have provided guidance
for the certification and accreditation of federal information systems.
This guidance includes new guidelines just issued by NIST, which
emphasize a model of continuous monitoring, as well as compliance with
FISMA-required standards for minimum-security controls. Many agencies
report that they have begun to use the new guidance in their
certification and accreditation processes.
The reported percentage of systems certified and accredited for
operation as of the first half of 2004 was 63 percent for 24 major
federal agencies. However, the picture is not uniform across the
government, with 7 of the agencies reporting greater than 90 percent
of their systems certified and accredited but 6 reporting fewer than
half. GAO‘s analyses also highlighted instances in which agencies do
not consistently report FISMA performance measurement data, as well as
other factors that lessen the usefulness of these data, such as the
limited assurance of data reliability and quality.
All the agencies GAO surveyed reported that their certification and
accreditation processes met criteria consistent with those identified
in federal guidance, such as a current risk assessment and security
control evaluation. However, our review of documentation for the
certification and accreditation of 32 selected systems at four of
these agencies showed that these criteria were not always met (see
chart)”results similar to those found by agency inspectors general.
Further, three of these four agencies did not have routine quality
review processes to determine whether such criteria are met”processes
that could help agency accrediting officials receive consistent
information on which to base their decisions. Several agencies cited
obstacles in implementing their certification and accreditation
processes, including resource and staffing limitations. Some agencies
have taken actions to improve their processes, such as redefining
system boundaries to better manage systems.
Number and Percentage of 32 Selected Agency Systems Meeting Specific
Certification and Accreditation Criteria:
[See PDF for image]
[End of figure]
What GAO Recommends:
GAO is making recommendations to the Director, Office of Management
and Budget, to help ensure that agencies‘ certification and
accreditation processes consistently provide adequate and effective
information security controls. In oral comments on a draft of this
report, OMB officials generally agreed with GAO‘s recommendations.
www.gao.gov/cgi-bin/getrpt?GAO-04-376.
To view the full product, including the scope and methodology, click
on the link above. For more information, contact Robert F. Dacey at
(202) 512-3317 or daceyr@gao.gov.
[End of section]
Contents:
Letter:
Objectives, Scope, and Methodology:
Results in Brief:
Background:
Certification and Accreditation Guidance Is Provided by NIST and Other
Responsible Agencies:
Reported Percentages of Systems Certified and Accredited Vary Widely:
Processes at Selected Agencies Do Not Ensure Consistent or Adequate
Information:
Conclusions:
Recommendations for Executive Action:
Agency Comments:
Appendixes:
Appendix I: Comments from the Department of Commerce:
Appendix II: GAO Contact and Staff Acknowledgments:
GAO Contact:
Acknowledgments:
Tables Tables:
Table 1: Agency Systems Reported as Authorized After Certification and
Accreditation:
Table 2: Certification and Accreditation Criteria Required to Be Met by
Processes at 24 Major Agencies:
Table 3: Number and Percentage of 32 Selected Agency Systems Meeting
Specific Certification and Accreditation Criteria:
Figure:
Figure 1: NIST Security Certification and Accreditation Process :
Abbreviations
CIO: chief information officer:
DITSCAP: DOD Information Technology Security Certification and
Accreditation Process:
DOD: Department of Defense:
EPA: Environmental Protection Agency:
FIPS: Federal Information Processing Standard:
FISMA: Federal Information Security Management Act:
IG: inspector general:
IT: information technology:
NASA: National Aeronautics and Space Administration:
NIST: National Institute of Standards and Technology:
OMB: Office of Management and Budget:
Letter June 28, 2004:
The Honorable Tom Davis:
Chairman:
Committee on Government Reform:
House of Representatives:
The Honorable Adam H. Putnam:
Chairman:
Subcommittee on Technology, Information Policy, Intergovernmental
Relations and the Census:
Committee on Government Reform:
House of Representatives:
Office of Management and Budget (OMB) information security policy
requires agency management officials to formally authorize each of
their information systems to process, store, or transmit information,
and to accept the risk associated with their operation. This
authorization (accreditation) decision is to be supported by a formal
technical evaluation (certification) of the management, operational,
and technical controls established in an information system's security
plan. As required by OMB, agencies are also to reaccredit their systems
prior to a significant change in processing, but at least every 3 years
(more often where there is a high risk and potential magnitude of
harm).
The Federal Information Security Management Act of 2002 (FISMA)
provides the overall framework for ensuring the effectiveness of
information security controls that support federal operations and
assets and requires agencies and OMB to report annually to the Congress
on their information security programs.[Footnote 1] As part of its
responsibilities under FISMA, OMB requires agencies to report the
number of systems authorized for processing following certification and
accreditation as one of the key performance measures for their
information security programs. Although not required by FISMA, OMB
considers certification and accreditation to be an important
information security quality control, and this process reinforces
several of the act's requirements, including those for a system risk
assessment, a security plan, control testing, and contingency planning.
Further, OMB emphasized the significance of this process in its FY 2003
Report to Congress on Federal Government Information Security
Management,[Footnote 2] in which it noted that most security weaknesses
could be found in operational systems that either have never been
certified or accredited, or whose certification and accreditation is
out of date.
OMB's information technology policies and its authorities under FISMA
generally do not apply to national security systems.[Footnote 3]
However, the head of each agency operating or exercising control of a
national security system is responsible for complying with FISMA
requirements, and agencies such as the Department of Defense (DOD) have
established policies requiring certification and accreditation of
national security systems.
Objectives, Scope, and Methodology:
In response to your request, our objectives were to:
* identify existing governmentwide requirements and guidelines for
certifying and accrediting federal information systems,
* determine the extent to which federal agencies have reported that
their information systems are certified and accredited, and:
* assess whether agencies' certification and accreditation processes
provide (1) consistent and comparable results, and (2) adequate
information for authorizing officials to understand risks and make
informed decisions.
To determine what requirements and guidelines exist for agencies to
follow in certifying and accrediting their systems, we obtained and
reviewed information security policies and guidance issued by OMB, the
National Institute of Standards and Technology (NIST), and DOD,
including its National Security Agency and the Committee on National
Security Systems, which is chaired by DOD's Chief Information Officer.
We also met with representatives from these agencies to discuss these
policies and guidance, as well as to identify any planned revisions or
additional guidance. This included guidance for both non-national
security and national security systems.In addition, to help address all
three of our objectives, we conducted a survey of 24 major departments
and agencies, which included questions on the guidance they follow in
certifying and accrediting their systems.[Footnote 4]
To determine the extent to which agencies have certified and accredited
their systems, we analyzed performance measurement data reported to OMB
by the agencies for their fiscal year 2002 and 2003 annual reporting
and for their March 2004 quarterly updates, which was due to OMB on
March 15, 2004. This performance measurement data largely reflects non-
national security systems, but some agencies also included data on
national security systems.
To assess whether agencies' certification and accreditation processes
provide consistent and comparable results and adequate information for
authorizing officials, we analyzed the results of our survey to
determine the extent to which agencies reported that their processes
addressed specific criteria identified in federal certification and
accreditation guidance, such as a current risk assessment and evidence
of control testing. In addition, for selected systems at four agencies-
-the Departments of Commerce and Energy, the Environmental Protection
Agency (EPA), and the National Aeronautics and Space Administration
(NASA)--we also analyzed certification and accreditation documentation
to determine whether the certification and accreditation criteria were
met. We selected these agencies based primarily on the high percentages
of certified and accredited systems they reported to OMB in their
annual reports for fiscal years 2002 and 2003.
We did not validate the accuracy of the data in agencies' FISMA
reports, survey responses, or system certification and accreditation
documentation. However, we considered the data within the context of a
significant body of existing knowledge and evidence about agency
certification and accreditation practices and, to the extent that they
addressed their agencies' certification and accreditation efforts,
reviewed and compared the results of agencies' inspectors general (IG)
fiscal year 2003 FISMA independent evaluations.
We performed our work in the Washington, D.C., metropolitan area from
September 2003 to June 2004, in accordance with generally accepted
government auditing standards.
Results in Brief:
With NIST's recent issuance of new guidelines, certification and
accreditation processes for federal information systems continue to
evolve. To be used for non-national security systems, the new
guidelines update previous NIST guidance to reflect today's more
distributed computing environment in which systems are constantly
evolving and require real-time, on-going monitoring. These guidelines
also incorporate other recent NIST standards and guidance required by
FISMA, including those to categorize and provide recommended security
controls for federal information systems. Other agencies have also
developed certification and accreditation guidance, particularly for
national security systems.
For the 24 agencies we surveyed, the average percentage of systems
authorized after certification and accreditation was 63 percent for the
first half of fiscal year 2004. However, the status of individual
agencies was mixed, with 7 agencies reporting certification and
accreditation for 90 percent or more of their systems, but 6 reporting
that fewer than half of their systems were certified and accredited.
Our analysis also highlighted inconsistencies in the way agencies
report such certification and accreditation performance data. For
example, national security systems are included in some reported agency
totals, but not in others. Further, there are other factors that lessen
the usefulness of these and other FISMA performance data, including the
limited assurance of data reliability and quality and the need to
refine reporting requirements to provide better information on the
status of agencies' information security efforts.
All the agencies we surveyed reported that their certification and
accreditation processes met criteria consistent with those identified
in federal guidance, such as a current risk assessment, security
control evaluation, and an accreditation statement that indicates the
level of residual risk being accepted by the authorizing official.
However, our review of certification and accreditation documentation
for selected systems at four agencies showed that these criteria were
not always met--results similar to those found by inspectors general
(IGs) in their FISMA evaluations. Further, three of these four agencies
did not have processes to routinely review the quality of their
certification and accreditation efforts--processes that could help
agencies ensure that accrediting officials consistently receive
sufficient information on which to base their decisions.
Survey results also identified potential challenges and obstacles to
agencies' certification and accreditation processes, particularly
regarding funding and staffing issues. The new NIST guidelines suggest
ways to help address resource issues, such as reusing and sharing of
security control development, implementation, and assessment-related
information. Some agencies had also undertaken successful practices in
implementing their certification and accreditation processes that can
help address such challenges, such as redefining system boundaries to
better organize their efforts and manage systems.
This report contains recommendations to the Director of OMB, including
that OMB's information security policy and guidance encourage agencies
to ensure that periodic testing and evaluation of information security
controls, as required by FISMA, include assessing the quality of
security certifications and accreditations to ensure that decisions are
based on consistent consideration of key criteria outlined in federal
guidance. We also recommend that OMB consider changes to its FISMA
reporting guidance, including requiring reporting on the quality and
consistency of certifications and accreditations and encouraging the
IGs to assess agency processes and test agency-reported performance
data as part of their FISMA-mandated independent evaluations.
In oral comments on a draft of this report, OMB representatives in its
Office of Information and Regulatory Affairs and Office of General
Counsel agreed that the quality of agency certification and
accreditation processes varies, and generally agreed with our
recommendations. In addition to the recent issuance of certification
and accreditation guidance by NIST, OMB believes that existing
guidance, including its Circular A-130 and FISMA implementing guidance,
is adequate to ensure that implementation of certification and
accreditation is effective. Further, OMB stated that its planned fiscal
year 2004 FISMA guidance to the agencies would address many of the
issues in our report. The Department of Commerce provided written
comments on a draft of this report (see app. I), and we also received
written and oral technical comments from the Departments of Defense and
Energy, EPA, NASA, and NIST. Comments from all these agencies have been
incorporated into the report, as appropriate.
Background:
FISMA permanently authorized information security program, evaluation,
and reporting requirements for federal agencies. As a key element of
agencies' implementation of FISMA requirements, OMB has continued to
emphasize its longstanding policy of requiring a management official to
formally authorize an information system to process information and
accept the risk associated with its operation based on a formal
evaluation of the system's security controls. Further, compliance with
new FISMA-required standards and guidance will become important
considerations in the certification and accreditation of agency
systems.
FISMA Establishes Federal Information Security Requirements:
Enacted into law on December 17, 2002, as title III of the E-Government
Act of 2002, FISMA assigns specific information security
responsibilities to OMB, NIST, agency heads, chief information officers
(CIO), and IGs. For OMB, these responsibilities include developing and
overseeing the implementation of policies, principles, standards, and
guidelines on information security; and reviewing at least annually,
and approving or disapproving, agency information security programs.
FISMA continues to delegate OMB responsibilities for national security
systems to the Secretary of Defense and the Director of Central
Intelligence. Therefore, OMB's information technology policies and its
authorities under FISMA, as well as federal information system
standards and guidelines developed by NIST, generally do not apply to
national security systems. However, according to FISMA, the head of
each agency operating or exercising control of a national security
system is responsible for providing information security protections
commensurate with the risk and magnitude of harm, implementing
information security policies and practices as required by standards
and guidelines for national security systems, and complying with FISMA
requirements.
FISMA requires each agency, including agencies with national security
systems, to develop, document, and implement an agencywide information
security program to provide information security for the information
and information systems that support the operations and assets of the
agency, including those provided or managed by another agency,
contractor, or other source. Specifically, this program is to include:
* periodic assessments of the risk and magnitude of harm that could
result from the unauthorized access, use, disclosure, disruption,
modification, or destruction of information or information systems;
* risk-based policies and procedures that cost-effectively reduce
information security risks to an acceptable level and ensure that
information security is addressed throughout the life cycle of each
information system;
* subordinate plans for providing adequate information security for
networks, facilities, and systems or groups of information systems;
* security awareness training for agency personnel, including
contractors and other users of information systems that support the
operations and assets of the agency;
* periodic testing and evaluation of the effectiveness of information
security policies, procedures, and practices, performed with a
frequency depending on risk, but no less than annually, and that
includes testing of management, operational, and technical controls for
every system identified in the agency's required inventory of major
information systems;
* a process for planning, implementing, evaluating, and documenting
remedial action to address any deficiencies in the information security
policies, procedures, and practices of the agency;
* procedures for detecting, reporting, and responding to security
incidents; and:
* plans and procedures to ensure continuity of operations for
information systems that support the operations and assets of the
agency.
In addition to these information security program requirements, FISMA
also requires each agency to develop, maintain, and annually update an
inventory of major information systems (including major national
security systems) operated by the agency or that are under its control.
This inventory is to include an identification of the interfaces
between each system and all other systems or networks, including those
not operated by or under the control of the agency.
Under FISMA, each agency must have an annual independent evaluation of
its information security program and practices, including control
testing and compliance assessment. Evaluations of non-national security
systems are to be performed by the agency IG or by an independent
external auditor, while evaluations related to national security
systems are to be performed only by an entity designated by the agency
head.
Other major FISMA provisions require NIST to develop, for systems other
than national security systems, (1) standards to be used by all
agencies to categorize all their information and information systems
based on the objectives of providing appropriate levels of information
security according to a range of risk levels; (2) guidelines
recommending the types of information and information systems to be
included in each category; and (3) minimum information security
requirements for information and information systems in each category.
In conjunction with DOD and the National Security Agency, NIST is
responsible for developing guidelines for identifying an information
system as a national security system.
OMB Continues to Emphasize Certification and Accreditation:
Since the mid-1980s, OMB policy for information technology (IT)
management has required that an agency official attest to the adequacy
of an information system's security safeguards. As currently described
in its Circular A-130,[Footnote 5] OMB requires federal agencies to
ensure that a management official authorizes in writing the use of each
general support system or major application based on implementation of
its security plan before beginning or significantly changing its
processing.[Footnote 6] This management approval, or accreditation, is
the authorization of an IT system to process, store, or transmit
information that provides a form of quality control and challenges
managers and technical staff to find the best fit for security given
technical constraints, operational constraints, and mission
requirements. The accreditation decision is based on the implementation
of an agreed-upon set of management, operational, and technical
controls for a system, and is supported by a comprehensive evaluation
or certification of these security controls that provides the necessary
information for a management official to formally declare that a system
is approved to operate at an acceptable level of risk. OMB policy also
specifies the following:
* Security staff should not make the accreditation decision. In
general, the security official is closer to the day-to-day operation of
the system and will direct or perform security tasks, while the
authorizing official will normally have general responsibility for the
organization supported by the system.
* Agencies are required to reaccredit their systems prior to a
significant change in processing, but at least every 3 years (more
often where there is a high risk and potential magnitude of harm).
With the implementation of FISMA, OMB has continued to emphasize system
certification and accreditation by requiring agencies to report the
number of systems certified and accredited as one of the key
performance measures for reporting under these laws. Continuing this
requirement as part of its overall authority under FISMA to develop and
oversee the implementation of policies, principles, standards, and
guidelines on information security, OMB has taken other steps to help
integrate certification and accreditation into agencies' information
security programs. For example, in the President's fiscal year 2004
budget, OMB established a governmentwide goal that 80 percent of
federal IT systems be certified and accredited by the end of calendar
year 2003. According to OMB, it also monitors the certification and
accreditation of major systems through the budget process with the
possibility that funding could be denied for those IT investments that
do not meet security requirements, such as not being fully certified
and accredited prior to becoming operational. In addition, in its
fiscal year 2003 report to the Congress, OMB outlined a plan of action
to improve performance in IT security that identifies specific steps it
will pursue to assist agencies. One such step concerns the President's
Management Agenda Scorecard, where one criterion that agencies must
meet to "get to green" under the Expanding:
E-Government Scorecard for IT security is to attain certification and
accreditation for 90 percent of their operational IT systems.
FISMA-Required Standards and Guidance Are Important Considerations:
NIST has issued a number of information security standards and guidance
documents that contribute to the certification and accreditation
process, such as its guidance on conducting risk assessments and on the
format and content of security plans.[Footnote 7] In addition, as part
of its statutory responsibilities under FISMA, NIST has issued
additional standards and guidance that will be important considerations
in agencies' future certification and accreditation efforts. As we
reported in our March 2004 testimony,[Footnote 8] these included the
following:
* In December 2003 NIST issued the final version of its Standards for
Security Categorization of Federal Information and Information Systems
(FIPS Publication 199). NIST was required to submit these
categorization standards to the Secretary of Commerce for promulgation
no later than 12 months after FISMA was enacted. These standards are
intended to provide a common framework and understanding for expressing
security that promotes effective management and oversight of
information security programs, and consistent reporting to OMB and the
Congress on the adequacy and effectiveness of information security
policies, procedures, and practices. To help establish security
categories for both information and information systems, the standards
establish three levels of potential impact on organizational
operations, assets, or individuals should a breach of security occur--
high (severe or catastrophic), moderate (serious), and low (limited)--
and are used to determine the impact for each of the FISMA-specified
security objectives of confidentiality, integrity, and
availability.[Footnote 9] Once determined, security categories are to
be used in conjunction with vulnerability and threat information in
assessing the risk to an organization.
* In October 2003 NIST issued an initial public draft of Recommended
Security Controls for Federal Information Systems (Special Publication
800-53) to provide guidelines for selecting and specifying security
controls for information systems categorized in accordance with FIPS
Publication 199. This draft includes baseline security controls for
low-and moderate-impact information systems, with controls for high-
impact systems to be provided in subsequent drafts. This publication,
when completed, will serve as interim guidance until December 2005 (36
months after FISMA enactment), which is the statutory deadline to
publish minimum standards for all non-national security systems. In
addition, testing and evaluation procedures used to verify the
effectiveness of security controls are to be provided this summer in
NIST's Guide for Assessing the Security Controls in Federal Information
Systems (Special Publication 800-53A).
* In August 2003 NIST issued Guideline for Identifying an Information
System as a National Security System (Special Publication 800-59). This
document provides guidelines developed in conjunction with DOD,
including the National Security Agency, to ensure that agencies receive
consistent guidance on the identification of systems that should be
governed by national security system requirements. Except for national
security systems as defined by FISMA, the Secretary of Commerce is
responsible for prescribing standards and guidelines developed by NIST.
DOD and the Director of Central Intelligence have authority to develop
policies, guidelines, and standards for national security systems. The
Director of Central Intelligence is also responsible for policies
relating to systems processing intelligence information.
Certification and Accreditation Guidance Is Provided by NIST and Other
Responsible Agencies:
For more than 20 years, NIST guidance has provided a basic framework
for federal agencies to establish a certification and accreditation
process. As part of its efforts to support FISMA, NIST has recently
issued new certification and accreditation guidance intended, in part,
to create more complete, reliable, and trustworthy information for
accreditation decisions. In addition, other agencies responsible for
national security systems, such as DOD, have also developed
certification and accredition guidance.
Early NIST Guidance Provides Basic Framework:
In September 1983, the National Bureau of Standards, the predecessor to
NIST, issued Federal Information Processing Standards (FIPS)
Publication 102, Guideline for Computer Security Certification and
Accreditation. Identified by OMB in its Circular A-130, this guidance
provided federal agencies with a basic framework for establishing a
certification and accreditation process intended to help improve
management control over computer security and increase computer
security awareness throughout the organization.
FIPS Publication 102 focused on establishing a certification and
accreditation process for sensitive applications, that is, those
applications that require a measure of protection because they process
sensitive information or because of the risk or magnitude of loss or
harm that could result from the improper operation or deliberate
manipulation of them.[Footnote 10] For less sensitive applications, the
guidance advised that a less elaborate process could be used. Elements
of the certification and accreditation process described by this
guidance include the following:
* Roles and responsibilities. Several roles and responsibilities were
identified for the certification and accreditation process, including
the following key roles:
* Accrediting officials are the agency officials who have authority to
accept an application's security safeguards and issue an accreditation
statement that records the decision. These officials must possess
authority to allocate resources to achieve acceptable security and to
remedy security deficiencies. An accrediting official or group of
officials may be responsible for several applications, but there is
typically only one official or group assigned to each application. In
general, the more sensitive the application, the higher the accrediting
officials are in the organization.
* The application certification manager is responsible for managing a
specific certification effort, including planning the effort and
overseeing the production of the security evaluation report. To help
ensure an objective evaluation, this person is to be as independent as
possible from the application being certified.
* Security evaluators are reponsible for performing the technical
security evaluation tasks and providing expert technical judgements in
their areas of specialization. The required specializations vary with
each application, and the more detailed the evaluation, the greater the
specialization required. Useful specialties identified included
application analysts, system analysts, engineers, application
programmers, and system programmers. Security evaluators are to be as
independent as possible from the application.
* Evaluation techniques for security certification. This element
describes the various computer security evaluations that can use
security requirements as criteria and, thus, can be used for
certification. Specifically, these include (1) an analysis of risk to
understand the security problem; (2) validation, verification, and
testing performed in developing the application and throughout its
lifecycle; (3) a security safeguard evaluation performed by people
independent of the application, but internal to the organizational
division in which the application resides (which may include a security
officer); and (4) an electronic data processing audit performed within
internal audit to assess the controls in an organization's system that
rely on computers.
* Performing a certification. The certification process described
consists of five steps: (1) planning the effort to understand the
issues for the entire system and to place boundaries on the work; (2)
collecting critical data and information such as the risk analysis,
inputs, processing steps, outputs, and a listing of application system
controls; (3) performing a basic evaluation of security requirements
and functions, control implementation, and the implementation method;
(4) in the event that a basic evaluation does not provide enough
evidence for certification, performing a detailed evaluation to analyze
the quality of security safeguards; and (5) preparing a security
evaluation report--the primary product of a certification--that
includes both technical and management security recommendations and a
proposed accreditation statement.
* Security evaluation report. The format and contents of the security
evaluation report are described, including major findings, recommended
corrective actions, and a proposed accreditation statement. In
particular, the major findings are to include both proposed residual
vulnerabilities and proposed vulnerabilities requiring correction.
Depending on the seriousness of the security flaw identified,
implementation of an application under development may be delayed or an
operational application may require removal from service. However,
other intermediate alternatives were also identified, such as
withholding accreditiation pending completion of corrections, adding
procedural security controls, restricting the application to process
only nonsensitive or minimally sensitive data, or removing especially
vulnerable application functions or components.
* The accreditation decision and statement. The accrediting official
essentially uses the security evaluation report to evaluate the
certification evidence, decides on the acceptability of security
safeguards, approves corrective actions, signs the accreditation
statement, and ensures that corrective actions are accomplished. The
accreditation statement officially documents the explicit acceptance of
responsibility for computer security, and should identify any
restrictions of operation for the application, as well as any
corrective actions.
* Recertification and reaccreditation. The guidance explains that
certification and accreditation are not permanent, and may need to be
performed again for reasons including changes to the application,
changes in requirements, passage of a time interval (such as the 3-year
interval established by OMB), the occurrence of a significant
violation, or audit or evaluation findings.
New NIST Guidance Intended to Improve the Process:
In May 2004, NIST issued its Guide for the Security Certification and
Accreditation of Federal Information Systems (Special Publication 800-
37) to be used in certifying and accrediting non-national security
systems.[Footnote 11] Developed as part of NIST's project to promote
the development of standards and guidelines to support FISMA, this new
guide is to replace FIPS Publication 102 when it is rescinded (which,
according to a NIST official, should take place in the next six
months). At the time of our survey, all 24 agencies reported that they
planned to adopt or modify their existing guidance to be consistent
with Special Publication 800-37, and 14 agencies reported they already
used a draft version of the guidance.
As discussed in the guide, its overall purpose is to help achieve more
secure information systems within the federal government by:
* enabling more consistent, comparable, and repeatable evaluations of
security controls applied to federal information systems;
* promoting a better understanding of agency-related mission risks
resulting from the operation of information systems; and:
* creating more complete, reliable, and trustworthy information for
authorizing officials to facilitate more informed security
accreditation decisions.
Further, NIST encourages state, local, and tribal governments, as well
as private-sector organizations comprising the critical infrastructure
of the United States, to consider the use of these guidelines, as
appropriate.
The new NIST guidance updates the process described in FIPS Publication
102. For example, according to a NIST official, the certification
process in FIPS Publication 102 was a static evaluation of systems
where systems were tested at a given, single point in time to determine
the overall risk. Further, this official stated that although this
process was an adequate measure 20 years ago, in today's more
distributed computing environment where systems are constantly
evolving, real-time, ongoing monitoring is required. As a result, the
new process described in Special Publication 800-37 identifies four
phases, which includes a continuous monitoring phase. Each of these
phases--initiation, security certification, security accreditation,
and continuous monitoring--consists of a set of defined tasks and
subtasks that are to be carried out by the various roles assigned for
the process. To help illustrate this process, figure 1 provides a high-
level view, along with the key tasks associated with each phase.
Figure 1: NIST Security Certification and Accreditation Process:
[See PDF for image]
[End of figure]
The new guidance continues to emphasize the assessment of risk and the
development of system security plans as two important activities in an
agency's information security program that directly support the
security accreditation process. It also emphasizes the importance of
the security assessment (certification) in the accreditation process to
help ensure that agency officials have the most complete, accurate, and
trustworthy information possible on the security status of their
information systems in order to make timely, credible, risk-based
decisions on whether to authorize operation of those systems. The guide
also emphasizes several new concepts and includes other significant
changes from FIPS Publication 102, such as the incorporation of FISMA-
mandated standards and guidelines into the process. This and other new
concepts and changes are discussed below.
FISMA-Required Standards Are Incorporated:
FISMA-required standards issued by NIST are incorporated as an integral
part of the new certification and accreditation process. The
certification and accreditation guideline identifies specific examples
of how these standards are considered, including the following:
* The security category of an information system (overall potential
impact level of high, moderate, or low) assigned based on FIPS
Publication 199 influences the initial selection of security controls
from NIST Special Publication 800-53 and the initial selection of
assessment methods and procedures from NIST Special Publication 800-
53A. The level of effort applied to the certification and accreditation
tasks and subtasks should be commensurate with the strength of the
security controls selected and the rigor and formality of the
assessment methods and procedures selected. Further, because of the
limited adverse effect expected for low-impact systems, the scalability
of the certification and accreditation process for these systems
results in the elimination of the independent certification agent, the
incorporation of self-assessment activities, and a reduction in the
associated level of supporting documentation and paperwork.
* The security category of the information system should guide the
degree of independence of the certification agent. When the potential
impact on agency operations, agency assets, or individuals is low, a
self-assessment activity may be reasonable and appropriate and not
require an independent certification agent. When the potential agency-
level impact is moderate or high, certification agent independence is
needed and justified.
* Security categories can play an important part in helping to define
the accreditation boundary for an information system by partitioning
the agency's information systems according to the criticality or
sensitivity of the systems and the importance of those systems in
accomplishing the agency's mission.
* Information systems, especially mission-critical or high-impact
systems, should not be operating with significant security
vulnerabilities requiring extended remediation time.
Additional Roles and Responsibilities:
The guide defines additional participants in the certification and
accreditation process and provides further clarification of the
responsibilities of others. For example, it identifies the roles played
by the chief information officer, senior agency information security
officer, information system owner, information system security officer,
certification agent, and user representative(s). The guide also creates
a new role of authorizing official's designated representative to act
on the authorizing official's behalf in coordinating and carrying out
the necessary activities required during the security certification and
accreditation process. The designated representative interacts with
other participants in the process; can be empowered by the authorizing
official to make certain decisions, such as acceptance of the system
security plan; and may also be called upon to prepare the final
security accreditation package. However, the authority to make the
security accreditation decision and to sign the associated decision
letter remains with the authorizing official and cannot be delegated to
the designated representative.
The guide continues to identify the authorizing official as the
official who, through the accreditation decision, assumes
responsibility and is accountable for the risks associated with
operating an information system. It also indicates that this official
should have the authority to oversee the budget or business operations
of the information system within the agency and is often called upon to
approve system security requirements and system security plans.
Further, in addition to authorizing system operation, the authorizing
official can issue an interim authorization to operate the system under
specific terms and conditions or deny authorization to operate the
system (or if the system is already operational, halt operations) if
unacceptable security risks exist.
Common Security Controls:
The NIST guideline describes common security controls that can apply to
all agency information systems, a group of information systems at a
specific site (sometimes associated with the terms site certification/
accreditation), or common information systems, subsystems, or
applications (that is, common hardware software, and/or firmware)
deployed at multiple operational sites (sometimes associated with the
terms type certification/accreditation). Common security controls are
typically identified during a collaborative agencywide process with the
involvement of the senior agency information security officer,
authorizing officials, information system owners, and information
system security officers. The results from the assessment of such
controls can be used to support the security certification and
accreditation processes of agency information systems where those
controls have been applied. Further, many of the management and
operational controls (e.g., contingency planning controls, incident
response controls, security training and awareness controls, personnel
security controls, and physical security controls) needed to protect an
information system may be excellent candidates for common security
control status.
Conditions for Interim Authorization to Operate:
If, after assessing the results of the security certification, the
authorizing official deems that the risk to agency operations, agency
assets, or individuals is unacceptable, but there is an overarching
mission necessity to place the information system into operation or
continue its operation, an interim authorization to operate may be
issued. An interim authorization to operate is provided when the
identified security vulnerabilities in the information system resulting
from deficiencies in the planned or implemented security controls are
significant, but can be addressed in a timely manner. Further, an
interim authorization provides a limited authorization to operate the
information system under specific terms and conditions and acknowledges
greater risk to the agency for a specified period of time. These terms
and conditions are established by the authorizing official and convey
limitations on information system operations.
Documentation of Security Accreditation:
The accreditation package documents the results of the security
certification and provides the authorizing official with the essential
information needed to make a credible, risk-based decision on whether
to authorize operation of the information system. The package is
generally compiled and submitted by the information system owner, who
receives inputs from the information system security officer,
certification agent, and senior agency information security officer.
The package contains the approved system security plan, security
assessment report, and plan of action and milestones,[Footnote 12] and
is submitted to the authorizing official or designated representative.
The accreditation decision letter is used to transmit the decision from
the authorizing official to the information system owner. Prepared for
the authorizing official by the designated representative, the final
letter should contain the accreditation decision, supporting rationale
for the decision, and terms and conditions for the authorization. It
also indicates whether the system is fully authorized to operate,
authorized to operate on an interim basis under strict terms and
conditions, or not authorized to operate. The accreditation decision
letter is attached to the original accreditation package and returned
to the information system owner, who maintains this documentation.
Transition to New Certification and Accreditation Guidance:
Although OMB representatives state that its Circular A-130 is being
revised, the current version does not reflect FISMA requirements or
recent guidance issued by NIST. Although OMB requires agencies to
ensure that their policies, standards, and procedures are consistent
with NIST guidance, specifically requiring security certification and
accreditation processes consistent with NIST's Special Publication 800-
37 guidance in OMB policy and guidance would help ensure consistency in
implementing such processes. To help with the transition to NIST's
Special Publication 800-37, in July 2003 OMB issued interim guidance
summarizing the minimum activities that agencies should implement to
comply with the certification and accreditation requirement in OMB
Circular A-130, as well as to facilitate easy alignment when the NIST
guideline is finalized. Among other things, the interim guidance
encouraged the use of NIST's Security Self-Assessment Guide for
Information Technology Systems for conducting certification reviews,
which uses an extensive questionnaire containing specific control
objectives and techniques against which an unclassified system or group
of interconnected systems can be tested and measured.[Footnote 13]
Responsible Agencies Provide Guidance for National Security Systems:
Because OMB's authorities and NIST guidance are not applicable to
national security systems, agencies responsible for these systems have
also issued certification and accreditation guidance. The processes and
criteria established by this guidance are similar to those required by
NIST guidance for non-national security systems, that is, they require
risk assessments, verification of security requirements in a security
plan or other document, testing of security controls, and formal
authorization by an authorizing official (or designated approving/
accrediting authority, as referred to by some agencies). Guidance
issued by other agencies include the following:
* DOD Directive 8500.1 on information assurance requires the heads of
all components to comply with established accreditation processes
required for all DOD information systems, and DOD Instruction Number
5200.40 creates the DOD Information Technology Security Certification
and Accreditation Process (DITSCAP) for both unclassified and
classified automated information systems, networks, and sites in the
department.[Footnote 14] Organized within four phases--definition,
verification, validation, and post accreditation--a key element of
DITSCAP is the development of an agreement among the program manager,
the designated approving authority, the certification authority, and
the user representative during the definition phase. This agreement
(the System Security Authorization Agreement) is used throughout the
entire DITSCAP to guide actions, document decisions, specify security
requirements, document certification tailoring and level of effort,
identify potential solutions, and maintain operational systems
security.
* The National Information Assurance Certification and Accreditation
Process, issued by the DOD-chaired National Security Telecommunications
and Information Systems Security Committee (now the Committee on
National Security Systems), establishes minimum national standards for
certifying and accrediting national security systems.[Footnote 15] A
key element of this guidance is the agreement among the program
manager, designated approving authority (accreditor), certification
agent (certifier), and user representative, who resolve critical
schedule, budget, security, functionality, and performance issues.
Agreements are documented in a System Security Authorization Agreement,
which is used to guide and document the results of the certification
and accreditation.
* Director of Central Intelligence Directive 6/3, Protecting Sensitive
Compartmented Information Within Information Systems, and its
implementation manual provide policy and procedures for the security
and protection of systems that create, process, store, and transmit
intelligence information, as well as define and mandate the use of a
risk management process and a certification and accreditation
process.[Footnote 16] The certification process described by this
guidance includes validation that appropriate levels of concern for
integrity and availability and an appropriate confidentiality
protection level have been selected from tables and descriptions
provided in the implementation manual, and that required safeguards
have been implemented as described in the system security plan. This
process also considers other factors associated with the information
system and its operational environment, including mission criticality,
functional requirements, information system security boundaries,
threat and vulnerability assessments, and other intelligence-related
factors.
Reported Percentages of Systems Certified and Accredited Vary Widely:
For the 24 agencies we surveyed, the average percentage of systems
authorized after certification and accreditation was 63 percent for the
first half of fiscal year 2004. However, the status at individual
agencies was mixed, with six reporting that they have certified and
accredited less than half of their systems. Our analysis also
highlighted inconsistencies in the way agencies report such
certification and accreditation performance data. For example, national
security systems are included in some reported agency totals, but not
in others. Further, there are other factors that lessen the usefulness
of these and other FISMA performance data, including the limited
assurance of data reliability and quality and the need to refine
reporting requirements to provide better information on the status of
agencies' information security efforts.
Progress by Individual Agencies Is Mixed:
The average percentage of systems authorized after certification and
accreditation reported by the 24 agencies was 63 percent for the first
half of fiscal year 2004. This compares to 48 percent for fiscal year
2002 and to 62 percent for fiscal year 2003. Despite this reported
overall progress, the status of individual agencies varies widely. For
example, 7 agencies reported more than 90 percent of their systems were
certified and accredited for the first half of fiscal year 2004,
including the Nuclear Regulatory Commission, which reported 100
percent. In contrast, 6 agencies reported less than half of their
systems were certified and accredited, including the Department of
Housing and Urban Development, which reported none. Table 1 summarizes
the percentages reported by the agencies for the 2 fiscal years and for
the first half of fiscal year 2004.
Table 1: Agency Systems Reported as Authorized After Certification and
Accreditation:
Department or agency: Agency for International Development;
Percentage by fiscal year: 2002: 100%;
Percentage by fiscal year: 2003: 88%;
Percentage by fiscal year: 1st Half 2004: 70%.
Department or agency: Agriculture;
Percentage by fiscal year: 2002: 8%;
Percentage by fiscal year: 2003: 14%;
Percentage by fiscal year: 1st Half 2004: 0[B]%.
Department or agency: Commerce;
Percentage by fiscal year: 2002: 77%;
Percentage by fiscal year: 2003: 97%;
Percentage by fiscal year: 1st Half 2004: 96%.
Department or agency: Defense;
Percentage by fiscal year: 2002: 55%;
Percentage by fiscal year: 2003: 80%;
Percentage by fiscal year: 1st Half 2004: 77%.
Department or agency: Education;
Percentage by fiscal year: 2002: 0%;
Percentage by fiscal year: 2003: 13%;
Percentage by fiscal year: 1st Half 2004: 61%.
Department or agency: Energy;
Percentage by fiscal year: 2002: 46%;
Percentage by fiscal year: 2003: 83%;
Percentage by fiscal year: 1st Half 2004: 86%.
Department or agency: Environmental Protection Agency;
Percentage by fiscal year: 2002: 87%;
Percentage by fiscal year: 2003: 94%;
Percentage by fiscal year: 1st Half 2004: 94%.
Department or agency: General Services Administration;
Percentage by fiscal year: 2002: 13%;
Percentage by fiscal year: 2003: 22%;
Percentage by fiscal year: 1st Half 2004: 58%.
Department or agency: Health and Human Services;
Percentage by fiscal year: 2002: 11%;
Percentage by fiscal year: 2003: 41%;
Percentage by fiscal year: 1st Half 2004: 59%.
Department or agency: Homeland Security;
Percentage by fiscal year: 2002: [A]%;
Percentage by fiscal year: 2003: 42%;
Percentage by fiscal year: 1st Half 2004: 59%.
Department or agency: Housing and Urban Development;
Percentage by fiscal year: 2002: 72%;
Percentage by fiscal year: 2003: 9%;
Percentage by fiscal year: 1st Half 2004: 0[B]%.
Department or agency: Interior;
Percentage by fiscal year: 2002: 22%;
Percentage by fiscal year: 2003: 10%;
Percentage by fiscal year: 1st Half 2004: 19%.
Department or agency: Justice;
Percentage by fiscal year: 2002: 76%;
Percentage by fiscal year: 2003: 79%;
Percentage by fiscal year: 1st Half 2004: 88%.
Department or agency: Labor;
Percentage by fiscal year: 2002: 70%;
Percentage by fiscal year: 2003: 58%;
Percentage by fiscal year: 1st Half 2004: 85%.
Department or agency: National Aeronautics and Space Administration;
Percentage by fiscal year: 2002: 89%;
Percentage by fiscal year: 2003: 98%;
Percentage by fiscal year: 1st Half 2004: 98%.
Department or agency: National Science Foundation;
Percentage by fiscal year: 2002: 30%;
Percentage by fiscal year: 2003: 95%;
Percentage by fiscal year: 1st Half 2004: 95%.
Department or agency: Nuclear Regulatory Commission;
Percentage by fiscal year: 2002: 50%;
Percentage by fiscal year: 2003: 90%;
Percentage by fiscal year: 1st Half 2004: 100%.
Department or agency: Office of Personnel Management;
Percentage by fiscal year: 2002: 0%;
Percentage by fiscal year: 2003: 91%;
Percentage by fiscal year: 1st Half 2004: 94%.
Department or agency: Small Business Administration;
Percentage by fiscal year: 2002: 65%;
Percentage by fiscal year: 2003: 74%;
Percentage by fiscal year: 1st Half 2004: 87%.
Department or agency: Social Security Administration;
Percentage by fiscal year: 2002: 100%;
Percentage by fiscal year: 2003: 100%;
Percentage by fiscal year: 1st Half 2004: 100%.
Department or agency: State;
Percentage by fiscal year: 2002: 0%;
Percentage by fiscal year: 2003: 36%;
Percentage by fiscal year: 1st Half 2004: 38%.
Department or agency: Transportation;
Percentage by fiscal year: 2002: 8%;
Percentage by fiscal year: 2003: 33%;
Percentage by fiscal year: 1st Half 2004: 49%.
Department or agency: Treasury;
Percentage by fiscal year: 2002: 43%;
Percentage by fiscal year: 2003: 24%;
Percentage by fiscal year: 1st Half 2004: 58%.
Department or agency: Veterans Affairs;
Percentage by fiscal year: 2002: 31%;
Percentage by fiscal year: 2003: 39%;
Percentage by fiscal year: 1st Half 2004: 12%.
Average percentage;
Percentage by fiscal year: 2002: 48%;
Percentage by fiscal year: 2003: 62%;
Percentage by fiscal year: 1st Half 2004: 63%.
Sources: OMB, agencies (data), and GAO (analysis).
[A] The Department of Homeland Security began its FISMA reporting in
fiscal year 2003. However, the fiscal year 2002 percentage included the
Federal Emergency Management Agency, which became part of the new
department. Components of other agencies also became part of the
department, including the U.S. Coast Guard and U.S. Customs Service,
which were formerly within the Departments of Transportation and the
Treasury, respectively.
[B] Agriculture and Housing and Urban Development officials indicated that
concerns over the quality and consistency of their certification and
accreditation processes were the basis for reporting no certified and
accredited systems during the first half of 2004. Both agencies have
sought the services of contractors to assist them in establishing a
certification and accreditation process and in ensuring that most, if
not all, of their agencies' systems are certified and accredited by the
end of calendar year 2004.
[End of table]
As shown in table 1, in comparing fiscal year 2003 results with those
shown for the first half of 2004, agencies showing the greatest
increase included Education (+48 percentage points) and the General
Services Administration (+ 36 percentage points). On the other hand,
some showed decreasing percentages, including Veterans Affairs (-27
percentage points) and Agriculture (-14 percentage points).
In responding to our survey, agencies cited several reasons why not all
of their systems were certified and accredited. These reasons included
systems' being decommissioned or retired; agency efforts' being focused
on the most critical systems, with the less critical systems' being
scheduled later; higher priority operational requirements and limited
funding; and legacy systems' being unable to support required technical
controls.
Our analysis of survey responses also highlighted instances in which
agencies report performance measurement data differently. For example,
some agencies, such as Energy, include both non-national security and
national security systems in their reported performance data, while
others, such as NASA, do not include their national security systems.
As another example, DOD includes systems with interim authorization to
operate among those systems reported as certified and accredited
because, according to DOD officials, interim authorizations still
represent a management approval to operate. In contrast, the National
Science Foundation does not report systems with interim authorization
to operate as certified and accredited. OMB instructions for fiscal
year 2003 FISMA reporting were not specific regarding whether national
security systems should be reflected in agency performance measurement
data nor did they address how to report systems with interim
authorization to operate. OMB representatives indicated that national
security systems are to be reflected in reporting performance
measurement data and that only systems granted full authorization to
operate should be considered in reporting the number of systems
certified and accredited. Clarification of such issues in future FISMA
guidance would improve consistency and comparability of agency-reported
FISMA information.
In analyzing these and future results indicated by agency-reported
percentages of systems authorized after certification and
accreditation, it is also important to consider several factors that
lessen the usefulness of performance measurement data being reported by
the agencies for FISMA. As first discussed in our March 2004
testimony,[Footnote 17] these factors include the following:
* Limited assurance of data reliability and quality. The FISMA
performance measures reported by the agencies are primarily based on
self-assessments and are not independently validated. OMB did not
require IGs to validate agency responses to the performance measures,
but did instruct them to assess the reliability of the data for the
subset of systems they evaluate as part of their independent
evaluations. Nonetheless, some IG evaluations did identify problems
with data reliability and quality that could affect agency performance
data. For example, for the performance measurement on the number of
agency systems authorized for processing after certification and
accreditation, six IGs indicated different results from those reported
by their agencies, for reasons such as out-of-date certifications and
accreditations. Further, as we discuss later in more detail, other IGs
identified problems with the quality of the certifications and
accreditations, such as security control reviews not being performed.
OMB's requirement for IGs to assess the reliability of such information
as part of their FISMA responsibilities could provide valuable
information on the quality of reported FISMA information and assist
management and Congress in their FISMA oversight. For example, for
certifications and accreditations for the subset of systems they
review, the IGs could determine whether the agencies met specific
criteria, including a current risk assessment and security plan,
control testing, and contingency planning and determine whether such
information is accurately reflected in the agencies' compilation of
related performance measures.
* Accuracy of agency system inventories. The total number of agency
systems is a key element in OMB's performance measures, in that agency
progress is indicated by the percentage of total systems that meet
specific information security requirements. Thus, inaccurate or
incomplete data on the total number of agency systems affects the
percentage of systems shown as meeting the requirements. FISMA requires
that each agency develop, maintain, and annually update an inventory of
major information systems operated by the agency or under its control.
However, according to their fiscal year 2003 FISMA reports, only 13 of
the 24 agencies reported that they had completed their system
inventories. Further, independent evaluations by IGs for 3 of these 13
agencies did not agree that system inventories were complete. Although
we recently reported that all 24 agencies now report they develop and
maintain the FISMA-required inventory of major information
systems,[Footnote 18] maintaining an accurate inventory will continue
to be a key element of agency performance measures and in ensuring that
information security programs cover all agency systems.
* Further refinement of performance measures. Refinement of FISMA
performance measurement data is needed to provide better information on
the status of agencies' information security efforts. For example, OMB
currently requires agencies to report performance data in aggregate for
the total number of agency systems, but does not require information
that could be used to better assess the quality of certifications and
accreditations performed, such as reporting systems according to their
risk or security category, which would help indicate whether agencies
are prioritizing their efforts according to risk and focusing on their
most important systems. All the agencies responding to our survey
indicated that they did prioritize their certification and
accreditation efforts to focus on their most important systems.
However, during our review of certifications and accreditations
processes at the four agencies we visited, we noted that system
prioritization was not always used to monitor overall activity. In
fact, at one agency, system priority was not indicated in its overall
inventory of systems, and one system identified by the agency as a
national critical asset for critical infrastructure protection purposes
had not been certified and accredited.[Footnote 19] The agency has
since acted to certify and accredit this system, recently reporting its
full accreditation as of June 2004. OMB has also recognized the need
for further information on agencies' certification and accreditation
processes. According to its fiscal year 2003 report to the Congress, in
fiscal year 2004 FISMA guidance, OMB planned to further emphasize
security performance measurement, including evolving performance
measures to move beyond status reporting to also identify the quality
of the work done, such as determining both the number of systems
certified and accredited and the quality of certification and
accreditation conducted.
Processes at Selected Agencies Do Not Ensure Consistent or Adequate
Information:
Although agencies responding to our survey indicated that their
certification and accreditation processes required that specific
criteria identified in federal guidance be met, our review of
certification and accreditation documentation for selected systems at
four agencies, as well as IG FISMA evaluations for fiscal year 2003,
noted instances in which agencies do not consistently meet such
criteria as a current risk assessment and security control evaluation.
Further, three of the four agencies we reviewed had no routine
processes to ensure that such criteria are met. In describing their
processes, agencies identified challenges and obstacles to implementing
an effective certification and accreditation program, such as resource
and staffing constraints. They also identified successful practices to
help mitigate such challenges.
Agencies Report Using Consistent Criteria:
Agency responses to our survey showed that their certification and
accreditation processes were generally consistent in how they defined
system boundaries for certification and accreditation, with all 24
agencies reporting that they identified systems using OMB's definitions
of a general support system and a major application. In addition,
essentially all the agencies reported that their certification and
accreditation processes for both new and existing systems required
documentation or evidence to show that specific criteria found in
federal guidance are met, such as requiring a current risk assessment
and a security control evaluation. However, in one area--contingency
plan testing--4 agencies (17 percent) reported that their processes did
not require documentation that plans were tested. Two of these agencies
reported that contingency plan testing was not required because either
they thought it was inappropriate for new systems or their security
program did not require such testing.[Footnote 20] Table 2 summarizes
the agency responses for specific certification and accreditation
criteria.
Table 2: Certification and Accreditation Criteria Required to Be Met by
Processes at 24 Major Agencies:
Criterion: Current risk assessment?
Agency Responses: Yes: 24;
Agency Responses: No: 0.
Criterion: Current security plan updated to reflect certification
results?
Agency Responses: Yes: 23;
Agency Responses: No: 1.
Criterion: Evaluated and documented management, operational, and
technical security controls/requirements?
Agency Responses: Yes: 23;
Agency Responses: No: 1.
Criterion: A plan with milestones prepared to correct weaknesses
identified during security control evaluation?
Agency Responses: Yes: 24;
Agency Responses: No: 0.
Criterion: Written management authorization that details the rules of
behavior for systems that interface/interconnect with other agencies or
contractors?
Agency Responses: Yes: 23;
Agency Responses: No: 1.
Criterion: A current and adequate contingency plan?
Agency Responses: Yes: 22;
Agency Responses: No: 2.
Criterion: System contingency plan has been tested?
Agency Responses: Yes: 20;
Agency Responses: No: 4.
Criterion: Results of certification tests attested to by the certifier?
Agency Responses: Yes: 22;
Agency Responses: No: 2.
Criterion: Residual risk identified by the certifier?
Agency Responses: Yes: 23;
Agency Responses: No: 1.
Criterion: Specific corrective actions identified and recommended by
the certifier?
Agency Responses: Yes: 22;
Agency Responses: No: 2.
Criterion: An accreditation statement authorizing the system to process
information and signed by the authorizing official?
Agency Responses: Yes: 23;
Agency Responses: No: 1.
Criterion: An accreditation statement that indicates the level of
residual risk being accepted by the authorizing official?
Agency Responses: Yes: 23;
Agency Responses: No: 1.
Criterion: Authorizing official is a management official with general
responsibility for the organizational mission supported by the system?
Agency Responses: Yes: 22;
Agency Responses: No: 2.
Source: Agency responses to GAO survey.
[End of table]
Processes at Selected Agencies Do Not Ensure that Criteria Are Met:
Although the 24 agencies reported that they require specific criteria
to be met, our analyses of documentation at 4 agencies for the
certification and accreditation of a total of 32 mission-or national-
critical systems showed that such documentation did not always
demonstrate that specific criteria were met. For example, only 22 of
the 32 systems showed results of control testing and only 19 systems
had contingency plans. In addition, documentation for only 17 of the
systems identified the actual residual risk being accepted by the
accrediting official. Table 3 summarizes results for these and other
criteria for the agencies.
Table 3: Number and Percentage of 32 Selected Agency Systems Meeting
Specific Certification and Accreditation Criteria:
Criterion: Current risk assessment?
Number of systems meeting criterion (percentage): 23 (72%).
Criterion: Current security plan?
Number of systems meeting criterion (percentage): 26 (81%).
Criterion: Controls tested?
Number of systems meeting criterion (percentage): 22 (69%).
Criterion: Contingency plan?
Number of systems meeting criterion (percentage): 19 (59%).
Criterion: Contingency plan tested?
Number of systems meeting criterion (percentage): 8 (42%)[A].
Criterion: Plan with milestones prepared for weaknesses?
Number of systems meeting criterion (percentage): 17 (81%)[B].
Criterion: Residual risk identified?
Number of systems meeting criterion (percentage): 17 (53%).
Source: GAO analysis of agency data.
[A] Percentage based on the total of 19 systems with contingency plans.
[B] Percentage based on 21 systems for which plans were required to
correct identified weaknesses.
[End of table]
As we recently testified, results of IG FISMA independent evaluations
have also demonstrated deficiencies in agencies' certifications and
accreditations.[Footnote 21] Some of their fiscal year 2003 FISMA
reports identified instances in which certifications and accreditations
were not current and controls were not tested. Others also recommended
improvements in agency processes. For example, for the Office of
Personnel Management, the IG recommended that the agency develop a
procedure to ensure that all documented findings and corrective actions
are reviewed by both the certification and accreditation officials and
included in the certification statement, accreditation statement, and
plan of action and milestones report.
At the four agencies we reviewed, only the IGs at Commerce and Energy
specifically addressed certification and accreditation as part of their
fiscal year 2003 FISMA reporting. The Commerce IG recognized that the
department was undergoing changes in implementing new certification and
accreditation guidance, but reported cases in which system
certification was granted without evidence of testing. The Energy IG
reported findings that included lack of security control reviews and
management authorizations to operate systems, as well as risk
assessments that were incomplete or outdated and system security plans
that were missing critical elements or did not cover changes to their
IT environment.
Lastly, CIO offices at the four agencies we reviewed monitored the
status of system certifications and accreditations agencywide, but only
one--Commerce--routinely assessed the quality of its efforts. Largely
to facilitate FISMA reporting to OMB, the agencies all had processes to
update the status of system certification and accreditation activities,
ranging from periodic data calls at Energy to EPA's use of its
Automated Security Self-Evaluation and Remediation Tracking tool to
centrally track Web-enabled plans of action and milestones
reports.[Footnote 22] These processes do not ensure the quality of the
certifications and accreditations, such as whether the criteria
identified in guidance are met. Such a quality control process could
facilitate accrediting officials consistently receiving sufficient
information on which to base their decisions, yet only Commerce had an
agencywide process to routinely ensure quality. As described by a
Commerce IT security official, the department has a continuous,
comprehensive control review process that includes annual program and
system evaluations through both self-assessments by component program
managers and compliance reviews by IT security officials under the
Commerce CIO. Specifically with regard to certification and
accreditation, the process includes the use of a checklist on the
content and quality of the documentation. Further, as part of the
compliance review process, in fiscal year 2003, Commerce conducted
reviews to ensure that all the department's classified, mission-
critical, and national-critical systems met legal and departmental
requirements. These reviews included checks for compliance with
certification and accreditation criteria, such as risk assessments,
contingency plans, certifier's statements, and accreditation letters.
According to the Commerce official, such reviews will continue to be
conducted on a sample basis with all systems reviewed at least once
over a 3-year review cycle. An official at Energy also identified a
process to independently verify and validate that department's
certification and accreditation packages, but explained that due to the
large number of systems, this process has been limited to reviews for
its headquarters systems. This official added that to help address this
issue, they are working with the IG's office to have it begin
conducting random reviews of certification and accreditation packages
this fall.
Challenges and Obstacles to Agency Processes:
Through our survey and interviews with agency staff, agencies noted
several overall challenges or obstacles to efforts to certify and
accredit their systems. Funding and staffing issues were most commonly
indicated, including those associated with implementing the new NIST
guidance.
According to OMB's March 2004 report to the Congress, funding for IT
security has increased from $2.7 billion in FY 2002 to $4.2 billion in
FY 2003. Nevertheless, a total of 18 agencies identified funding as a
challenge to performing their certifications and accreditations. For
example, Commerce noted that certification and accreditation was an
expensive process and that in order to develop and implement its
program, it had to reprogram and reprioritize internal funds and absorb
costs in existing funding levels. In another case, the Department of
Health and Human Services stated that because of limited funding,
higher emphasis is placed on using funds to certify and accredit new
systems as opposed to existing systems. Energy also noted that funding
was a challenge because security costs were not integrated into the
overall life-cycle costs for all of its systems. Despite these and
other concerns related to security cost funding, most agencies did not
know how much they spent on certification and accreditation. For
example, only 11 agencies could identify their actual or estimated
costs for fiscal year 2003, which totaled $75.5 million for these
agencies.
Nineteen of the agencies we surveyed also reported that they had
encountered staffing challenges for their certification and
accreditation activities that essentially consisted of the need for
full-time staff with the appropriate backgrounds, specialized skills,
and security clearances. In addition, 13 agencies reported challenges
in providing training to staff or officials responsible for certifying
or accrediting agency systems.
In Special Publication 800-37, NIST acknowledges that the cost of
conducting certifications and accreditations on large numbers of
information systems with varying degrees of complexity is a critical
issue facing agencies today. NIST suggests part of the solution is
promoting the reuse and sharing of security control development,
implementation, and assessment-related information in the agency's
agencywide information security program, including:
* employment of standardized security controls and methods for
assessing those controls;
* development of standardized assessment plans, methods and procedures
to be used in security certifications and accreditations;
* adoption, specification, and promulgation of standardized policies,
procedures, and documentation for common security program areas (e.g.,
rules of behavior, system administration, auditing, system monitoring,
vulnerability scanning, management of user accounts, configuration
management, incident response, contingency planning, and system
maintenance);
* refinement of policies, procedures, and documentation on a system-by-
system basis, as needed, by preparing amendments or adding system-
specific appendixes;
* adoption, publication, and distribution (preferably in an online
database) of agency-prescribed or -developed security implementation
guidance;
* establishment of a protected central repository, preferably online,
for all certification and accreditation documentation, acquisition-
related information, risk and vulnerability assessments, compliance
surveys, security incident reporting and remediation results, external
security audits, and making these easily accessible by appropriate
agency personnel; and:
* procurement of agencywide licenses for automated tools such as
vulnerability scanners, online security monitoring tools, audit
reduction tools, and certification and accreditation support tools.
As another means to help address the cost of certification and
accreditation, the NIST guideline also highlights the importance of
leveraging the results of previous assessments and audits conducted on
an agency's information system or the particular products comprising
that system. Potential sources identified include commercial product
testing and evaluation programs, privacy impact assessments, physical
security assessments, self-assessments, and internal and external
audits. According to the guideline, these assessments and audits can
support the security certification and accreditation process by helping
to gauge the preparedness of an information system for security
certification and accreditation by examining the status of key security
controls in the system and by potentially being reused as evidence,
when appropriate, during the security certification and accreditation
process. Further, evidence from other assessments and audits can help
reduce the potential cost of security certification and accreditation,
as well as increase the overall confidence in the final certification
and accreditation results.
Although the NIST guideline emphasizes leveraging the results of
previous assessments and audits, it is important that agencies note the
difference between the level of control testing envisioned for annual
FISMA testing and that performed for system certification and
accreditation. FISMA requires agencies to periodically test and
evaluate the effectiveness of information security policies,
procedures, and practices for each system with a frequency depending on
risk, but no less than annually. In contrast, current OMB policy
requires agencies to reaccredit their systems (which also includes
control testing) at least every 3 years. In its fiscal year 2003 FISMA
reporting guidance, OMB distinguished between these two requirements,
explaining that annual FISMA testing is not of the complexity required
for certification and accreditation of systems as described in NIST
guidance. Rather, the FISMA provision recognizes the importance of
maintaining a continuous process of assessing risk and ensuring that
security controls maintain risk at an acceptable level and underscores
the need to understand the security status of each system in order to
accurately maintain system-level plans of action and milestones and
report annually on the overall health of an agency's IT security
program.
During our review, agencies also identified some actions that can help
address identified challenges and contribute to more efficient and
effective certification and accreditation processes. In particular,
citing proactive senior management support as critical to the success
of its program, Commerce identified several actions, including that it
has:
* informed program mangers of their responsibilities and held them
accountable for the security of IT resources;
* redefined system boundaries to better organize certification and
accreditation efforts and manage systems;
* collaborated to solve common obstacles and to optimize available
internal departmental resources both in the central security program
office and in other bureaus to overcome skills gaps and staff
shortages;
* provided role-based training that tailors certification and
accreditation requirements and responsibilities to those with IT
security roles; and:
* reviewed mission critical and national critical systems to ensure
that they are in compliance with the department's security policy and
guidance.
Other identified actions included those by Transportation, which
maintains a dedicated, trained, experienced staff of contractors as
part of its centralized certification process and provides training to
system owners during the certification process. In addition, as
mentioned previously, EPA has developed a tool to annually evaluate the
risk in computer systems and to produce and centrally track Web-enabled
plans of action and milestones reports. EPA is offering this tool to
other agencies, including hosting the tool for them at its National
Computer Center. Lastly, 21 of the 24 agencies surveyed reported that
they used automated tools as part of their certification and
accreditation process for a number of functions, including managing the
process and developing documentation, tracking corrective actions,
configuration management, vulnerability scanning, penetration testing,
and technical controls testing.
Conclusions:
Certification and accreditation has become a key measure in determining
the status of agencies' information security programs, and NIST and
other agencies have provided overall guidance to assist agencies in
establishing effective certification and accreditation. Agencies are
reporting increasing numbers of systems certified and accredited, but
some still have not certified a significant percentage of their
systems. Further, agency certifications and accreditations do not
always meet criteria identified in federal guidance. Unless such
criteria are met, agencies cannot ensure that accrediting officials are
receiving consistent information on which to base their decisions, and
the value of this process as a management control for ensuring
information system security is limited. In addition, unverified agency-
reported performance data may not accurately reflect the status of an
agency's efforts to implement this requirement. Consistent reporting of
performance measurement data by agencies on their certifications and
accreditations, as well as additional information on the quality of
agency processes provided through both management oversight and
independent evaluation, would provide increased assurance for the
administration and the Congress that critical federal systems are
meeting FISMA requirements and do not contain significant security
weaknesses that could threaten essential federal operations. It would
also assist the administration and the Congress in their oversight
responsibilities by helping to identify and respond to challenges in
effectively and efficiently implementing this requirement for the
federal government.
Recommendations for Executive Action:
To help ensure that federal agencies' certification and accreditation
processes consistently provide adequate and effective security controls
in their information systems, we recommend that the Director of the
Office of Management and Budget take the following five actions. First,
we recommend that the OMB Director revise policy and guidance on the
security of automated information resources to require federal agencies
to:
* continue to implement security certification and accreditation
processes consistent with guidance and standards issued by NIST for
non-national security systems, including specific reference to the new
certification and accreditation guidance as well as FISMA-required
standards such as those for system security categorization and minimum
security controls; and:
* ensure that periodic testing and evaluation of information security
controls, as required by FISMA, include assessing the quality of
security certifications and accreditations to facilitate decisions that
are based on consistent consideration of key criteria outlined in
federal guidance, including a current risk assessment, appropriate
control testing and evaluation, a tested contingency plan, and the
identification of the specific residual risk being accepted.
Further, to improve the consistency and reliability of agency FISMA
reporting for administration and congressional oversight, we recommend
that the OMB Director consider changes to OMB's FISMA reporting
guidance that would:
* provide additional clarification that national security systems are
to be reflected in reporting performance measurement data and that only
systems granted full authorization to operate should be considered in
reporting the number of systems certified and accredited;
* require reporting on key aspects of agencies' certification and
accreditation processes and efforts, such as how agencies ensure the
quality and consistency of their certifications and accreditations and
the status of their efforts according to levels of risk or impact
established for their systems; and:
* encourage the IGs to assess agency FISMA reporting processes and test
agency-reported performance data as part of their FISMA-mandated
independent evaluations; for example, the IGs could review the quality
of agency certifications and accreditations for the subset of systems
they evaluate to determine whether they meet appropriate criteria and
determine whether such information is accurately reflected in the
agencies' compilation of related performance measures.
Agency Comments:
We received oral comments on a draft of this report from
representatives of OMB's Office of Information and Regulatory Affairs
and Office of General Counsel. The representatives agreed with our
findings that the quality of agency certification and accreditation
processes varies, and generally agreed with our recommendations to
improve certification and accreditation processes. OMB stated that it
plans to address key certification and accreditation practices in its
upcoming FISMA reporting guidance to agencies, and believes the recent
completion of NIST Special Publication 800-37 and reviews by designated
accrediting authorities are fundamental drivers for improving the
quality of the certification and accreditation process. In addition,
OMB stated its belief that existing guidance, including its Circular A-
130 and FISMA implementing guidance, helps ensure that implementation
of certification and accreditation is effective, and that its planned
agency guidance for fiscal year 2004 FISMA reporting will address many
of the issues in our report. The Department of Commerce provided
written comments on a draft of this report (see app. I). In these
comments, the department generally agreed with our report and provided
certain technical comments. We also received written and oral technical
comments from the Departments of Defense and Energy, EPA, NASA, and
NIST. Comments from all these agencies have been incorporated into the
report, as appropriate.
As agreed with your offices, unless you publicly announce the contents
of this report earlier, we plan no further distribution of it until 30
days from the date of this letter. At that time, we will send copies of
the report to other interested congressional committees; the Director,
Office of Management and Budget; and the heads of the agencies
discussed in the report. In addition, the report will be available at
no charge on the GAO Web site at [Hyperlink, http://www.gao.gov].
Copies will also be made available to others upon request.
Should you or your offices have any questions concerning this report,
please call me at (202) 512-3317 or Ben Ritt, Assistant Director, at
(202) 512-6443. We can also be reached by e-mail at [Hyperlink,
daceyr@gao.gov] and [Hyperlink, rittw@gao.gov], respectively. Key
contributors to this report are listed in appendix II.
Signed by:
Robert F. Dacey:
Director, Information Security Issues:
[End of section]
Appendixes:
Appendix I: Comments from the Department of Commerce:
THE SECRETARY OF COMMERCE:
Washington, D.C. 20230:
June 23, 2004:
Mr. Robert F. Dacey:
Director, Information Security Issues:
United States General Accounting Office:
Washington, DC 20548:
Dear Mr. Dacey:
Thank you for the opportunity to comment on the GAO draft report
"Information Security: Agencies Need to Implement Consistent Processes
in Authorizing Systems for Operation." The report presents a realistic
representation of the state of certification and accreditation (C&A)
practices in the Federal Government today, and the information
attributed to the Department of Commerce is accurate, except as noted
in the enclosure.
The Department of Commerce recognizes the value of establishing sound,
repeatable, consistent practices to ensure the quality of the C&A
process for federal information technology (IT) systems. We appreciate
the support of Congress and OMB for establishing requirements for
these practices, and are pleased with GAO's recognition of the National
Institute of Standards and Technology (NIST) as it provides
comprehensive federal guidance in this important area.
We have accorded IT security a high priority in the Commerce
Department, and are also pleased to support other federal agencies and
the private sector through our NIST IT security products.
Sincerely,
Signed by:
Donald L. Evans:
Enclosure:
[End of section]
Appendix II: GAO Contact and Staff Acknowledgments:
GAO Contact:
William B. Ritt, (202) 512-6443:
Acknowledgments:
In addition to the person named above, Larry Crosland, Mark Fostek,
Michael P. Fruitman, Danielle Hollomon, Elizabeth Johnston, Anjalique
Lawrence, Min Lee, Tracy Pierson, and Monica Wolford made key
contributions to this report.
(310504):
FOOTNOTES
[1] Federal Information Security Management Act of 2002, Title III, E-
Government Act of 2002, P.L. 107-347, December 17, 2002.
[2] Office of Management and Budget, FY 2003 Report to Congress on the
Federal Government Information Management, March 1, 2004.
[3] As currently defined in FISMA, the term "national security system"
means any information system (including any telecommunications system)
used or operated by an agency or by a contractor of an agency, or other
organization on behalf of an agency (1) the function, operation, or use
of which involves intelligence activities, cryptologic activities
related to national security, command and control of military forces,
equipment that is an integral part of a weapon or weapons system, or is
critical to the direct fulfillment of military or intelligence missions
(excluding systems used for routine administrative and business
applications); or (2) is protected at all times by procedures
established for information that have been specifically authorized
under criteria established by an executive order or an act of Congress
to be kept classified in the interest of national defense or foreign
policy.
[4] These 24 departments and agencies are the Departments of
Agriculture, Commerce, Defense (DOD), Education, Energy, Health and
Human Services, Homeland Security, Housing and Urban Development,
Interior, Justice, Labor, State, Transportation, Treasury, and Veterans
Affairs, the Environmental Protection Agency, General Services
Administration, Office of Personnel Management, National Aeronautics
and Space Administration, National Science Foundation, Nuclear
Regulatory Commission, Small Business Administration, Social Security
Administration, and U.S. Agency for International Development.
[5] Office of Management and Budget, Management of Federal Information
Resources, Circular No. A-130, Transmittal Memorandum No. 4, Appendix
III, Security of Federal Automated Information Resources, November 28,
2000.
[6] Per OMB Circular No. A-130, a general support system is an
interconnected set of information resources under the same direct
management control that shares common functionality. It normally
includes hardware, software, information, data, applications,
communications, and people. An application means the use of information
resources to satisfy a specific set of user requirements, and a major
application is an application that requires special attention to
security due to the risk and magnitude of the harm resulting from the
loss, misuse, or unauthorized access to or modification of the
information in the application.
[7] National Institute of Standards and Technology, Risk Management
Guide for Information Technology Systems, Special Publication 800-30
(July 2002); and Guide for Developing Security Plans for Information
Technology Systems Special Publication 800-18 (December 1998).
[8] U.S. General Accounting Office, Information Security: Continued
Efforts Needed to Sustain Progress in Implementing Statutory
Requirements, GAO-04-483T (Washington, D.C.: March 16, 2004).
[9] The loss of confidentiality is the unauthorized disclosure of
information, the loss of integrity is the unauthorized modification or
destruction of information, and the loss of availability is the
disruption of access to or use of information or an information system.
[10] FIPS Publication 102 defined a computer application as the use(s)
for which a computer system is intentionally employed. Further, an
application broadly represents a variety of certification entities,
including software programs, hardware components, applications,
systems, terminals, networks, installations, and other entities.
[11] National Institute of Standards and Technology, Guide for the
Security Certification and Accreditation of Federal Information
Systems, Special Publication 800-37-Final (May 2004).
[12] Used by OMB to monitor the status of remediation efforts for
FISMA, plans of action and milestones are required for all programs and
systems where an IT security weakness has been found. The plan lists
the weaknesses and shows estimated resource needs or other challenges
to resolving them, key milestones and completion dates, and the status
of corrective actions.
[13] National Institute of Standards and Technology, Security Self-
Assessment Guide for Information Technology Systems, Special
Publication 800-26 (November 2001).
[14] Department of Defense, Information Assurance (IA), Directive
8500.1 (Oct. 24, 2002); and DOD Information Technology Security
Certification and Accreditation Process (DITSCAP), Instruction Number
5200.40 (Dec. 30, 1997).
[15] National Security Telecommunications and Information Systems
Security Committee, National Information Assurance Certification and
Accreditation Process (NIACAP), NSTISSI No. 1000 (April 2000).
[16] Director of Central Intelligence, Protecting Sensitive
Compartmented Information Within Information Systems, Directive 6/3
(DCID 6/3) (June 5, 1999); and Protecting Sensitive Compartmented
Information Within Information Systems (DCID 6/3)--Manual (Aug. 1,
2000).
[17] GAO-04-483T.
[18] U.S. General Accounting Office, Information Security: Continued
Action Needed to Improve Software Patch Management, GAO-04-706
(Washington, D.C.: June 2, 2004).
[19] Critical infrastructure protection activities called for in
federal policy and law are intended to enhance the security of cyber
and physical, public and private infrastructures that are essential to
national security, national economic security, or national public
health and safety.
[20] According to the National Institute of Standards and Technology's
Special Publication 800-34, Contingency Planning Guide for Information
Technology Systems (June 2002), elements of contingency planning should
be undertaken throughout the system development life cycle, including
the development phase. Regarding testing, the guide notes that during
the implementation phase for a new system, contingency strategies
should be tested to ensure that technical features and recovery
procedures are accurate and effective. Further, during the operations
and maintenance phase of the system, exercises and tests should be
conducted to ensure that the contingency plan procedures continue to be
effective.
[21] GAO-04-483T.
[22] ASSERT is an Internet-based, automated version of NIST's Self-
Assessment Guide for Information Technology Systems (Special
Publication 800-26) that annually evaluates the risk in computer
systems at EPA and produces and centrally tracks Web-enabled plans of
action and milestones reports.
GAO's Mission:
The Government Accountability Office, the investigative arm of
Congress, exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office
441 G Street NW, Room LM
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm
E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director,
NelliganJ@gao.gov
(202) 512-4800
U.S. Government Accountability Office,
441 G Street NW, Room 7149
Washington, D.C. 20548: