Secure Border Initiative
DHS Needs to Address Significant Risks in Delivering Key Technology Investment
Gao ID: GAO-08-1086 September 22, 2008
The Department of Homeland Security's (DHS) Secure Border Initiative (SBI) is a multiyear, multibillion-dollar program to secure the nation's borders through, among other things, new technology, increased staffing, and new fencing and barriers. The technology component of SBI, which is known as SBInet, involves the acquisition, development, integration, and deployment of surveillance systems and command, control, communications, and intelligence technologies. GAO was asked to determine whether DHS (1) has defined the scope and timing of SBInet capabilities and how these capabilities will be developed and deployed, (2) is effectively defining and managing SBInet requirements, and (3) is effectively managing SBInet testing. To do so, GAO reviewed key program documentation and interviewed program officials, analyzed a random sample of requirements, and observed operations of a pilot project.
Important aspects of SBInet remain ambiguous and in a continued state of flux, making it unclear and uncertain what technology capabilities will be delivered, when and where they will be delivered, and how they will be delivered. For example, the scope and timing of planned SBInet deployments and capabilities have continued to change since the program began and, even now, are unclear. Further, the program office does not have an approved integrated master schedule to guide the execution of the program, and GAO's assimilation of available information indicates that the schedule has continued to change. This schedule-related risk is exacerbated by the continuous change in and the absence of a clear definition of the approach that is being used to define, develop, acquire, test, and deploy SBInet. The absence of clarity and stability in these key aspects of SBInet impairs the ability of the Congress to oversee the program and hold DHS accountable for program results, and it hampers DHS's ability to measure program progress. SBInet requirements have not been effectively defined and managed. While the program office recently issued guidance that defines key practices associated with effectively developing and managing requirements, such as eliciting user needs and ensuring that different levels of requirements and associated verification methods are properly aligned with one another, the guidance was developed after several key activities had been completed. In the absence of this guidance, the program has not effectively performed key requirements definition and management practices. For example, it has not ensured that different levels of requirements are properly aligned, as evidenced by GAO's analysis of a random probability sample of component requirements showing that a large percentage of them could not be traced to higher-level system and operational requirements. Also, some of SBInet's operational requirements, which are the basis for all lower-level requirements, were found by an independent DHS review to be unaffordable and unverifiable, thus casting doubt on the quality of lower-level requirements that are derived from them. As a result, the risk of SBInet not meeting mission needs and performing as intended is increased, as are the chances of expensive and time-consuming system rework. SBInet testing has not been effectively managed. For example, the program office has not tested the individual system components to be deployed to the initial deployment locations, even though the contractor initiated integration testing of these components with other system components and subsystems in June 2008. Further, while a test management strategy was drafted in May 2008, it has not been finalized and approved, and it does not contain, among other things, a clear definition of testing roles and responsibilities; a high-level master schedule of SBInet test activities; or sufficient detail to effectively guide project-specific test planning, such as milestones and metrics for specific project testing. Without a structured and disciplined approach to testing, the risk that SBInet will not satisfy user needs and operational requirements, thus requiring system rework, is increased.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-08-1086, Secure Border Initiative: DHS Needs to Address Significant Risks in Delivering Key Technology Investment
This is the accessible text file for GAO report number GAO-08-1086
entitled 'Secure Border Initiative: DHS Needs to Address Significant
Risks in Delivering Key Technology Investment' which was released on
September 22, 2008.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to Congressional Requesters:
United States Government Accountability Office:
GAO:
September 2008:
Secure Border Initiative:
DHS Needs to Address Significant Risks in Delivering Key Technology
Investment:
GAO-08-1086:
GAO Highlights:
Highlights of GAO-08-1086, a report to congressional requesters.
Why GAO Did This Study:
The Department of Homeland Security‘s (DHS) Secure Border Initiative
(SBI) is a multiyear, multibillion-dollar program to secure the
nation‘s borders through, among other things, new technology, increased
staffing, and new fencing and barriers. The technology component of
SBI, which is known as SBInet, involves the acquisition, development,
integration, and deployment of surveillance systems and command,
control, communications, and intelligence technologies. GAO was asked
to determine whether DHS (1) has defined the scope and timing of SBInet
capabilities and how these capabilities will be developed and deployed,
(2) is effectively defining and managing SBInet requirements, and (3)
is effectively managing SBInet testing. To do so, GAO reviewed key
program documentation and interviewed program officials, analyzed a
random sample of requirements, and observed operations of a pilot
project.
What GAO Found:
Important aspects of SBInet remain ambiguous and in a continued state
of flux, making it unclear and uncertain what technology capabilities
will be delivered, when and where they will be delivered, and how they
will be delivered. For example, the scope and timing of planned SBInet
deployments and capabilities have continued to change since the program
began and, even now, are unclear. Further, the program office does not
have an approved integrated master schedule to guide the execution of
the program, and GAO‘s assimilation of available information indicates
that the schedule has continued to change. This schedule-related risk
is exacerbated by the continuous change in and the absence of a clear
definition of the approach that is being used to define, develop,
acquire, test, and deploy SBInet. The absence of clarity and stability
in these key aspects of SBInet impairs the ability of the Congress to
oversee the program and hold DHS accountable for program results, and
it hampers DHS‘s ability to measure program progress.
SBInet requirements have not been effectively defined and managed.
While the program office recently issued guidance that defines key
practices associated with effectively developing and managing
requirements, such as eliciting user needs and ensuring that different
levels of requirements and associated verification methods are properly
aligned with one another, the guidance was developed after several key
activities had been completed. In the absence of this guidance, the
program has not effectively performed key requirements definition and
management practices. For example, it has not ensured that different
levels of requirements are properly aligned, as evidenced by GAO‘s
analysis of a random probability sample of component requirements
showing that a large percentage of them could not be traced to higher-
level system and operational requirements. Also, some of SBInet‘s
operational requirements, which are the basis for all lower-level
requirements, were found by an independent DHS review to be
unaffordable and unverifiable, thus casting doubt on the quality of
lower-level requirements that are derived from them. As a result, the
risk of SBInet not meeting mission needs and performing as intended is
increased, as are the chances of expensive and time-consuming system
rework.
SBInet testing has not been effectively managed. For example, the
program office has not tested the individual system components to be
deployed to the initial deployment locations, even though the
contractor initiated integration testing of these components with other
system components and subsystems in June 2008. Further, while a test
management strategy was drafted in May 2008, it has not been finalized
and approved, and it does not contain, among other things, a clear
definition of testing roles and responsibilities; a high-level master
schedule of SBInet test activities; or sufficient detail to effectively
guide project-specific test planning, such as milestones and metrics
for specific project testing. Without a structured and disciplined
approach to testing, the risk that SBInet will not satisfy user needs
and operational requirements, thus requiring system rework, is
increased.
What GAO Recommends:
GAO is recommending that DHS assess and disclose the risks associated
with its planned SBInet development, testing, and deployment
activities, and address the system deployment, requirements management,
and testing weaknesses that GAO identified. DHS agreed with all but one
of GAO‘s eight recommendations and described actions completed,
underway, and planned to address them.
To view the full product, including the scope and methodology, click on
[hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-08-1086]. For more
information, contact Randolph C. Hite at (202) 512-3439 or
hiter@gao.gov.
[End of section]
Contents:
Letter:
Results in Brief:
Background:
Limited Definition of SBInet Deployments, Capabilities, Schedule, and
Life Cycle Management Process Increases Program's Exposure to Risk:
Limitations of SBInet Requirements Development and Management Efforts
Increase Program Risk:
Limitations in Key SBInet Testing and Test Management Activities
Increase Program Risk:
Conclusions:
Recommendations for Executive Action:
Agency Comments and Our Evaluation:
Appendix I: Objectives, Scope, and Methodology:
Appendix II: Comments from the Department of Homeland Security:
Appendix III: GAO Contact and Staff Acknowledgments:
Tables:
Table 1: System-Level Reviews and Their Purpose:
Table 2: Project-Level Reviews and Their Purpose:
Table 3: SBInet Requirements Types:
Table 4: SBInet Tests:
Table 5: SBInet Requirements Traceability Results:
Table 6: Roles and Responsibilities of Entities Identified in the Draft
SBInet Test and Evaluation Master Plan:
Figures:
Figure 1: High-Level, Conceptual Depiction of Long-Term SBInet
Operations:
Figure 2: COP in a Command Center and Agent Vehicle:
Figure 3: SBInet Building Block Approach:
Figure 4: Relationships among Requirements:
Figure 5: Changes in Planned Deployments over Time:
Figure 6: Changes in Schedules for Reviews and Deployment:
Figure 7: Changes in Schedules for Testing:
Abbreviations:
CBP: U.S. Customs and Border Protection:
CDR: Critical Design Review:
COP: Common Operating Picture:
C3I: command, control, communications and intelligence:
DDR: Deployment Design Review:
DHS: Department of Homeland Security:
DOORS: Dynamic Object-Oriented Requirements System:
DRR: Deployment Readiness Review:
IT: information technology:
PDR: Preliminary Design Review:
RAD/JAD: Rapid Application Development and Joint Application Design:
SBI: Secure Border Initiative:
SRR: System Requirements Review:
TRR: Test Readiness Review:
[End of section]
United States Government Accountability Office:
Washington, DC 20548:
September 22, 2008:
The Honorable Bennie G. Thompson:
Chairman:
Committee on Homeland Security:
House of Representatives:
The Honorable Christopher P. Carney:
Chairman:
The Honorable Mike Rogers:
Ranking Member:
Subcommittee on Management, Investigations, and Oversight:
Committee on Homeland Security:
House of Representatives:
The Honorable Kendrick B. Meek:
House of Representatives:
Securing the nation's borders from illegal entry of aliens and
contraband continues to be a major challenge because much of the 6,000
miles[Footnote 1] of international borders with Canada and Mexico
remains vulnerable to unlawful activities. Although the Department of
Homeland Security (DHS) apprehends hundreds of thousands of people
entering the country illegally each year, many more unauthorized
entrants go undetected.
As we have reported, previous attempts to acquire and deploy
surveillance technologies along the nation's borders to assist in
detecting and responding to illegal entries have not been successful.
Specifically, the former Immigration and Naturalization Service's
Integrated Surveillance Intelligence System, begun in the late 1990s,
was difficult and expensive to maintain; it provided limited command,
control, and situational awareness capability; and its component
systems were not integrated.[Footnote 2] In response, DHS established
the America's Shield Initiative in 2004, but this program was halted in
2005 because of the program's limited scope and the department's shift
in strategy for achieving border security and interior enforcement
goals.[Footnote 3] In November 2005, DHS launched the Secure Border
Initiative (SBI), a multiyear, multibillion-dollar program to secure
the nation's borders through enhanced use of surveillance technologies,
increased staffing levels, improved infrastructure, and increased
domestic enforcement of immigration laws. One component of SBI, known
as SBInet, is focused on the acquisition and deployment of surveillance
and communications technologies. This program is managed by the SBInet
System Program Office within U.S. Customs and Border Protection (CBP).
Because of the size and complexity of SBInet, and the problems
experienced by its predecessors, you asked us to determine whether DHS
(1) has defined the scope and timing of planned SBInet capabilities and
how these capabilities will be developed and deployed, (2) is
effectively defining and managing SBInet requirements, and (3) is
effectively managing SBInet testing.
To accomplish our objectives, we reviewed key program documentation,
including guidance, plans, and requirements and testing documentation.
In cases where such documentation was not available, we interviewed
program officials about the development of capabilities and the
management of requirements and testing. We then compared this
information to relevant federal system acquisition guidance. We also
analyzed a random probability sample of system requirements and
observed operations of the initial SBInet project.
We conducted this performance audit from August 2007 to September 2008
in accordance with generally accepted government auditing standards.
Those standards require that we plan and perform the audit to obtain
sufficient, appropriate evidence to provide a reasonable basis for our
findings and conclusions based on our audit objectives. We believe that
the evidence obtained provides a reasonable basis for our findings and
conclusions based on our audit objectives. Further details of our
objectives, scope, and methodology are included in appendix I.
Results in Brief:
Important aspects of SBInet--the scope and schedule, and the
development and deployment approach--remain ambiguous and in a
continued state of flux, making it unclear and uncertain what
technology capabilities will be delivered, when and where they will be
delivered, and how they will be delivered, as the following examples
illustrate:
* The scope of what is to be achieved has become more limited without
becoming more specific. In particular, in December 2006, the department
committed to having an initial set of capabilities operational along
the entire southwest border by late 2008 and having a full set of
capabilities operational along the entire southwest and northern
borders by late 2009. In March 2008, the SBInet System Program Office
had reduced its commitment to deploying a to-be-determined set of
technology capabilities to three out of nine sectors[Footnote 4] along
the southwest border by 2011 and to only two locations in one of nine
sectors by the end of 2008. As of July 2008, the program office
reported that the dates for the two locations would slip into 2009;
however, specific dates were not available and thus remain uncertain.
* The timing and sequencing of the work, activities, and events that
need to occur have continued to be unclear. Specifically, the program
office does not have an approved integrated master schedule to govern
the execution of the program. Further, our assimilation of available
information from multiple program sources indicates that the schedule
has continued to change.
* This schedule-related risk is exacerbated by the continuous change in
and absence of clear definition around the system life cycle management
approach that is being used to develop and deploy SBInet. In
particular, important details about key life cycle processes are not
documented, and what is defined in various documents is not fully
consistent across them. Moreover, in discussions with agency officials
to obtain clarification regarding the processes, new information has
been introduced routinely, as recently as late July 2008, which differs
from available documentation or earlier statements by these officials.
SBInet requirements have not been effectively developed and managed.
While the program office recently issued guidance that defines key
practices associated with effectively developing and managing
requirements--such as eliciting user needs, documenting and approving
the requirements to establish a baseline,[Footnote 5] and ensuring that
requirements are traceable--the guidance was not used in developing
SBInet requirements because it was issued after their development. In
the absence of well-defined guidance, the program's efforts to
effectively define and manage requirements have been mixed. For
example, the program has taken credible steps to include users in the
definition of requirements. However, several requirements development
and management limitations exist. For example, all requirements have
not been finalized, such as the requirements for the command, control,
and communication subsystem of SBInet and the requirements specific to
the first two deployment locations. Also, an independent review found
some of the operational requirements to be unverifiable or
unaffordable, indicating that these requirements had not been properly
defined and validated. Moreover, the different levels of requirements
are not properly aligned (i.e., traceable), as evidenced by our
analysis of a random probability sample of requirements where we found
large percentages that were not traceable back to higher level
requirements or forward to more detailed system design specifications
and verification methods.
SBInet testing has not been effectively managed. Specifically, critical
testing activities, as called for in federal guidance[Footnote 6] and
described in the program office's own test documents, have yet to be
performed, and the infrastructure for managing testing has not been
fully established. For example, the program office has not tested the
individual system components to be deployed to the initial two
locations, even though the contractor initiated integration testing of
multiple components with the command, control, and communication
subsystem in June 2008. Further, while a test management strategy,
known as the Test and Evaluation Master Plan, was drafted in May 2008,
it has not been finalized and approved, and it does not contain, among
other things, clear definitions of testing roles and responsibilities;
a high-level master schedule of SBInet test activities; or sufficient
detail to effectively guide project-specific test planning, such as
milestones and metrics for specific project testing.
Collectively, the above limitations in the scope and timing of SBInet
to-be-deployed capabilities, the ambiguity surrounding the schedule and
approach for accomplishing these deployments, and the weaknesses in
requirements development and management and test management, introduce
considerable risks to the program. To address these risks, we are
making recommendations to DHS to immediately re-evaluate its plans and
approach in relation to the status of the system and related
development, acquisition, and testing activities, as discussed in this
report, and to disclose to CBP and DHS leadership, as well as
appropriate congressional committees, the results of this assessment,
including proposed changes to its planned schedule of activities to
mitigate the associated risks. In addition, we are making
recommendations to address each of the system acquisition and
development problems discussed in this report, including those
associated with finalizing an integrated master schedule, having a well-
defined and stable life cycle management approach, and implementing
effective requirements development and management and testing.
In comments on a draft of this report, reprinted in appendix II, the
department stated that the report was factually sound and that it
agreed with seven of our eight recommendations and partially disagreed
with one aspect of the remaining recommendation. Specifically, DHS
disagreed with that aspect of one recommendation for conducting
appropriate component-level testing prior to integrating system
components, stating that its current test strategy provides the
appropriate degree of confidence in these commercially available
components, as evidenced by either component manufacturer certificates
of conformance, independent government laboratory test documentation,
or the prime contractor's component-integration level testing. We
support DHS's current test strategy as it is consistent with our
recommendation. Specifically, it expands on the department's prior
strategy for component testing, which was limited to manufacturer self-
certification of component conformance and informal observations of
system components, by adding the use of independent government
laboratories to test the components. We would emphasize, however, that
regardless of the method used, it is important that confidence be
gained in components prior to integrating them, which our
recommendation recognizes. As our report states, such a hierarchical
approach to testing allows for the source of any system defects to be
discovered and isolated sooner rather than later, and thus helps to
avoid the potential for expensive and time-consuming system rework.
With respect to all of our recommendations, the department added that
it is working to address our recommendations and resolve the management
and operational challenges identified in the report as expeditiously as
possible, and it described actions recently completed, underway, and
planned that it said would address them. The department also provided
technical comments that we incorporated in the report, as appropriate.
Background:
CBP's SBI program is to leverage technology, tactical infrastructure,
[Footnote 7] and people to allow CBP agents to gain control of the
nation's borders. Within SBI, SBInet is the program for acquiring,
developing, integrating, and deploying an appropriate mix of (1)
surveillance technologies, such as cameras, radars, and sensors, and
(2) command, control, communications, and intelligence (C3I)
technologies.
The initial focus of SBInet has been on addressing the requirements of
CBP's Office of Border Patrol, which is responsible for securing the
borders between the established ports of entry.[Footnote 8] The longer-
term SBInet systems solution also is to address requirements of CBP's
two other major components--the Office of Field Operations, which
controls vehicle and pedestrian traffic at the ports of entry, and the
Office of Air and Marine Operations, which operates helicopters, fixed-
wing aircraft, and marine vessels used in securing the borders. Figure
1 provides a high-level, operational concept of the long-term SBInet
systems solution.
Figure 1: High-Level, Conceptual Depiction of Long-Term SBInet
Operations:
[See PDF for image]
This figure is an illustration of a high-level, conceptual depiction of
long-term SBInet Operations, showing the following:
* Marine support;
* Air support;
* Unmanned aerial surveillance;
* Air and marine station;
* Command Center:
- Border patrol sector headquarters;
- Border patrol station;
* Agent vehicles;
- Mobile data terminal;
* Mobile surveillance system;
* Port of entry;
* Unattended ground sensors;
* Radio/camera tower.
Source: GAO analysis of agency data, Art Explosion (clip at).
[End of figure]
Surveillance technologies are to include a variety of sensor systems
that improve CBP's ability to detect, identify, classify, and track
items of interest along the borders. Unattended ground sensors are to
be used to detect heat and vibrations associated with foot traffic and
metal associated with vehicles. Radars mounted on fixed and mobile
towers are to detect movement, and cameras on fixed and mobile towers
are to be used to identify, classify, and track items of interest
detected by the ground sensors and the radars. Aerial assets are also
to be used to provide video and infrared imaging to enhance tracking of
targets.
The C3I technologies are to include software and hardware to produce a
Common Operating Picture (COP)--a uniform presentation of activities
within specific areas along the border. The sensors, radars, and
cameras are to gather information along the border, and the system is
to transmit this information to the COP terminals located in command
centers and agent vehicles and assemble this information to provide CBP
agents with border situational awareness. More specifically, the COP
technology is to allow agents to (1) view data from radars and sensors
that detect and track movement in the border areas, (2) control cameras
to help identify and classify illegal entries, (3) correlate entries
with the positions of nearby agents, and (4) enhance tactical decision
making regarding the appropriate response to apprehend an entry, if
necessary.
Initially, COP information is to be distributed to terminals in command
centers. We observed that these terminals look like a standard computer
workstation with multiple screens. From this workstation, an operator
is to be able to view an area of interest in several different ways.
For example, the operator is to see different types of maps, satellite
images, and camera footage on the multiple screens. The operator is
also to be able to move the cameras to track images on the screen.
According to program officials, eventually, when the radars detect
potential items of interest, the system is to automatically move the
cameras so the operator does not always need to initiate the search in
the area.
We observed that COP data are also available on laptop computers, known
as mobile data terminals, mounted in select agent vehicles in the
field. These terminals are to enable field agents to see information
similar to that seen by command center operators. Eventually, the COP
technology is to be capable of providing distributed surveillance and
tactical decision-support information to other DHS agencies and
stakeholders external to DHS, such as local law enforcement. Figure 2
shows examples of COP technology in a command station and an agent
vehicle.
Figure 2: COP in a Command Center and Agent Vehicle:
[See PDF for image]
This figure is an illustration of COP in a Command Center and Agent
Vehicle. The following are depicted:
Agent vehicles:
- photograph of mobile data terminal;
Command center (Border patrol sector headquarters and Border patrol
station) (photograph of commend center): receiving data from:
- Unattended ground sensors;
- Radar/camera tower.
Source: GAO analysis of agency data, GAO (photos), Art Explosion (clip
art).
[End of figure]
The first SBInet capabilities were deployed under a pilot or prototype
effort known as "Project 28." Project 28 is currently operating along
28 miles of the southwest border in the Tucson Sector of Arizona.
Project 28 was accepted by the government for deployment 8 months
behind schedule (in February 2008); this delay occurred because the
contractor-delivered system did not perform as intended. As we have
previously reported,[Footnote 9] reasons for Project 28 performance
shortfalls and delays include the following:
* System requirements were not adequately defined, and users were not
involved in developing the requirements.
* System integration testing was not adequately performed.
* Contractor oversight was limited.
* Project scope and complexity were underestimated.
To manage SBInet, DHS established a program office within CBP. The
program office is led by a program manager and deputy program managers
for program operations and mission operations. The program manager is
responsible for the execution of the program, including developing,
producing, deploying, and sustaining the system to meet the users'
needs. Among other things, this includes developing and analyzing
requirements and system alternatives, managing system design and
development, evaluating the system's operational effectiveness, and
managing program risk.
SBInet Program Office Description of Its Life Cycle Management
Approach:
A system life cycle management approach typically consists of a series
of phases, milestone reviews, and related processes to guide the
acquisition, development, deployment, and operation and maintenance of
a system. Among other things, the phases, reviews, and processes cover
such important life cycle activities as requirements development and
management, design, software development, and testing. Based on
available program documentation, augmented by program official
briefings and statements, key aspects of the SBInet system life cycle
management approach are described below.
Approach Is Intended to Deliver Incremental Capabilities through a
Series of Blocks:
In general, SBInet surveillance systems are to be acquired through the
purchase of commercially available products, while the COP systems
involve development of new, customized systems and software. Together,
both categories are to form a deployable increment of the SBInet
capabilities, which the program office refers to as a "block." Each
block is to include a release or version of the COP.
SBInet documentation shows that the program office is acquiring the
blocks incrementally using a "spiral" approach, under which an initial
system capability is to be delivered based on a defined subset of the
system's total requirements. This approach is intended to allow CBP
agents access to new technological tools sooner rather than later for
both operational use and feedback on needed enhancements or changes.
Subsequent spirals or iterations of system capability are to be
delivered based on feedback and unmet requirements, as well as the
availability of new technologies. Figure 3 illustrates conceptually how
the different capabilities are to come together to form a block and how
future blocks are to introduce more capabilities.
Figure 3: SBInet Building Block Approach:
[See PDF for image]
This figure is an illustration of the SBInet Building Block Approach,
as follows:
Building block: Surveillance system components;
Building block: C3I systems;
Building block: COP software.
Together these building blocks form:
Block 1 capability.
In the SBInet Building Block approach, future capabilities attached on
both sides of the Block 1 capability.
Source: GAO analysis of agency data from multiple sources.
[End of figure]
The approach used to design and develop SBInet system capabilities for
each block includes such key activities as requirements development,
system design, system acquisition and development, and testing. The
approach, as explained by program officials and depicted in part in
various documents, also includes various reviews, or decision points,
to help ensure that these activities are being done properly and that
the system meets user needs and requirements. These reviews are to be
used in developing both the overall SBInet Block 1 capability and the
COP software. Table 1 provides a high-level description of the major
reviews that are to be performed in designing and developing the system
prior to deployment to the field and in the order that they occur.
[Footnote 10]
Table 1: System-Level Reviews and Their Purpose:
Review: System Requirements Review;
Purpose: Ensures that system requirements have been completely and
properly identified and that a mutual understanding exists between the
DHS CBP SBInet System Program Office and the contractor on the
requirements to be met.
Review: Preliminary Design Review;
Purpose: Confirms that sufficient design has been accomplished to
verify the completeness and achievability of system requirements.
Review: Critical Design Review;
Purpose: Demonstrates that the detailed designs are complete; meet
requirements; provide for external interfaces; and are ready for
fabrication, coding, assembly, and integration.
Review: System Qualification Test Readiness Review;
Purpose: Ensures that the test objectives are clear and test procedures
are adequate to test that the system requirements are being met.
Review: System Production Readiness Review;
Purpose: Demonstrates stakeholder concurrence that the system is ready
for deployment and examines the program to determine if the design is
ready for production.
Source: GAO analysis of SBInet and CBP documents.
[End of table]
Before a set of capabilities (i.e., block) is deployed to a specific
area or sector of the border, activities such as site selection,
surveys, and environmental impact assessments are conducted to
determine the area's unique environmental requirements. The border area
that receives a given block, or set of system capabilities, is referred
to as a "project." Each project is to have a given block configured to
its unique environmental requirements, referred to as a project
"laydown."
The deployment approach is to include such key activities as
requirements development, system design, project laydown, integration,
testing, and installation. The deployment approach is also to entail
various reviews, or decision points, to help ensure that these
activities are being done properly and that the system meets user needs
and requirements. Table 2 provides a high-level description of the
major reviews that are to be part of project laydown in the order that
they occur.[Footnote 11]
Table 2: Project-Level Reviews and Their Purpose:
Review: Project Requirements Review[A];
Purpose: Demonstrates contractor understanding and analysis of the
project-level requirements.
Review: Deployment Design Review;
Purpose: Ensures that the SBInet System Program Office and the
contractor concur that the proposed system design meets the project-
level requirements.
Review: Deployment Readiness Review;
Purpose: Provides an assessment of the design maturity and ensures that
the contractor is ready to begin construction, such as site
preparation, building the access roads, laying the tower foundations,
and installing the tower power supplies.
Review: System Acceptance Test Readiness Review;
Purpose: Ensures that the test objectives are clear and test procedures
are adequate to test whether or not the system is ready to be accepted
by the government.
Review: Operational Test Readiness Review;
Purpose: Ensures that the system can proceed to operational testing
with a high probability of success.
Review: Operational Readiness Review;
Purpose: Documents stakeholder concurrence that the system is ready for
full operation.
Source: GAO analysis of SBInet and CBP documents.
[A] Currently referred to by the program office as "Deployment Planning
Review."
[End of table]
Approach Includes Key Processes for Developing and Managing
Requirements and for Managing Testing:
Among the key processes provided for in the SBInet system life cycle
management approach are processes for developing and managing
requirements and for managing testing activities. With respect to
requirements development and management, SBInet requirements are to
consist of a hierarchy of six types of requirements, with the high-
level operational requirements at the top. These high-level
requirements are to be decomposed into lower-level, more detailed
system, component, design, software, and project requirements. Having a
decomposed hierarchy of requirements is a characteristic of complex
information technology (IT) projects. The various types of SBInet
requirements are described in table 3. Figure 4 shows how each of these
requirements relate to or are derived from the other requirements.
Table 3: SBInet Requirements Types:
Type: Operational requirements;
Description: Describe the missions and operational capabilities that
the resulting system must satisfy. These are the user requirements for
the SBInet system.
Type: System requirements;
Description: Describe the SBInet performance and system functional and
nonfunctional characteristics; Used for the design, development,
integration, verification, and deployment of the SBInet system.
Type: Component requirements;
Description: Describe required features of various surveillance
components, such as cameras and radars, sufficient to guide system
design; Also associated with one or more verification (test) methods
that will be used to ensure that the system is in compliance with
component requirements.
Type: Design requirements;
Description: Describe the performance, design, and acceptance features
for various component products, such as a short-range or long-range
camera.
Type: COP software requirements;
Description: Describe the functionality and capability of the COP
software, such as allowing the user to control and view information
from the sensors.
Type: Project requirements;
Description: Describe the unique environmental requirements and
capabilities that are deployed for a project (geographic area).
Source: GAO analysis of SBInet documents.
[End of table]
Figure 4: Relationships among Requirements:
[See PDF for image]
This figure shows the relationships among requirements, as follows:
Operational Requirements:
System requirements;
* Compound requirements;
- Design requirements;
* C3I/COP requirements;
* Project requirements.
Source: DHS.
[End of figure]
With respect to test management, SBInet testing consists of a sequence
of tests that are intended to verify first that individual system parts
meet specified requirements, and then verify that these combined parts
perform as intended as an integrated and operational system. Such an
incremental approach to testing is a characteristic of complex IT
system acquisition and development efforts. Through such an approach,
the source of defects can be isolated more easily and sooner, before
they are more difficult and expensive to address. Table 4 summarizes
these tests.
Table 4: SBInet Tests:
Test: Developmental Testing;
Purpose: Verifies and validates the system's engineering process;
Government/contractor role: [Empty];
Location: [Empty].
Test: System Integration Testing;
Purpose: Consists of three types of tests (below);
Government/contractor role: Contractor performs;
Location: Laboratory.
Test: Component-level testing;
Purpose: Verifies the functional performance of individual components
against component requirements;
Government/contractor role: Contractor performs;
Location: Laboratory.
Test: Interim-level integration testing;
Purpose: Verifies compatibility of individual interfaces of hardware
and software components;
Government/contractor role: Contractor performs;
Location: Laboratory.
Test: System-level integration testing;
Purpose: Verifies that system requirements are met when subsystems are
integrated with the COP software;
Government/contractor role: Contractor performs;
Location: Mostly laboratory, some field.
Test: System Verification Testing[A];
Purpose: Verifies that the design being tested is compliant with the
component or system requirements;
Government/contractor role: Contractor performs;
Location: Mostly laboratory, some field.
Test: System Acceptance Testing;
Purpose: Verifies that the installed system meets system requirements
(i.e., the system functions as designed, performs as predicted in the
deployed environment, and is ready for operational testing). Provides
the basis for government accepting the system;
Government/contractor role: Contractor performs;
Location: Field.
Test: Operational Testing;
Purpose: Determines system operational effectiveness and suitability
for the intended use by representative users in a realistic
environment;
Government/contractor role: Government users and independent testers
perform;
Location: Field.
Source: GAO analysis of draft Test and Evaluation Master Plan.
[A] Currently referred to by the program office as "System
Qualification Testing."
[End of table]
Limited Definition of SBInet Deployments, Capabilities, Schedule, and
Life Cycle Management Process Increases Program's Exposure to Risk:
Important aspects of SBInet remain ambiguous and in a continued state
of flux, making it unclear and uncertain what technology capabilities
will be delivered, when and where they will be delivered, and how they
will be delivered. For example, the scope and timing of planned SBInet
deployments and capabilities have continued to change since the program
began and, even now, remain unclear. Further, the approach that is
being used to define, develop, acquire, test, and deploy SBInet is
similarly unclear and has continued to change. According to SBInet
officials, schedule changes are due largely to an immature system
design, and the lack of a stable development approach is due to
insufficient staff and turnover. The absence of clarity and stability
in these key aspects of SBInet introduces considerable program risks,
hampers DHS's ability to measure program progress, and impairs the
ability of the Congress to oversee the program and hold DHS accountable
for program results.
Scope and Timing of Planned Deployments and Capabilities Are Not Clear
and Stable:
One key aspect of successfully managing large IT programs, like SBInet,
is establishing program commitments, including what capabilities are to
be deployed and when and where they are to be deployed. Only when such
commitments are clearly established can program progress be measured
and can responsible parties be held accountable.
The scope and timing of planned SBInet deployments and capabilities
that are to be delivered have not been clearly established, but rather
have continued to change since the program began. Specifically, as of
December 2006, the SBInet System Program Office planned to deploy an
"initial" set of capabilities along the entire southwest border by late
2008 and planned to deploy a "full" set of operational capabilities
along the southern and northern borders (a total of about 6,000 miles)
by late 2009.[Footnote 12] As of March 2007, the program office had
modified its plans, deciding instead to deploy the initial set of
capabilities along the southwest border by the end of fiscal year 2007
(almost a year earlier than originally planned) and delayed the
deployment of the final set of capabilities for the southern and
northern borders until 2011.
In March 2008, the program office again modified its deployment plans,
this time significantly reducing the area to which SBInet capabilities
are to be deployed. At this time, DHS planned to complete deployments
to three out of nine sectors along the southwest border--specifically,
to Tucson Sector by 2009, Yuma Sector by 2010, and El Paso Sector by
2011. According to program officials, other than the dates for the
Tucson, Yuma, and El Paso Sectors, no other deployment dates have been
established for the remainder of the southwest or northern borders.
(Figure 5 shows the changes in the planned deployment areas.)
Figure 5: Changes in Planned Deployments over Time:
[See PDF for image]
This figure contains a chart with two maps associated with chart data.
The following information is depicted:
Source: Systems Engineering Management Plan (Dec. 21, 2006);
Milestone and status: Initiated, mid-2007;
Milestone and status: Initial operational capability, late 2008;
Milestone and status: Full operational capability, late 2009;
Area of deployment: Entire northern and southwest borders.
Source: Operational Requirements Document (Mar. 6, 2007);
Milestone and status: Initial operational capability, mid-2007;
Milestone and status: Full operational capability, mid-2011;
Area of deployment: Entire northern and southwest borders.
Source: SBI Expenditure Plan (March 2008);
Milestone and status: Initiate Block 1, late 2008;
Milestone and status: Tucson completed, mid-2009;
Milestone and status: Yuma completed, mid-2010;
Milestone and status: El Paso completed, mid-2011;
Area of deployment: Southern borders of Yuma, Tucson, and El Paso
sectors (first Block 1 deployment: Tucson 1; second Block 1 deployment:
Ajo 1).
Source: GAO analysis of SBInet data from multiple sources; Map Art
(map).
[End of figure]
The figure also shows the two sites within the Tucson Sector, Tucson 1
and Ajo 1, at which an initial Block 1 capability is to be deployed.
Together, these two deployments cover 53 miles of the 1,989-mile-long
southern border.[Footnote 13] According to the March 2008 SBI
expenditure plan and agency documentation as of June 2008, these two
sites were to have been operational by the end of 2008. However, as of
late July 2008, program officials reported that the deployment schedule
for these two sites has again been modified, and they will not be
operational until "sometime" in 2009. According to program officials,
the slippage in the deployment schedule is due to the need to complete
environmental impact assessment documentation for these locations. The
slippages in the dates for the first two Tucson deployments, according
to a program official, will, in turn, delay subsequent Tucson
deployments, although revised dates for these subsequent deployments
have not been set.
Just as the scope and timing of planned deployments have not been clear
and have changed over time, the specific capabilities that are to be
deployed have been unclear. For example, in April 2008, program
officials stated that they would not know which of the SBInet
requirements would be met by Block 1 until the Critical Design Review,
which at that time was scheduled for June 2008. At that time, program
officials stated that the capabilities to be delivered would be driven
by the functionality of the COP. In June, the review was held, but
according to available documentation, the government did not consider
the design put forth by the contractor to be mature. As a result, the
system design was not accepted in June as planned. Among the design
limitations found was a lack of evidence that the system requirements
were used as the basis for the Tucson 1 and Ajo 1 design, lack of
linkage between the performance of surveillance components and the
system requirements, and incomplete definition of system interfaces. As
of late July 2008, these issues were unresolved, and thus the design
still had not been accepted. In addition, in late July 2008, agency
officials stated that the capabilities to be delivered will be driven
by the functionality of the surveillance components, not the COP.
In addition, the design does not provide key capabilities that are in
requirements documents and were anticipated to be part of the Block 1
deployments to Tucson 1 and Ajo 1. For example, the first deployments
of Block 1 will not have the mobile data terminals in border patrol
vehicles, even though (1) such terminals are part of Project 28
capabilities and (2) workshops were held with the users in February
2008 and June 2008 to specifically define the requirements for these
terminals for inclusion in Block 1. According to program officials,
these terminals will not be part of Block 1 because the wireless
communications infrastructure needed to support these terminals will
not be available in time for the Tucson 1 deployment. Rather, they
expect the wireless infrastructure to be ready "sometime" in 2009 and
said that they will include the mobile data terminals in Block 1
deployments when the infrastructure is ready. Without the mobile data
terminals, agents will not be able to obtain key information from their
vehicles, such as maps of activity in a specific area, incident
reports, and the location of other agents in the area. Instead, the
agents will have to use radios to communicate with the sector
headquarters to obtain this information.
In addition, program officials told us that a number of other
requirements cannot be included in the Block 1 version of the COP,
referred to as version 0.5, due to cost and schedule issues. However,
we have yet to receive a list of these requirements. According to
program officials, they hope to upgrade the COP in 2009 to include
these requirements.
Without clearly establishing program commitments, such as capabilities
to be deployed and when and where they are to be deployed, program
progress cannot be measured and responsible parties cannot be held
accountable.
Program Schedule Is Unsettled:
Another key aspect of successfully managing large programs like SBInet
is having a schedule that defines the sequence and timing of key
activities and events and is realistic, achievable, and minimizes
program risks. However, the program office does not yet have an
approved integrated master schedule to guide the execution of SBInet,
and according to program officials, such a schedule has not been in
place since late 2007. In the absence of an approved integrated master
schedule, program officials stated in mid-August 2008 that they have
managed the program largely using task-order-specific baselined
schedules[Footnote 14], and have been working to create a more
integrated approach. A program official also stated that they have
recently developed an integrated master schedule but that this schedule
is already out of date and undergoing revision. For example, the
deployment of the SBInet system to the Tucson Sector will not be
completed in 2009 as planned.
To understand where the program is relative to established commitments,
we analyzed schedule-related information obtained from other available
program documents, briefings, and interviews. In short, our analysis
shows a schedule in which key activities and events are subject to
constant change, as depicted in the following two figures. Figure 6
shows the changes to the schedule of planned and held reviews and
anticipated deployment dates, and figure 7 shows the changes to the
schedule of testing activities.
Figure 6: Changes in Schedules for Reviews and Deployment:
[See PDF for image]
This figure depicts changes in schedules for reviews and deployment as
follows:
Block 1 System: COP;
SRR, held, January 2008;
PDR/CDR, held, February 2008;
TRR, planned, June 2008; slippage to planned, July, 2008, slippage to
planned August 2008;
Block 1 System:
Combination of surveillance systems, COP, communications, vehicles,
etc.;
CDR, planned, April 2008; slippage to planned May, 2008; slippage to
held June 2008;
TRR, planned June 2008, slippage to planned August 2008.
Project: Tucson 1;
DDR, held, April 2008; slippage to planned June 2008;
DDR, planned July 2008; slippage to planned August 2008;
TRR, planned October 2008;
Tucson 1 deployment, planned November 2008; slippage to: to be
determined.
Project: Ajo 1;
DDR, held May 2008;
DRR, planned 2008;
TRR, planned November 2008;
Ajo 1 deployment, planned December 2008; slippage to: to be determined.
CDR - Critical Design Review;
DDR - Deployment Design Review;
DRR - Deployment Readiness Review;
PDR - Preliminary Design Review;
SRR - Systems Requirements Review;
TRR - Test Readiness Review.
Source: GAO analysis of SBInet data from multiple sources.
[End of figure]
Figure 7: Changes in Schedules for Testing:
[See PDF for image]
This figure depicts changes in schedules for testing as follows:
System integration testing - Block 1:
Started, June, 2008.
System verification testing - Tucson 1:
Planned, July 2008; slippage to planned September, 2008; slippage to:
to be determined.
System acceptance testing:
Planned October, 2008; slippage to January 2009.
Operational testing - Tucson 1:
Planned, October, 2008; slippage to planned November, 2008; slippage
to: to be determined.
Source: GAO analysis of SBInet data from multiple resources.
[End of figure]
In May 2008, program officials stated that the schedule changes were
due largely to the fact that the contractor had not yet provided a
satisfactory system-level design. They also noted that the contractor's
workforce has experienced considerable turnover, including three
different program managers and three different lead system engineers.
They also stated that the System Program Office has experienced
attrition, including turnover in the SBInet Program Manager position.
Without stability and certainty in the program's schedule, program cost
and schedule risks increase, and meaningful measurement and oversight
of program status and progress cannot occur, in turn, limiting
accountability for results.
SBInet Life Cycle Management Approach Has Not Been Clearly Defined and
Has Continued to Change:
System quality and performance are in large part governed by the
processes followed in developing and acquiring the system. To the
extent that a system's life cycle management approach and related
development and acquisition processes are well-defined, the chances of
delivering promised system capabilities and benefits on time and within
budget are increased. To be well-defined, the approach and processes
should be fully documented, so that they can be understood and properly
implemented by those responsible for doing so.
The life cycle management approach and processes being used by the
SBInet System Program Office to manage the definition, design,
development, testing, and deployment of system capabilities has not
been fully and clearly documented. Rather, what is defined in various
program documents is limited and not fully consistent across these
documents. Moreover, in discussions with agency officials to clarify
our understanding of these processes, new terms and processes have been
routinely introduced, indicating that the processes are continuing to
evolve. Agency officials acknowledge that they are still learning about
and improving their processes. Without a clearly documented and
universally understood life cycle management approach and supporting
processes, the program is at increased risk of not meeting
expectations.
Key program documentation that is to be used to guide acquisition and
development activities, including testing and deployment activities, is
incomplete, even though SBInet acquisition and development are already
under way. For example, officials have stated that they are using the
draft Systems Engineering Plan, dated February 2008, to guide the
design, development, and deployment of system capabilities, and the
draft Test and Evaluation Master Plan, dated May 2008, to guide the
testing process--but both of these documents are lacking sufficient
information to clearly guide system activities, as the following
examples explain:
* The Systems Engineering Plan includes a diagram of the engineering
process; however, the steps of the process and the gate reviews are not
defined or described in the text of the document. For example, this
document does not contain sufficient information to understand what
occurs at key reviews, such as the Preliminary Design Review, the
Critical Design Review, and the Test Readiness Review.
* The Test and Evaluation Master Plan describes in more detail some of
the reviews that are not described in the Systems Engineering Plan, but
the reviews included are not consistent between the documents. For
example, the Test and Evaluation Master Plan includes a System
Development and Demonstration Review that is not listed in the Systems
Engineering Plan. In addition, it is not clear from the Test and
Evaluation Master Plan how the reviews fit into the overall engineering
process.
Statements by program officials responsible for system development and
testing activities, as well as briefing materials and diagrams that
these officials provided, did not add sufficient clarity to describe a
well-defined life cycle management approach. Moreover, these
descriptions were not always consistent with what was contained in the
documentation, as the following examples demonstrate:
* Component testing is not described in the Test and Evaluation Master
Plan in a manner consistent with how officials described this testing.
Specifically, while the plan states that components will be tested
against the corresponding component requirements to ensure all
component performance can be verified, program officials stated that
not all components will undergo component testing. Instead, they said
that testing is not required if component vendors submit a certificate
of compliance for certain specifications.
* Functional qualification testing was described to us by program
officials in July 2008 as a type of testing to be performed during
software development activities. However, this type of testing is not
defined in available program documentation, and it was not included in
any versions of the documentation associated with the life cycle
management approach and related engineering processes.
* Certain reviews specified in documentation of the life cycle
management process are not sufficiently defined. For example, the
Systems Engineering Plan shows a Production Readiness Review as part of
the system-level process and an Operational Readiness Review as part of
the project-level process. However, program officials stated that these
reviews are not completely relevant to SBInet because they are targeted
for informational systems rather than tactical support systems, such as
SBInet. According to the officials, they are in the process of
determining how to apply the reviews to SBInet. For example, in July
2008, officials reported that they may move the Production Readiness
Review from the system-level set of activities, as shown in the Systems
Engineering Plan, to the project-level set of activities. Program
officials also stated that they are working to better define these
reviews in the Systems Engineering Plan.
Program officials told us that the SBInet life cycle management
approach and related engineering processes are understood by both
government and contractor staff through the combination of the draft
Systems Engineering Plan and government-contractor interactions during
design meetings. Nevertheless, they acknowledged that the approach and
processes are not well documented, citing a lack of sufficient staff to
both document the processes and oversee the system's design,
development, testing, and deployment. They also told us that they are
adding new people to the project with different acquisition
backgrounds, and that they are still learning about, evolving, and
improving the approach and processes. According to these officials, a
revised and updated Systems Engineering Plan should be finalized by
September 2008.
The lack of definition and stability in the approach and related
processes being used to define, design, develop, acquire, test, and
deploy SBInet introduce considerable risk that both the program
officials and contractor staff will not understand what needs to be
done when, and thus that the program will not consistently employ
disciplined and rigorous methods. Without the use of such methods, the
risk of delivering a system that does not meet operational needs and
does not perform as intend is increased. Moreover, without a well-
defined approach and processes, it is difficult to gauge progress and
thus promote performance and accountability for results.
Limitations of SBInet Requirements Development and Management Efforts
Increase Program Risk:
Well-defined and managed requirements are a cornerstone of effective
system development and acquisition. According to recognized guidance,
[Footnote 15] documenting and implementing a disciplined process for
developing and managing requirements can help reduce the risks of
developing a system that does not meet user needs, cannot be adequately
tested, and does not perform or function as intended. Such a process
includes, among other things, eliciting user needs and involving users
in the development process; ensuring that requirements are complete,
feasible, verifiable, and approved by all stakeholders; documenting and
approving the requirements to establish a baseline for subsequent
development and change control; and ensuring that requirements are
traceable both back to operational requirements and forward to detailed
system requirements and test cases.
To the program office's credit, it recently developed guidance for
developing and managing requirements that is consistent with recognized
leading practices. For example, the program's guidance states that a
requirements baseline should be established and that requirements are
to be traceable both back to higher-level requirements and forward to
verification methods. However, this guidance was not finalized until
February 2008 and thus was not used in performing a number of key
requirements-related activities.
In the absence of well-defined guidance, the program's efforts to
implement leading practices for developing and managing requirements
have been mixed. For example, while the program has elicited user needs
as part of its efforts to develop high-level operational requirements,
it has not baselined all requirements. Further, it has not ensured that
the operational requirements were, for example, verifiable, and it has
not made certain that all of the different levels of requirements are
aligned to one another. As a result, the risk of SBInet not meeting
mission needs and performing as intended is increased, as are the
chances of expensive and time-consuming system rework.
Requirements Development and Management Guidance Is Consistent with
Leading Practices but Was Developed After Many Requirements Activities
Were Performed:
The SBInet program office has developed guidance for developing and
managing requirements that is generally consistent with recognized
leading practices. According to these practices, effectively developing
and managing requirements includes, among other things, eliciting
users' needs early in the development process and involving them
throughout the process; ensuring that requirements are complete,
feasible, verifiable, and approved by all stakeholders; documenting and
approving the requirements to establish a baseline for subsequent
development and change control; and ensuring that requirements are
traceable both back to operational requirements and forward to detailed
system requirements and test cases.
In February 2008, the program office approved its SBInet Requirements
Development and Management Plan. According to the plan, its purpose is
to describe a comprehensive approach to developing and managing
requirements for the SBInet program. Our analysis of this plan shows
that it is consistent with leading practices. For example, the plan
states that:
* users should provide input to the requirements definition process
through early and ongoing participation in integrated product teams,
user conferences, and other requirements-gathering and verifying
activities;
* requirements developers are to ensure that requirements are complete,
unambiguous, achievable, verifiable, and not redundant;
* a requirements baseline should be created to provide a common
understanding of the system to be built and to prevent deviations to
the requirements from entering during design, development, or testing;
and:
* bidirectional traceability both back to higher-level requirements and
forward to detailed test methods should be established and maintained
and that the requirements management team is responsible for
maintaining the requirements management database and holding the
contractor responsible for any traceability issues.
Moreover, the plan defines procedures and steps for accomplishing each
of these goals. For example, the procedure for requirements development
outlines the purpose of the procedure, who is involved, what
documentation is necessary to begin the procedure, and what outputs are
expected at the end. The procedure then describes 16 steps necessary to
achieve the requirements development activities. A separate procedure
is provided for requirements verification and validation.
However, the plan was not approved until after key SBInet requirements
documents were written and baselined. Specifically, user operational
requirements were approved and baselined in March 2007; system
requirements were first baselined in March 2007; and several component
requirements were baselined in June, August, and October 2007. As a
result, the plan was not used to guide these requirements development
and management efforts.
Program Office Has Taken Steps to Involve Users in Developing High-
Level Requirements:
As noted above, one of the leading practices associated with effective
requirements development and management is engaging system users early
and continuously. In doing so, the chances of defining, designing, and
delivering a system that meets their needs and performs as intended are
increased.
In developing the operational requirements, the System Program Office
involved SBInet users in a manner that is consistent with leading
practices and that reflects lessons learned from Project 28.
Specifically, it conducted requirements-gathering workshops from
October 2006 through April 2007 to ascertain the needs of Border Patrol
agents. In addition, it established work groups in September 2007 to
inform the next revision of the operational requirements by soliciting
input from both the Office of Air and Marine Operations and the Office
of Field Operations. Further, to develop the COP technology for SBInet,
the program office is following a software development methodology that
allows end users to be directly involved in software development
activities and, thereby, permits software solutions to be tailored to
users' needs.[Footnote 16]
Through such efforts to identify and elicit user needs in developing
high-level requirements, the chances of developing a system that will
meet user needs are increased.
Not All Levels of Requirements Have Been Adequately Baselined:
The creation of a requirements baseline is important for providing a
stable basis for system design, development, and testing. Such a
baseline establishes a set of requirements that have been formally
reviewed and agreed on and thus serve as the basis for further
development or delivery. Until requirements are baselined, they remain
unclear and subject to considerable and uncontrolled change, which in
turn makes system design, development, testing, and deployment efforts
equally uncertain. According to SBInet program officials, the SBInet
Requirements Development and Management Plan, and leading practices,
requirements should be baselined before key system design activities
begin, since the requirements are intended to inform, guide, and
constrain the system's design.
For SBInet, while many of the requirements have been baselined, two
types have not yet been baselined. According to the System Program
Office, the operational requirements and the system requirements were
approved and baselined in March 2007. In addition, various system
component requirements were baselined in June, August, and October of
2007. However, the program had not baselined its COP software
requirements as of July 2008, although according to program officials,
the COP has been designed and is under development. Further, it has yet
to baseline its project-level requirements, which define the
requirements for the system configuration to be deployed to a specific
geographical area, such as Tucson 1.
With respect to the COP, requirements for the software and hardware had
not been baselined as of the end of July 2008, despite the fact that a
combined Preliminary Design Review and Critical Design Review for the
COP was held in February 2008 and a Critical Design Review for the
system as whole was held in June 2008. According to agency officials
and the SBInet Requirements Development and Management Plan,
requirements should be baselined before the Critical Design Review.
Regardless, program officials state that the contractor has developed
several "builds" (i.e., versions) of the COP, which are currently being
tested. According to program officials, the requirements were not
complete because certain interface requirements[Footnote 17] had not
yet been completely identified and defined. Without baselined
requirements, the basis of the system design and the degree to which it
satisfies requirements are unclear. Moreover, the risk of the design
not aligning to requirements is increased. According to the results of
the Critical Design Review, this risk was realized. Specifically, the
System Program Office notified the contractor that there was no
evidence linking the performance of surveillance components to the
system requirements, that the review could not be completed until the
interface requirements had been finalized, and that a mature design had
not been presented at the review.
With respect to project-level (i.e., geographic area) deployment
requirements, baselined requirements do not yet exist. Specifically,
requirements for the Tucson Sector, which includes Tucson 1 and Ajo 1,
have yet to be baselined. According to the SBInet Requirements
Development and Management Plan, requirements should be baselined
before the Project Requirements Review, and a new requirements baseline
should be created following the subsequent Deployment Design Review.
However, project-level requirements were not baselined at a Project
Requirements Review held for the Tucson Sector Project in March 2007 or
at a Deployment Design Review in June 2007. Officials stated that this
is because the plan was not approved until February 2008 and thus was
not in effect. However, since the plan became effective, Deployment
Design Reviews were held for Tucson 1 and Ajo 1 in April 2008 and May
2008, respectively, but the project-level requirements were not
baselined.
Despite the absence of baselined requirements, the System Program
Office has proceeded with development, integration, and testing
activities for the Block 1 capabilities to be delivered to Tucson 1 and
Ajo l. As a result, it faces an increased risk of deploying systems
that do not align well with requirements and thus may require
subsequent rework. The lack of project requirements has already had an
effect on testing activities. Specifically, the draft system
integration test plan notes that, without project requirements, testing
will have to be guided by a combination of other documents, including
engineering development requirements, related component requirements,
and architectural design documents.
Baselined Operational Requirements Used to Inform Lower-Level
Requirements Are Limited:
As stated above, one of the leading practices for developing and
managing requirements--which is reflected in the program office's own
plan--is that requirements should be sufficiently analyzed to ensure
that the requirements are, among other things, complete, unambiguous,
and verifiable. However, an independent review of SBInet operational
requirements reported numerous problems. Specifically, a review of the
SBInet program commissioned by the Office of the Deputy Secretary of
Homeland Security found that several requirements were unaffordable and
unverifiable. Examples of these requirements include the following:
* Allow for complete coverage of the specified area or zone to be
surveilled.
* Maximize intended deterrence and minimize countermeasure
effectiveness.
* Function with high reliability under reasonably foreseeable
circumstances.
* Reliably provide the appropriate power and bandwidth at the least
cost that will support the demand.
In April 2008, a program official stated that the operational
requirements document is currently being rewritten to address concerns
raised by this review. For example, we were told that certain system
performance requirements are being revised in response to a finding
that the requirements are insufficient to ensure delivery of a properly
functioning system. Program officials stated that they expect to
finalize the revised operational requirements document in October 2008.
However, given the number and types of problems associated with the
operational requirements--which are the program's most basic customer
requirements and form the basis for all lower-level requirements--it is
unclear how the system, component, and software requirements can be
viewed as verifiable, testable, or affordable. Until these problems are
addressed, the risk of building and deploying a system that does not
meet mission needs and customer expectations is increased, which in
turn increases the chances of expensive and time-consuming system
rework.
SBInet Requirements Have Not Been Sufficiently Aligned:
As noted above, one of the leading practices associated with developing
and managing requirements is maintaining bidirectional traceability
from high-level operational requirements through detailed low-level
requirements to test cases. The SBInet Requirements Development and
Management Plan recognizes the importance of traceability, stating that
a traceability relationship should exist among the various levels of
requirements. For example, it states that operational requirements
should trace to the system requirements, which in turn should trace to
component requirements. Further, it states that component requirements
should trace to design requirements and to a verification method. In
addition, the SBInet System Program Office established detailed
guidance[Footnote 18] for populating and maintaining the requirements
database for maintaining linkages among the various levels of
requirements and test verification methods.
To provide for requirements traceability, the prime contractor
established such a requirements management database. However, the
reliability of the requirements in this database is questionable, and
the SBInet System Program Office has not effectively overseen the
contractor's management of requirements through this database.
Specifically, we attempted to trace requirements in the version of this
database that the program office received in March 2008 and were unable
to trace large percentages of component requirements to either higher-
level or lower-level requirements. For example, an estimated 76 percent
(with a 95 percent degree of confidence of being between 64 and 86
percent) of the component requirements that we randomly sampled could
not be traced to the system requirements and then to the operational
requirements. In addition, an estimated 20 percent (with a 95 percent
degree of confidence of being between 11 and 33 percent) of the
component requirements in our sample failed to trace to a verification
method. See table 5 for the failure rates for each of our tracing
analyses, along with the related confidence intervals.
Table 5: SBInet Requirements Traceability Results:
Traceability links from component requirement: To system requirement
and then to operational requirement;
Estimated failure rate: 76%;
95 percent confidence interval: 64-86%.
Traceability links from component requirement: To system requirement;
Estimated failure rate: 48%;
95 percent confidence interval: 34-61%.
Traceability links from component requirement: To verification method;
Estimated failure rate: 20%;
95 percent confidence interval: 11-33%.
Traceability links from component requirement: To design requirement;
Estimated failure rate: 100%;
95 percent confidence interval: 95-100%.
Source: GAO analysis of program office data.
[End of table]
While program officials could not explain the reason for this lack of
traceability in most cases, they did attribute the 100-percent failure
in tracing component requirements to the design requirements to the
absence of any design requirements in the program office's copy of the
database.
A contributing factor to the program office's inability to explain why
requirements were not traceable is its limited oversight of the
contractor's efforts to manage requirements through this database.
According to program officials, the contractor created the SBInet
requirements management database in December 2006, but the program
office did not receive a copy of the database until March 2008, despite
requests for it beginning in fall 2007. In early May 2008, the Chief
Engineer told us the contractor had been reluctant to provide the
database because it viewed the database's maturity level as low.
Moreover, the program office's direct access to the database had not
been established because of security issues, according to this
official.
Following our efforts to trace requirements, the program office
obtained direct access to the contractor's database and initiated
efforts with the contractor to resolve the traceability gaps. However,
program officials told us that they are still not certain that the
database currently contains all of the system requirements or even the
reduced Block 1 system requirements. As a result, they did not rely on
it during the recent Critical Design Review to verify requirements
traceability. Instead, they said that manual tracking methods were
used.
Without ensuring that requirements are fully traceable, the program
office does not have a sufficient basis for knowing that the scope of
the contractor's design, development, and testing efforts will produce
a system solution that meets operational needs and performs as
intended. As a result, the risk of expensive and time-consuming system
rework is increased.
Limitations in Key SBInet Testing and Test Management Activities
Increase Program Risk:
To be effectively managed, testing should be planned and conducted in a
structured and disciplined fashion. This includes, among other things,
having an overarching test plan or strategy as a basis for managing
system testing, developing well-defined and approved plans for
executing testing activities, and testing individual system components
to ensure that they satisfy defined requirements prior to integrating
them into the overall system.
The SBInet System Program Office is not effectively managing its
testing activities. Specifically, it has not tested individual system
components prior to integrating these components with other components
and the COP software. In addition, although the program's draft Test
and Evaluation Master Plan is currently being used as the program's
overall strategy to manage SBInet testing, the plan is incomplete and
unclear with respect to several test management functions. As a result,
the chances of SBInet testing being effectively performed are reduced,
which in turn increases the risk that the delivered and deployed system
will not meet operational needs and not perform as intended.
System Integration Testing Has Not Been Adequately Planned or Executed:
To be effectively managed, relevant federal guidance[Footnote 19]
states that testing should, among other things, be governed by a well-
defined and approved plan, and it should be executed in accordance with
this plan. Further, integration testing should be preceded by tests of
system components (whether acquired or developed) that are to be
integrated to form the overall system. Once the components are tested
to ensure that they satisfy defined requirements, the integrated system
can be tested to verify that it performs as required. For SBInet, this
has not occurred. As a result, the risk of the system not meeting
operational needs and not performing as intended is increased, which in
turn is likely to introduce the need for expensive and time-consuming
system rework.
The SBInet Systems Program Office reports that it began Block 1 system
integration testing in June 2008. However, it still does not have an
approved system integration test plan. Specifically, the system's
Critical Design Review, which was held in June 2008, found numerous
problems with the contractor's plan for system integration testing, and
as a result, the program office did not accept the plan. Examples of
problems were that the test plan refers to other test plans that the
contractor had yet to deliver and the plan identified system components
that were not expected to be part of Block 1. As a result, the program
office decided that the system integration plan needed to be revised to
document and describe the full set of actual testing activities that
were to occur, including identifying where and to what level the
different phases of integration tests would occur.
Notwithstanding these problems and the absence of an approved plan, the
contractor began integration testing in June 2008. According to program
officials, this was necessary to meet the tight time frames in the
schedule. However, without an accepted system integration test plan in
place, testing cannot be effectively managed. For example, the adequacy
of the test scope cannot be assured, and the progress in completing
test activities cannot be measured. As a result, there is an increased
risk that the delivered system will not meet operational needs and will
not perform as intended and that expensive and time-consuming system
rework will be required.
Moreover, the SBInet draft Test and Evaluation Master Plan describes
system integration testing as first testing individual components to
verify that the smallest defined module of a system works as intended
(i.e., meets functional and performance requirements). This allows
defects with individual components to be identified and corrected
before they are integrated with other system components. Once the
components are tested, their respective hardware and software
interfaces are to be tested before subsystems are tested in combination
with the COP software. Such an incremental approach to testing permits
system defects to be found and addressed before system components are
integrated and component problems become more expensive and time-
consuming to correct.
However, the SBInet System Program Office has not performed individual
component testing as part of integration testing. As of July 2008,
agency officials reported that component-level tests had not been
completed and were not scheduled to occur. Instead, officials stated
that Block 1 components were evaluated based on what they described as
"informal tests" (i.e., contractor observations of cameras and radar
suites in operation at a National Guard facility in the Tucson Sector)
and stated that the contractors' self-certification that the components
meet functional and performance requirements was acceptable. However,
this approach is not consistent with the Test and Evaluation Master
Plan. Moreover, program officials acknowledged that this approach did
not verify if the individual components in fact met requirements.
Nevertheless, they said that they have recently modified their
definition of component testing to allow the contractor's certification
to be used.
In our view, relying solely on contractor certification is not a
sufficient substitute for component testing--as defined in the Test and
Evaluation Master Plan--because it increases the risk that components,
and thus the entire system, will not perform as intended. This risk is
already starting to be realized. Specifically, the results of the Block
1 Critical Design Review performed in early in June 2008 show that
design documents did not link components' performance to the system
requirements.
Key Test Management Activities Have Not Been Adequately Defined and
Addressed:
To ensure that system testing is effectively performed, federal
guidance provides for having an overarching test plan or strategy to
use as a basis for managing system testing. Among other things, this
test management plan should define the schedule of high-level test
activities in sufficient detail to allow for more detailed test
planning and execution to occur and to ensure that test progress can be
tracked and results can be reported and addressed. The plan should also
define the roles and responsibilities of the various groups responsible
for different levels of testing and include a description of how the
test organization manages and oversees these groups in their
activities.
The SBInet Test and Evaluation Master Plan, which documents the
program's test strategy and is being used to manage system testing, has
yet to be approved by the SBInet Acting Program Manager. As of July
2008, program officials told us that they did not expect the draft plan
to be approved until August 2008, even though testing activities began
in June 2008.[Footnote 20]
Moreover, the draft Test and Evaluation Master Plan is not complete.
For example, it does not contain an accurate and up-to-date test
schedule with milestones and completion dates for all levels of test
activities. Rather, the schedule information included in the plan has
been overtaken by events, to the point that program officials stated
that many of the dates on the schedules have changed or are not
accurate. Moreover, they described attempting to have an accurate
schedule of testing activities and events as "futile" because the
program's schedule is constantly changing. As a another example, the
draft Test and Evaluation Master Plan does not identify any metrics for
measuring testing progress for any type of testing to be performed.
According to federal guidance, an accurate schedule is necessary to
inform planning for and sequencing of each type of testing to be
performed, including ensuring that test resources are available when
needed and that predecessor test events occur before successor events
begin. This guidance also states that without performance metrics, it
is difficult to understand where test activities stand and what they
show in a manner that can inform program decision making.
As another example, the draft Test and Evaluation Master Plan does not
clearly define the roles and responsibilities of various entities that
are involved in system testing. Specifically, the plan identifies seven
entities, but it only provides vague descriptions of their respective
roles and responsibilities that are not meaningful enough to
effectively guide their efforts. For example, the plan identifies two
entities that are to be involved in operational testing: the DHS
Science and Technology Test and Evaluation Office and the U.S. Army
Test and Evaluation Command. According to the plan, the DHS office is
to function as the operational test authority and will be responsible
for initial planning of "dedicated initial" and "follow-on" operational
testing and evaluation, and the Army group is to conduct operational
testing. With no further clarification, it is not clear what is
expected of each of these entities, including how they are to interact.
Table 6 lists each of the identified entities and provides their
respective roles and responsibilities copied from the draft plan.
Table 6: Roles and Responsibilities of Entities Identified in the Draft
SBInet Test and Evaluation Master Plan:
Entity: System Prime Contractor;
Roles and responsibilities as defined in plan:
* Conducts developmental testing (e.g., component acceptance testing,
system integration testing, and verification and validation of
performance).
Entity: U.S. Army Test and Evaluation Command;
Roles and responsibilities as defined in plan:
* Conducts sample System Prime verification testing;
* Conducts operational testing;
* Gathers field data to assess the degree to which SBInet achieves
mission needs.
Entity: U.S. Customs and Border Protection;
Roles and responsibilities as defined in plan:
* Provides operational feedback using defined metrics that are to be
used to identify and evaluate necessary SBInet Concept of Operations
modifications.
Entity: Independent Verification and Validation agent;
Roles and responsibilities as defined in plan:
* Provides CBP with objective third party evaluation of the C3I
development to verify system design and applications meet C3I
requirements.
Entity: DHS Science and Technology Test and Evaluation Office;
Roles and responsibilities as defined in plan:
* Functions as the Operational Test Authority, including initial
planning of dedicated Initial Operational Test and Evaluation and
Follow-on Operational Test and Evaluation.
Entity: Program Test and Evaluation Integrated Project Teams;
Roles and responsibilities as defined in plan:
* Provides technical expertise to the System Prime Integrated Project
Teams;
* Provides expertise in test plan review and test observation.
Entity: SBInet System Program Office Test and Evaluation Division;
Roles and responsibilities as defined in plan:
* Develops test and evaluation reports.
Source: SBInet data.
[End of table]
Besides being vague, the descriptions of roles and responsibilities are
also incomplete and not consistent with other program documents. For
example, according to the draft plan, the Test and Evaluation Division
of the SBInet System Program Office is responsible only for developing
test and evaluation reports. However, according to the draft SBInet
Systems Engineering Plan, this entity is to act as a subject matter
expert for the oversight or conduct of various testing activities.
Beyond this lack of clearly defined roles and responsibilities, there
are other problems with the groups assigned testing roles. First, some
of the entities identified in the draft plan are not yet operational
and thus are unavailable to participate and perform their assigned
roles and responsibilities. According to program officials, the
independent verification and validation agent has not been selected,
the Integrated Project Teams have not been chartered, and DHS is still
in the process of establishing the DHS Science and Technology Test and
Evaluation Office. Second, although CBP has an interagency agreement
with the U.S. Army Test and Evaluation Command for operational testing,
no such agreement exists with the SBInet program specifically for Block
1 testing.
Finally, neither the draft Test and Evaluation Master Plan nor the
draft Systems Engineering Plan clearly defines the program office's
role and responsibilities for managing and overseeing each of the other
six test entities and their respective test activities. The SBInet
System Program Office is responsible and accountable for ensuring that
the system is successfully deployed and operates as intended. Given the
criticality of testing in ensuring a successful program, this means
that the program office must ensure that each of these entities
executes its assigned roles effectively. However, the draft Test and
Evaluation Master Plan does not recognize this role and its associated
responsibilities. Further, while the draft Systems Engineering Plan
states that the program office is responsible for engaging external
test agents, it provides no further description of the program office's
roles and responsibilities.
Without clearly defined roles and responsibilities for all entities
involved in SBInet testing, the risk of test activities not being
effectively and efficiently performed increases. As a result, the
chances are increased that the deployed system will not meet
operational requirements and perform as intended. Ultimately, this
could lead to expensive and time-consuming system rework.
Conclusions:
A fundamental aspect of successfully implementing a large program like
SBInet is establishing program commitments, including what capabilities
will be delivered and when and where they will be delivered. Only
through establishing such commitments and by adequately defining the
approach and processes to be used in delivering these commitments, can
DHS effectively position itself for measuring progress, ensuring
accountability for results, and delivering a system solution with its
promised capabilities and benefits on time and within budget
constraints. For SBInet, this has not occurred to the extent that it
needs to for the program to have a meaningful chance of succeeding. In
particular, commitments to the timing and scope of system capabilities
remain unclear and continue to change, with the program committing to
far fewer capabilities than originally envisioned. Further, how the
SBInet system solution is to be delivered has been equally unclear and
inadequately defined. Moreover, while the program office has defined
key practices for developing and managing requirements, these practices
were developed after several key requirements activities were
performed. In addition, efforts performed to date to test whether the
system meets requirements and functions as intended have been limited.
Collectively, these limitations are significant in that they increase
the risk that the delivered system solution will not meet user needs
and operational requirements and will not perform as intended. These
consequences, in turn, increase the chances that the system will
require expensive and time-consuming rework. In light of these
circumstances and risks surrounding SBInet, it is important for the
program office to reassess its approach to and plans for the program--
including its associated exposure to cost, schedule, and performance
risks--and to disclose these risks and alternative courses of action
for addressing them to DHS and congressional decision makers. It is
also important for the program to correct the weaknesses discussed in
this report surrounding the program's unclear and constantly changing
commitments and its life cycle management approach and processes,
including the processes and efforts performed to date relating to
requirements development and management and testing.
While doing so will not guarantee a successful program, it will
minimize the program's exposure to risk and thus decrease the
likelihood that it will fall short of expectations. For SBInet, living
up to expectations is important because the program is a large,
complex, and integral component of DHS's border security and
immigration control strategy.
Recommendations for Executive Action:
To improve DHS's efforts to acquire and implement SBInet we are making
eight recommendations.
To permit meaningful measurement and oversight of and accountability
for the program, we recommend that the Secretary of Homeland Security
direct the CBP Commissioner to ensure that (1) the risks associated
with planned SBInet acquisition, development, testing, and deployment
activities are immediately assessed and (2) the results, including
proposed alternative courses of action for mitigating the risks, are
provided to the Commissioner and DHS's senior leadership, as well as to
the department's congressional authorization and appropriation
committees.
We further recommend that the Secretary of Homeland Security direct the
CBP Commissioner to have the Acting SBInet Program Manager take the
following additional actions:
* Establish and baseline the specific program commitments, including
the specific system functional and performance capabilities that are to
be deployed to the Tucson, Yuma, and El Paso Sectors, and establish
when these capabilities are to be deployed and are to be operational.
* Finalize and approve an integrated master schedule that reflects the
timing and sequencing of the work needed to achieve these commitments.
* Revise and approve versions of the SBInet life cycle management
approach, including the draft Systems Engineering Plan and draft Test
and Evaluation Management Plan, and in doing so, ensure that these
revised and approved versions are consistent with one another, reflect
program officials' recently described changes to the engineering and
testing approaches, and reflect relevant federal guidance and
associated leading practices.
* Ensure that the revised and approved life cycle management approach
is fully implemented.
* Implement key requirements development and management practices to
include (1) baselining requirements before system design and
development efforts begin; (2) analyzing requirements prior to being
baselined to ensure that they are complete, achievable, and verifiable;
and (3) tracing requirements to higher-level requirements, lower-level
requirements, and test cases.
* Implement key test management practices to include (1) developing and
documenting test plans prior to the start of testing; (2) conducting
appropriate component level testing prior to integrating system
components; and (3) approving a test management strategy that, at a
minimum, includes a relevant testing schedule, establishes
accountability for testing activities by clearly defining testing roles
and responsibilities, and includes sufficient detail to allow for
testing and oversight activities to be clearly understood and
communicated to test stakeholders.
Agency Comments and Our Evaluation:
In written comments on a draft of this report, signed by the Director,
Departmental GAO/Office of Inspector General Liaison and reprinted in
appendix II, the department stated that it agrees with seven of our
eight recommendations, and partially disagrees with one aspect of the
remaining recommendation. The department also stated that our report is
factually sound and that it is working to address our recommendations
and resolve the management and operational challenges identified in the
report as expeditiously as possible. In this regard, it described
actions recently completed, underway, and planned that it said
addresses our recommendations. It also provided technical comments that
we have incorporated in the report, as appropriate.
Regarding our recommendation to implement key test management
practices, including conducting appropriate component-level testing
prior to integrating system components, DHS commented that its current
test strategy provides for the appropriate degree of technical
confidence for commercially available products, as evidenced by either
certificates of conformance from the original equipment manufacturer,
test documentation from independent government laboratories, or the
prime contractor's component/integration level testing. We support
DHS's current test strategy, as it is consistent with our
recommendation. Specifically, it expands on the department's prior
strategy for component testing, which was limited to manufacturer self-
certification of component conformance and informal observations of
system components, by adding the use of independent government
laboratories to test the components. We would emphasize, however, that
regardless of the method used, it is important that confidence be
gained in components prior to integrating them, which our
recommendation recognizes. As our report states, component-level
testing was not performed for Block 1 components prior to initiating
integration testing. Federal guidance and the SBInet program office's
own Test and Evaluation Master Plan recognize the need to first test
individual components to verify that the system modules work as
intended (i.e., meet functional and performance requirements) before
conducting integration testing. By adopting such a hierarchical
approach to testing, the source of any system defects can be discovered
and isolated sooner rather than later, thus helping to avoid the
potential for expensive and time-consuming system rework.
We are sending copies of this report to the Chairmen and Ranking
Members of the Senate and House Appropriations Committees and other
Senate and House committees and subcommittees that have authorization
and oversight responsibilities for homeland security. We will also send
copies to the Secretary of Homeland Security, the Commissioner of U.S.
Customs and Border Protection, and the Director of the Office of
Management and Budget. In addition, this report will be available at no
cost on the GAO Web site at [hyperlink, http://www.gao.gov].
Should your offices have any questions on matters discussed in this
report, please contact me at (202) 512-3439 or at hiter@gao.gov.
Contact points for our Office of Congressional Relations and Public
Affairs may be found on the last page of this report. Key contributors
to this report are listed in appendix III.
Signed by:
Randolph C. Hite:
Director, Information Technology Architecture and Systems Issues:
[End of section]
Appendix I: Objectives, Scope, and Methodology:
Our objectives were to determine whether the Department of Homeland
Security (DHS) (1) has defined the scope and timing of planned SBInet
capabilities and how these capabilities will be developed and deployed,
(2) is effectively defining and managing SBInet requirements, and (3)
is effectively managing SBInet testing.
To determine the extent to which DHS has defined the scope and timing
of planned SBInet capabilities and how these capabilities will be
developed and deployed, we reviewed program documentation, such as the
draft Systems Engineering Plan, the Systems Engineering Management
Plan, the Operational Requirements Document, the Mission Engineering
Process, the draft Test and Evaluation Master Plan, and the 2008 SBI
Expenditure Plan to understand the SBInet engineering process and the
scope and timing of planned deployments. We also interviewed SBInet
officials and contractors to gain clarity beyond what was included in
the program documentation and to obtain schedule information in the
absence of an integrated master schedule for the program.
To determine if DHS is effectively defining and managing SBInet
requirements, we reviewed relevant documentation, such as the
Requirements Development and Management Plan, the Requirements
Management Plan, the Configuration and Data Management Plan, the
Operational Requirements Document, System of Systems A-Level
Specification, B-2 Specifications, and Vendor Item Control Drawings,
and compared them to industry best practices[Footnote 21] to determine
the extent to which the program has effectively managed the systems
requirements and maintained traceability backwards to high-level
operational requirements and system requirements, and forward to system
design and verification methods.
To assess reliability of the requirements data, we reviewed quality and
access controls of the requirements database. We then randomly selected
59 requirements from a sample of 1,666 component requirements and
traced them backwards to the system requirements and then to the
operational requirements and forward to design requirements and
verification methods. Because we followed a probability procedure based
on random selection, we are 95 percent confident that each of the
confidence intervals in this report will include the true values in the
study population. We used statistical methods appropriate for audit
compliance testing to estimate 95 percent confidence intervals for the
traceability of requirements in our sample. In addition, we interviewed
program and contractor officials involved in requirements management to
understand their roles and responsibilities. We also visited a
contractor development facility in Huntsville, Alabama, to understand
the contractor's role in requirements management and development and
the use of its requirements management tool, known as the Dynamic
Object-Oriented Requirements System (DOORS). In addition, we attended a
demonstration of SBInet Rapid Application Development/Joint Application
Design to understand how the users are involved in developing
requirements.
To determine if DHS is effectively managing SBInet testing, we reviewed
relevant documentation, such as the SBInet Test and Evaluation Master
Plan, the Systems Integration Test Plan, the Quality Assurance
Surveillance Plan, the Requirements Verification Plan, the
Characterization Test Plan, and the Prime Mission Product Design, and
compared them to relevant federal guidance[Footnote 22] to determine
the extent to which the program has effectively managed its testing
activities. We also interviewed SBInet officials to gain clarity beyond
what was included in the program documentation and to obtain schedule
information in the absence of a formal testing schedule. In addition,
we visited a contractor facility in Huntsville, Alabama, to better
understand the contractor's role in testing activities and to observe
the test lab and how testing is performed.
In addition, we visited the Tucson Sector Border Patrol Headquarters in
Tucson, Arizona, to see the technology that was deployed as a prototype
to understand the scope of the technology, how the Border Patrol agents
use the technology, and future plans.
To assess data reliability, we reviewed related program documentation
to substantiate data provided in interviews with knowledgeable agency
officials, where available. For the information contained in the DHS
independent study on SBInet, we interviewed the individuals responsible
for conducting the review to understand their methodology, and
determined that the information derived from this study was
sufficiently reliable for the purposes of this report. We have made
appropriate attribution indicating the data's sources.
We performed our work at the U.S. Customs and Border Protection
headquarters and contractor facilities in the Washington, D.C.,
metropolitan area; the Tucson Sector Border Patrol headquarters in
Tucson, Arizona; and a contractor facility in Huntsville, Alabama. We
conducted this performance audit from August 2007 to September 2008 in
accordance with generally accepted government auditing standards. Those
standards require that we plan and perform the audit to obtain
sufficient, appropriate evidence to provide a reasonable basis for our
findings and conclusions based on our audit objectives. We believe that
the evidence obtained provides a reasonable basis for our findings and
conclusions based on our audit objectives.
[End of section]
Appendix II: Comments from the Department of Homeland Security:
U.S. Department of Homeland Security:
Washington, DC 20528:
September 9, 2008:
Mr. Randolph C. Hite:
Director:
Information Technology Architecture and Systems Issues:
U.S. Government Accountability Office:
441 G Street, NW:
Washington, DC 20548:
Dear Mr. Hite:
Re: Draft Report GAO-08-1086, Secure Border Initiative: DHS Needs to
Address Significant Risks in Delivering Key Technology Investment (GAO
Job Code 310644):
The Department of Homeland Security (DHS) appreciates the opportunity
to review and comment on the draft report referenced above that focuses
on the technology component ("SBInet") of the Department's Secure
Border Initiative. The GAO report addresses three broad areas of
concern: (1) limited definition of SBInet deployments, capabilities,
schedule, and lifecycle management processes; (2) limitations of SBInet
requirements development and management efforts; and (3) limitations in
key SBInet testing and test management activities.
U.S. Customs and Border Protection (CBP) acknowledges that the GAO
report is factually sound. CBP concurs with recommendations 1 through 7
and partially non-concurs with recommendation 8 for Executive Action.
The draft report highlights several management and operational
challenges that CBP is working to resolve as expeditiously as possible.
To address the challenges described in the draft report, CBP has
initiated a series of actions to improve both the technical and
management processes for SBInet.
Beginning with the DHS's Deep Dive Review and other internal
assessments and continuing through GAO's review, CBP has identified
many of the same concerns as GAO.
On August 19, 2008, CBP met with the DHS Investment Review Board to
formally discuss program risk and agree on courses of action to best
mitigate this risk. The most significant result of this meeting was
DHS's direction to delay the TUS-1 and AJO-1 deployments into 2009.
This recognizes the need to manage the technology risk before the
deployment, and also to divert funds intended to be used for these
deployments in Fiscal Year (FY) 2008 to address more urgent vehicle and
pedestrian fencing project funding requirements. CBP believes this risk
assessment, executive briefings, and discussion not only satisfy the
requirements of Recommendations 1 and 2 but also help to ensure that
program goals are met.
Following the decision to delay the two deployments, SBI was also
directed to provide programmatic documentation to DHS. This
comprehensive programmatic documentation will include:
* Acquisition Program Baseline (APB) including:
- A clear articulation of the overall program scope.
- Out-year Future Years Homeland Security Program (FYHSP) profile.
- A description of the program's evolutionary approach, including Key
Performance Parameters (KPPs) by Block release.
- A definition of Block 1, including Common Operating Picture (COP)
Release 0.5 usable segment.
- A Block deployment approach.
- The Block 1 phased deployment approach.
- Realistic cost and schedule thresholds for Block 1 including
individual phases of deployment within the Block.
- Establishment of a Key Decision Point (KDP) 2 date.
* Test and Evaluation Master Plan (TEMP).
* Life Cycle Cost Estimate (LCCE).
* Integrated Logistics Sustainment Plan (ILSP).
* Systems Engineering Plan (SEP).
* Revised Operational Requirements Document (ORD) - allocated Block 1
requirements.
* SBInet Integrated Master Schedule (IMS).
* SBInet Operational Capabilities Document.
We are confident that this programmatic documentation will satisfy
Recommendations 3, 4, 5, and 7. Recommendation 6 is addressed within
this document. CBP does not, however, fully concur with Recommendation
8. As noted above, CBP will provide testing requirements and a testing
plan that will satisfy most of this recommendation.
The eight recommendations and corrective actions to address them are
included below:
Recommendations 1 and 2:
To improve DHS's efforts to acquire and implement SBInet, as well as to
permit meaningful measurement and oversight of and accountability for
the program, we recommend that the Secretary of Homeland Security
direct the CBP Commissioner to ensure that (1) the risks associated
with planned SBInet acquisition, development, testing, and deployment
activities are immediately assessed, and (2) the results, including
proposed alternative courses of action for mitigating them, are
provided to the Commissioner and DHS's senior leadership.
Response:
CBP concurs with Recommendations 1 and 2.
CBP identified and briefed DHS's senior leadership on the risks
associated with the planned SBInet acquisition, development, testing,
and deployment activities. As a result, SBInet will develop a detailed
program re-plan with supporting program documentation and a SBInet
Block I APB. It should be noted that DHS and CBP have also begun
collaborating on SBInet program re-planning alternatives to reduce
schedule concurrency (program risk) between system testing and
deployment. Proposed alternative courses of action for mitigating them
will continue to be briefed to the Commissioner and DHS's senior
leadership.
Due Date: December 31, 2008.
Recommendation 3:
Establish and baseline the specific program commitments, including the
specific system functional and performance capabilities that are to be
deployed to the Tucson, Yuma, and El Paso sectors, and when these
capabilities are to be deployed and are to be operational.
Response:
CBP concurs with Recommendation 3.
SBInet will develop detailed projections as part of the SBInet Block 1
APB that will establish and baseline the specific Block 1 program
commitments, including the specific system functional and performance
capabilities that are to be deployed to the Tucson and Yuma Border
Patrol Sectors. Because of the current re-planning, CBP is now planning
to field SBInet to the El Paso Sector as part of SBInet Block 2, which
will likely be delivered over FYs 2011 through 2012.
Due Date: December 31, 2008.
Recommendation 4:
Finalize and approve an integrated master schedule that reflects the
timing and sequencing of the work needed to achieve these commitments.
Response:
CBP concurs with Recommendation 4.
A finalized and approved integrated master schedule that reflects the
timing and sequencing of the work needed will be completed, re-
baselined, and included in the SBInet Block 1 APB.
Due Date: December 31, 2008.
Recommendation 5:
Revise and approve versions of the SBInet life cycle management
approach, including the draft System Engineering Plan and draft Test
and Evaluation Management Plan, and in doing so, ensure that these
revised and approved versions are consistent with one another, reflect
program officials' recently described changes to the engineering and
testing approaches, and reflect relevant federal guidance and
associated leading practices.
Response:
CBP concurs with Recommendation 5.
SBI will include in the SBInet Block 1 APB a revised SBInet life cycle
management approach, including a System Engineering Plan and a Test and
Evaluation Management Plan (TEMP). SBI will ensure that the approved
versions reflect program officials' recently described changes to the
engineering and testing approaches, and reflect relevant federal
guidance and associated leading practices. The SBInet Block I APB will
include supporting documentation for key program management processes
related to program requirements and performance projections, testing
and deployment plans, integrated schedules, cost estimates and risk
assessments.
Due Date: December 31, 2008.
Recommendation 6:
Ensure that the revised and approved life cycle management approach is
fully implemented.
Response:
CBP concurs with Recommendation 6.
SBInet will develop a quality management program to ensure that the
revised and approved life cycle management approach is fully
implemented.
Due Date: December 31, 2008.
Recommendation 7:
Implement key requirements development and management practices to
include (1) baselining requirements before system design and
development efforts begin; (2) analyzing requirements prior to being
baselined to ensure that they are complete, achievable, and verifiable;
and (3) tracing requirements to higher-level requirements, lower-level
requirements, and test cases.
Response:
CBP concurs with Recommendation 7.
CBP understands and will comply with GAO's recommendation for following
a disciplined set of activities for the analysis and documentation
(i.e., "baselining") of key system requirements and the definition and
documentation of detailed performance specifications and test plans
underpinning the key system requirements. The SBInet Block I APB will
be based on completed system requirements, system design
specifications, and detailed test plans. Also in conjunction with the
SBInet Block I Qualification Test Readiness Review, CBP will complete
its ongoing effort to update the appropriate databases for tracing all
of these detailed requirements (i.e., the DOORS database).
While CBP officials concur with this recommendation, CBP officials
would like to note that SBInet personnel believe that rather than a
sequential "heel-to-toe" approach for these efforts (i.e. identifying
baseline requirements before beginning design and test) SBInet is
employing a "spiral" approach of iterative design-prototype-test-learn
cycles. This spiral development approach provides for an initial
definition of requirements (under formal configuration control), and
then a period of development and testing to gain user feedback and
engineering confidence with initial (sometimes draft) designs. Final
requirements, as well as final designs, are worked together and often
in parallel. The initial test spirals also help to complete detailed
test plans and procedures needed for final qualification and acceptance
testing.
Due Date: December 31, 2008.
Recommendation 8:
Implement key test management practices to include (1) developing and
documenting test plans prior to the start of testing; (2) conducting
appropriate component level testing prior to integrating system
components; and (3) approving a test management strategy that, at a
minimum, includes a relevant testing schedule, establishes
accountability for testing activities by clearly defining testing roles
and responsibilities, and includes sufficient detail to allow for
testing and oversight activities to be clearly understood and
communicated to test stakeholders.
Response:
CBP partially non-concurs with Recommendation 8.
This partial non-concurrence results from the recommendation language
requiring CBP to "Implement key test management practices to include...
(2) conducting appropriate component level testing prior to integrating
system components..."
In general, CBP understands and will comply with GAO's recommendation
for following a disciplined set of activities for planning, executing,
and reporting SBInet program testing. While CBP presented to GAO
significant progress in maturing SBInet testing strategies and plans,
to include the hierarchical approach for System Integration Tests,
System Qualification Tests, and System Acceptance Tests, CBP
acknowledges program management documentation is lagging. CBP will
update the SBInet Block I APB, which will be based on improved test
plans and system test progress to-date. Additionally, CBP will provide
an updated TEMP with the SBInet Block I APB to highlight specific
testing roles, responsibilities, and those parties accountable for the
revised test program. Moreover, this TEMP is to be signed by the DHS
principal office for SBInet test oversight (DHS Science and Technology
Directorate).
Regarding "appropriate" component-level testing, the SBInet approach
provides for disciplined, hierarchical testing at the hardware/software
component level, integrated component level, and ultimately at the
integrated system level, in the laboratory as well as in operationally
representative field conditions. For component-level testing, SBInet
leadership believes the current test strategy provides the appropriate
degree of technical confidence for commercial off the shelf (COTS)
components, as evidenced by either Certificates of Conformance from the
Original Equipment Manufacturer, test documentation from independent
government laboratories, or through Boeing component/integration level
testing. SBInet believes this approach is consistent with COTS
intensive program best practices and avoids the added cost of
duplicative component level testing with little added utility.
Notwithstanding the differences between the GAO recommendation and
SBInet implementation, SBInet will continue to improve and update
program documentation that is comprehensive, consistent, and keeps
stakeholders abreast of all activities.
Due Date: December 31, 2008.
CBP conducted a sensitivity review of the draft report and did not
identify any information that would require a "For Official Use Only"
designation. Technical comments have been provided under separate cover
and should help clarify particular statements prior to finalizing the
report.
Sincerely,
Signed by:
Jerald E. Levine:
Director:
Departmental GAO/OIG Liaison Office:
[End of section]
Appendix III: GAO Contact and Staff Acknowledgments:
GAO Contact:
Randolph C. Hite, (202) 512-3439 or hiter@gao.gov:
Staff Acknowledgments:
In addition to the contact named above, Deborah Davis (Assistant
Director), Carl Barden, Neil Doherty, Lee McCracken, Jamelyn Payan,
Karl Seifert, Sushmita Srikanth, Karen Talley, and Merry Woo made key
contributions to this report.
[End of section]
Footnotes:
[1] The scope of SBInet is the contiguous United States' land border
with Mexico and Canada.
[2] GAO, Border Security: Key Unresolved Issues Justify Reevaluation of
Border Surveillance Technology Program, [hyperlink,
http://www.gao.gov/cgi-bin/getrpt?GAO-06-295] (Washington, D.C.: Feb.
22, 2006).
[3] [hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-06-295].
[4] CBP divides the United States' borders with Mexico and Canada into
20 sectors responsible for detecting, interdicting, and apprehending
those who attempt illegal entry or to smuggle contraband across U.S.
borders.
[5] Carnegie Mellon Software Engineering Institute's Capability
Maturity Model® Integration for Development defines a baseline as a set
of specifications or work products that has been formally reviewed and
agreed on, which thereafter serves as the basis for further development
or delivery, and that can be changed only through change control
procedures.
[6] See, for example, GAO, Year 2000 Computing Crisis: A Testing Guide,
[hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-10.1.21]
(Washington, D.C.: November 1998).
[7] Tactical infrastructure includes roads, vehicle barriers,
pedestrian fences, etc.
[8] At a port of entry location, CBP officers secure the flow of people
and cargo into and out of the country, while facilitating legitimate
travel and trade.
[9] GAO, Secure Border Initiative: Observations on Selected Aspects of
SBInet Program Implementation, [hyperlink, http://www.gao.gov/cgi-
bin/getrpt?GAO-08-131T] (Washington, D.C.: Oct. 24, 2007) and GAO,
Secure Border Initiative: Observations on the Importance of Applying
Lessons Learned to Future Projects, [hyperlink, http://www.gao.gov/cgi-
bin/getrpt?GAO-08-508T] (Washington, D.C.: Feb. 27, 2008).
[10] The list of reviews is from the draft SBInet Systems Engineering
Plan, dated February 12, 2008. The actual descriptions of the reviews
are from other SBInet documents.
[11] The list of reviews is from the draft SBInet Systems Engineering
Plan, dated February 12, 2008. The actual descriptions of the reviews
are from other SBInet documents.
[12] DHS did not report these timeframes in its December 4, 2006, SBI
Expenditure Plan. Rather, it reported that it planned to deploy SBInet
to the southwest border by 2011 and did not provide a timeframe for
deployment to the northern border.
[13] The area that will be covered by Tucson 1 is similar to the area
covered by Project 28 (sometimes referred to as "Block 0"); it is to
include 23 of the 28 miles associated with Project 28. According to the
System Program Office, the Project 28 capabilities (surveillance
systems and COP) will be replaced with Block 1 capabilities as part of
the Tucson 1 deployment.
[14] The program office is monitoring the individual schedules for five
SBInet task orders.
[15] The Capability Maturity Model® Integration for Development,
developed by the Software Engineering Institute of Carnegie Mellon
University, defines key practices that are recognized hallmarks for
successful organizations that, if effectively implemented, can greatly
increase the chances of successfully developing and acquiring software
and systems. Carnegie Mellon Software Engineering Institute, Capability
Maturity Model® Integration for Development, Version 1.2 (Pittsburgh,
Penn., August 2006).
[16] This method, Rapid Application Development and Joint Application
Design (RAD/JAD), uses graphical user interfaces and direct-end-user
involvement in a collaborative development approach.
[17] Interface requirements describe the capabilities that must be in
place in order to integrate components and products together.
[18] SBInet Requirements Management Plan, January 15, 2007.
[19] See, for example, GAO/AIMD-10.1.21.
[20] DHS stated in agency comments that it plans to revise and approve
the Test and Evaluation Master Plan by December 31, 2008.
[21] Carnegie Mellon Software Engineering Institute, Capability
Maturity Model® Integration for Development, Version 1.2 (Pittsburgh,
Penn., August 2006).
[22] GAO, Year 2000 Computing Crisis: A Testing Guide, [hyperlink,
http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-10.1.21] (November 1998).
[End of section]
GAO's Mission:
The Government Accountability Office, the audit, evaluation and
investigative arm of Congress, exists to support Congress in meeting
its constitutional responsibilities and to help improve the performance
and accountability of the federal government for the American people.
GAO examines the use of public funds; evaluates federal programs and
policies; and provides analyses, recommendations, and other assistance
to help Congress make informed oversight, policy, and funding
decisions. GAO's commitment to good government is reflected in its core
values of accountability, integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each
weekday, GAO posts newly released reports, testimony, and
correspondence on its Web site. To have GAO e-mail you a list of newly
posted products every afternoon, go to [hyperlink, http://www.gao.gov]
and select "E-mail Updates."
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office:
441 G Street NW, Room LM:
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]:
E-mail: fraudnet@gao.gov:
Automated answering system: (800) 424-5454 or (202) 512-7470:
Congressional Relations:
Ralph Dawn, Managing Director, dawnr@gao.gov:
(202) 512-4400:
U.S. Government Accountability Office:
441 G Street NW, Room 7125:
Washington, D.C. 20548:
Public Affairs:
Chuck Young, Managing Director, youngc1@gao.gov:
(202) 512-4800:
U.S. Government Accountability Office:
441 G Street NW, Room 7149:
Washington, D.C. 20548: