Defense Transportation

Study Limitations Raise Questions about the Adequacy and Completeness of the Mobility Capabilities Study and Report Gao ID: GAO-06-938 September 20, 2006

The Department of Defense (DOD) issued the Mobility Capabilities Study (MCS), which was intended to identify and quantify the mobility capabilities needed to support U.S. strategic objectives into the next decade. The MCS found that projected capabilities are adequate to achieve U.S. objectives with an acceptable level of risk--that is, current U.S. inventory of aircraft, ships, prepositioned assets, and other capabilities are sufficient, in conjunction with host nation support, and assuming planned investments take place. The Senate report accompanying the bill for the fiscal year 2005 Defense Authorization Act required GAO to report on the adequacy and completeness of the MCS. GAO assessed the extent to which the MCS met generally accepted research standards that this type of study would be expected to meet to be considered sound and complete.

DOD used an innovative approach in conducting the study and acknowledged methodological limitations in its report; however, it did not fully disclose how these limitations could affect the MCS conclusions and recommendations. Therefore, it is not transparent how the analyses done for the study support DOD's conclusions. Measured against relevant generally accepted research standards, GAO has identified limitations in the MCS and its report that raise questions about their adequacy and completeness. GAO suggests that Congress and other decision makers exercise caution in using the MCS to make investment decisions. Among GAO's findings: Aspects of modeling and data were inadequate in some areas because data were lacking and the models used could not simulate all relevant aspects of the missions. The report did not explain how these limitations could affect the study results or what the impact on projected mobility capabilities might be. Generally accepted research standards require that models used are adequate for the intended purpose, represent a complete range of conditions, and that data used are properly generated and complete. For example, the MCS modeled hypothetical homeland defense missions rather than homeland defense demands derived from a well defined and approved concept of operations for homeland defense, because the specific details of the missions were still being determined and the data used may be incomplete. The MCS also was unable to model the flexible deterrent options/deployment order process to move units and equipment into theater because of lack of data, but the study assumed a robust use of this process. In addition, the MCS report contains over 80 references to the need for improved modeling or data. While the MCS concluded that combined U.S. and host nation transportation assets were adequate, in describing the use of warfighting metrics in its analyses, the report does not provide a clear understanding of the direct relationship of warfighting objectives to transportation capabilities. Additionally, the report stated that further analysis is required to understand the operational impact of increased or decreased strategic lift on achieving warfighting objectives. Relevant generally accepted research standards require that conclusions be supported by analyses. The use of both warfighting and mobility metrics would allow decision makers to know whether combat tasks were achieved and how much strategic transportation is needed to accomplish those tasks. In some cases, the MCS results were incomplete, unclear, or contingent on further study, making it difficult to identify findings and evaluate evidence. Relevant research standards require results to be presented in a complete, accurate, and relevant manner. For example, the report contains recommendations for further studies and assessments, five of which are under way. However, DOD has no plans to report the impact of these studies on the MCS results after the studies are complete. In addition, the report contains qualified information that is not presented clearly, such as varying assessments of intra-theater assets in three different places.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-06-938, Defense Transportation: Study Limitations Raise Questions about the Adequacy and Completeness of the Mobility Capabilities Study and Report This is the accessible text file for GAO report number GAO-06-938 entitled 'Defense Transportation: Study Limitations Raise Questions about the Adequacy and Completeness of the Mobility Capabilities Study and Report' which was released on September 20, 2006. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Committees: United States Government Accountability Office: GAO: September 2006: Defense Transportation: Study Limitations Raise Questions about the Adequacy and Completeness of the Mobility Capabilities Study and Report: Defense Transportation: GAO-06-938: GAO Highlights: Highlights of GAO-06-938, a report to congressional committees Why GAO Did This Study: The Department of Defense (DOD) issued the Mobility Capabilities Study (MCS), which was intended to identify and quantify the mobility capabilities needed to support U.S. strategic objectives into the next decade. The MCS found that projected capabilities are adequate to achieve U.S. objectives with an acceptable level of risk”that is, current U.S. inventory of aircraft, ships, prepositioned assets, and other capabilities are sufficient, in conjunction with host nation support, and assuming planned investments take place. The Senate report accompanying the bill for the fiscal year 2005 Defense Authorization Act required GAO to report on the adequacy and completeness of the MCS. GAO assessed the extent to which the MCS met generally accepted research standards that this type of study would be expected to meet to be considered sound and complete. What GAO Found: DOD used an innovative approach in conducting the study and acknowledged methodological limitations in its report; however, it did not fully disclose how these limitations could affect the MCS conclusions and recommendations. Therefore, it is not transparent how the analyses done for the study support DOD‘s conclusions. Measured against relevant generally accepted research standards, GAO has identified limitations in the MCS and its report that raise questions about their adequacy and completeness. GAO suggests that Congress and other decision makers exercise caution in using the MCS to make investment decisions. Among GAO‘s findings: * Aspects of modeling and data were inadequate in some areas because data were lacking and the models used could not simulate all relevant aspects of the missions. The report did not explain how these limitations could affect the study results or what the impact on projected mobility capabilities might be. Generally accepted research standards require that models used are adequate for the intended purpose, represent a complete range of conditions, and that data used are properly generated and complete. For example, the MCS modeled hypothetical homeland defense missions rather than homeland defense demands derived from a well defined and approved concept of operations for homeland defense, because the specific details of the missions were still being determined and the data used may be incomplete. The MCS also was unable to model the flexible deterrent options/deployment order process to move units and equipment into theater because of lack of data, but the study assumed a robust use of this process. In addition, the MCS report contains over 80 references to the need for improved modeling or data. * While the MCS concluded that combined U.S. and host nation transportation assets were adequate, in describing the use of warfighting metrics in its analyses, the report does not provide a clear understanding of the direct relationship of warfighting objectives to transportation capabilities. Additionally, the report stated that further analysis is required to understand the operational impact of increased or decreased strategic lift on achieving warfighting objectives. Relevant generally accepted research standards require that conclusions be supported by analyses. The use of both warfighting and mobility metrics would allow decision makers to know whether combat tasks were achieved and how much strategic transportation is needed to accomplish those tasks. * In some cases, the MCS results were incomplete, unclear, or contingent on further study, making it difficult to identify findings and evaluate evidence. Relevant research standards require results to be presented in a complete, accurate, and relevant manner. For example, the report contains recommendations for further studies and assessments, five of which are under way. However, DOD has no plans to report the impact of these studies on the MCS results after the studies are complete. In addition, the report contains qualified information that is not presented clearly, such as varying assessments of intra- theater assets in three different places. What GAO Recommends: GAO recommends that the Secretary of Defense, in future mobility s capabilities studies beginning with any study currently under way, develop models and data for critical missions and processes; include in study reports an explanation of how stated limitations might impact results; and, incorporate both mobility and warfighting metrics to determine capabilities. In comments, DOD concurred with two of the recommendations and claimed they did not understand the third, which GAO clarified. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-06-938]. To view the full product, including the scope and methodology, click on the link above. For more information, contact William M. Solis at (202)512-8365 or solisw@gao.gov. [End of Section] Contents: Letter: Results in Brief: Background: Limitations in the MCS Study and Report Raise Questions about Adequacy and Completeness: Conclusions: Recommendation for Executive Action: Agency Comments and Our Evaluation: Appendix I: Scope and Methodology: Appendix II: Comments from the Department of Defense: Appendix III: GAO Contact and Staff Acknowledgments: Table: Table 1: Generally Accepted Research Standards Relevant to MCS Requirements: Figure: Figure 1: MCS Hypothetical Ongoing Contingencies during 7-year Baseline Security Posture Time Frame: United States Government Accountability Office: Washington, DC 20548: September 20, 2006: Congressional Committees: The National Security Strategy of the United States requires global mobility through rapid, effective, and efficient projections of power at home and abroad to deploy and sustain America's armed forces. To improve its mobility capabilities, the Department of Defense (DOD) plans to spend more than $50 billion from fiscal years 2006 through 2011 for aircraft, ships, ground transportation, prepositioned assets, and other mobility assets. DOD has conducted several studies to determine mobility requirements and recently completed a study of its mobility capabilities and issued a report in December 2005. The intent of the Mobility Capabilities Study (MCS) was to identify and quantify the mobility capabilities needed to support U.S. strategic objectives into the next decade. The MCS determined that the projected mobility capabilities are adequate to achieve U.S. objectives with an acceptable level of risk during the period from fiscal years 2007 through 2013; that is, the current U.S. inventory of aircraft, ships, prepositioned assets, and other capabilities are sufficient, in conjunction with host nation support. The MCS emphasized that continued investment in the mobility system, in line with current departmental priorities and planned spending, is required to maintain these capabilities in the future. This includes, for example, fully funding Army prepositioned assets as planned and completing a planned reengineering of the C-5 aircraft. The MCS report also made recommendations to conduct further studies, develop plans and strategies, and improve data collection and mobility models. In fact, DOD officials told us that a Mobility Capabilities Study--2006 is underway. In the Senate report accompanying the bill for the fiscal year 2005 Defense Authorization Act, you asked us to monitor the process used to conduct the MCS and report on the adequacy and completeness of the study.[Footnote 1] Specifically, our objective was to determine whether the MCS was adequate and complete. On March 1, 2006, we briefed your staff on our preliminary observations. This report expands on that briefing and makes recommendations to the Secretary of Defense. To conduct our review of the MCS, we analyzed the final MCS report, the MCS Terms of Reference and MCS Study Plan, as well as other DOD policies and guidance concerning how DOD would conduct the MCS and the databases and models used in the study. We identified generally accepted research standards that define a sound and complete quality study that were relevant to the MCS, and assessed the extent to which the MCS report met these standards. We interviewed study officials, study participants, and subject matter experts from several DOD entities, including the combatant commands and the military services. As we monitored the development of the MCS, we requested that DOD provide documentation supporting and verifying key analytical and decision-making processes.[Footnote 2] DOD officials could not produce this documentation during the development of the MCS or following issuance of the report. Consequently, we were unable to fully determine whether the analytical and decision-making processes that we believe are significant to the credibility of the study supported the MCS effort and its conclusions. Our scope and methodology are discussed in more detail in appendix I. We conducted our work from July 2004 through July 2006 in accordance with generally accepted government auditing standards. Results in Brief: DOD used an innovative approach in conducting the study and acknowledged some methodological limitations in its report, as any sound study should. However, it did not fully disclose how these limitations could affect the MCS conclusions and recommendations. Therefore, it is not transparent how the analyses done for the study support DOD's conclusions. As measured against relevant generally accepted research standards, we identified limitations in the MCS study and report that raise questions about their adequacy and completeness. Among our findings: * Aspects of modeling and data were inadequate in some areas because data were lacking and some of the models used could not simulate all relevant aspects of the missions. The report did not explain how these limitations could affect the study results or what the effect on the projected mobility capabilities might be. Relevant research standards require that models used are adequate for the intended purpose, represent a complete range of conditions, and that data used are properly generated and complete. For example, the MCS modeled hypothetical homeland defense missions rather than homeland defense demands derived from a well defined and approved concept of operations for homeland defense, because the specific details of the missions were still being determined, and DOD acknowledged that the data used may be incomplete. The MCS also was unable to model the flexible deterrent options/deployment order process to move units and equipment into theater due to lack of data, but the study assumed a robust use of this process, which in one scenario accounted for approximately 60 percent of the airlift prior to beginning combat operations.[Footnote 3] In addition, the MCS report contains more than 80 references to the need for improved modeling, and 12 of these references call for additional data or other refinements. Additionally, the MCS modeled the year 2012 to determine the transportation capabilities needed for the years 2007 through 2013. The year 2012 did not place as much demand for mobility assets in support of smaller military operations, such as peacekeeping, as other years. However, DOD officials considered 2012--the year modeled--as "most likely" to occur and stated that statistically it was not different from other years in the 2007 to 2013 period even though the number of smaller military operations is the least of any of the years reviewed. * While the MCS concluded that combined U.S. and host nation transportation assets were adequate to meet U.S. objectives with acceptable risk, the report, in describing the use of warfighting metrics in its analyses, does not provide a clear understanding of the direct relationship of warfighting objectives to transportation capabilities. Acknowledging this point, the report stated that further analysis is required to understand the operational impact of increased or decreased strategic lift on achieving warfighting objectives. Relevant generally accepted research standards require that conclusions be supported by analyses. The use of warfighting metrics is a measure to determine whether combat tasks, such as achieving air superiority, are achieved. However, they do not measure whether appropriate personnel, supplies, and equipment arrived in accordance with timelines. As a result, we could not determine how the study concluded that planned transportation assets were adequate because the study did not contain a transparent analysis to support its conclusion or a clear roadmap in the report to help decision makers understand what that conclusion meant in terms of type and number of mobility assets needed. Previous DOD mobility studies primarily used mobility metrics, which measured success in terms of tons of equipment and personnel moved per day to accomplish military objectives. The use of both warfighting and mobility metrics to measure success would allow decision makers to know whether combat tasks were achieved and how much strategic transportation is needed to accomplish those tasks. * In some cases, the MCS results were incomplete, unclear, or contingent on further study, making it difficult to identify findings and evaluate evidence. Relevant research standards require results to be presented in a complete, accurate, and relevant manner. For example, the report contains several recommendations for further studies and assessments, five of which are under way. However, DOD has no plans to report the effect of these studies on the MCS results after the studies are complete. In addition, the report contains qualified information that is not presented clearly, such as varying assessments of intratheater assets in three different places in the report. The lack of clarity and conciseness of the reported results can limit the study's usefulness to decision makers and stakeholders. * Verification, validation, and accreditation (VV&A) of models and data used to conduct the study was not complete because it was not done in accordance with DOD policy or relevant research standards. Moreover, relevant research standards state that a study report should include a VV&A accreditation report that is signed by the study director and addresses the models and data certification. DOD officials acknowledged that they did not comply with DOD VV&A policy when using legacy models in the MCS because they contended that long-term use of models and data constitutes an equivalent VV&A process. Other than a description of the process contained in the MCS report, DOD officials could provide no additional documentation to verify and validate this equivalent process to provide the assurance that models and data used in the MCS reduced the risk inherent in modeling and simulation and added to the credibility of the results. Moreover, officials could not provide documentation to support key analytical and decision-making processes used by senior DOD leadership, thus undermining the credibility of the reported study results. These limitations to the study's methodology raise questions concerning the accuracy of the study's finding that projected capabilities are adequate to achieve U.S. objectives with an acceptable level of risk. Until DOD conducts an adequate and complete future MCS and clearly discloses all limitations and their effects on the study results, decision makers may be unable to clearly understand the operational implications of the study results and make fully informed programmatic investment decisions concerning mobility capabilities. We are recommending that the Secretary of Defense, when conducting future mobility capabilities studies, beginning with any study currently under way, develop and use models and data for critical missions and processes that are verified, validated, and accredited as required; include in study reports an explanation of how stated limitations might impact the study results and, at a minimum, describe how recommended future studies might be conducted to enhance the results of the original study; and incorporate both mobility and warfighting metrics in determining capabilities. In commenting on a draft of this report, DOD concurred with the first and third recommendations and claimed that they did not understand the second. We have clarified that recommendation to the Secretary of Defense to include in study reports an explanation of how stated limitations might impact the study results and, at a minimum, describe how recommended future studies might be conducted to enhance the results of the original study. In its comments, DOD also stated that the report contained misleading information and factual errors. We disagree with DOD's assertion. We did modify our report to respond to a DOD technical comment related to homeland defense missions. DOD's comments and our evaluation of them are discussed in the agency comments section of this report. Background: The MCS was the first assessment of DOD's mobility system since 2000. The study was designed to identify changes in DOD's transportation force structure due to changes in threats and national security and military strategies. The MCS is the fourth in a series of major mobility studies that DOD has conducted since the end of the Cold War. The first study, the Mobility Requirements Study, conducted in 1992, was undertaken because of concern about the DOD's strategic mobility capabilities in the wake of Operation Desert Shield and Operation Desert Storm. That study established mobility requirements for the post- Cold War era; defined baseline requirements for intertheater, or strategic, mobility; and proposed a long-range investment plan to meet these requirements. The Mobility Requirements Study Bottom-Up Review Update, conducted in 1994, reaffirmed the need for increases in key mobility components and validated the prior study's recommendation for the procurement of additional ships for afloat prepositioning and for surge deployments of forces based in the continental United States. The Mobility Requirements Study--2005, issued in 2001, projected future mobility requirements based on two nearly simultaneous major regional contingencies. It included a broader range of factors, including host nation support and enemy use of weapons of mass destruction, than the previous studies. The current MCS, which began in May 2004, reassessed DOD's mobility capabilities against the backdrop of a revised National Military Strategy that included the ongoing war against violent extremism, an evolving global defense posture, a new force-sizing construct, revised campaign scenarios, and ongoing departmentwide transformation efforts. The study results were intended to support decisions on future strategic airlift, aerial refueling aircraft, and sealift procurements needed to meet varying military requirements. The study used an innovative "capabilities-based" approach, measuring existing and currently projected mobility capabilities against warfighting demands that could be expected in fiscal year 2012 while also considering mobility demands during the 7-year period from fiscal year 2007 through fiscal year 2013. According to DOD officials, the Secretary of Defense believed this approach would give him greater flexibility in deciding which capabilities to fund in a constrained budget environment. In considering each aspect of the National Military Strategy,[Footnote 4] the MCS modeled warfighting scenarios in the year 2012 using different premises with varying assumptions to develop and evaluate mobility capability mix alternatives. The models were used to evaluate transportation alternatives, including variations in alternative transportation modes (air, land, sea) and sources (military, civilian, foreign), as well as factors that affect transportation mode and source decisions. The scope of the MCS described the study as an assessment of the full range of transportation needs required to support (1) combat operations; (2) smaller military efforts, such as peacekeeping or overseas evacuation of American citizens; (3) homeland defense/civil support, such as disaster relief and antiterrorism response; and (4) other strategic missions, such as strategic nuclear and global strike missions. The study was coauthored by the Office of the Secretary of Defense, Office of the Director, Program Analysis and Evaluation (PA&E), and the Chairman, Joint Chiefs of Staff, Office of the Director of Logistics. Other DOD components involved in the study included the U.S. Transportation Command and its subordinate commands, the Office of the Under Secretary of Defense (Acquisition, Technology and Logistics), the combatant commanders, the military services, and others. The final report was signed on December 19, 2005, by the Deputy Secretary of Defense. Limitations in the MCS Study and Report Raise Questions about Adequacy and Completeness: As measured against relevant generally accepted research standards, limitations in the MCS study and report raise questions about their adequacy and completeness. For example, aspects of modeling and data were inadequate in some areas because data were lacking and some of the models used could not simulate all relevant aspects of the missions. Furthermore, the exclusive use of warfighting metrics in the MCS analyses limited the usefulness of the report. Moreover, in some cases the MCS results were incomplete, unclear, or contingent on further study, making it difficult to identify findings and evaluate evidence. Finally, verification, validation, and accreditation of models and data used to conduct the study were incomplete because they were not done in accordance with DOD policy or relevant research standards, and supporting documentation for key processes could not be provided. Aspects of Modeling and Data Were Inadequate: Aspects of modeling and data were inadequate in some areas because data were lacking and some of the models used could not simulate all relevant aspects of the missions. Relevant research standards require that models used are adequate for the intended purpose, represent a complete range of conditions, and that data used are properly generated and complete. As DOD acknowledged in the MCS report as a study limitation, some modeling tools were not available to analyze key missions. The MCS cited deficiencies in several existing mobility models and the need for follow-on MCS analyses. The MCS report also acknowledged that the identified deficiencies in data hindered analysis of future requirements and transportation system performance. However, the report did not explain how these limitations could affect the study results or what the effect on the projected mobility capabilities might be. For example, the MCS modeled hypothetical homeland defense missions rather than homeland defense demands derived from a well defined and approved concept of operations for homeland defense, because the specific details of the missions were still being determined, and DOD acknowledged that the data used are incomplete. The MCS report recommended further analysis of mobility capabilities after homeland defense needs are refined. However, the report did not identify the potential effect that using these hypothetical scenarios might have on the MCS results. The MCS also was unable to model the flexible deterrent options/deployment order process to move units and equipment into theater because of lack of data on how deployment orders have been issued in the past for major combat operations. However, the MCS assumed a robust use of the flexible deterrent option/deployment order process, which in one scenario accounted for approximately 60 percent of the early airlift movement prior to the beginning of combat operations. Instead, the MCS modeled the flow of forces and equipment contained in the time-phased force deployment data process.[Footnote 5] Based on the scenarios provided for the MCS analyses, we could not determine how the deployment order process would affect the mobility assets required for major combat operations. The MCS report noted that additional analysis is required to determine the implications of the deployment order process and to provide sufficient information for decision makers concerning the amount of future mobility assets actually needed. In addition to these modeling and data issues, the MCS report contains more than 80 references to the need for improved modeling, and 12 of these references call for additional data or other refinements, such as follow-on analyses, further assessments, future evaluations, additional study, and investigation of alternatives in a wide range of areas, such as antiterrorism response, infrastructure availability, intratheater airlift, strategic sealift, air refueling, and availability of civil reserve aircraft. Some of these studies are currently underway, as discussed later in this report. Moreover, our analysis of the MCS report showed that the year modeled (2012) to determine the DOD transportation capabilities needed for the years 2007 through 2013 did not place as much demand for mobility assets in support of smaller military operations, such as peacekeeping, as other years. To establish transportation demands for mission requirements, DOD developed and used a baseline security posture[Footnote 6] that covered a 7-year period. This baseline was developed, in part, using a historical analysis of DOD's movement of personnel, equipment, supplies, and other items. According to DOD officials, Office of the Under Secretary of Defense for Policy, which developed the baseline security posture, selected the year modeled in the MCS because it was deemed the "most likely" to occur in terms of transportation demands and because it was not statistically different from other years in the 7-year period. However, our analysis showed that 2012 involved the least demand for transportation assets in support of smaller military efforts than any year in the 7-year period and did not fully stress the defense transportation system. Figure 1 depicts the number of hypothetical ongoing contingencies for each year in the baseline as shown in the MCS. Figure 1: MCS Hypothetical Ongoing Contingencies during 7-year Baseline Security Posture Time Frame: [See PDF for image] Source: GAO's analysis of MCS notional lesser contingency schedule. Note: A particular contingency may be ongoing in more than 1 year. Each contingency has unique cargo and passenger requirements. For example, a contingency that may be ongoing over a 2-or 3-year time frame may require more or less mobility capability than a 6-month contingency. [End of figure] Although not transparent in the study, DOD officials said scenarios in the year modeled were not intended to fully stress the defense transportation system. DOD officials provided no further explanation for the year selected to develop the DOD transportation capabilities other than it was directed by Office of the Under Secretary of Defense for Policy and agreed to by the study leadership. We believe that selection of a different year that placed an increased demand on transportation assets for smaller military efforts may have revealed gaps in mobility requirements. Therefore, we found that the selection of 2012 as the modeling year was a limitation in the MCS with respect to smaller military efforts. Because of these modeling and data limitations, the MCS may have incorrectly estimated the future mobility requirements needed to support homeland defense missions, major combat operations, and smaller contingencies. Until DOD improves aspects of the modeling and data used to conduct the MCS--to include defining its homeland defense mission needs, developing models for the deployment order process, and explaining how identified modeling and data limitations could affect the study results--decision makers may not have adequate and complete information about DOD's mobility capabilities. Exclusive Use of Warfighting Metrics in MCS Analyses Limited Usefulness of Report: While the MCS concluded that combined U.S. and host nation transportation assets were adequate to meet U.S. objectives with acceptable risk, the report, in describing the use of warfighting metrics in its analyses, does not provide a clear understanding of the direct relationship of warfighting objectives to transportation capabilities. The report acknowledged that further analysis is required to understand the operational impact of increased or decreased strategic lift on achieving warfighting objectives. Relevant generally accepted research standards require that conclusions be supported by analyses. The use of warfighting metrics is a measure to determine whether combat tasks, such as establishing air superiority, are achieved. However, they do not measure success in terms of whether appropriate personnel, supplies, and equipment arrived in accordance with timelines. As a result, we could not determine how the study concluded that planned transportation assets were adequate because the study did not contain a transparent analysis to support its conclusion. In our opinion, it is important for decision makers to have an understanding of both combat tasks that must be achieved and the amount of transportation assets needed to achieve those tasks with some level of success. This understanding would allow creation of a clear roadmap for investment decisions. However, we could not determine how the study calculated the specific numbers of transportation assets needed or whether there are specific gaps, overlaps, or excesses in transportation assets, a key objective of the study. Previous DOD mobility studies, including the Mobility Requirements Study--2005, primarily used mobility metrics, which measured success in terms of tons of equipment and personnel moved per day to accomplish military objectives. Million-ton-miles per day is a commonly accepted measure of airlift performance and reflects how much cargo can be delivered over a given distance in a given period of time based on the capability of each type of aircraft. A similar mobility metric--short tons--is used to measure ship capability. However, these studies did not fully integrate combat tasks as a metric. The use of both warfighting and mobility metrics to measure success would allow decision makers to know whether there is sufficient capability to achieve warfighting objectives, as well as to understand the number, type, and mix of mobility assets that are actually needed. Results Are Not Always Complete or Presented Clearly and Are Qualified or Contingent on Further Study or Analysis: In some cases, the MCS results were incomplete, unclear, or contingent on further study, making it difficult to identify findings and evaluate evidence. Relevant research standards require results to be presented in a complete, accurate, and relevant manner; conclusions to be sound and complete; and recommendations to be supported by analyses. Our analysis of the MCS report found that it contains several recommendations for further studies and assessments, five of which are under way. The five studies address intratheater lift capabilities; sealift petroleum, oil, and lubricants; logistics contingency operations capabilities; aerial refueling; and integrated global presence and basing. However, the report does not explain the potential effect of these ongoing studies on the MCS results after the studies are complete, nor does DOD have plans to report the effect of these studies on the MCS results. In addition, the report contains qualified information that is not presented clearly in the report, such as varying assessments of intratheater assets in three different places. For example, the report states in the assessment section of the executive summary that projected transportation assets are sufficient to address intratheater demands in the fiscal years 2007 through 2013 time frame. However, in the recommendations section of the executive summary, the report states that DOD should take action to determine the proper mix of intratheater assets needed to meet requirements. Then, in the part of the report that discusses intratheater airlift, the report states that a detailed analysis of intratheater airlift needs would require improved modeling tools to accurately capture interactions among land, sealift, and airlift capabilities and that data sets must be developed that accurately describe the requirement in light of emerging concepts. VV&A of Models and Data Was Not Complete: VV&A of models and data used to conduct the study was not complete because it was not done in accordance with DOD policy or relevant research standards. DOD policy issued by the Under Secretary of Defense for Acquisition, Technology and Logistics requires that DOD models and data go through a VV&A process. Moreover, relevant research standards state that a study report should include a VV&A accreditation report that is signed by the study director and addresses the models and data certification. DOD officials acknowledged that they did not comply with the VV&A policy when using legacy models in the MCS because they believed such an approach was not warranted for legacy models that have been used for many years and have proved reliable. Moreover, these officials believe that such long-term use constitutes a VV&A process equivalent to that required in the policy. However, the DOD policy does not specify that the actual use of a model constitutes an equivalent VV&A process. VV&A of models and data reduces the risk inherent in the use of models and simulations by improving the credibility of modeling and simulation results. We previously reported our concerns that DOD did not follow its policy in executing the MCS and had little documentation to support the VV&A process used.[Footnote 7] We found that the final MCS report contained a description of the equivalent VV&A process, but DOD officials could provide no further documentation to verify and validate this equivalent process other than the description included in the report. We also found no documentation in the study report to support DOD claims that the models have proven reliable. Furthermore, DOD officials were unable to provide documentation to support and verify key analytical and decision-making processes used by senior DOD leadership throughout the study. Relevant research standards support documenting the study's analytical and decision-making processes to ensure the accuracy, completeness, and credibility of study results. DOD officials told us that the study's key analytical and decision-making processes were validated and approved by study participants during working group meetings and by senior leadership during General Officer Steering Committee meetings and Executive Committee meetings. PA&E officials could not produce documentation of these meetings because they said documentation did not exist. Nor could they produce other documents we requested during the development of the MCS or following issuance of the report. Consequently, we were unable to determine the adequacy and completeness of the analytical and decision-making processes that supported the MCS effort and that we believe are significant to the credibility of the study and its conclusions. Conclusions: The methodological limitations in the MCS that we identified--some of which were acknowledged by DOD in the MCS report--raise questions about the adequacy and completeness of the study and its report. Until DOD improves aspects of the modeling and data used to conduct the MCS--such as defining its homeland defense mission needs and developing models for the deployment order process--decision makers may not have adequate and complete information about DOD's mobility capabilities to enable them to make fully informed investment decisions. Furthermore, in the absence of an explanation of how identified modeling and data limitations could affect the study results or how such limitations could affect projected mobility capability requirements, the accuracy of the study's finding that projected capabilities are adequate to achieve U.S. objectives with an acceptable level of risk during the period from fiscal years 2007 through 2013 is unclear. Moreover, without a transparent comparison between existing mobility assets and projected needed assets, decision makers will be unable to use study results to identify and quantify the specific types and mix of mobility assets needed to address mobility gaps, overlaps, and excesses. Until DOD conducts an adequate and complete future MCS and clearly discloses all limitations and their effects on the study results, decision makers likely will not have full information concerning DOD's mobility capabilities. As a result, we suggest that Congress and other decision makers exercise caution in using the MCS to make programmatic investment decisions. Recommendation for Executive Action: To provide decision makers with adequate and complete information concerning mobility capabilities so they are able to clearly understand the operational implications of the study and make fully informed programmatic investment decisions, and to improve the usefulness of future mobility capabilities studies, we recommend that the Secretary of Defense take the following three actions, when conducting future mobility capabilities studies beginning with any study currently underway: * develop models and data for all critical missions, such as homeland defense, and processes, such as the flexible deterrent options/ deployment order process; * include in study reports an explanation of how stated limitations might impact the study results and, at a minimum, describe how recommended future studies might be conducted to enhance the results of the original study; and: * incorporate both mobility and warfighting metrics in determining capabilities. Agency Comments and Our Evaluation: In commenting on a draft of this report, DOD concurred with our first and third recommendations. DOD stated it did not understand our second recommendation that the Secretary of Defense, when conducting future mobility studies, beginning with any study currently underway, include in study reports an explanation of how ongoing and follow-on studies and modeling and data limitations that are referenced in the report could affect the reported results. DOD also noted that it plans to continue its ongoing efforts to enhance the models and data collection processes used to assess mobility capabilities across the full range of strategic missions and supports the notion that continual improvements are needed. As we noted throughout our report, the MCS report contains numerous and repeated references to the need for improved modeling and additional data or other refinements, such as follow-on analyses, further assessments, future evaluations, additional study, and investigation of alternatives in a wide range of areas. DOD further commented that while a completed study can recommend that follow-on studies be conducted, it cannot explain how future studies might affect the results of the current study. We acknowledge that DOD cannot quantitatively predict the outcome of an ongoing or future study. However, we believe DOD should be able to explain what ongoing follow- on studies or evaluations seek to determine, what changes are being made to the data inputs and modeling tools that are being used to conduct the studies, and how DOD expects the results may differ from current study results. While the explanation may be hypothetical, as are many operations research study hypotheses, it can provide decision makers with a better understanding of the current study's limitations and results and how an ongoing or future study's results may differ. Therefore, we refined our recommendation to recommend that the Secretary of Defense, when conducting future mobility studies, beginning with any study currently under way, include in study reports an explanation of how stated limitations might impact the study results and, at a minimum, describe how recommended future studies might be conducted to enhance the results of the original study. For example, if modeling and data are limitations to a study, the report should discuss the ways in which the results might change with better models and data. DOD also commented that our report contained misleading information and factual errors and that it stands by the adequacy and completeness of the MCS. The department provided examples in its technical comments where it believed our report contained misleading information and factual errors. We disagree with the department's comments regarding the facts in our report and have addressed each of the department's comments in appendix II. Lastly, DOD stated that the MCS and its conclusions are well accepted by the civilian and military leadership of the department, and pointed out that in March 2006 testimony before the House Armed Services Committee, the Commander, U.S. Transportation Command, stated that the planned strategic airlift fleet determined by the MCS is "about the right capacity". However we note that in the same hearing, the Commander also stated that he thought DOD needed "somewhere in the neighborhood of" 20 C-17 cargo aircraft beyond what is planned. We also note that in the Commander's April 2006 testimony before the Senate Armed Services Committee, Subcommittee on Sealift, he stated that, in an internal Focused Mobility Analysis to study strategic mobility from a Transportation Command perspective, the MCS will be the baseline, "but we will explore how changes in key assumptions may impact the analytical outcome." We are sending copies of this report to the Secretary of Defense; the Director of PA&E; and the Office of the Chairman, Joint Chiefs of Staff. We will also make copies available to others upon request. In addition, the report is available at no charge on the GAO Web site at [Hyperlink, http://www.gao.gov]. If you or your staff have any questions regarding the briefing or this report, please contact me at (202) 512-8365 or solisw@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made key contributions to this report are listed in appendix III. Signed by: William M. Solis: Director, Defense Capabilities and Management: List of Congressional Committees: The Honorable John Warner: Chairman: The Honorable Carl Levin: Ranking Minority Member: Committee on Armed Services: United States Senate: The Honorable Ted Stevens: Chairman: The Honorable Daniel K. Inouye: Ranking Minority Member: Subcommittee on Defense: Committee on Appropriations: United States Senate: The Honorable Duncan L. Hunter: Chairman: The Honorable Ike Skelton: Ranking Minority Member: Committee on Armed Services: House of Representatives: The Honorable C.W. Bill Young: Chairman: The Honorable John P. Murtha: Ranking Minority Member: Subcommittee on Defense: Committee on Appropriations: House of Representatives: [End of section] Appendix I: Scope and Methodology: To conduct our review of the Mobility Capabilities Study (MCS), we reviewed and analyzed the final MCS report; the MCS Terms of Reference; the MCS Study Plan; applicable Department of Defense (DOD) strategic planning guidance; as well as other DOD guidance, directives, instructions, and memos that describe how DOD would conduct its MCS. We also reviewed the National Security Strategy of the United States of America and the National Military Strategy of the United States of America; DOD guidance concerning data collection, development, and management in support of strategic analysis; DOD modeling and simulation instruction; Defense Modeling and Simulation Office guidance; descriptions of models used to conduct the study; and the databases used in the models. We interviewed study officials from the Office of the Secretary of Defense, Program Analysis and Evaluation (PA&E), and the office of the Chairman, Joint Chiefs of Staff, Logistics, as well as study participants and subject matter experts from the U.S. Transportation Command, Air Mobility Command, Surface Deployment and Distribution Command, the combatant commands, and the military services concerning the extent of their input to the study. We also interviewed officials from the Office of the Secretary of Defense, Acquisition, Technology and Logistics, and the Modeling and Simulation Technical Director at the Defense Modeling and Simulation Office. Additionally, we reviewed research literature and DOD guidance and identified frequently occurring, generally accepted research standards that are relevant for defense studies such as the MCS that define a quality or sound and complete study. The following were our sources for these standards: * GAO, Government Auditing Standards: 2003 Revision, GAO-03-673G (Washington, D.C.: June 2003); * GAO, Designing Evaluations, GAO/PEMD-10.1.4 (Washington, D.C.: March 1991); * GAO, Dimensions of Quality, GAO/QTM-94-1 (Washington, D.C.: February 2004); * RAND Corporation, RAND Standards for High-Quality Research and Analysis (Santa Monica, Calif.: June 2004); * Air Force Office of Aerospace Studies, Analysts Handbook: On Understanding the Nature of Analysis (January 2000); * Air Force, Office of Aerospace Studies, Air Force Analysis Handbook, A Guide for Performing Analysis Studies: For Analysis of Alternatives or Functional Solution Analysis (July 2004); * Department of Defense, DOD Modeling and Simulation (M&S) Verification, Validation, Accreditation (VV&A), Instruction 5000.61 (Washington, D.C.: May 2003); * Department of Defense, Data Collection, Development, and Management in Support of Strategic Analysis, Directive 8260.1 (Washington, D.C.: December 2, 2003); and: * Department of Defense, Implementation of Data Collection, Development, and Management for Strategic Analyses, Instruction 8260.2 (Washington, D.C.: January 21, 2003). During the process of identifying generally accepted research standards we noted that not all studies are conducted the same way. For example, while all studies use data, not all use baseline data. Likewise, all studies require analyses, but not all use models or simulation to conduct analyses. We tailored the research standards we identified as relevant to the MCS, as shown in table 1. Table 1: Generally Accepted Research Standards Relevant to MCS Requirements: Design: The Study is well designed: I; Design: The Study is well designed: Study plan, scope, and objectives follow Defense Planning Guidance. I.a; Design: The Study is well designed: (Do the study scope and objectives fully address the charter presented in the 2004 Defense Planning Guidance?). I.a.1; Design: The Study is well designed: Does the study plan address specified guidance?. I.b; Design: The Study is well designed: Was the study plan followed?. I.c; Design: The Study is well designed: Were deviations from the study plan explained and documented?. I.d; Design: The Study is well designed: Was the study plan updatedover the course of the study and the updates explicitly identified in the study and updated study plan?. II; Design: The Study is well designed: Assumptions and constraints are reasonable and consistent: II.a; Design: The Study is well designed: Are assumptions and constraints explicitly identified?. II.a.1; Design: The Study is well designed: (Are the study assumptions necessary and reasonable?). II.b; Design: The Study is well designed: Do the study assumptions support a sound analysis?. II.c; Design: The Study is well designed: Are the assumptions used in analyses common throughout the study and models?. II.d; Design: The Study is well designed: Do the assumptions contribute to an objective and balanced research effort?. III; Design: The Study is well designed: Scenarios and threats are reasonable: III.a; Design: The Study is well designed: Are scenarios traceable back to formal guidance?. III.b; Design: The Study is well designed: Were the threat scenarios validated and Defense Intelligence Agency approved and documented?. III.c; Design: The Study is well designed: Do scenarios represent a reasonably complete range of conditions?. III.d; Design: The Study is well designed: (Were the threats varied to allow for the conduct of sensitivity analysis?). Execution: The study is well executed. IV; Design: The Study is well designed: Methodology is successfully executed: IV.a; Design: The Study is well designed: Was the study methodology executed consistent with the (MCS) study plan and schedule?. IV.b; Design: The Study is well designed: (Does the methodology support accomplishing the objectives presented in the study plan?). IV.c; Design: The Study is well designed: Were the models used to support the analyses adequate for their intended purpose?. IV.d; Design: The Study is well designed: Were the model input data properly generated to support the methodology?. V; Design: The Study is well designed: (Analytical ) Baseline data and other data used to support study and analyses validated, verified, and approved: V.a; Design: The Study is well designed: Is the (analytical) baseline fully and completely identified and used consistently throughout the study for the various analyses?. V.b; Design: The Study is well designed: Were data limitations identified (and the impact of the limitations fully explained?). V.c; Design: The Study is well designed: Were the (baseline security posture) data verified and validated?. V.d; Design: The Study is well designed: Was the data verification and validation process documented?. VI; Design: The Study is well designed: Models, simulations, and verification, validation, and accreditation are reasonable: VI.a; Design: The Study is well designed: Was a VV&A accreditation report that addresses the models and data certification signed by the study director and included in the report?. VI.b; Design: The Study is well designed: Were modeling and simulation limitations identified and explained?. VI.c; Design: The Study is well designed: Has each model in the study been described?. VI.d; Design: The Study is well designed: Are the model processes clearly explained, documented and understood?. VII; Design: The Study is well designed: Measures of effectiveness (MOEs) and essential elements of analysis (EEAs) are addressed: VII.a; Design: The Study is well designed: (Do MOEs adhere to the guidance in the study terms of reference?). VII.b; Design: The Study is well designed: (Are the MOEs fully addressed in the study?). VII.c; Design: The Study is well designed: (Are the EEAs addressed in the study?). Presentation of results: Timely, complete, accurate, concise, and relevant to the client and stakeholders; VIII; Design: The Study is well designed: Presentation of results support findings: VIII.a; Design: The Study is well designed: Does the report address the objectives?. VIII.b; Design: The Study is well designed: Does the report present an assessment that is well documented and conclusions that are supported by the analyses?. VIII.c; Design: The Study is well designed: Are conclusions sound and complete?. VIII.d; Design: The Study is well designed: Are recommendations supported by analyses?. VIII.e; Design: The Study is well designed: Is a realistic range of options provided?. VIII.f; Design: The Study is well designed: Are the study results presented in the report in a clear manner?. VIII.g; Design: The Study is well designed: Are study participants/ stakeholders (i.e., services and Combatant Commands) informed of the study results and recommendations?. Source: GAO analysis of industry and DOD study and research standards. [End of table] We used these relevant standards as our criteria to assess the reported MCS results. All eight key areas of the study process were considered to have equal importance relative to the soundness and completeness of the study; that is, a sufficiently serious concern in any category could raise questions concerning the adequacy and completeness of the report. The analysts independently reviewed evidence relevant to each subquestion, including the study itself, the study Terms of Reference, and its strategic planning guidance. For each of the subquestions in the key study process areas, the analysts determined whether (1) the evidence had no limitations or raised no concerns, (2) the evidence had some limitations or raised some concerns, (3) the evidence had significant limitations or raised significant concerns, or (4) we could not determine the extent of limitations or concerns because there was not sufficient information. The analysts then met, compared, and discussed their individual assessments, and reached an overall assessment for each subquestion. Areas of the study where we identified either "some" or "significant" limitations or concerns were considered to affect the adequacy or completeness of the study. Additionally, areas of the study that could not be assessed because of the lack of supporting documentation were considered to affect the credibility of the study. Throughout our review PA&E officials told us that the documentation needed to support and verify the key analytical and decision-making processes used to conduct the MCS, documentation that was vetted and approved by DOD leadership and all of the study participants, would not be completed and available for our review until the study report was issued. However, after the report was issued, we were told that the report provides all of the supporting documentation needed and that the other documentation we requested could not be provided. As a result, we were unable to determine the adequacy and completeness of the analytical and decision-making processes that supported the MCS effort to evaluate the credibility of the study. We believe these processes are significant to the credibility of the study and its results. We conducted our review from July 2004 through July 2006 in accordance with generally accepted government auditing standards. [End of section] Appendix II: Comments from the Department of Defense: Deputy Secretary Of Defense: 1010 Defense Pentagon: Washington, DC 20301-1010: Aug 30 2006: The Honorable David M. Walker: Comptroller General of the United States: U.S. Government Accountability Office: 441 G Street, N.W. Washington, DC 20548: Dear Mr. Walker: This is the Department of Defense (DoD) response to the GAO draft report, GAO-06-938, "Defense Transportation: Study Limitations Raise Questions about the Adequacy and Completeness of the Mobility Capabilities Study (MCS) and Report," dated August 21, 2006 (GAO Code 350558). DoD's responses to the report's recommendations are enclosed. The GAO draft report contains misleading information and factual errors. The Department stands by the adequacy and completeness of the MCS, which is one of the most comprehensive capability studies ever conducted by the Department of Defense. The data and models used by the study are sound, and the results of the study are valid. Furthermore, in its groundbreaking assessment of homeland defense mission needs, the MCS accurately reflects the Department's most current understanding of this developing national mission. Lastly, MCS insights - drawn from the study's broad array of mobility metrics and detailed operational simulations - significantly enhanced senior leaders' deliberations during the 2005 Quadrennial Defense Review. The MCS was a collaborative effort led by the Office of the Secretary of Defense and the Joint Staff, and its conclusions are well accepted by the civilian and military leadership of the Department. In his March 2006 Congressional testimony, the Commander, U.S. Transportation Command acknowledged the study's finding by stating that the planned strategic airlift fleet of 292 aircraft "is about the right capacity." Likewise, during a press conference in November 2005, the Vice Chairman of the Joint Chiefs of Staff expressed his support for the study, citing the MCS finding that the Department has "a very capable and adequate airlift fleet." Signed: Enclosure: As stated: Unclassifed: GAO Draft Report - Dated August 21, 2006 GAO CODE 350558/GAO-06-938: "Defense Transportation: Study Limitations Raise Questions about the Adequacy and Completeness of the Mobility Capabilities Study and Report" Department Of Defense Comments To The Recommendations: Recommendation 1: The GAO recommended that the Secretary of Defense, when conducting future mobility capabilities studies, beginning with any study currently underway, develop models and data for all critical missions, such as homeland defense, and processes, such as the flexible deterrent options/deployment order process. DOD Response: DoD concurs. The Department plans to continue its ongoing efforts to enhance the models and data collection processes used to assess mobility capabilities across the full range of strategic missions. The models used in the MCS are sound and have consistently produced valid results in the four mobility studies conducted by DoD since the end of the Cold War. Furthermore, as recommended in the MCS report, DoD supports the notion that continual improvements are needed to provide enhanced analytic tools for the Department as it strives to address complex real-world processes. Recommendation 2: The GAO recommended that the Secretary of Defense, when conducting future mobility studies, beginning with any study currently underway, include in study reports an explanation of how ongoing and follow-on studies and modeling and data limitations that are referenced in the report could affect the reported results. DOD Response: Dod does not understand this recommendation. While a completed study can recommend that follow-on studies be conducted, it cannot explain how future studies might affect the results of the current study. Recommendation 3: The GAO recommended that the Secretary of Defense, when conducting future mobility capabilities studies, beginning with any study currently underway, incorporate both mobility and warfighting metrics in determining capabilities. DOD Response: DoD concurs. In fact, the MCS employed a rigorous process by which the Services, Combatant Commands, and the Joint Staff collaboratively developed warfighting metrics. These metrics were used along with a comprehensive set of mobility metrics for airlift, sealift, surface transport, and prepositioned equipment to determine the adequacy of the full spectrum of mobility capabilities. GAO Draft Report --Dated August 21, 2006 GAO CODE 350558/GAO-06-938: "Defense Transportation: Study Limitations Raise Questions about the Adequacy and Completeness of the Mobility Capabilities Study and Report" Department Of Defense Technical Comments: The following are three examples of factual errors and misleading information contained in the draft GAO report: 1. Page 3: "The [modeled] year of 2012 did not place as much demand for mobility assets in support of smaller military operations, such as peacekeeping, as other years." Response: False. The seven-year demand (2007-2013) developed as part of the Baseline Security Posture, and used by the MCS, does not have significant variance from year to year. 2012 demand is not significantly less than 2009, and is larger than 2013. It is important to note the fact that the MCS modeled a surge demand on the mobility system in 2012 that far exceeds anything this nation has experienced since World War II, In doing so, the study correctly applied the Department's strategic planning framework (1-4-2-1) and its guidance with respect to the conduct of concurrent lesser contingencies during overlapping war fights. GAO has incorrectly focused on the number of operations, not the level of effort. 2. Page 4: "The use of war fighting metrics is a measure to determine whether combat tasks, such as achieving air superiority, are achieved. However, they do not measure whether appropriate personnel, supplies, and equipment arrived in accordance with timelines." Response: False. The war fighting metrics developed by the MCS do measure whether appropriate personnel, supplies, and equipment arrived in accordance with timelines. In fact, the MCS employed a rigorous process to develop war fighting metrics that were used along with a comprehensive set of mobility metrics for airlift, sealift, and surface transport. These metrics were used to determine if the modeled war fights accomplished the commander's objectives within the right timelines. As we explained to GAO, being able to achieve a desired task within the desired timeline requires the appropriate personnel, supplies and equipment to be in place on time, indicating that the transportation capabilities are adequate. 3_ Page 3: "The MCS also was unable to model the flexible deterrent options/deployment order process . . Response: False. The MCS DID analyze flexible deterrent option (FDO) movements to the theater. As part of the MCS analysis, the study modeled the flow of forces in response to rising tensions in various regions of the world, Forces were flowed to the region in anticipation of operations before war plans were executed. These flexible deterrent options were included in the MCS analysis. What the study did not do was model the deployment order process used in OIF. Rather, the study used the time-phased force deployment data (TPFDD) process as the accepted methodology for flowing follow-on forces. 4. Page 3: "The MCS modeled hypothetical homeland defense missions rather than actual homeland defense demands because the specific details of the missions were still being determined, and DoD acknowledged that the data used may be incomplete." Response: The statement is misleading. We are not sure what the report means by "actual homeland defense demands." Actual demands are encountered in response to actual events. The MCS was intended to inform the DoD leadership concerning the impact of potential demand on the mobility system, and to assess the risks associated with different potential demand levels. The study used the latest approved homeland defense scenarios developed by NORTHCOM and PACOM, as well as the Department's current planning guidance, to determine the range of mobility assets needed to support a range of missions. Given the many unknowns associated with homeland defense, the study assessed low, moderate, and high levels of DoD support for these missions. 5. Page 3: "Aspects of modeling and data were inadequate." Response: This statement is misleading. These are the same models that the Department has relied on to complete three previous mobility studies since the end of the Cold War. Furthermore, the Department has dedicated significant resources to make sure that these models and the data collection processes are adequate. The fact that the study recommends future improvements was meant to focus future enhancements as the Department strives to address increasingly complex real world issues. The MCS models and data were adequate to assess relevant aspects of the missions required to support the National Military Strategy. GAO's Responses to DOD's Technical Comments: 1. DOD disagreed with our assessment that the modeled year--2012--did not place as much demand for mobility assets in support of smaller military operations, such as peacekeeping, as other years. DOD also stated that we incorrectly focused on the number of operations, not the level of effort. We disagree. The MCS report (Annex A to Appendix F) made no distinction between the number of lesser contingencies and the level of effort. Specifically, the Vignettes for Baseline Security Posture Analysis did not report the level of effort by year and instead aggregated the data, in many instances across several modeled years. Consequently, we compared the number of operations conducted in the model year. Throughout our review, PA&E officials consistently told us that the completed MCS report would contain all the documentation needed to support its analyses. Furthermore, although demand in the modeled year may exceed previous efforts, the MCS was chartered to assess the ability of the mobility system to support the National Military Strategy into the next decade. The size of the selected model year in relation to efforts conducted between 1941 and 2006 is not at issue. As our report makes clear, our concern is that modeling what appears to be the least demanding year does not address whether the United States has sufficient capability to support national objectives during a peak demand period and may underestimate and underreport demands to senior decision makers. 2. DOD disagreed with our observation that the MCS report does not provide a clear understanding of the direct relationship of warfighting objectives to transportation capabilities. We disagree. We understand that achieving a combat task requires delivering the right commodity to the right place at the right time. However, the specific combat tasks (e.g., attaining air superiority) necessary to satisfy the commander's campaign objectives are not a direct measure of mobility capability. For example, the problems in using a single metric are reflected in the MCS Appendix H, where the MCS report states that "the study itself still had difficulty in evaluating the operational impact of the delivery of theater support elements," adding that "we [DOD] were unable to develop a satisfactory mechanism to capture the linkage of the closely related, but delayed, follow-on support needed." Finally, the MCS concludes that "there was no way to model a decrease in [Air Force] squadron effectiveness if this support was late. Additional effort is required to develop a methodology for evaluation the operational impact of support equipment availability." We continue to believe, and DOD agreed with our recommendation, that warfighting metrics, in conjunction with mobility metrics, can give decision makers a full picture of the capabilities needed to meet a specific warfighting goal. 3. DOD disagreed with our statement that the MCS was unable to model the flexible deterrent options/deployment order process and that the study in fact analyzed flexible deterrent option (FDO) movements to theater. We do not dispute that DOD analyzed FDO movements as part of the MCS analysis. However, the degree to which the MCS analyses successfully modeled FDOs is in question. The MCS report, Appendix H, stated that an individual FDO is essentially the same as a deployment order. It also states in that section that "Deployment orders [DEPORDS] are issued to deploy specific capabilities as commitment decisions are made. This was not modeled due to lack of data on how DEPORDS would have been issued for an MCO [major combat operation] deployment." In the same paragraph, the MCS concludes that "the impact on the mobility system of the DEPORD process should be assessed in follow-on MCS analyses," adding that "there is a data deficit on how to model and execute a DEPORD process." Furthermore, the MCS report states that "additional analysis is required to investigate the implications of the DEPORD process decisions and provide data for future decision-makers to develop a DEPORD execution process." 4. DOD believes our statement concerning homeland defense missions is misleading and is not sure what the report means by "actual homeland defense demands." We removed the word "actual" and clarified our report to discuss "demands derived from a well defined and approved concept of operations for homeland defense", which were not available for the study according to the MCS report. Furthermore, in chapter 4, the MCS report states that "maintaining a dedicated capability to support multiple, nearly simultaneous homeland defense/civil support events concurrent with the peak demand period of two overlapping warfights, greatly exceeds programmed lift capabilities". This raises questions about the conclusions of the MCS that there are adequate mobility capabilities to meet national security objectives. Also, in Chapter 3 of the MCS report, it states that the DOD homeland defense concept of operations required refinement and was one of nine issue areas cited within the homeland defense portion of the study that "need to be addressed and investigated in more detail." All of these nine areas potentially impact mobility support for homeland defense operations. Moreover, the MCS Executive Summary notes that reassessment of these missions is required as DOD's role in homeland defense evolves. The MCS report, chapter 4, concludes by calling for further refinement of mission requirements, continuing risk assessments, and an effort to determine corresponding mobility solutions. We continue to believe that the MCS conclusion that adequate mobility capability exists is conditional given the results of the homeland defense portion of the study and that the accuracy and completeness of the data, modeling, and results for this portion of the MCS remain in question. 5. We disagree with DOD's characterization that our information was "misleading" regarding the adequacy of some aspects of the MCS' modeling and data. Furthermore, we continue to disagree with DOD's statement that the models and data used by the study were sound and adequate to assess relevant aspects of missions required to support the National Military Strategy, and that the results of the study are valid. In this report, as in our September 14, 2005 report,[Footnote 8] we reaffirm our concern that the data and models used by the study may not be sound and the results may not be valid since the verification, validation, and accreditation (VV&A) of the models and data used to conduct the study was not done in accordance with DOD policy or relevant research standards. VV&A of models and data reduces the risk inherent in the use of models and simulations by improving the credibility of modeling and simulation results. We do not dispute DOD's assertion that it has relied upon the same models to produce mobility studies done "since the end of the Cold War". However, as we discuss in our report, the MCS report fails to explain or qualify the impact that identified data or modeling limitations might have on its results. For example, in the MCS chapter 4, entitled Operational Data, the MCS states that "data deficiencies negatively affected MCS's ability to use current execution data to project future requirements and assess system performance." Unclear is the extent to which these deficiencies impacted the MCS ability to meet the objective of identifying mobility capability gaps, overlaps, or excesses and provide associated risk assessments, an MCS objective. Similarly, in the section of chapter 4 entitled Analysis Tools, the report states that "MCS analysis revealed several deficiencies in existing mobility models." The section concludes with five recommended tool enhancements but it does not explain the impact that the absence of these enhanced tools may have. We continue to believe that because of these modeling and data limitations, the MCS may have incorrectly estimated the future mobility requirements needed to support homeland defense missions, smaller contingencies, and major combat operations. [End of section] Appendix III: GAO Contact and Staff Acknowledgments: GAO Contact: William M. Solis, (202) 512-8365 or solisw@gao.gov: Acknowledgments: Ann Borseth, Assistant Director; Brian Lepore, Assistant Director; Nabajyoti Barkakati; Renee Brown; Claudia Dickey; Ron La Due Lake; Oscar Mardis; Deborah Owolabi; Kenneth Patton; and Stephen Woods made significant contributions to this report. FOOTNOTES [1] S. Rep. 108-260, at 126 (2004). [2] Examples of the documentation we requested to support and verify key analytical and decision-making processes used by DOD to conduct the MCS included (1) the accreditation report and supporting documentation or evidence of the verification, validation, and accreditation process for the models and data used in the MCS; (2) copies of MCS working group meeting minutes that verify and validate the analytical processes the various MCS study teams and study participants used to vet and agree upon data, scenarios, assumptions, models, and associated risk; and (3) copies of MCS General Officer Steering Committee and Executive Committee meeting minutes that verify and validate the analytical and decision-making processes the DOD senior leadership used to vet and agree upon the key data, scenarios, assumptions, models, and associated risk used to conduct the MCS, as well as agreement with the study results. [3] Deployment orders are issued to deploy specific capabilities as commitment decisions are made, rather than a deploying unit's full set of equipment or capabilities. Flexible Deterrent Options (FDOs) provide escalation options during the initial stages of a conflict. FDOs are employed under certain conditions to deter adversarial actions contrary to U.S. interests. [4] The 2004 National Military Strategy of the United States calls for a force sized to defend the homeland, proactively dissuade adversaries in and from four global regions, and conduct two overlapping "swift defeat" campaigns. Even when committed to a limited number of lesser contingencies, the force must be able to "win decisively" in one of the two campaigns. This "1-4-2-1" force-sizing construct places a premium on increasingly innovative and efficient methods to achieve objectives. [5] In joint military planning, time-phased force deployment data are defined as a computer database that contains detailed personnel and cargo planning data; it usually includes priority and sequencing of deploying forces. [6] The baseline security posture projects the position from which combatant commanders will perform future missions, including how they will address the global war on terrorism, ongoing operations, and other day-to-day activities to which U.S. forces remain committed and from which they are not likely to disengage entirely. [7] See GAO, Defense Transportation: Opportunities Exist to Enhance the Credibility of the Current and Future Mobility Capabilities Studies, GAO-05-659R (Washington, D.C.: Sept. 14, 2005), for a more detailed discussion. [8] See GAO, Defense Transportation: Opportunities Exist to Enhance the Credibility of the Current and Future Mobility Capabilities Studies, GAO-05-659R (Washington, D.C., September 14, 2005), for a more detailed discussion. GAO's Mission: The Government Accountability Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.