Military Education

DOD Needs to Develop Performance Goals and Metrics for Advanced Distributed Learning in Professional Military Education Gao ID: GAO-04-873 July 30, 2004

As part of its transformation to prepare the armed forces to meet current and future challenges, the Department of Defense (DOD) is expanding its use of advanced distributed learning (ADL) techniques in senior- and intermediate-level officer professional military education (PME). ADL instruction does not require an instructor's presence, and it facilitates the use of varied learning management systems. To date, the application of ADL has been targeted to nonresident students. To determine whether DOD uses a systematic process for evaluating the results of ADL application, GAO was asked to examine DOD's metrics for assessing program effectiveness, to compare DOD's criteria for converting courses to ADL with those of private-sector institutions, and to identify the challenges to ADL implementation.

DOD does not have specific performance goals and metrics with which to assess ADL effectiveness in PME. Furthermore, although GAO and private-sector organization have established frameworks for assessing the effectiveness of educational programs by focusing on metrics for learning outcomes--that is, the knowledge, skills, and abilities that students attain through learning activities--DOD's oversight focuses instead on educational inputs such as facilities, student to faculty ratios, and student body composition. Since ADL is still a new and evolving tool, systematic evaluative processes have not yet been required. Without clear goals and an effective process for evaluating the results of ADL application, DOD cannot ensure that its program is achieving an appropriate return on investment and other goals. The criteria for converting PME courses and curricula to ADL vary by school and by military service, are based on subjective choices as to which content is suited for online delivery, and are focused solely on nonresident programs. The private sector similarly lacks systematic criteria in its use of ADL. However, DOD's implementation of ADL programs for PME compares favorably with private-sector institutions. Cultural, technological, and resource challenges affect ADL implementation. For example, some military policies reflect a lower estimation of the value of nonresident PME, and many respondents to a survey of ADL students and alumni indicated that its quality and achievement of outcomes did not compare favorably, in their view, with those of resident education programs. The technological challenges of balancing computer access with network security, along with resource challenges of funding and increased burdens on limited administrative staff, are additional concerns.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-04-873, Military Education: DOD Needs to Develop Performance Goals and Metrics for Advanced Distributed Learning in Professional Military Education This is the accessible text file for GAO report number GAO-04-873 entitled 'Military Education: DOD Needs to Develop Performance Goals and Metrics for Advanced Distributed Learning in Professional Military Education' which was released on July 30, 2004. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to the Ranking Minority Member, Committee on Armed Services, House of Representatives: July 2004: MILITARY EDUCATION: DOD Needs to Develop Performance Goals and Metrics for Advanced Distributed Learning in Professional Military Education: GAO-04-873: GAO Highlights: Highlights of GAO-04-873, a report to the Ranking Minority Member, Committee on Armed Services, House of Representatives Why GAO Did This Study: As part of its transformation to prepare the armed forces to meet current and future challenges, the Department of Defense (DOD) is expanding its use of advanced distributed learning (ADL) techniques in senior- and intermediate-level officer professional military education (PME) (see table at right). ADL instruction does not require an instructor‘s presence, and it facilitates the use of varied learning management systems. To date, the application of ADL has been targeted to nonresident students. To determine whether DOD uses a systematic process for evaluating the results of ADL application, GAO was asked to examine DOD‘s metrics for assessing program effectiveness, to compare DOD‘s criteria for converting courses to ADL with those of private-sector institutions, and to identify the challenges to ADL implementation. What GAO Found: DOD does not have specific performance goals and metrics with which to assess ADL effectiveness in PME. Furthermore, although GAO and private-sector organization have established frameworks for assessing the effectiveness of educational programs by focusing on metrics for learning outcomes”that is, the knowledge, skills, and abilities that students attain through learning activities”DOD‘s oversight focuses instead on educational inputs such as facilities, student to faculty ratios, and student body composition. Since ADL is still a new and evolving tool, systematic evaluative processes have not yet been required. Without clear goals and an effective process for evaluating the results of ADL application, DOD cannot ensure that its program is achieving an appropriate return on investment and other goals. The criteria for converting PME courses and curricula to ADL vary by school and by military service, are based on subjective choices as to which content is suited for online delivery, and are focused solely on nonresident programs. The private sector similarly lacks systematic criteria in its use of ADL. However, DOD‘s implementation of ADL programs for PME compares favorably with private-sector institutions. Cultural, technological, and resource challenges affect ADL implementation. For example, some military policies reflect a lower estimation of the value of nonresident PME, and many respondents to a survey of ADL students and alumni indicated that its quality and achievement of outcomes did not compare favorably, in their view, with those of resident education programs. The technological challenges of balancing computer access with network security, along with resource challenges of funding and increased burdens on limited administrative staff, are additional concerns. [See PDF for figure] [End of figure] What GAO Recommends: GAO recommends that the Secretary of Defense promote (1) the development of specific performance effectiveness goals for ADL in PME schools and (2) the use of ADL technologies to provide and establish metrics for learning outcomes. DOD partially concurred with the first recommendation and fully concurred with the second. DOD supports the use of specific effectiveness goals for PME, but believes such goals are not appropriate for any specific delivery method. www.gao.gov/cgi-bin/getrpt?GAO-04-873. To view the full product, including the scope and methodology, click on the link above. For more information, contact Neal P. Curtin at (757) 552-8100 or curtinn@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: DOD Does Not Have Specific Metrics for Assessing Performance Goals or Learning Outcomes: ADL Conversion Varied by School and by Service Based on Subjective Assessments of Content Suitability: Cultural, Technological, and Resource Barriers and Challenges Affect ADL Implementation in PME Programs: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendixes: Appendix I: Scope and Methodology: Appendix II: Methodology for Our Survey of Nonresident PME Students and Graduates: The Study Population: Developing the Survey: The Sample Design and Administration: Sampling Error: Nonsampling Error and Data Quality: Appendix III: Survey Responses: Introduction: Frame of Reference: Appendix IV: ADL Applications and Additional Features of Nonresident Programs: U.S. Army War College: Naval War College: Air Command and Staff College: Joint Forces Staff College: Additional Features of Nonresident PME Programs: Appendix V: Comments from the Department of Defense: GAO's Comment: Tables: Table 1: Enrollment Statistics for Resident and Nonresident Students for Each Senior-and Intermediate-Level PME School for the Academic Year 2003-2004: Table 2: Disposition of Sample: Table 3: Levels-of-Learning Definitions: Abbreviations: ACSC: Air Command and Staff College: ADL: advanced distributed learning: DOD: Department of Defense: JFSC: Joint Forces Staff College: NWC: Naval War College: PME: professional military education: USAWC: U.S. Army War College: United States General Accounting Office: Washington, DC 20548: Letter July 30, 2004: The Honorable Ike Skelton: Ranking Minority Member: Committee on Armed Services: House of Representatives: Dear Mr. Skelton: As part of its transformation to prepare the armed forces to meet current and future challenges, the Department of Defense (DOD) is transforming the way it trains to favor more rapid and responsive deployment. A significant element of its training transformation strategy is the application of advanced distributed learning (ADL), a technique of instruction that does not require an instructor's presence; can use more than one form of media; and emphasizes the use of reusable content, networks, and learning management systems. DOD has been expanding its use of ADL in its program of professional military education (PME). PME provides military officers with a wide array of college-level academic courses in both resident and nonresident settings; to date, the application of ADL has been targeted to nonresident students. As a new tool, ADL is being examined to determine whether DOD is applying a systematic performance evaluation approach, particularly in light of the increased rate at which servicemembers are being deployed worldwide. Without clear goals and an effective process for evaluating the results of ADL application, DOD cannot ensure that its program is achieving an appropriate return on investment and other goals. We were asked to review DOD's use of ADL in senior-and intermediate- level officer PME, and specifically: 1. to examine the metrics DOD uses to assess the effectiveness of ADL in PME, 2. to determine what processes and criteria DOD uses to select the courses or curricula it converts to ADL and how these criteria compare with those of other institutions in meeting ADL objectives in nonresident education, and: 3. to identify what barriers and challenges exist for implementing ADL in PME. We also reviewed and assessed the policies and guidance of several DOD offices responsible for providing oversight for PME activities. These offices included the Under Secretary of Defense for Personnel and Readiness and the Joint Staff's Joint Education Branch. We also studied experience in the private education sector and in other parts of the government in measuring the effectiveness of education programs. In addition, we surveyed 437 current students and graduates of senior-and intermediate-level PME programs to obtain their perspectives on their PME experience. Appendixes I and II describe our scope and methodology in more detail. We conducted our review from March 2003 through June 2004 in accordance with generally accepted government auditing standards. Results in Brief: DOD does not have specific performance goals or metrics with which to assess the effectiveness of ADL in PME, and its oversight activities focus on educational inputs rather than on learning outcomes. While DOD believes ADL has had a positive impact, its views are based on anecdotal information; clear goals and an effective process for evaluating results of ADL implementation are absent. Although numerous organizations have roles in providing oversight of PME activities, with several specifically responsible for ensuring that PME meets general standards of accreditation, DOD accreditation activities, like those in the private sector, focus primarily on educational process inputs--for example, facilities or student to faculty ratios. But we and a private- sector organization have established guidelines and frameworks for assessing the effectiveness of educational programs that stress a focus on measurable outcomes--that is, the knowledge, skills, and abilities a student acquires from a course. Furthermore, ADL has a unique ability to capture, retain, store, and document interactions in an online environment, thus providing the opportunity to demonstrate student skill improvements and to customize performance metrics. However, we found no evidence to indicate that DOD is using this ability. The processes for converting PME courses and curricula to ADL vary by school and by military service, and they feature a mixture of in-house and contractor approaches. PME schools generally focus their ADL applications on nonresident education programs, and they tend to convert an entire curriculum as a package rather than in a modular, course-by-course manner. No systematic criteria inform PME schools' decisions about which courses or curricula to convert to ADL. Instead, schools make individual, subjective choices as to which content is best suited for online rather than another delivery method. Notably, we found that nonmilitary educational institutions also lack systematic criteria when converting courses or curricula to ADL. DOD's approaches are in fact consistent with mainstream practice, and in some cases, compare favorably with the best implementations. Numerous cultural, technological, and resource challenges affect ADL implementation in PME programs, some which may affect ADL expansion or maintenance. Cultural issues include concerns by PME school officials about ADL's acceptance as an appropriate learning method and the appropriate extent of its use for nonresident education. In our survey, nonresident students expressed concerns about the quality of their courses, regardless of nonresident delivery method, as compared with those taken in residence. Technological challenges, particularly those concerning the optimal balance between student access (computer availability and freedom of information) and network security (protection of sensitive information and use of military installation firewalls), remain to be addressed. With respect to resources, there are concerns about ADL's ability to compete for limited funding and about the potentially burdensome administrative impact on nonresident program staff. To better assess the effectiveness of ADL in professional military education, we recommend that DOD promote (1) the development of specific performance effectiveness goals for ADL in PME schools and (2) the use of ADL technologies to provide and establish metrics for learning outcomes. In commenting on a draft of this report, DOD partially concurred with our first recommendation and fully concurred with the second. DOD supports the use of specific effectiveness goals for PME, but believes such goals are not appropriate for any specific delivery method. DOD stated that current accreditation practices are already promoting the data collection capabilities of ADL technologies for assessing multiple delivery methods. Background: Each military service has separate PME schools for senior-and intermediate-level officers. As defined by the Joint Staff's Officer Professional Military Education Policy,[Footnote 1] the senior-level schools, typically for O-5 and O-6 ranked officers, focus on warfighting within the context of strategy. The intermediate-level schools, typically for O-4 ranked officers, focus on warfighting within the context of operations.[Footnote 2] (See table 1 for a list of PME schools and enrollment totals.) The senior-and intermediate-level PME schools are not alike in terms of program offerings for resident and nonresident students. As indicated in table 1, while all senior-level PME schools offer resident programs, only the Army War College and the Air War College have analogous nonresident programs. Also as indicated in table 1, all intermediate- level PME schools offer resident and nonresident programs. DOD has approximately 39,318 students enrolled in its senior-and intermediate-level PME schools. The vast majority of these enrollees are nonresident students. Of the total enrolled, approximately 3,788, or 10 percent, are taking course work as resident students; the rest, or 90 percent, are nonresident enrollees. Table 1: Enrollment Statistics for Resident and Nonresident Students for Each Senior-and Intermediate-Level PME School for the Academic Year 2003-2004: PME institutions: Joint Senior-Level Schools: National War College; Resident students: 200; Nonresident students: N/A. PME institutions: Joint Senior-Level Schools: Industrial College of the Armed Forces; Resident students: 309; Nonresident students: N/A. PME institutions: Joint Combined-Level School: Joint Forces Staff College; Resident students: 229; Nonresident students: N/A. PME institutions: Senior-Level Schools: Air War College; Resident students: 265; Nonresident students: 6,100. PME institutions: Senior-Level Schools: Army War College; Resident students: 340; Nonresident students: 654. PME institutions: Senior-Level Schools: College of Naval Warfare-Naval War College; Resident students: 209; Nonresident students: N/A. PME institutions: Senior-Level Schools: Marine Corps War College; Resident students: 16; Nonresident students: N/A. PME institutions: Intermediate-Level Schools: Air Command and Staff College; Resident students: 587; Nonresident students: 12,069. PME institutions: Intermediate-Level Schools: Army Command and General Staff College; Resident students: 1,183; Nonresident students: 10,000[A]. PME institutions: Intermediate-Level Schools: College of Naval Command and Staff-Naval War College; Resident students: 256; Nonresident students: 1,799[ B]. PME institutions: Intermediate-Level Schools: Marine Corps Command and Staff College; Resident students: 194; Nonresident students: [C]. PME institutions: Intermediate-Level Schools: Marine Corps College of Continuing Education; Resident students: [C]; Nonresident students: 4,908. Total; Resident students: 3,788; Nonresident students: 35,530. Source: DOD. Note: N/A-School without a nonresident component. [A] According to Army Command and General Staff College officials, the nonresident student total fluctuates and could be plus or minus 2,000 students on any given day. [B] Naval War College's nonresident programs are offered at the intermediate level (equivalent to the Naval Command and Staff College) through its College of Distance Education. [C] The Marine Corps College of Continuing Education is the nonresident component of the Marine Corps Command and Staff College. [End of table] Nonresident PME exists to provide PME to a larger population than can be supported in resident institutions. Nonresident PME provides alternative learning-style options for officers not selected for residence or unable to participate in residence due to operational commitments. The military services have had nonresident PME programs for many years. The Naval War College (NWC) has had a department for correspondence courses since 1914. The U.S. Army War College (USAWC) has provided a nonresident course offering since 1968. The Air Force's nonresident programs were created in 1947 for its senior-level PME school and 1948 for its intermediate-level PME school. Paper-based correspondence is the traditional nonresident PME delivery mode. Students complete correspondence courses individually with limited faculty contact. Course materials and submissions are exchanged between students and faculty primarily by mail. PME schools have implemented other delivery modes, including seminars conducted at remote sites by PME faculty and CD-ROM distribution. Increasingly, PME schools are using ADL[Footnote 3] techniques in their nonresident program offerings. Several ADL applications are currently in use at senior-and intermediate-level PME schools, and all of them are focused on nonresident programs. They are offered at the U.S. Army War College, the Naval War College, and the Air Command and Staff College (ACSC). A planned ADL offering for reserve component staff is under development at the Joint Forces Staff College. See appendix IV for details on these programs. The addition of an ADL application for the Army Command and General Staff College nonresident PME course is anticipated for fiscal year 2005. DOD Does Not Have Specific Metrics for Assessing Performance Goals or Learning Outcomes: DOD does not have specific performance goals and metrics to assess the effectiveness of ADL in PME. While DOD believes ADL has had a positive impact, its views are based on anecdotal information, rather than a systematic performance measurement. Thus, DOD cannot determine whether ADL is meeting performance goals in comparison to other delivery methods. Although numerous organizations are providing oversight of PME activities, with several specifically responsible for ensuring that PME meets general standards of accreditation, these organizations do not focus on student learning outcomes--that is, the knowledge, skills, and abilities a student acquires from a course. Instead, DOD accreditation activities, like those in the private sector, focus primarily on educational process inputs, such as quality of facilities and student faculty ratios. We and a private-sector organization have recently established guidelines and frameworks for assessing the effectiveness of educational programs that stress a focus on measurable outcomes. ADL is a new and evolving tool for which systematic evaluation requirements have not been established. ADL has a unique ability to capture, retain, store, and document interactions in an online environment, which provides the opportunity to demonstrate student skill improvements, and thus to customize performance metrics. However, we have found no evidence to indicate that DOD is utilizing this ability. Numerous Organizations Have Roles in Providing Oversight of PME Activities: Numerous oversight organizations review PME activities, with several organizations specifically designed to ensure that PME conforms to general standards of accreditation. The preeminent mechanism for oversight is the Joint Chiefs of Staff's Process for Accreditation of Joint Education. The process is designed to provide oversight and assessment of PME institutions for purposes of strengthening and sustaining Joint Professional Military Education.[Footnote 4] It is a peer-review process involving a self-study component and a team assessment. The review sequence includes certification, accreditation, and reaffirmation of accreditation status. Accreditation can currently be granted for up to 5 years, and all PME programs with current ADL applications are Joint Staff-accredited. The Joint Staff also sponsors the Military Education Coordinating Council, an advisory body composed of high-ranking PME leadership. The purpose of the Council is to address key issues of interest for joint military education, to promote cooperation and collaboration among member institutions, and to coordinate joint education initiatives. The military services have responsibility for the service PME institutions in terms of managing PME content and quality and conducting all levels within the guidelines of the military educational framework. Consistent with Title 10 of the United States Code, the Secretary of Defense requires that each PME institution periodically review and revise curriculum to strengthen focus on joint matters and on preparing officers for joint duty assignments. PME is also reviewed by other internal and external organizations. Each PME institution has a Board of Visitors/Advisors that provides guidance over PME activities. The Board of Visitors/Advisors is composed of military and/or civilian academic officials who are nominated by PME schools and appointed by service secretaries to provide advice on educational and institutional issues. Service PME institutions have other internal and external advisory committees that perform activities such as providing advice, communicating feedback from major commands, and conducting curriculum review. Service Inspector General offices have conducted periodic reports and assessments on PME schools. The military services' education and training commands also provide oversight of PME activities, though not day-to-day administration. Additionally, private-sector regional accreditation agencies assess senior-and intermediate-level PME programs. Their accrediting activities generally guide the Joint Staff's review process. Performance-Effectiveness Metrics for ADL Implementation Are Lacking: PME schools have not established, and oversight organizations have not reviewed, specific goals or metrics of performance effectiveness for ADL implementation. As was stated in our recently issued guide for establishing a framework for assessing training and development efforts in the federal government, "it is increasingly important for agencies to be able to evaluate training and development programs and demonstrate how these efforts help develop employees and improve the agencies' performance."[Footnote 5] The Sloan Consortium--a private- sector organization that maintains a repository of information on distance education--views metrics as crucial for assessing program effectiveness. For example, metrics can (1) demonstrate that the "learning effectiveness" of nonresident education is at least as good as that of its resident counterpart, (2) identify cost comparisons that can be used to develop better strategic plans, and (3) provide information on student retention and completion rates. As was stated in our report on oversight for the military academies, such elements embody the principles of effective management, in which achievements are tracked in comparison with plans, goals, and objectives, and the differences between actual performance and planned results are analyzed.[Footnote 6] PME schools identified advantages of ADL over other means of delivery, but the advantages appeared to be anecdotally derived. PME school officials stated that ADL has resulted in quality improvements in PME delivery, especially when compared with paper-based correspondence. These advantages include (1) better facilitation of student and faculty interaction; (2) increased flexibility in modifying course material; (3) reductions in time required to complete programs; (4) better leveraging of resources for administrative support; and (5) establishment of learning management systems that monitor student progress and produce management reports. But there were no indications that evidence for these advantages were based on an evaluative effort to compare differences between ADL and paper-based correspondence courses. Since PME schools have not detailed a comprehensive process for evaluating ADL benefits over paper-based correspondence, it cannot be determined whether ADL is meeting performance goals based on appropriate returns on investment, student retention, student access to courses, or other goals that schools use to measure program effectiveness. Additionally, we did not observe any oversight agency focus on specific metrics of ADL effectiveness. According to Joint Staff officials, they perform reviews of nonresident programs as part of their accreditation activities. However, their reports focus on the nonresident program as a whole and not on particular methods of delivery. ADL is a new and evolving tool, and a systematic assessment of these applications has not yet been required. The three regional accreditation agencies that review PME schools with ADL implementations show variances in nonresident program evaluation policy.[Footnote 7] One agency stated that nonresident programs are not separately evaluated, although the programs may be included within the scope of the institution's existing accreditation. Another agency stated that additional procedures must be performed before nonresident programs are included within the scope of the institution's accreditation. The third agency required schools to evaluate its nonresident programs to ensure comparability to resident programs. In addition, we have not observed any Office of the Secretary of Defense or Board of Visitors/Advisors reviews in relation to ADL effectiveness for nonresident PME. While we did not observe measures of effectiveness specifically geared toward ADL applications, PME schools with ADL applications did perform program effectiveness assessments for nonresident education by the way of student satisfaction assessments as part of the Joint Staff accreditation process. These assessments used in-course student surveys, graduate surveys, and supervisory surveys to obtain feedback as part of a systematic approach to instructional design and to update and improve curriculum offerings. * USAWC performs surveys of students, alumni, and general officers with USAWC graduates in their commands. Students are surveyed for each course regarding particular aspects of the course and general degrees of satisfaction. A survey of alumni is conducted every 2 years. A general officer survey, designed to assess general officer impressions of alumni, will now be conducted annually instead of every 3 years, as in the past. Prior feedback from general officer surveys reported that the curriculum should emphasize application of strategic thinking to national security issues. USAWC also performs internal course evaluations as part of its curriculum assessment process. USAWC faculty members are required to undergo training to provide a degree of standardization in instruction and evaluation. This standardization, especially for evaluation, is more stringent for nonresident education. USAWC can conduct trend analyses for student performance and student satisfaction to determine statistical significances. * NWC uses student and alumni surveys to assess the academic program's effectiveness. Depending on the department, student assessments include daily sessions critiques, lecture critiques, end-of-course critiques, major exercise critiques, and exam critiques. Alumni are sent questionnaires 2 years after graduation asking for feedback on their educational experience. All academic departments conduct an extensive analysis of various student surveys to determine areas of the curriculum that are not meeting student needs so that these areas can be improved. Surveys are based on standards promulgated by accrediting agencies and external organizations to help objectively measure institutional excellence. Resident and nonresident student programs are measured the same since a single faculty is responsible for both. Peer evaluation of faculty members is used to sustain teaching method quality. * ACSC uses internal and external evaluations at all phases of its curriculum development process. It conducts end-of-course surveys that focus on delivery and educational support and end-of-year surveys for students to provide feedback about whether they believed the school (1) prepared them to lead commands, (2) accomplished its mission, (3) was institutionally effective, and (4) was beneficial to professional development. Surveys are also given to graduates and graduate supervisors to obtain perspectives on whether the school (1) accomplished its mission and institutional effectiveness; (2) enhanced graduates' ability to think operationally and critically; (3) prepared graduates to assume leadership duties, and (4) made the experience valuable in professional development. Metrics for Learning Outcomes Are Lacking: Student learning outcomes, as stated by the Council for Higher Education Accreditation--a national association representing accrediting organizations--are "properly defined in terms of the knowledge, skills, and abilities, that a student has attained at the end (or as a result) of his or her engagement in a particular set of higher education experiences."[Footnote 8] PME schools generally are not assessed for student learning outcomes as a means of determining program effectiveness. The Joint Staff's accreditation organization responsible for assessing PME schools has primarily focused on inputs to the educational process. As detailed in its policy, its educational standard assessment and self-study requirements focus on internal aspects such as organizational structure, facilities, curricula, student to faculty ratios, student body composition/mix, and faculty qualifications. However, as stated in our recently published guide for assessing training and development programs, the focus on evaluating activities and processes takes away from evaluating training and development's contribution to improved performance, reduced costs, or greater capacity to meet new and emerging transformation challenges.[Footnote 9] The Joint Staff has identified the usefulness of student learning outcomes and is currently in the process of developing student learning outcomes for PME and procedures to include them in the accreditation process. Our recently published report on distance education states that there is increased interest in using outcomes more extensively as a means of ensuring quality in all forms of education, including nonresident education.[Footnote 10] The Council for Higher Education Accreditation has issued guidelines on nonresident education and campus-based programs that call for greater attention to student learning outcomes, and the congressionally appointed Web-based Education Commission[Footnote 11] has also called for greater attention to student outcomes. The Commission said that a primary concern related to program accreditation is that "quality assurance has too often measured educational inputs (e.g., number of books in the library, etc.) rather than student outcomes." Private-sector educational institutions are just beginning to emphasize the evaluation of learning outcomes as a viable measure of program effectiveness. For example, the University of Maryland University College, a school with a comparably large distance education program and which serves a large number of military personnel, is piloting a project to identify and measure learning outcomes in five general areas--writing efficiency and oral communication, technology fluency, information literacy, quantitative literacy, and scientific literacy. The university will use knowledge captured by its distance education database to serve as a basis for this determination. Accrediting agencies and our recent report on training and development program assessments are also emphasizing the evaluation of learning outcomes as a measure of program effectiveness. Some of the regional agencies that accredit programs at the senior-and intermediate-level PME schools generally recognize the importance of student learning outcomes and have instituted practices that reflect some aspects of a systematic, outcome-based approach called for in [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-279].[Footnote 12] However, these agencies vary in the extent to which standards and policies address student learning outcomes for distance education. Our training and development assessment guide states that agencies need credible information on how training and development programs affect organizational performance, and that decision makers will likely want to compare the performance of these programs with that of other programs. Furthermore, programs lacking outcome metrics will be unable to demonstrate how they contribute to results. We surveyed nonresident PME current students and graduates to obtain their perspectives on the achievement of PME learning objectives and PME's impact on their career objectives. (See appendix III for presentation of survey results.) Because we only surveyed nonresident students, we could not compare the results with those of resident students. However, we believe the data can be useful for DOD to consider in its continuing study of program effectiveness. ADL Can Be Used to Capture Valuable Measurement Data: ADL has a unique ability to capture, retain, store, and document interactions in an online environment, which provides the opportunity to demonstrate student skill improvements, and thus to customize performance metrics. Since work is done on a computer, various data points are automatically collected as a student works, including the time spent, specific pages of the text visited, use of online help, and communication with others. University of Maryland University College officials pointed out ADL's unique ability when compared with other delivery methods to retain, capture, store, and document baseline data that can be used as the basis for performance metrics. These officials said they would use such data in designing performance measures for learning outcomes. However, we found no evidence to indicate that DOD is using this ability. DOD may be missing an opportunity to enhance its ability to measure effectiveness. ADL Conversion Varied by School and by Service Based on Subjective Assessments of Content Suitability: The processes for converting PME courses to ADL varied by school and by military service, and they feature a mixture of in-house and contractor approaches. PME schools generally focus their ADL applications on nonresident programs to improve their efficiency and effectiveness. In most cases, conversion decisions were made in collaboration with school academic boards. PME schools did not identify any systematic criteria as the basis for their decisions as to which courses or curricula to convert to ADL. They subjectively focused ADL conversion on the suitability of content for Web-based applications. Curriculum conversions were made because of a DOD-wide need (1) to improve access to a diverse officer corps and (2) to increase the efficiency of educational delivery. Since nonmilitary educational institutions also lack systematic criteria for converting courses or curricula to ADL for nonresident education, DOD's approaches are in fact consistent with mainstream practice, and in some cases, compare favorably with best practices in nonmilitary education. DOD Processes and Criteria for ADL Conversion Varied: The processes for converting PME courses to ADL varied by school and by military service, and they feature a mixture of in-house and contractor approaches. However, the conversions were focused on the schools' entire nonresident programs. USAWC's and ACSC's ADL applications were developed and managed by in-house staff and faculty. USAWC and ACSC used staff consisting of instructional designers and courseware developers interacting with respective faculty to develop courses. NWC's ADL application combined the use of contractor and in-house support. Contractor staff created Web-based applications for two of the three courses in NWC's curriculum. NWC officials learned enough from contractor efforts to create a Web-based application for the remaining course with in-house staff. In all cases, the ADL applications were applied to affect the entire nonresident curriculum and, in most cases, were preceded by reviews and final decisions from the schools' academic boards. The PME schools did not identify any systematic criteria that inform their selection of course for conversion to ADL. Rather, they made subjective decisions as to the appropriate parts of the curriculum that should be delivered online based on content suitability. While USAWC's application is delivered mostly through the Internet, print media delivers a portion of the course content as well. USAWC's application also includes two 2-week resident components in which students are brought together to achieve learning objectives best suited to resident instruction. These objectives include verbal communication, interaction in live settings, interpersonal skills used in direct relationships, and other skills that are important components of the resident experience. USAWC's application also uses asynchronous "threaded discussions," in which faculty members initiate online discussions with students on various academic topics. NWC's and ACSC's applications, which do not include resident components, blend content delivery using print media, CD-ROM, and the Internet. In NWC's application, print media is used for material that is mostly text and requires limited interactive capabilities; CD-ROMs are delivered to students for material that is not routinely updated; and students are assigned to cohort teams that allow online interactive opportunities and group discussions. In ACSC's application, almost all nonresident materials are provided to students on CD-ROMs as well as on the Internet to allow as much flexibility as possible to complete nonresident courses. Generally, PME school officials stated that ADL conversions were made because of nonresident PME's need (1) to respond to multiple learning styles in a diverse officer corps, (2) to increase the efficiency of educational delivery, and (3) to improve the quality of the educational offering. Additionally, NWC saw ADL as a means of potentially affecting retention and increasing enrollment. USAWC saw ADL as a means of responding to student demands for a more efficient and relevant educational experience. ACSC saw ADL as an improved means for delivering course material to increasing numbers of deployed officers. Nonmilitary ADL Conversion Decisions Were Similar: In nonmilitary applications of ADL, we observed processes and criteria for conversion decisions that were similar to DOD's. Course conversions followed analogous processes and criteria decisions, driven by factors such as interest and student enrollment that were not systematic. University of Maryland University College officials stated that their conversion process includes a dedicated staff of instructional designers and subject matter experts (usually faculty) who produce course conversion to distance learning content within an established framework to ensure maintenance of standards. Their criteria for conversion focus on high-student demand and high levels of interest and on course work that required less "hands-on" training, such as business and information technology courses. Further research of private-sector practices supports the observation that the lack of systematic criteria is consistent with mainstream practice in ADL adoption for nonresident education. DOD's approaches for course conversion are thus consistent with mainstream practice, and in some cases, compare favorably with the best practices in nonmilitary education. For example, NWC's ADL application received the Crystal Award in 2002 from the Association for Educational Communications and Technology based on "innovative and creative use of the medium, instructional value and relevance, soundness of instructional strategy, quality of production, and evidence of successful outcomes." As of June 2004, USAWC's nonresident education program is fully accredited by the Middle States Commission of Higher Education for the awarding of Master's of Strategic Studies Degrees. USAWC is the first military institution to achieve degree-granting authority for its nonresident ADL-based program. Cultural, Technological, and Resource Barriers and Challenges Affect ADL Implementation in PME Programs: PME schools identified a number of cultural, technological, and resource challenges that affect ADL implementation and may affect future maintenance or expansion of ADL efforts. Cultural issues such as appropriate extent of ADL incorporation, general perceptions about nonresident education, and limited ADL research in military education affect the degree of ADL implementation. Technological trade-offs and nonresident program resourcing also affect continued ADL efforts. Cultural: PME officials question the appropriate extent to which ADL should be used in nonresident education and how closely it can, or should, enable nonresident education to approximate resident education. It is generally recognized that resident programs are better in providing acculturation,[Footnote 13] interactive skills, and simulations that are critical for professional development of officers, and that there are challenges in providing such aspects in nonresident education. But some observers believe that nonresident education should not be compared to resident education and that nonresident education represents a vital broadening experience in its own right. In addition, there are indications that an ADL approach could significantly enrich nonresident content by excluding teaching methods that do not work in residence, allowing students the flexibility to focus on material that requires further study without class disruption, and serving as the basis for applications that can be used to upgrade teaching methods in resident programs. ADL implementation could also be affected by certain negative perceptions concerning nonresident education that are held by students, and in some cases, reflected in policy. Our survey of nonresident PME current students and recent graduates indicated that about 49 percent of current students and 48 percent of graduates believe they are not as well prepared as are their resident student counterparts, regardless of nonresident delivery method.[Footnote 14] While not universal between the services, we observed instances of military education policies that reinforce the perception that nonresident education is not as desirable as resident education. The Air Force's PME Instruction AFI36-2301 states that "ideally, all officers will attend PME in residence," and that limited resources restrict resident attendance to the "best qualified." Furthermore, "completing nonresident PME programs will not affect eligibility for resident PME programs." Indeed, we were told of instances where officers who after completing the nonresident PME program subsequently enrolled in the corresponding senior-or intermediate-level course in residence. The extent of ADL implementation in nonresident education is affected by the role that PME completion plays in promotional consideration. Programs maintained to foster promotional consideration--that is, "personnel-oriented"--might not be compatible with programs emphasizing the learning outcomes brought about by ADL--that is, "education-oriented." Our survey shows that promotional considerations, rather than learning outcomes, are the focus for students in nonresident education. An estimated 73 percent of current students and 84 percent of recent graduates listed as their predominant reason for participating in nonresident PME a desire to improve their promotional chances or to meet promotional requirements. In addition, of an estimated 22 percent of PME graduates who were promoted to a higher rank after PME completion, an estimated 88 percent stated that PME contributed to a "great" or "very great extent" in their promotions.[Footnote 15] But ADL adoption goals should focus more on learning outcomes than on simply "checking the boxes" for promotional opportunity enhancement. PME officials state that there are concerns that ADL advantages could be oversold to the point that ADL may be used to supersede resident programs and shift the burden of PME administration. DOD officials, already viewing ADL delivery methods as attractive from a cost savings perspective, are desirous to expand such programs, even at the expense of resident programs. However, PME officials believe that ADL expansion should be considered only after completely understanding its impact on military operations and recognizing the resident program's role as the basis for a nonresident program. In addition, PME officials noted that ADL could be used as a means of shifting the burden from the school to the student and the student's commands without providing appropriate command support, compensation, or resources. While ADL research exists for military training courses, there is only limited research on ADL's impact on military education, especially in terms of its impact on learning that requires interactive elements. A DOD official stated that it is well-known in studies that distance learning (usually teleconferencing) instruction offers no significant differences as compared with classroom instruction. However, officials believe that more work should be done to collect information that ADL improves learning in military education and that these studies should focus on collaborative learning environments and the extent of their translation. In addition, further efforts should look at commercial education studies (undergraduate and graduate education) and its transfer to military education. Technological: PME school officials have stated that decisions are needed on trade- offs between increased demands for student access (more servers, more bandwidth, or reduced firewalls) and the maintenance for network security. Such decisions are complicated by what is viewed as a lack of standardization in DOD and within individual services on security requirements. In our nonresident survey, approximately 19 percent of current students and 16 percent of recent graduates experienced computer/Internet related problems affecting their PME experience. Some identified problems included servers, bandwidth issues, and security firewalls. PME schools considered using ".edu" domains to make courseware more available to students because of the difficulty of interacting ".mil" domains with systems outside of military organizations. However, such moves would be expensive and would conflict with increasing requirements to reduce the number of servers and personnel needed to operate these systems. Some reported problems involved limited bandwidth issues. While such problems can be overcome, they require time and money to resolve. Firewalls maintained for security purposes have caused schools to limit library access to nonresident students due to perceived security threats. Resources and Funding: Most ADL efforts at PME schools were fielded independently with limited budgets and staffing. USAWC's and ACSC's ADL applications were developed and supported with in-house staff responsible for managing resident and nonresident programs. These applications were fielded independently within the services. PME officials stated that PME schools' ability to fund ADL applications is limited due to DOD's priority to focus more on its training and operational activities. An emerging funding issue involves the use of copyrighted material in ADL applications. Increasing costs in using copyrighted material for course work could result in limiting course flexibility and methodologies. New technologies, such as ADL, create new requirements for faculty personnel with higher technical expertise and more equipment and structure than traditional programs. Faculty members must be skilled to perform in online and classroom settings. PME schools are beginning to observe they must offer faculty opportunities to teach courses in multiple medias or risk losing qualified faculty to competitors. Conclusions: Although PME schools receive oversight from a number of organizations, we observed that neither the schools nor the oversight agencies had focused on (1) establishing specific performance effectiveness goals for ADL implementation or (2) measuring learning outcomes as a means of evaluating program effectiveness. The Joint Staff's accreditation reports on nonresident education do not detail performance goals for any particular delivery method. The military services, which have primary responsibility for PME oversight, view the accreditation process provided by the Joint Staff as the primary means of ensuring the effectiveness of nonresident education. DOD is not alone in this problem--the lack of metrics for performance effectiveness and learning outcomes is pervasive throughout all educational institutions. Our prior efforts indicate that most public and private institutions lack a framework with which to assess implementation of training and development efforts. However, agencies need clear information on how training and development efforts affect organizational performance, and decision makers will likely want to compare the performance of these programs with that of other programs. Without clear goals and an effective process for evaluating the results of ADL application, DOD cannot ensure that ADL is achieving appropriate return on investment, student retention, student access, and other goals in comparison with prior efforts. Furthermore, programs lacking outcome metrics will be unable to demonstrate how they contribute to results. Moreover, by not capturing and using student data that are uniquely available through ADL techniques, DOD is missing the opportunity to develop the basis for effectiveness metrics and knowledge about learning outcomes. Recommendations for Executive Action: We recommend that the Secretary of Defense direct the Under Secretary of Defense for Personnel and Readiness, in concert with the Joint Staff, service headquarters, and the PME schools, to take the following two actions: * promote the development of specific performance effectiveness goals for ADL and: * promote the use of ADL technologies to capture data to provide knowledge about learning outcomes. Agency Comments and Our Evaluation: DOD partially concurred with our first recommendation and fully concurred with the second. DOD supports the use of specific effectiveness goals for PME, but believes such goals are not appropriate for any specific delivery method. While we acknowledge DOD's concerns with focusing on a specific delivery method, we believe that ADL is unlike other means of delivery because of its potential to modernize the educational experience and because its use is rapidly expanding in other areas of PME. We believe it would be worthwhile for DOD to know specifically how well ADL performs, especially in comparison with other delivery methods, in order to better understand its appropriate use for PME. DOD concurred with our second recommendation and stated that current accreditation practices are already promoting the data collection capabilities of ADL technologies for assessing multiple delivery methods. DOD's comments are included in this report as appendix V. DOD also provided technical changes, which we incorporated as appropriate. We are sending copies of this report to congressional members as appropriate. We will also send copies to the Secretary of Defense; the Secretaries of the Air Force, Army, and Navy; the Commandant of the Marine Corps; and the Chairman of the Joint Chiefs of Staff. We will make copies available to others on request. In addition, this report will be available at no charge on the GAO Web site at http://www.gao.gov. If you or your staff have any questions, please call me on (757) 552- 8100 or Clifton Spruill, Assistant Director, on (202) 512-4531. Major contributors to this report were Arnett Sanders, Maewanda Michael- Jackson, Jean Orland, David Dornisch, Terry Richardson, and Cheryl Weissman. Sincerely yours, Signed by: Neal P. Curtin: Director, Defense Capabilities and Management: [End of section] Appendix I: Scope and Methodology: We reviewed Department of Defense's (DOD) implementation of advanced distributed learning (ADL) in senior-and intermediate-level professional military education (PME) to determine processes and criteria used for converting courses, metrics to assess ADL effectiveness and its fulfillment of learning objectives, and barriers and challenges in ADL implementation. We collected, reviewed, and analyzed relevant program information and conducted interviews with DOD officials responsible for ADL implementation in PME programs and oversight responsibilities. We initially obtained data from the Principal Deputy Under Secretary of Defense for Personnel and Readiness to identify PME programs with ADL applications. A review of the data indicated that there were three existing programs. We identified, interviewed, and obtained data from officials from PME schools with ADL applications. Those schools were: * U.S. Army War College, Carlisle, Pennsylvania; * Naval War College, Newport, Rhode Island; and: * Air Command and Staff College, Montgomery, Alabama. We also interviewed and obtained data from officials at the Joint Forces Staff College in Norfolk, Virginia, on their pending ADL application. We also interviewed and obtained data from agencies within DOD responsible for oversight of PME activities. Those agencies included: * The Joint Chiefs of Staff's Operational Plans and Joint Force Development Directorate (J-7), Joint Doctrine, Education, and Training Division, Joint Education Branch, Arlington, Virginia; * The Office of the Under Secretary of Defense for Personnel and Readiness, Deputy to the Under Secretary for Readiness, Office of Readiness and Training Policy and Programs, Arlington, Virginia; * The Office of the Under Secretary of Defense for Personnel and Readiness, Deputy to the Under Secretary for Military Personnel Policy, Arlington, Virginia; * Department of the Army, Office of the Deputy Chief of Staff, Operations and Plans, Arlington, Virginia; * Department of the Navy, Office of the Chief of Naval Operations, Personal Development and Accessions Division, Washington, D.C; * Department of the Air Force, Office of the Deputy Chief of Staff for Personnel, Learning, and Force Development, Arlington, Virginia; * U.S. Marine Corps Combat Development Center, Training and Education Command, Quantico, Virginia; * U.S. Army War College Board of Visitors, Carlisle, Pennsylvania; * Naval War College Board of Advisors, Newport, Rhode Island; and: * Air University Board of Visitors, Montgomery, Alabama. To determine sufficient metrics to assess ADL effectiveness, we provided PME program officials with a detailed list of questions that included those relating to effectiveness and learning objectives. We reviewed written responses, if provided, and followed up with site visits and correspondence with oversight agencies to clarify or obtain additional information if necessary. We also obtained and analyzed data from a survey of nonresident PME current students and graduates, which included questions designed to obtain perceptions on program effectiveness. Details of the survey methodology are presented in appendix II. To determine processes and criteria DOD used for ADL conversion, we provided PME program officials with a detailed list of questions that included those relating to process and criteria decisions. We reviewed written responses, if provided, and followed up with site visits to clarify or obtain additional information if necessary. To determine whether criteria were consistent with those of other institutions performing distance education, we researched prior literature on this topic and conducted a site visit to the University of Maryland University College in Adelphi, Maryland. The school was identified in our prior reports on distance education as having a program with a large distance education population, as well as educating a significant number of military officers. We also contacted and received data from the Sloan Consortium, an organization designed to encourage collaborative sharing of knowledge and effective practices to improve online education. To determine barriers and challenges to ADL implementation, we provided PME program officials with a detailed list of questions that included those relating to barriers and challenges. We reviewed written responses, if provided, and followed up with site visits and correspondence with DOD oversight agencies to clarify or obtain additional information if necessary. We also obtained and analyzed data from a survey of nonresident PME current students and graduates, which included questions designed to obtain perceptions on barriers and challenges in completing PME courses. Details of the survey methodology are presented in appendix II. [End of section] Appendix II: Methodology for Our Survey of Nonresident PME Students and Graduates: To obtain military officers' perspectives on nonresident PME in terms of impact on careers, achievement of learning objectives, and obstacles and challenges, we conducted a statistically representative survey of current nonresident senior-and intermediate-level PME students and graduates of these schools from April 1999 to March 2003, roughly the period coinciding with initial ADL implementation at several senior-and intermediate-level schools. We present the survey questions and response results in appendix III. The Study Population: The population for the nonresident PME survey consisted of current students and graduates who fulfilled the following criteria: 1. Respondents were identified as enrolled in a senior-and intermediate-level nonresident PME program of study from April 1999 to March 2003. We decided on this time period to ensure that our respondents would have begun their programs after Web-based PME had been clearly established as a mode of instruction or have been in PME long enough to have meaningful responses to our questions. 2. Respondents participated in a senior-and intermediate-level nonresident PME program of study, as opposed to individual PME courses undertaken via continuing education programs. 3. Respondents are currently active in the U.S. military services or reserves, excluding U.S. civilians; U.S. Coast Guard members; and international members, either military or civilian. 4. Respondents participated (i.e., currently enrolled or graduated) in a nonresident PME program in one of the six senior-and intermediate-level PME schools: U.S. Army War College, Army Command and General Staff College, Air War College, Air Command and Staff College, Naval War College, and Marine Corps College of Continuing Education. The survey asked respondents about PME's impact on furthering career objectives, their achievement of learning objectives, and obstacles and challenges of the programs. Specific questions concerned students' satisfaction with their overall program, various modes of program delivery, and with technologies used; students' time and duty management concerns; and reasons for participation in nonresident PME. Developing the Survey: To develop areas of inquiry for the survey, we reviewed our previous work related to distance education and PME. We reviewed a series of survey questionnaires developed by us and by DOD. We used these sources and our own analysis to develop an initial set of questions. We further developed and refined the questionnaire by obtaining and incorporating written comments regarding the initial questions from administrators and other representatives of the senior-and intermediate-level PME schools. In addition to an internal expert technical review by our Survey Coordination Group, we pretested the survey with five individuals whose personal characteristics corresponded to our eligibility criteria. We identified pretest subjects through our contacts who were current military personnel or who knew military personnel and our PME points of contact. The Sample Design and Administration: We conducted the survey between January and April of 2004 on a random sample of 437 current students and graduates of nonresident PME programs using a self-administered Web-based questionnaire. We drew the names of our respondents from an overall population data set we constructed that combined separate data sets received from each of the senior-and intermediate-level PME schools. For each data set, we requested the officer's name, school attended, month and year of initial enrollment, component (defined as either active duty or reservist), and mode of PME delivery. We requested e-mail addresses and, if needed, phone numbers from potential respondents after they were drawn from the population data sets. We stratified our sample by component in order to better understand any differences between these components. We activated the survey Web site and informed our sample respondents of the Web site, their logon name, and passwords by e-mail on January 30, 2004. To maximize the response rate, we sent five subsequent follow-up e-mail reminders to nonrespondents in February and March 2004. At the time of the third mailing, we also telephoned many of the nonrespondents to encourage them to complete the survey. We ended data collection activities on April 30, 2004. Of the 437 selectees included in our sample, we received 273 useable questionnaires. We defined useable as respondents who completed the survey and were not identified as out-of-scope. During the survey, we deemed 67 of the 437 to be outside the scope of our survey after determining that they did not meet at least one of our eligibility criteria. Disregarding these 67 responses, our overall response rate was 73.8 percent (273/370). Table 2 shows the final disposition of the sample (the 437 respondent accounts activated) by strata. Table 2: Disposition of Sample: Stratum: Active Duty; Sample: 219; Useable: 136; Out of scope: 31; Number of nonrespondents: 52. Stratum: Reserve Duty; Sample: 218; Useable: 137; Out of scope: 36; Number of nonrespondents: 45. Source: GAO. [End of table] The estimates we make in this report are the result of weighting the survey responses to account for effective sampling rates in each stratum. These weights reflect both the initial sampling rate and the response rate for each stratum. As with many surveys, our estimation method assumes that nonrespondents would have answered like the respondents. Sampling Error: For the estimates we present in this report, we are 95 percent confident that the results we would have obtained had we studied the entire population are within +/-10 or fewer percentage points of our estimates (unless otherwise noted). Because we surveyed a sample of recent nonresident PME students, our results are estimates of student and graduate characteristics and thus are subject to sampling errors. Our confidence in the precision of the results from this sample is expressed in 95 percent confidence intervals, which are expected to include the actual results for 95 percent of the samples of this type. We calculated confidence intervals for our results using methods appropriate for a stratified probability sample. Nonsampling Error and Data Quality: We conducted in-depth pretesting of the questionnaire to minimize measurement error. However, the practical difficulties in conducting surveys of this type may introduce other types of errors, commonly known as nonsampling errors. For example, measurement errors can be introduced if (1) respondents have difficulty interpreting a particular question, (2) respondents have access to different amounts of information in answering a question, or (3) errors in data processing occur. We took extensive steps to minimize such errors in developing the questionnaire, collecting the data, and editing and analyzing the information. The Web-based data management system we used provides a systematized process for processing, transferring, and analyzing data that also protects against nonsampling errors. In addition, we performed tests to ensure the reliability and usefulness of the data provided by the PME schools. These included computer analyses to identify inconsistencies both within and across the data sets and other errors in the data sets from which we developed our overall sampling frame. We also interviewed agency officials knowledgeable about the data. We determined that the data were sufficiently reliable for the purposes of this report. [End of section] Appendix III: Survey Responses: Introduction: Welcome DoD Distance Learning Student or Graduate. The U.S. General Accounting Office (GAO) - an agency of the U.S. Congress has been requested to review various aspects of nonresident professional military education (PME). Part of that effort is to evaluate whether the appropriate decisions are being made in regards to nonresident education for intermediate and senior-level military officers. As part of this effort, we are assessing opinions of nonresident PME graduates and current students on (1) the achievement of PME learning objectives, (2) obstacles and challenges in completing PME, and (3) PME's impact on furthering career objectives. The survey should take approximately 10 minutes to complete. Participation in this survey is voluntary but encouraged. Your responses will be confidential and the results of the survey will be reported in aggregate form only. Before choosing an answer, please read the full question and all response choices carefully. Thank you in advance for your participation. Frame of Reference: Please consider only the following when answering the questions on this survey: 1. Nonresident Professional Military Education programs of study that you have participated in, as opposed to individual PME courses undertaken via continuing education programs. AND: 2. Nonresident Professional Military Education programs of study that you began in April 1999 or after. Our survey of nonresident Professional Military Education students was divided into two main parts, one with questions appropriate for current students and one with similar questions worded slightly differently for graduates. There are also questions on demographics to which both current students and graduates responded. Survey questions and responses for graduates are indicated in italics and those for current students are in plain text. The information provided here represents weighted data. For information on weighting, see appendix II. Except where noted by the following, all percentage estimates have 95% confidence intervals within +/-10 percentage points: [A] Confidence interval exceeds +10 percentage points: [B] Confidence interval exceeds +25 percentage points and estimate is unreliable: Questions 1 and 31 are intentionally omitted because they contained instructions telling respondents which questions to answer. Survey of Nonresident Professional Military Education Graduates and Current Students: Q2. The name of program in which currently enrolled: Q32. From which school did you graduate? Current Students; Air Command and Staff College Nonresident Program (percent): 30%; Air War College Nonresident Program (percent): 25%; Army Command and General Staff College Nonresident Program (percent): 31%; Army War College Distance Education Program (percent): 1%; Marine Corps College of Continuing Education (percent): 10%; Naval College of Distance Education (percent): 2%. Graduates; Air Command and Staff College Nonresident Program (percent): 40%; Air War College Nonresident Program (percent): 27%; Army Command and General Staff College Nonresident Program (percent): 27%; Army War College Distance Education Program (percent): 2%; Marine Corps College of Continuing Education (percent): 1%; Naval College of Distance Education (percent): 3%. [End of table] Q3. In what month and year did you begin your PME program? Q33. In what month and year did you begin your PME program (if you graduated from more than one program, answer for the most recent one)? Current Students; 1999 (percent): 1%; 2000 (percent): 3%; 2001 (percent): 14%; 2002 (percent): 30%; 2003 (percent): 47%; Other (percent): 5%. Graduates; 1999 (percent): 15%; 2000 (percent): 26%; 2001 (percent): 34%; 2002 (percent): 21%; 2003 (percent): 5%; Other (percent): 0%. [End of table] Q4. What mode of instruction have you used most often in your nonresident Professional Military program? Q34. What mode of instruction did you use most often in your nonresident Professional Military program? Current Students; Seminar or Classroom Instruction (percent): 24%; Web- Based Correspondence (percent): 25%; Paper-Based Correspondence (percent): 51%. Graduates; Seminar or Classroom Instruction (percent): 42%; Web-Based Correspondence (percent): 18%; Paper-Based Correspondence (percent): 40%. [End of table] Q5. In what month and year do you expect to complete your PME studies program? Q35. In what month and year did you complete your Professional Military Education program? Current Students; 2003 (percent): 1%; 2004 (percent): 68%; 2005 (percent): 21%; 2006 (percent): 4%; 2007 (percent): 0%; Other (percent): 6%. Graduates; 2003 (percent): 1%; 2004 (percent): 10%; 2005 (percent): 18%; 2006 (percent): 31%; 2007 (percent): 39%; Other (percent): 1%. [End of table] Q6. In a typical week, approximately how many hours did you spend in Professional Military Education-related activities, including preparation, study, working on-line, and time in class? Q36. In a typical week, approximately how many hours did you spend in Professional Military Education-related activities, including preparation, study, working on-line, and time in class? Current Students; Mean: 5.8%. Graduates; Mean: 8.4%. [End of table] Q7. Does the military or your employer afford you time during your work-week for Professional Military Education? Q37. Did the military or your employer afford you time during your work-week for Professional Military Education? Current Students; Yes (percent): 23%; No (percent): 77%. Graduates; Yes (percent): 42%; No (percent): 58%. [End of table] Q8. How many hours do you work in paid employment in a typical work- week (outside of Professional Military Education-related activities)? Q38. During the period of time that you were completing your Professional Military Education program, how many hours did you work in paid employment in a typical work-week (outside of Professional Military Education-related activities)? Current Students; Mean: 52.0%. Graduates; Mean: 47.5%. [End of table] Q9. Listed below are various reasons why someone would participate in a nonresident PME program. What is your greatest reason for participating in a nonresident Professional Military Education program? Q39. Listed below are various reasons why someone would participate in a nonresident PME program. What was your greatest reason for participating in a nonresident Professional Military Education program? To develop professionally; Current Students (percent): 20%; Graduates (percent): 13%. To gain access to better assignments; Current Students (percent): 1%; Graduates (percent): 0%. To gain knowledge in my field or in fields of interest to me; Current Students (percent): 2%; Graduates (percent): 0%. To improve my chances of, or meet the requirements for, promotion; Current Students (percent): 73%; Graduates (percent): 84%. To network with other officers; Current Students (percent): 1%; Graduates (percent): 0%. To obtain college credit; Current Students (percent): 0%; Graduates (percent): 0%. Other; Current Students (percent): 4%; Graduates (percent): 3%. [End of table] Q10. To this point in time, how satisfied or dissatisfied are you with your Professional Military Education program? Q40. Overall, how satisfied or dissatisfied are you with the Professional Military Education program in which you graduated? Current Students; Very satisfied (percent): 16%; Somewhat satisfied (percent): 35%; Neither satisfied nor dissatisfied (percent): 21%; Somewhat dissatisfied (percent): 21%; Very dissatisfied (percent): 8%. Graduates; Very satisfied (percent): 20%; Somewhat satisfied (percent): 45%; Neither satisfied nor dissatisfied (percent): 15%; Somewhat dissatisfied (percent): 15%; Very dissatisfied (percent): 5%. [End of table] Q11. To what extent, if any, has your Professional Military Education program benefited your military career to this point in time? Q41. To what extent, if any, has graduation from your Professional Military Education program benefited your military career to this point in time? Current Students; Very great extent (percent): 1%; Great extent (percent): 21%; Moderate extent (percent): 42%; Little extent (percent): 20%; No extent (percent): 16%. Graduates; Very great extent (percent): 15%; Great extent (percent): 21%; Moderate extent (percent): 30%; Little extent (percent): 22%; No extent (percent): 12%. [End of table] Q12. To what extent, if any, do you believe the knowledge you are acquiring in your Professional Military Education program will improve your effectiveness in future assignments? Q42. To what extent, if any, do you believe the knowledge you acquired in your Professional Military Education program has improved your effectiveness in job assignments? Current Students; Very great extent (percent): 5%; Great extent (percent): 22%; Moderate extent (percent): 42%; Little extent (percent): 25%; No extent (percent): 6%. Graduates; Very great extent (percent): 9%; Great extent (percent): 16%; Moderate extent (percent): 39%; Little extent (percent): 24%; No extent (percent): 12%. [End of table] No Parallel Question for Current Students: Q43. Have you been promoted to a higher rank since you completed your Professional Military Education program? Graduates; Yes (percent): 78%; No (percent): 22%. [End of table] No Parallel Question for Current Students: Q44. To what extent, if any, do you believe that completion of your Professional Military Education program contributed to your promotion? Graduates; Very great extent (percent): 53%[A]; Great extent (percent): 34%[A]; Moderate extent (percent): 3%[A]; Little extent (percent): 9%[A]; No extent (percent): 0%. [End of table] Q13. To what extent, if any, does your Professional Military Education program enable you to acquire the knowledge you are expected to obtain? Q45. To what extent, if any, did your Professional Military Education program enable you to acquire the knowledge you were expected to obtain? Current Students; Very great extent (percent): 5%; Great extent (percent): 25%; Moderate extent (percent): 43%; Little extent (percent): 21%; No extent (percent): 6%. Graduates; Very great extent (percent): 9%; Great extent (percent): 28%; Moderate extent (percent): 40%; Little extent (percent): 19%; No extent (percent): 4%. [End of table] Q14. Was/is any part of your Professional Military Education program taken through seminar/classroom-based instruction? Q46. Was any part of your Professional Military Education program taken through seminar/classroom-based instruction? Current Students; No (percent): 67%; Yes (percent): 33%. Graduates; No (percent): 50%; Yes (percent): 50%. [End of table] Q15. Overall, how would you rate the quality of this seminar/classroom- based learning? Q47. Overall, how would you rate the quality of this seminar/classroom- based learning? Current Students; Excellent (percent): 8%[A]; Very good (percent): 66%[A]; Good (percent): 11%[A]; Fair (percent): 13%[A]; Poor (percent): 3%%[A]. Graduates; Excellent (percent): 18%[A]; Very good (percent): 35%[A]; Good (percent): 31%[A]; Fair (percent): 14%; Poor (percent): 3%. [End of table] Q16. Was/is any part of your Professional Military Education program taken through paper-based correspondence? Q48. Was any part of your Professional Military Education program taken through paper-based correspondence? Current Students; No (percent): 36%; Yes (percent): 64%. Graduates; No (percent): 40%; Yes (percent): 60%. [End of table] Q17. Overall, how would you rate the quality of this paper-based correspondence learning? Q49. Overall, how would you rate the quality of this paper-based correspondence learning? Current Students; Excellent (percent): 3%; Very good (percent): 27%[A]; Good (percent): 40%[A]; Fair (percent): 26%[A]; Poor (percent): 4%. Graduates; Excellent (percent): 6%; Very good (percent): 31%; Good (percent): 33%; Fair (percent): 20%; Poor (percent): 10%. [End of table] Q18. Was/is any part of your Professional Military Education program taken through the World-Wide Web or Internet?): Q50. Was any part of your Professional Military Education program taken through the World-Wide Web or Internet? Current Students; No (percent): 64%; Yes (percent): 36%. Graduates; No (percent): 60%; Yes (percent): 40%. [End of table] Q19. Overall, how would you rate the quality of this web/Internet- based learning? Q51. Overall, how would you rate the quality of this web/Internet- based learning? Current Students; Excellent (percent): 17%[A]; Very good (percent): 22%[A]; Good (percent): 40%[A]; Fair (percent): 15%[A]; Poor (percent): 5%[A]. Graduates; Excellent (percent): 8%[A]; Very good (percent): 39%[A]; Good (percent): 41%[A]; Fair (percent): 10%[A]; Poor (percent): 2%. [End of table] Q20. How easy or difficult has it been for you to use web/Internet- based learning? Q52. How easy or difficult was it for you to use web/Internet-based learning? Current Students; Very easy (percent): 34%[A]; Somewhat easy (percent): 21%[A]; Neither easy nor difficult (percent): 32%[A]; Somewhat difficult (percent): 8%[A]; Very difficult (percent): 5%[A]. Graduates; Very easy (percent): 40%[A]; Somewhat easy (percent): 37%[A]; Neither easy nor difficult (percent): 10%[A]; Somewhat difficult (percent): 12%[A]; Very difficult (percent): 0%. [End of table] Q21. How easy or difficult have you found interaction with faculty during your web/Internet-based learning? Q53. How easy or difficult did you find interaction with faculty during your web/Internet -based learning? Current Students; Very easy (percent): 26%[A]; Somewhat easy (percent): 11%[A]; Neither easy nor difficult (percent): 35%[A]; Somewhat difficult (percent): 11%[A]; Very difficult (percent): 16%[A]. Graduates; Very easy (percent): 18%[A]; Somewhat easy (percent): 48%[A]; Neither easy nor difficult (percent): 29%[A]; Somewhat difficult (percent): 5%[A]; Very difficult (percent): 0%. [End of table] Q22. How easy or difficult have you found interaction with other students during your web/Internet-based learning? Q54. How easy or difficult did you find interaction with other students during your web/Internet-based learning? Current Students; Very easy (percent): 16%[A]; Somewhat easy (percent): 31%[A]; Neither easy nor difficult (percent): [B]; Somewhat difficult (percent): 16%[A]; Very difficult (percent): 11%[A]. Graduates; Very easy (percent): 15%[A]; Somewhat easy (percent): 41%[A]; Neither easy nor difficult (percent): 26%[A]; Somewhat difficult (percent): 15%[A]; Very difficult (percent): 4%[A]. [End of table] Q23. How well does the courseware/course software work on the computer equipment to which you have access for taking web/Internet-based learning? Q55. How well did the courseware/course software work on the computer equipment to which you had access for taking web/Internet-based learning? Current Students; Excellent (percent): 8%[A]; Very good (percent): 41%[A]; Good (percent): 32%[A]; Fair (percent): 19%[A]; Poor (percent): 0%. Graduates; Excellent (percent): 21%[A]; Very good (percent): 43%[A]; Good (percent): 28%[A]; Fair (percent): 5%; Poor (percent): 2%. [End of table] Q24. How reliable is your network access for taking web/Internet-based learning (e.g. ability to connect to upload and download assignments, etc)? Q56. How reliable was your network access for taking web/Internet- based learning (e.g. ability to connect to upload and download assignments, etc)? Current Students; Very reliable (percent): 41%[A]; Somewhat reliable (percent): 40%[A]; As reliable as unreliable (percent): 14%[A]; Somewhat unreliable (percent): 3%[A]; Very unreliable (percent): 3%[A]. Graduates; Very reliable (percent): 52%[A]; Somewhat reliable (percent): 45%[A]; As reliable as unreliable (percent): 2%; Somewhat unreliable (percent): 2%; Very unreliable (percent): 0%. [End of table] Q25. Compared to resident Professional Military Education students in the school in which you are enrolled, of the following options, do you believe you are prepared? Q57. Compared to resident Professional Military Education program graduates of your school, of the following options, do you believe you are prepared? Current Students; better than resident students. (percent): 2%; as well as resident students. (percent): 27%; worse than resident students. (percent): 49%; Don't know (percent): 23%. Graduates; better than resident students. (percent): 3%; as well as resident students. (percent): 30%; worse than resident students. (percent): 48%; Don't know (percent): 19%. [End of table] Q26. Overall, what has been the primary challenge, if any, affecting your Professional Military Education program? Q58. Overall, what was the primary challenge, if any, affecting your Professional Military Education program? Computer/Internet-related problem%; Current Students (percent): 3%; Graduates (percent): 0%. Deployment cycle; Current Students (percent): 12%; Graduates (percent): 5%. Domestic circumstances; Current Students (percent): 12%; Graduates (percent): 10%. Maintaining focus; Current Students (percent): 16%; Graduates (percent): 12%. Present job duties; Current Students (percent): 45%; Graduates (percent): 53%. Not applicable, I am not experiencing & have not experienced any challenges to this point in time; Current Students (percent): 6%; Graduates (percent): 17%. Other; Current Students (percent): 8%; Graduates (percent): 3%. [End of table] Q27. Have you experienced any computer/Internet-related problems affecting your Professional Military Education program? Q59. Did you experience any computer/Internet-related problems affecting your Professional Military Education program? Current Students; No (percent): 81%; Yes (percent): 19%. Graduates; No (percent): 84%; Yes (percent): 16%. [End of table] Q28. What specific computer/Internet-related problems have you incurred? Q60. What specific computer/Internet-related problems have you incurred? Current Students: a. Bandwidth/ network speed: Yes (percent); Graduates: a. Bandwidth/ network speed: Yes (percent). a. Bandwidth/ network speed; Current Students: 63%[A]; Graduates: [B]. b. Inadequate uploading or downloading ability/lack of high-speed internet equipment; Current Students: 57%[A]; Graduates: [B]. c. Inadequate technical support; Current Students: [B]; Graduates: [B]. d. Defective/ incompatible equipment; Current Students: 25%[A]; Graduates: [B]. e. Lack of computer skills; Current Students: 5%[A]; Graduates: [B]. f. Lack of network availability/access to internet; Current Students: 52%[A]; Graduates: [B]. g. Security/Firewall issues; Current Students: 26%[A]; Graduates: [B]. h. Other; Current Students: [B]; Graduates: [B]. [End of table] Q29. At any point during your Professional Military Education program, have you had to defer/disenroll from your studies? Q61. At any point during your Professional Military Education program, did you have to defer/disenroll from your studies? Current Students; No (percent): 66%; Yes (percent): 34%. Graduates; No (percent): 86%; Yes (percent): 14%. [End of table] Q30. What was the primary reason you had to defer/disenroll from your studies? Q62. What was the primary reason you had to defer/disenroll from your studies? Open-ended comments not shown here. Q63. What is the highest degree or level of school that you have completed? High school or equivalent; Current Students (percent): 0%; Graduates (percent): 0%. 1 or more years of college, no degree; Current Students (percent): 0%; Graduates (percent): 0%. Associate's degree; Current Students (percent): 1%; Graduates (percent): 0%. Bachelor's degree; Current Students (percent): 42%; Graduates (percent): 25%. Master's degree; Current Students (percent): 44%; Graduates (percent): 61%. Doctoral or professional school degree; Current Students (percent): 10%; Graduates (percent): 12%. Other; Current Students (percent): 3%; Graduates (percent): 2%. [End of table] Q64. In what branch of the military do you serve? Air Force; Current Students (percent): 54%; Graduates (percent): 66%. Army; Current Students (percent): 33%; Graduates (percent): 30%. Marines; Current Students (percent): 11%; Graduates (percent): 1%. Navy; Current Students (percent): 2%; Graduates (percent): 3%. [End of table] Q65. What duty capacity best describes you during the majority of your Professional Military Education program? Non-Active Duty; Current Students (percent): 43%; Graduates (percent): 26%. Active Duty; Current Students (percent): 57%; Graduates (percent): 74%. [End of table] Q66. What component best describes you during the majority of your Professional Military Education program? Active Component; Current Students (percent): 46; Graduates (percent): 67%. Reserve Component; Current Students (percent): 54%; Graduates (percent): 33%. [End of table] Q67. Are you a member of the National Guard? Yes; Current Students (percent): 35%[A]; Graduates (percent): 45%[A]. No; Current Students (percent): 65%[A]; Graduates (percent): 55%[A]. [End of table] Q68. What military occupational category best describes you during the majority of your Professional Military Education program? Administrative; Current Students (percent): 12%; Graduates (percent): 9%. Engineering & Maintenance Officers; Current Students (percent): 10%; Graduates (percent): 13%. General Officers & Executives; Current Students (percent): 4%; Graduates (percent): 7%. Health Care Officers; Current Students (percent): 18%; Graduates (percent): 13%. Intelligence Officers; Current Students (percent): 6%; Graduates (percent): 1%. Scientists & Professionals; Current Students (percent): 9%; Graduates (percent): 10%. Supply & Procurement & Allied Officers; Current Students (percent): 7%; Graduates (percent): 8%. Tactical Operation Officers; Current Students (percent): 28%; Graduates (percent): 31%. Non-Occupational; Current Students (percent): 0%; Graduates (percent): 1%. Not Listed Above/Other; Current Students (percent): 6%; Graduates (percent): 7%. [End of table] Q69. What was your rank when you began your Professional Military Education program? O-2; Current Students (percent): 1%; Graduates (percent): 1%. O-3; Current Students (percent): 18%; Graduates (percent): 18%. O-4; Current Students (percent): 65%; Graduates (percent): 59%. O-5; Current Students (percent): 13%; Graduates (percent): 21%. O-6; Current Students (percent): 2%; Graduates (percent): 1%. Other; Current Students (percent): 2%; Graduates (percent): 0%. [End of table] Q70. What is your current rank? O-2; Current Students (percent): 0%; Graduates (percent): 0%. O-3; Current Students (percent): 4%; Graduates (percent): 3%. O-4; Current Students (percent): 70%; Graduates (percent): 59%. O-5; Current Students (percent): 24%; Graduates (percent): 27%. O-6; Current Students (percent): 3%; Graduates (percent): 7%. O-7; Current Students (percent): 0%; Graduates (percent): 1%. Other; Current Students (percent): 0%; Graduates (percent): 3%. [End of table] Q71. If you have any other comments related to your PME education, training, assignments, distance learning, or any other matters related to this questionnaire, please note them here. Open-ended comments not shown here. [End of section] Appendix IV: ADL Applications and Additional Features of Nonresident Programs: We observed three current ADL applications at PME senior-and intermediate-level schools. These schools have geared their ADL efforts to their nonresident programs. The programs vary from service to service in terms of enrollment, structure, duration, and credits received for graduation. In addition, we observed additional features of nonresident programs that affect the nature of their ADL applications. U.S. Army War College: The U.S. Army War College (USAWC), the Army's senior-level PME school, initiated its Web-based nonresident education program in April 1999. The program went online in an evolutionary process until the spring of 2002, whereby students received both text and online versions. Since the spring of 2002, all nonresident students have received their education via a combination of ADL technology and appropriate text. Nonresident students are board selected to participate in the program. It is a 2-year Web-based program that is the only delivery method offered to nonresident students. The program has a "blended" component, whereby 2 of its 12 courses are taken in residence at USAWC. Also, distance courses are presented to students as a group or cohort; that is, students enroll at the beginning of the nonresident school year and must complete a sequenced load of 5 courses during the first year, followed by an additional 5 courses during the second year. The resident courses are of 2-week duration and are conducted at the end of each academic year. The nonresident program is designed to parallel the resident program, and graduates from both programs are awarded Master's Degrees in Strategic Studies. Naval War College: The Naval War College's (NWC) nonresident education programs are concentrated in its College of Distance Education, its only nonresident college and one of five colleges under the NWC umbrella. The College of Distance Education, an intermediate-level PME school, offers several nonresident options. The fleet seminar program has existed in various forms at the school since 1974; the Web-enabled correspondence program has been operating fully since October 2002; and the CD-ROM-based correspondence program, effective in April 2004, was designed to replace the phased-out paper-based correspondence course. Nonresident options are open to all officers and qualified personnel. The Web-based course can be completed in 18-24 months. While there is no formal resident portion to this course, students are assigned to cohort teams to facilitate team and faculty communication. This nonresident course is closely aligned with the resident course, and graduates are allowed to obtain graduate hour credits. In the case of several seminars of the fleet seminar program, students can apply for admission to a program of graduate study leading toward a Master's of Arts Degree in National Security and Strategic Studies. Air Command and Staff College: The Air Command and Staff College (ACSC), the Air Force's intermediate-level PME school, implemented its nonresident program in its present form in September 1999. There are two methods for completing the nonresident program: by seminar or by correspondence. The ACSC nonresident program is open to all officers and qualified personnel. The most recent version of the program consists of six courses organized into two semesters. The seminar method, which can take up to 11 months to complete, is conducted weekly, is typically composed of 3-18 students, and is led by assigned seminar leaders in order to facilitate group discussions. The correspondence program, a self-study program delivered in a balanced manner consisting of paper, CD-ROM, and Web-based delivery, requires up to 18 months to complete. Students move interchangeably between both programs, but they must achieve a minimum score of 70 percent on each of the six examinations and must complete four interactive Web-based exercises. The nonresident programs are designed to mirror resident programs, and there are multiple versions in use by ACSC nonresident students. These programs do not award master's degrees, but the American Council of Education recommends up to 21 semester hours of graduate credit for course completion. Joint Forces Staff College: National Defense University's Joint Forces Staff College (JFSC) is piloting an Advanced Joint Professional Military Education pilot course for senior-and intermediate-level reserve officers. Initially launched in September 2003, it is designed to last 38 weeks. The period consists of 35 weeks of Web-based education and 3 weeks of resident education with 1 week occurring after the first 8 weeks of Web-based education, and the last 2 weeks at the end of the 38-week period. JFSC, already responsible for the resident Joint PME Phase II course used to complete the education process for joint specialty officers, was tasked to develop a Joint PME course for reserve officers in response to the fiscal year 1999 National Defense Authorization Act and the Joint Staff Guidance in May 2000. While there is no joint specialty officer requirement for reserve officers, JFSC was required to prepare reserve officers for joint duty assignments by providing a course similar in content to its resident Joint PME course, and to do so by utilizing current distance learning applications. Additional Features of Nonresident PME Programs: There are additional features of PME programs that affect the nature of their ADL applications. Those features include: * Student Board Selection--Nonresident students are selected to attend the PME schools either through an annual board selection process or through open admissions. Only USAWC selects its nonresident students; the other programs with ADL applications have open-admission policies. * Joint Professional Military Education--A significant portion of the PME curriculum involves study of joint service issues along with service-specific issues. Officers who successfully complete senior-or intermediate-level PME course work are awarded Joint PME Phase I credit, which is required for those who wish to serve as joint specialty officers. All nonresident programs with ADL applications grant Joint PME Phase I credit. * Service Promotion Impact--PME officials stated that PME program completion and other forms of higher education are factors used in consideration for promotion and vary among the services. Generally, the Air Force requires completion of a corresponding PME level of study before a candidate is considered for the next promotion level. The Army, while not as strict as the Air Force, places a high value on PME and graduate education in promotion decisions. The Navy, placing a higher premium on operational experience, currently is less inclined to recognize PME as a credential for promotion. * Learning Objectives Between Programs--PME officials stated that, as outlined by Joint Staff policies, learning objectives for nonresident courses are required to be the same for resident courses, regardless of the method of delivery. PME schools have instituted internal control processes to ensure the achievement of learning objectives for all programs, irrespective of delivery method. Generally, PME schools apply similar evaluation systems and criteria to both resident and nonresident programs. * Levels-of-Learning--PME schools teach to differing achievement levels across and within the services, and they have designed their curricula accordingly. School officials refer to these achievement levels as "levels-of-learning" based on a taxonomy defined in the Joint Staff policy. (See table 3 for a detailed definition of levels-of-learning designations.) For the schools with ADL applications, their desired levels of learning for nonresident programs may or may not be equivalent to the corresponding resident programs: * USAWC--Synthesis/Analysis (same as for resident program). * NWC--Application (resident program calls for synthesis/analysis). * ACSC--Comprehension (resident program calls for synthesis/resident). * JFSC (Planned)--Application (same as for resident program). Table 3: Levels-of-Learning Definitions: Levels of learning: Knowledge; Definitions: The ability to remember previously learned material. This level involves recall of a wide range of material, from specific facts to complete theories, but all that is required is bringing to mind appropriate information. Terminology for achievement: defines, describes, identifies, labels, lists, matches, names, outlines, reproduces, selects, and states. Levels of learning: Comprehension; Definitions: The ability to grasp the meaning of material. Translating material from one form to another, interpreting material, or estimating future trends may show this level. Terminology for achievement: converts, defends, distinguishes, estimates, explains, extends, generalizes, gives examples, infers, paraphrases, predicts, rewrites, summarizes, translates, and understands. Levels of learning: Value; Definitions: The internalization and consistent display of a behavior. The levels of valuing consist of acceptance of a value, preference for a value, and commitment (conviction). Levels of learning: Application; Definitions: The ability to use learned material in new and concrete situations. This level includes application of rules, methods, concepts, principles, laws, and theories. Terminology for achievement: changes, computes, demonstrates, discovers, manipulates, modifies, operates, predicts, prepares, produces, relates, shows, solves, and uses. Levels of learning: Analysis; Definitions: The ability to break down material into its component parts so that its organizational structure may be understood. This level includes identification of the parts, analysis of the relationships between parts, and recognition of the organizational principles involved. Terminology for achievement: breaks down, diagrams, differentiates, discriminates, distinguishes, illustrates, infers, outlines, points out, selects, separates, and subdivides. Levels of learning: Synthesis; Definitions: The ability to put parts together to form a new whole. This level involves production of unique communications, a plan of operations, or a set of abstract relations. Terminology for achievement: categorizes, combines, compiles, composes, creates, devises, designs, explains, generates, modifies, organizes, plans, rearranges, reconstructs, relates, reorganizes, revises, rewrites, summarizes, tells, and writes. Levels of learning: Evaluation; Definitions: The ability to judge the value of material for a given purpose. Judgments are to be based on defined internal (organizational) or external (relevance to the purpose) criteria. Criteria are subject to value judgments. Terminology for achievement: appraises, criticizes, discriminates, explains, justifies, interprets, and supports. Source: DOD. Note: These terms, listed in increasing levels of achievement, are used to define the Joint PME learning objectives for PME schools. [End of table] [End of section] Appendix V: Comments from the Department of Defense: OFFICE OF THE UNDER SECRETARY OF DEFENSE: 4000 DEFENSE PENTAGON WASHINGTON, D.C. 20301-4000: PERSONNEL AND READINESS: JUL 23 2004: Mr. Neil P. Curtin: Director, Defense Capabilities and Management: U.S. Government Accountability Office: Washington, D.C. 20548: Dear Mr. Curtain: This is the Department of Defense (DoD) response to the Government Accountability Office Draft Report GAO-04-873 "MILITARY EDUCATION: DoD Needs to Develop Performance Goals and Metrics for Advanced Distributed Learning in Professional Military Education," July 2, 2004 (GAO Code 350327). I would like to make you aware of my assessment that the utility and validity of the current report is problematic. I have outlined my concerns below. This GAO report was initiated based on a request from Representative Skelton to assess the impact of Advanced Distributed Learning (ADL) on officer education. The current draft does not reflect the intent and does not acknowledge or address most of the identified objectives of the February 2003 engagement letter. The altered focus on assessment of ADL programs and the development of specific ADL performance goals and metrics do not correlate to the original intent and objectives. The Department appreciates the opportunity to comment on this draft. Of the two recommendations, we partially concur with the first and concur with the second. The Department's comments to the GAO draft recommendations are enclosed. Technical comments on the entire draft report were provided separately. Sincerely, Signed by: Paul W. Mayberry: Deputy Under Secretary: Readiness: Enclosure: As stated: GAO DRAFT REPORT - DATED JULY 2, 2004 GAO CODE 350327/GAO-04-873: "MILITARY EDUCATION: DoD Needs to Develop Performance Goals and Metrics for Advanced Distributed Learning in Professional Military Education": DEPARTMENT OF DEFENSE COMMENTS TO THE RECOMMENDATIONS: RECOMMENDATION 1: The GAO recommended that the Secretary of the Defense direct the Under Secretary of Defense for Personnel and Readiness, in concert with the Joint Staff, Service headquarters, and the professional military education schools, promote the development of specific performance effectiveness goals for advanced distributed learning. (Page 15-16/GAO Draft Report): DOD RESPONSE: Partially Concur. The Department does support the use of specific performance effectiveness goals for professional military education. However, development of specific performance effectiveness goals for any specific delivery method, such as advanced distributed learning (ADL), is not appropriate. Educational outcomes are based on common standards, as defined in the Officer Professional Military Education Policy, regardless of delivery method. RECOMMENDATION 2: The GAO recommended that the Secretary of the Defense direct the Under Secretary of Defense for Personnel and Readiness to promote the use of advanced distributed learning technologies to capture data to provide knowledge about learning outcomes. (Page 16/GAO Draft Report): DOD RESPONSE: Concur. Metrics for learning outcomes are established by the schools and ADL technologies can capture data that can be used to evaluate the metrics. Current accreditation practices are already promoting the data- collection capabilities of ADL technologies for assessing multiple delivery methods. GAO's Comment: The following is GAO's comment on the letter from the Department of Defense dated July 23, 2004. 1. When we initiated this engagement in February 2003, a key objective was to determine (1) the assumptions for DOD's decision to move officer senior-and intermediate-service schools from 1-year residency to shorter periods by using ADL and (2) which courses and schools would be affected. Immediately after fieldwork commenced, however, DOD informed us that it was no longer actively pursuing that approach. In April 2003, after consulting with the congressional requester, we informed our DOD point of contact regarding our pursuit of the engagement's remaining objectives. FOOTNOTES [1] Department of Defense, Chairman of the Joint Chiefs of Staff Instruction 1800.01A, Officer Professional Military Education Policy, December 2000. [2] There are also senior-and intermediate-level schools sponsored by DOD through its National Defense University. These schools are designed to educate officers on joint matters. The senior-level schools are the National War College, the Industrial College of Armed Forces, and the Joint and Combined Warfighting School-Senior at the Joint Forces Staff College. The intermediate-level school is the Joint and Combined Warfighting School-Intermediate at the Joint Forces Staff College. [3] Advanced distributed learning, as defined by DOD's April 1999 ADL Strategic Plan and May 2000 ADL Implementation Plan, expands distance learning by emphasizing computer-based instruction; common standards; and use of reusable content, networks, and learning management systems in an "anytime, anyplace" environment. [4] Joint Professional Military Education is a Joint Chiefs of Staff- approved body of objectives, policies, procedures, and standards supporting the educational requirements for joint officer management. Joint Professional Military Education is a portion of PME that supports fulfillment of the educational requirements of joint officer management. [5] U.S. General Accounting Office, Human Capital: A Guide for Assessing Strategic Training and Development Efforts in the Federal Government, GAO-04-546G (Washington, D.C.: March 2004). [6] U.S. General Accounting Office, Military Education: DOD Needs to Enhance Performance Goals and Measures to Improve Oversight of Military Academies, GAO-03-1000 (Washington, D.C.: Sept. 10, 2003). [7] The Middle States Association of College and Schools is the regional accrediting agency for the U.S. Army War College. The New England Association of Schools and Colleges is the regional accrediting agency for the Naval War College. The Southern Association of Colleges and Schools is the regional accrediting agency for the Air Command and Staff College. [8] Council for Higher Education Accreditation, Statement of Mutual Responsibilities for Student Learning Outcomes: Accreditation, Institutions, and Programs (Washington, D.C.: Sept. 2003). [9] See GAO-04-546G. [10] U.S. General Accounting Office, Distance Education: Improved Data on Program Costs and Guidelines on Quality Assessment Needed to Inform Federal Policy, GAO-04-279 (Washington, D.C.: Feb. 26, 2004). [11] The Congress established the Web-based Education Commission to prepare a report to the President and the Congress that contains recommendations for legislation and administrative actions, including those pertaining to the appropriate federal role in determining the quality of educational software products. Members of the Commission included senators, representatives, and leaders from postsecondary institutions. [12] See GAO-04-279. [13] Acculturation is defined as a developmental activity that involves the adoption of customs, protocols, and doctrine. The acculturation process is designed to prepare officers for shared leadership positions while reinforcing total force partnerships. [14] The percentages reported here are based on a sample of current students and graduates and are estimates. All percentage estimates from the survey reported have margins of error of plus or minus 10 percentage points or less, unless otherwise noted. [15] The confidence interval for this figure was +9 percent and -17 percent. GAO's Mission: The Government Accountability Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.