Academia.eduAcademia.edu

How Insurgencies End

This document and trademark(s) contained herein are protected by law as indicated in a notice appearing later in this work. This electronic representation of RAND intellectual property is provided for non-commercial use only. Unauthorized posting of RAND electronic documents to a non-RAND

CHILDREN AND FAMILIES EDUCATION AND THE ARTS ENERGY AND ENVIRONMENT HEALTH AND HEALTH CARE INFRASTRUCTURE AND TRANSPORTATION he RAND Corporation is a nonproit institution that helps improve policy and decisionmaking through research and analysis. his electronic document was made available from www.rand.org as a public service of the RAND Corporation. INTERNATIONAL AFFAIRS LAW AND BUSINESS NATIONAL SECURITY POPULATION AND AGING Skip all front matter: Jump to Page 16 PUBLIC SAFETY SCIENCE AND TECHNOLOGY TERRORISM AND HOMELAND SECURITY Support RAND Purchase this document Browse Reports & Bookstore Make a charitable contribution For More Information Visit RAND at www.rand.org Explore the RAND National Defense Research Institute View document details Limited Electronic Distribution Rights his document and trademark(s) contained herein are protected by law as indicated in a notice appearing later in this work. his electronic representation of RAND intellectual property is provided for non-commercial use only. Unauthorized posting of RAND electronic documents to a non-RAND website is prohibited. RAND electronic documents are protected under copyright law. Permission is required from RAND to reproduce, or reuse in another form, any of our research documents for commercial use. For information on reprint and linking permissions, please see RAND Permissions. This product is part of the RAND Corporation monograph series. RAND monographs present major research indings that address the challenges facing the public and private sectors. All RAND monographs undergo rigorous peer review to ensure high standards for research quality and objectivity. Embracing the Fog of War ASSESSMENT AND METRICS IN COUNTERINSURGENCY BEN CONNABLE Prepared for the U.S. Department of Defense Approved for public release; distribution unlimited NATIONAL DEFENSE RESEARCH INSTITUTE The research described in this report was prepared for the U.S. Department of Defense. The research was conducted within the RAND National Defense Research Institute, a federally funded research and development center sponsored by OSD, the Joint Staff, the Unified Combatant Commands, the Navy, the Marine Corps, the defense agencies, and the defense Intelligence Community under Contract W74V8H-06-C-0002. Library of Congress Cataloging-in-Publication Data Connable, Ben. Embracing the fog of war : assessment and metrics in counterinsurgency / Ben Connable. p. cm. Includes bibliographical references. ISBN 978-0-8330-5815-7 (pbk. : alk. paper) 1. Counterinsurgency—United States—Evaluation. 2. Counterinsurgency— Afghanistan—Evaluation. 3. Counterinsurgency—Vietnam—Evaluation. I. Title. U241C63 2012 355.02'180973—dc23 2012001598 The RAND Corporation is a nonprofit institution that helps improve policy and decisionmaking through research and analysis. R AND’s publications do not necessarily reflect the opinions of its research clients and sponsors. R® is a registered trademark. Cover design by Carol Earnest © Copyright 2012 RAND Corporation Permission is given to duplicate this document for personal use only, as long as it is unaltered and complete. Copies may not be duplicated for commercial purposes. Unauthorized posting of RAND documents to a non-RAND website is prohibited. RAND documents are protected under copyright law. For information on reprint and linking permissions, please visit the R AND permissions page (http://www.rand.org/ publications/permissions.html). Published 2012 by the RAND Corporation 1776 Main Street, P.O. Box 2138, Santa Monica, CA 90407-2138 1200 South Hayes Street, Arlington, VA 22202-5050 4570 Fifth Avenue, Suite 600, Pittsburgh, PA 15213-2665 RAND URL: http://www.rand.org To order RAND documents or to obtain additional information, contact Distribution Services: Telephone: (310) 451-7002; Fax: (310) 451-6915; Email: order@rand.org Preface In late 2009, RAND undertook an examination of military and nonmilitary assessment processes to help improve military methods for assessing counterinsurgency (COIN) operations. he original purpose of this research was to assist a military operational staf in shaping campaign design, in shifting priorities of efort during ongoing operations, and, ultimately, in meeting operational and strategic objectives in southern Afghanistan. he original intended audience for this resulting monograph was a single staf in Afghanistan. However, the study was expanded to include an examination of COIN assessment across the U.S. military. he indings of this broadened research efort will be of interest to military commanders and stafs who produce and use assessments, policymakers who rely on assessments to make decisions, and the U.S. public, the ultimate decisionmakers in any long-term U.S.-led COIN campaign. An assessment is a report or series of reports intended to inform operational and strategic decisionmaking. Military stafs build campaign assessments, and military commanders use them to make operational decisions in the ield. hese commanders add their comments to the assessments and then pass them along to policymakers. Assessments can be presented in a variety of forms, including narrative papers, quantitative graphs, maps, and brieing slides. Senior military leaders and policy stafs use these materials to assess the progress of military campaigns, determine how to allocate (or reallocate) resources, identify trends that may indicate success or failure, or ascertain whether and when it may be necessary to alter a given strategy. In support of this process, this monograph captures the complexity of the COIN operational environment, examines case studies of COIN assessment in context, explores critical weaknesses in the current assessment process, ofers recommendations for improvement, and presents an alternative assessment process. he manuscript for this publication was completed in mid-2011. Since that time, various elements in the U.S. Department of Defense have published new doctrine on assessment, some of which addresses criticisms raised here. Perhaps more importantly, the International Security Assistance Force in Afghanistan has completely revamped its assessment process, adopting many of the recommendations suggested in this monograph and in other subject-matter expert reports on COIN assessment. iii iv Embracing the Fog of War: Assessment and Metrics in Counterinsurgency his research was sponsored by the U.S. Department of Defense and conducted within the Intelligence Policy Center of the RAND National Defense Research Institute, a federally funded research and development center sponsored by the Oice of the Secretary of Defense, the Joint Staf, the Uniied Combatant Commands, the military services, the defense agencies, and the defense Intelligence Community. For more information on the RAND Intelligence Policy Center, see http://www. rand.org/nsrd/ndri/centers/intel.html or contact the director (contact information is provided on the web page). Contents Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix Tables and Text Boxes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xi Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xiii Acknowledgments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxvii Abbreviations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xxix ChAPTer One Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2 Diferences Between Campaign Assessment and Intelligence Analysis . . . . . . . . . . . . . . . . . . . . . . . 2 he Role and Purpose of Military Campaign Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Dissatisfaction with Counterinsurgency Campaign Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Research Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 Issues Not Addressed . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16 Organization of his Monograph . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17 ChAPTer TwO Concepts hat Shape Counterinsurgency Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 Policy Requirements and Counterinsurgency Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19 Wicked Problems and Counterinsurgency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27 Complex Adaptive Systems, Nonlinearity, and Chaos . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 Military Doctrine and the COIN Environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 36 ChAPTer Three Centralized Assessment heory and Pattern and Trend Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . 37 Centralized Quantitative Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38 Diferences Between Scientiic Research and Campaign Assessment. . . . . . . . . . . . . . . . . . . . . . . . . . . 41 Applying Scientiic Rigor to Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44 v vi Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Time-Series Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46 Time-Series Graphs in Military Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48 Retroactively Adjusting Time-Series Analysis and Assessment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51 Pattern and Trend Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 ChAPTer FOur he efects-Based Approach to Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 59 What Is Efects-Based Operations heory and How Does It Shape Assessment? . . . . . . . . . . . . 60 Introduction to Efects-Based Operations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 60 Elements of Efects-Based heory: Deinitions and System-of-Systems Analysis . . . . . . . . . . . 62 Efects-Based Assessment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 Measures and Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 Doctrinal Efects-Based Assessment Process: Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67 Selecting Measures and Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69 Weighting the Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71 Determining hresholds for Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 73 Examples of Color-Coded Reports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76 ChAPTer FIve Choosing Core Metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79 Choosing “Metrics” and Determining Values . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80 Considerations for Selecting Core Metrics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82 Example of a Core Metric: First-to-Fire Indicator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85 Example of Input Measures of Performance or Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89 Example of Output Measures of Efectiveness or Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91 Example of Measures of Efectiveness and Indicators from Doctrine . . . . . . . . . . . . . . . . . . . . . . . 92 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 93 ChAPTer SIx vietnam-era Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95 Orientation to Vietnam, Circa 1967–1973 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97 Overview of Vietnam-Era Assessment Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98 homas C. hayer and the Southeast Asia Analysis Reports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101 Limitations of hayer’s Data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102 he Concept of “Reasonable” Accuracy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106 Body Counts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106 he Hamlet Evaluation System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113 Hamlet Evaluation System Concept and Execution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113 Hamlet Evaluation System Data and Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122 Pattern and Trend Analysis of Vietnam War Data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131 Contents vii Casualty Data and Impressionistic Assessment of the Vietnam War . . . . . . . . . . . . . . . . . . . . . . . . . 140 Is Analysis of Vietnam Assessment Germane to Iraq and Afghanistan Assessment? . . . . . . . 146 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148 ChAPTer Seven Assessment in Afghanistan. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Overview of the Assessment Process in Afghanistan as of Late 2010 . . . . . . . . . . . . . . . . . . . . . . . . . Data Management Challenges: he Case of Afghanistan. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Problems with the Signiicant Activities Data Sets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Analysis of the Afghanistan Assessment Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153 155 158 161 166 169 ChAPTer eIghT why Does Centralized Assessment Fail in Counterinsurgency? . . . . . . . . . . . . . . . . . . . . . . . . . 171 Neither Transparent nor Credible . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172 Lack of Data Caveats and Citations Undermines Transparency and Credibility . . . . . . . . . 174 Weighting Metrics Reduces Transparency and Credibility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 175 Not Relevant to Policy Decisionmaking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 178 Unbalanced . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 178 Not Suiciently Analyzed . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180 Incongruent with Doctrinal Understanding of Counterinsurgency . . . . . . . . . . . . . . . . . . . . . . . . . . 182 Data Aggregation and the Counterinsurgency Mosaic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 186 Unparsimonious, Costly, Risky, and Contentious . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188 Additional Factors Contributing to the Failure of Centralized Assessment in Counterinsurgency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189 Overly Optimistic. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189 Inability to Adequately Address Missing Data in Time-Series Trend Analysis . . . . . . . . . . . 192 Time-Series hresholds Are Subjective and Unhelpfully Shape Behavior . . . . . . . . . . . . . . . . . 193 Incomplete and Inaccurate Data: U.S. Department of Defense Reports on Iraq . . . . . . . . . 195 Incomplete and Inaccurate Data: Civilian Casualties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 196 Misleading Metrics: Insurgent Recidivism . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 198 Aggregation and Color-Coding: he District Assessment Model. . . . . . . . . . . . . . . . . . . . . . . . . . . 200 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205 ChAPTer nIne Conclusions, recommendations, and Options. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 207 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 208 Recommendations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 215 Conduct a Joint and Service Review of Military Assessment Doctrine . . . . . . . . . . . . . . . . . . . . 215 Incorporate Counterinsurgency Assessment into Training and Education . . . . . . . . . . . . . . . . 216 Conduct an Interagency Review of Counterinsurgency Assessment . . . . . . . . . . . . . . . . . . . . . . . 216 viii Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Incorporate All-Source Analytic Methodology into the Campaign Assessment Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216 Options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 218 Option One: Improve the Existing Centralized Assessment Process . . . . . . . . . . . . . . . . . . . . . . . 218 Option Two: Build a Decentralized and Bottom-Up Assessment Process . . . . . . . . . . . . . . . . . 220 Practicality of a Bottom-Up Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 222 Delation, Inlation, and Inherent Subjectivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 224 ChAPTer Ten A Proposed Alternative to Centralized Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 227 he Contextual Assessment Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231 Step 1: Form an Assessment Working Group . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231 Step 2: Shape the Mission-Type Metric Requirements and Acquire Information. . . . . . . . . 232 Step 3: Write the Battalion-Level Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 233 Step 4: Assess Battalion Assessments at the Brigade or Regiment Level . . . . . . . . . . . . . . . . . . . 233 Step 5: Shape the Brigade or Regimental Mission-Type Metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . 234 Step 6: Write the Brigade or Regimental Assessment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 234 Step 7: Repeat the Process at the Regional Command Level . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235 Step 8: Conduct the heater Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235 Step 9: Write the heater Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236 Step 10: Write the heater Commander Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236 Step 11: Deliver the Report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236 Aggregation of Contextual Reports and Layered Summaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237 Contextual Assessment in Practice . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237 Potential Beneits and Constraints . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 239 Potential Risks or Shortcomings and Mitigating Factors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240 Chapter Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 242 APPenDIxeS Contextual Assessment Template . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 243 notional example of a Brigade/regimental-Level Contextual Assessment . . . . . . . 251 Phong Dinh Province report . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 261 Military Assistance Command, vietnam, Military Assessment, September 1967 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 273 e. Debate over efects-Based Operations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 279 A. B. C. D. Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 289 Figures S.1. S.2. 2.1. 2.2. 3.1. 3.2. 3.3. 4.1. 4.2. 4.3. 4.4. 4.5. 4.6. 4.7. 6.1. 6.2. 6.3. 7.1. 8.1. 8.2. 8.3. 8.4. 8.5. 8.6. 8.7. 8.8. 9.1. 9.2. 10.1. 10.2. Overview of the COIN Campaign Assessment Process . . . . . . . . . . . . . . . . . . . . . . . . . . . xiv he Contextual Assessment Reporting and Analysis Process . . . . . . . . . . . . . . . . . . . . xxv President Johnson and the Khe Sanh Model in the White House Situation Room . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22 Impressionistic Decisionmaking Using a Broad Array of Counterinsurgency Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 System Dynamics Model of Field Manual 3-24, Counterinsurgency. . . . . . . . . . . . . . 40 Attack Incidents Reported in Iraq, 2004–2007 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50 Iraq Brieing with Time-Series Graphs. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51 Efects-Based Operations and Cascading Efects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61 System-of-Systems Analysis Perspective of the Operational Environment. . . . . . . 64 he Efects-Based Assessment Process. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 69 Field Manual 5-0 Mathematical Assessment Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 72 Example of hresholds in Quantitative Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74 Example of a Color-Coded Efects-Based Assessment Report. . . . . . . . . . . . . . . . . . . . . . 75 Example of Color-Coded Assessment Reports on a Sliding Color Scale . . . . . . . . . . 76 Map of Vietnam, Including Military Sector Delineation . . . . . . . . . . . . . . . . . . . . . . . . . . 97 Photographs of Key Figures Cited or Referenced in his Chapter. . . . . . . . . . . . . . . . 98 Ratio of Enemy to Allied Casualties, Vietnam, 1965–1968 . . . . . . . . . . . . . . . . . . . . . . 142 Time-of-Day Attack Pattern Analysis for Regional Command South, Afghanistan. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 166 Relative Transparency of Aggregated and Disaggregated Data . . . . . . . . . . . . . . . . . . 173 Efects-Based Mathematical Model with Weighting Scheme . . . . . . . . . . . . . . . . . . . . 176 How Aggregation Hides Context . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187 Civilian Casualty Data, Iraq, 2007 and Early 2008. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197 he ISAF Joint Command District Assessment. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201 he Efect of Averaging Color Codes in the District Assessment Model . . . . . . . 202 he Efect of Aggregating Averaged Color Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203 Analysis of the Efect of Aggregating Averaged Color Codes . . . . . . . . . . . . . . . . . . . . 203 Centralized Impressionistic Decisionmaking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 210 Subjectivity in Commanders’ Assessments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 224 Contextual Impressionistic Decisionmaking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 231 he Contextual Assessment Reporting Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237 ix x Embracing the Fog of War: Assessment and Metrics in Counterinsurgency C.1. E.1. Map of Phong Dinh Province, Vietnam . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 262 Comparison of Systems Analysis Diagrams in Joint Doctrinal Publications . . . 287 Tables and Text Boxes Tables 1.1. 4.1. 4.2. 6.1. 7.1. 7.2. 9.1. 9.2. Campaign Assessment Versus Intelligence Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 Efects-Based Assessment Terminology in U.S. Army Doctrine . . . . . . . . . . . . . . . . . . . 67 Example of Measures of Efectiveness Spreadsheet with Indicators . . . . . . . . . . . . . . . 71 Performance of Vietnam-Era Assessment Against Assessment Standards . . . . . . 150 Regional Command Assessment Processes as of Late 2010. . . . . . . . . . . . . . . . . . . . . . . 158 Performance of Afghanistan Assessment Against Assessment Standards . . . . . . . 170 Efects-Based Assessment Compared to the Nature of Counterinsurgency . . . . 212 Comparison of Centralized Assessment Methods to Assessment Standards . . . 214 Text Boxes 4.1. 5.1. Example Progress Indicators in Counterinsurgency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 70 Measures of Efectiveness and Indicators from Field Manual 5-0 . . . . . . . . . . . . . . . . . 93 xi Summary his monograph examines the U.S. military assessment process for counterinsurgency (COIN) campaigns. It focuses on the methods employed to develop and present to policymakers these theater-level assessments of ongoing campaigns. In support of this process, it captures the complexity of the COIN operational environment, examines case studies of COIN assessment in context, explores critical weaknesses in the current assessment process, and ofers recommendations for improvement. Sound strategic decisionmaking in military campaigns relies, in part, on the quality of campaign assessments. he quality of an assessment, in turn, relects the methodology used to produce the assessment report or other materials: An assessment derived from a poorly conceived methodology might mislead decisionmakers, while a well-conceived assessment process might help shape a winning strategy. U.S. military campaign assessments are developed by a theater commander (for example, GEN William C. Westmoreland during the Vietnam War or Gen. John R. Allen during the current war in Afghanistan) for senior military and civilian decisionmakers. To produce a campaign assessment, military stafs currently collect and analyze information about a campaign’s progress and then develop narratives, brieings, and other materials that can be provided to decisionmakers. he U.S. Department of Defense (DoD) and policymakers in Congress and the executive branch rely on these ield reports to help them allocate (or reallocate) resources, identify trends that may indicate success or failure, and ascertain whether and when it may be necessary to alter a given strategy. Figure S.1 provides an overview of the U.S. military’s current approach to COIN campaign assessment. he assessment process begins with strategic direction from policymakers to the theater commander and staf.1 he theater commander then interprets the request and produces operational orders for the units under his command. 1 A more detailed schematic would show several layers of processing between the policymakers and the theater commander (e.g., the National Security Council, the combatant commander). his igure is purposefully simpliied to convey the general concept and process of assessment throughout the U.S. military. his process, the different layers of command, the relationships among actors, and the characteristics of the COIN environment are discussed in greater detail later in this monograph. xiii xiv Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure S.1 Overview of the COIN Campaign Assessment Process Strategic policy Policymaker Theater commander and staff Campaign assessment Feedback Input Intelligence agencies and advisors Input Operational commanders COIN environment Feedback Actions Operational orders Feedback Operational orders Tactical commanders RAND MG1086-S.1 hese orders are reinterpreted and executed at various organizational levels in theater, from the operational to the tactical. hese actions in the COIN operating environment produce feedback. his feedback is then passed back up through the chain of command to the theater commander and staf, who use the information, along with other inputs, to produce a campaign assessment. he commander delivers this assessment to the policymakers, who combine it with other resources (e.g., intelligence agency reports, advisory briefs) and reshape policy, beginning the cycle again.2 he military is most capable of producing clear and useful campaign assessments when provided with clearly articulated and understandable strategic objectives. In the absence of such guidance, military stafs are left scrambling to design operational objectives, and then assessments, that seem reasonable—and assessments derived from an unclear strategy are unlikely to satisfy anyone. With a clear, well-deined strategic policy in place, the military can develop an assessment process to answer questions that relate speciically to policymaker requirements. In a best-case scenario, assessments should be tied directly to national policy objectives, and policymakers should be able to view such assessments with relative conidence. he U.S. military has taken two broad approaches to assessing the progress of COIN campaigns. he irst approach—efects-based assessment (EBA)—attempts to 2 he igure also shows how information, or “inputs,” from other sources in the COIN environment (e.g., civilian agencies, host-nation organizations) can be incorporated into the assessment process. Feedback from policy action implemented in theater is only one type of information used for assessment; a great deal of information is derived from activities or observations that are not directly associated with friendly actions. Summary xv pinpoint and quantify events on the ground to produce centralized and highly accurate reports. he second approach—pattern and trend analysis—uses centralized quantitative analysis to produce a more-or-less impressionistic or, in some cases, deterministic understanding of campaign momentum. Both these approaches are centralized and rely to a great degree on quantitative measurement. No other comprehensive approach has been described or advocated in the literature on military COIN assessment.3 In practice, the military has relied on an ad hoc approach to COIN assessment that lies somewhere between EBA and pattern and trend analysis. Neither of these two centralized assessment methods is practical for COIN because, according to U.S. military doctrine, COIN is best practiced as a decentralized type of warfare predicated on “mission command.” Indeed, doctrine envisions the COIN battlespace as a “mosaic” of localized situations, challenges, and possible solutions.4 Decentralization and mission command necessitate a loosely structured, localized approach to prosecuting war. Such an approach allows commanders at the tactical level to assess local conditions and implement initiatives tailored to address local challenges. Due to the decentralized nature of COIN campaigns, few activities are generalizable or consistently implemented across a COIN theater of operations. his situation is exacerbated by the complexity and uncertainty inherent in the environment. It would be diicult (if not impossible) to develop a practical, centralized model for COIN assessment because complex COIN environments cannot be clearly interpreted through a centralized process that removes data from their salient local context. he incongruity between decentralized and complex COIN operations and centralized, decontextualized assessment has led military stafs to rely on ad hoc assessment methods that leave policymakers and the public dissatisied with U.S. COIN campaign assessments. his monograph examines and critiques EBA and pattern and trend analysis in an efort to explain why policymakers and the public tend to be dissatisied with U.S. military assessments of COIN campaigns. hat phase of the study involved 16 months of research, a detailed examination of two case studies and a review of a third case study, interviews and discussions with hundreds of operations and assessment oicials from the battalion to the DoD level, and exchanges with policymakers and their stafs. Among those who ofered their perspectives, there were signiicant concerns about the efectiveness and methodology of the current assessment process as it is being applied in support of COIN campaign assessment in Afghanistan. To provide historical con3 he few alternatives to EBA that have been ofered are centralized and quantitative and do not represent a signiicant departure from EBA. Most of the assessment literature relies on centralized, quantitative systems analysis methods to frame the COIN operating environment and to assess COIN campaigns. Although a review of COIN assessment methods did not uncover a viable alternative, this does not preclude the possibility that one exists but was simply not captured by the literature review. 4 According to Joint Publication (JP) 3-24, Counterinsurgency Operations (U.S. Joint Chiefs of Staf, 2009c, p. III-19), “he mosaic nature of COIN is ideally suited to decentralized execution.” xvi Embracing the Fog of War: Assessment and Metrics in Counterinsurgency text for an examination of the current assessment process, the study also considered the case of COIN campaign assessment and outcomes during the Vietnam War era. he literature on the U.S. military’s Vietnam-era assessment process was highly negative; even the most strident defenders of the assessment reports produced during that period expressed dissatisfaction with the process. Although, as discussed earlier, the assessment literature contains no clearly articulated alternative to the two accepted forms of COIN assessment (EBA and pattern and trend analysis), advocates of a traditionalist approach to warfare have critiqued the theories behind EBA and have voiced doubt that any quantitative approach could produce assessment materials with greater utility at the policymaking level.5 Traditionalists, including General James N. Mattis, adhere to strict interpretations of capstone service doctrine that describe warfare as inherently complex and chaotic. Some advocates of EBA and most advocates of pattern and trend analysis also acknowledge the complex and chaotic nature of war, but they argue that centralized assessment can suiciently overcome these challenges to inform policy. Traditionalists argue that complexity and chaos cannot be adequately overcome through centralized and quantitative analysis. Despite this debate, no viable alternative to centralized assessment has emerged from these traditionalist critiques or from other quarters. he study also produced a framework of standards for COIN assessment. hese standards were derived from a detailed literature review, interviews with experts and practitioners, exchanges with subject-matter experts, and direct observation of ield assessment activities. he irst two standards—transparency and credibility— are deemed requisite for successful COIN assessment in a democracy. All assessment reports should be relevant to policymakers; this is clear in the campaign assessment process as presented in Figure S.1. he other four standards are practical requirements that should be met to establish an efective methodology. he seven standards are as follows: 1. Transparent: Transparent assessment reports are widely releasable to oicial consumers and, ideally, are unclassiied and without distribution restrictions; the inal iteration of the report is suitable for distribution on the Internet. Such reports reveal both the methods and data used at each level of assessment, from tactical to strategic, and allow for detailed and in-depth analysis without requiring additional research or requests for information. Any subjectivity or justiication for disagreements between layers of command is explained clearly and comprehensively. 5 “Traditionalist” is a nondoctrinal, unoicial term used to help frame the discussion presented here. It should not be confused with “traditional warfare” as described in JP 3-24, Counterinsurgency Operations (U.S. Joint Chiefs of Staf, 2009c, p. I-8). Here, traditional warfare is equated with conventional warfare. Traditionalists adhere to the theories of capstone doctrine and maneuver warfare, as discussed in greater detail in Chapter Two. Summary xvii 2. Credible: Credible assessment reports are also transparent, because opacity devalues any report to the level of opinion: If the data and methods used to produce the report cannot be openly examined and debated, consumers have no reason to trust the reporter. In credible reports, all biases are explained, and laws in data collection, valuation, and analysis are clearly articulated. Data are explained in context. Such reports clearly explain the process used by commanders and stafs to select methods and are both accurate and precise to the greatest degree possible. 3. Relevant: Relevant reports efectively and eiciently inform policy consideration of COIN campaign progress. hey are suicient to help senior military leaders and policymakers determine resourcing and strategy, and they satisfy public demand for knowledge up to the point that they do not reveal sensitive or classiied information. 4. Balanced: Balanced reports relect information from all relevant sources available to military stafs and analysts, including both quantitative and qualitative data. Such reports relect input from military commanders at all levels of command and are broad enough in scope to incorporate nonmilitary information and open-source data. Balanced reports include countervailing opinions and analysis, as well as data that both agree with and contradict the overall indings. 5. Analyzed: Ideally, inished reports do not simply relect data input and the analysis of single data sets. hey also contain thorough analyses of all available data used to produce a uniied, holistic assessment. his analysis should be objective and predicated on at least a general methodological framework that can be modiied to it changing conditions and, if necessary, challenged by consumers. he requirement for holistic analysis is commonly voiced in the assessment literature. 6. Congruent: COIN assessment theory should be congruent with current U.S. joint and service understanding of warfare, the COIN environment, and the way in which COIN campaigns should be prosecuted. he standard for congruence was drawn from a combination of joint doctrine and service capstone doctrine, as discussed in greater detail in Chapter Two. 7. Parsimonious: Assessment cannot show all aspects of a COIN campaign, nor should it seek or claim to deliver omniscience. Collection and reporting requirements for assessment should be carefully considered relative to the demands and risk they may leverage on subordinate units. Assessment should rely to the greatest extent possible on information that is generated through intelligence and operational activities, without requiring additional collection and reporting. his framework shaped the study’s examination and critique of the current U.S. military COIN assessment process. It also provides a starting point for identifying potential alternatives or improvements to the current process. While the examination xviii Embracing the Fog of War: Assessment and Metrics in Counterinsurgency of EBA and pattern and trend analysis is critical, the ultimate purpose of this monograph is to help improve operational and strategic doctrine and practice for COIN assessment. he recommendations presented here are designed to support military commanders in their eforts to produce credible COIN assessments for policymakers and the public. he most important recommendation to improve COIN assessment is to ensure that the overall process includes contextual assessment. he concept of contextual assessment as discussed here is framed by the seven criteria identiied as useful standards for assessment. his monograph develops and uses this concept as a springboard for further exploration of alternative options to the current approach to COIN assessment. It is not intended to be the only option available to enhance the assessment process, nor is it necessarily the best option available. It is, however, intended to broaden the conceptual understanding of assessment methodology and pave the way for improvement. Findings he indings presented here call into question the use of centralized quantitative approaches to COIN assessment. An examination of theater-level assessments conducted in Vietnam, Iraq, and Afghanistan shows that U.S. COIN assessment has long relied on centralized and decontextualized quantitative methods at the expense of context, relevant qualitative data, and comprehensive analytic methods. his conclusion does not suggest, however, that quantitative data are not useful or that they are less valuable for COIN assessment than qualitative data. Rather, quantitative and qualitative data should be assessed equally, in context, and according to appropriate criteria. Further, because the indings presented here are narrowly focused on campaign COIN assessment, they have no proven bearing on the relative merit of quantitative methodologies used to assess other aspects of COIN operations (e.g., logistics), conventional warfare, or in any other military endeavor. Key Finding Context is critical: It is not possible to assess COIN campaign progress through centralized measurement because this approach cannot relect the contextual nuance of a COIN campaign. Assessment tells policymakers what has happened so they can try to gauge progress or failure. Analysis, as a function of either assessment or intelligence, tells policymakers why events occurred and how they might unfold in the future. Eforts to measure (rather than assess or analyze) COIN campaigns using aggregated and centralized quantitative methods provide neither of these services efectively. Centralized quantitative reports (e.g., graphs, color-coded maps, aggregated tables) tend to contain data that are inaccurate to a signiicant degree; they also lack context and fail Summary xix to account for qualitative inputs. Consequently, such reports often produce inaccurate or misleading indings. When they do identify recognizable patterns, these patterns may be misleading or insuicient to support policy decisions. Trends that are tied to mathematical thresholds can convey unrealistic accuracy and quantitative thresholds and milestones that are highly subjective. Patterns and trends can support assessment, but they cannot stand alone. Efective assessment depends on capturing and then relating contextual understanding in a way that is digestible to senior decisionmakers. The Realities of Policy At least for the duration of the Afghanistan campaign, policymakers will continue to require independent sets of aggregated quantitative metrics to support decisionmaking. Assessment stafs should make every efort to compensate for the inherent inadequacies and laws in these metrics to support policymaker requirements. his monograph describes best practices in assessment as identiied through empirical research, but it is important to note that best practices are not always attainable in the face of wartime friction and attendant bureaucratic realities. When required to measure the progress of a COIN campaign, civilian oicials and military oicers should strive to place metrics in context and to carefully tie visual representations of data to explanatory narratives. hey should balance quantitative metrics with probability and accuracy ratings and also identify and explain gaps in the available information. All information should be clearly sourced and should be retrievable by policy stafs seeking an in-depth understanding of speciic subjects. If possible, all quantitative reports should be presented as part of holistic, all-source analysis. Transparency and analytic quality might enhance the credibility of aggregated quantitative data. General Findings on Assessment in Counterinsurgency his research led to a number of indings and conclusions associated with the role of assessment in informing the future direction of COIN campaigns, as well as the beneits to those responsible for decisionmaking in support of such campaigns. An effective theater-level assessment is transparent, credible, and relevant to policymakers. Western democracies depend on sustainable public support to prosecute what are often lengthy COIN campaigns. Building and sustaining trust in oicial government assessments requires a process that is both transparent and credible. Transparent reports are (preferably) unclassiied and share with consumers the underlying data and the methods used at each level of assessment. Credibility is derived from transparency, accuracy, and the use of a sound methodology that captures the relevant context. Relevant reports are those that successfully and eiciently inform policy and satisfy the public need for information. Centralized assessments built from aggregated data do not produce transparent or credible reports. Centralized, uniform, top-down assessment requirements, aggregation, and decontextualization can undermine transparency by obscuring data xx Embracing the Fog of War: Assessment and Metrics in Counterinsurgency sources or the reasoning behind subordinate assessments. An assessment methodology that does not adhere to the tenet of transparency can undermine the credibility of a COIN campaign among consumers of the assessment report and the public. There is general agreement that information and data sets in COIN assessment materials are inaccurate and incomplete to a signiicant yet unknown degree. U.S. military doctrine, the COIN literature, at least two U.S. Secretaries of Defense, and senior assessment analysts reviewing case studies from Vietnam, Iraq, and Afghanistan concur that much of the information that is collected and fed into the COIN assessment process is both inaccurate and incomplete to a signiicant yet unknown degree. COIN information is best analyzed at the level at which it is collected. he COIN environment is complex and marked by sometimes-extreme variations in physical and human terrain, varying degrees of progress from area to area (and often from village to village), and varying levels of counterinsurgent presence and collection capability. herefore, information can have very diferent meanings from place to place and over time. Holistic military campaign assessment is not scientiic research. Sometimes, campaign assessment is likened to scientiic research. As this examination shows, assessment is not scientiic research, and most assessment data and methods cannot be held to strict scientiic standards. Because assessment is not scientiic research, campaign assessment reports should not be held to scientiic standards, nor should attendant data be presented as objective research indings. his conclusion relates to the study’s key inding that COIN assessment must be considered within the appropriate context. What holds true for one COIN campaign may not for another campaign being conducted elsewhere. It may be possible to identify some generalized patterns of behavior or trends in activity from aggregated quantitative metrics, but patterns and trends do not constitute assessments. Data can sometimes indicate very broad patterns of behavior or trends over time, but few patterns or trends can show clear cause and efect. hose that show correlation can be useful in some cases but usually only to help clarify more detailed assessments. Furthermore, inaccurate data can produce precise but inaccurate patterns and trends, which, in turn, can mislead both assessment stafs and policymakers. Many aggregated pattern and trend graphs can be interpreted as having diametrically opposed meanings by diferent consumers because these graphs do not relect causal evidence. Centralized, top-down assessment collection requirements are unparsimonious, they can unhelpfully shape behavior at the tactical level, and they undermine broadly accepted tenets of irregular warfare. he use of a centralized list of metrics (i.e., a “core metrics list”) for assessment sets what is essentially a standard for performance at the tactical level of operation—a standard that is often used to gauge the success or failure of military commanders. his informal standard of performance may encourage a misdirection of COIN eforts and an inappropriate expenditure of Summary xxi resources because not all areas across a COIN theater of operations require the same application of techniques or resources at any one point in time or over time. For example, a core metric may require subordinate units to report the number of schools in each area of operation, but while schools may be critical to success in one area, they may be far less meaningful in another. Regardless of conditions, both units might feel compelled to focus on building or supporting schools due to the inluence of the core metric. Trying to ind the “right” metrics is an exercise in futility. There are no inherently “good” or “bad” metrics in COIN. No single metric is necessarily good or bad. For example, it would not be correct to say, “A reduction in violence is always good,” because such an outcome could be the result of coalition withdrawal leading to defeat, a change in insurgent tactics, or insurgent victory. In speciic cases, the oft-maligned body count metric might be useful, but only if these data are supported by evidence of progress toward a speciic objective (e.g., a unit identiies 50 foreign ighters and the objective is to eliminate that unit of foreign ighters). It may be helpful to describe the potential strengths and weaknesses of various metrics in various contexts, but these descriptions should not be taken as prescription. Standing U.S. military doctrine on COIN assessment is, with few exceptions, incongruent with the literature and doctrine on COIN operations and environments. Joint and some service doctrine calls for the application of traditional COIN fundamentals in the form of bottom-up, distributed operations while simultaneously recommending a top-down, centralized assessment of those operations. Efects-based language that should have been removed by 2008 persists in joint doctrine and in practice in 2011. Efects-based theory has challenged the ability of the assessment process to usefully serve consumers at the tactical and policy levels. COIN assessment is a poorly understood process. Few military oicers, commanders, or policymakers have a strong understanding of the inner workings of COIN assessment. Neither military training nor military education adequately addresses COIN assessment, and there is little or no interagency collaboration on assessment in the U.S. government. Furthermore, assessment is rarely integral to U.S or NATO COIN campaign planning and execution. Counterinsurgency Assessment in Afghanistan Commanders and staffs must make due with the assessment tools they have on hand in the absence of a clearly deined, substantiated process. Because literature and experience with modern COIN assessment are limited and doctrine provides inadequate and sometimes contradictory guidance on assessment, commanders and stafs must make due with sometimes inventive but generally ad hoc solutions. The drive to create large data repositories for core metrics to meet requirements in Afghanistan has proved to be ineffective and, in some ways, counterproductive. Inequities generated by “one-size-its-all” collection requirements and weak xxii Embracing the Fog of War: Assessment and Metrics in Counterinsurgency communication infrastructure have prevented the development of a comprehensive, centralized report database. he centralized databases that exist are incomplete and contain an unknown number of inaccuracies. Even if these databases were sound, they would permit and encourage analysis of raw and aggregated data devoid of context. Assessment analysts who work with these databases spend an inordinate amount of time rectifying and categorizing data instead of analyzing data that have already been contextualized by subordinate units. As of early 2011, assessment was not synonymous with analysis. At least until early 2011, campaign assessment reports tended to contain very little comprehensive analysis and often included summaries of subordinate analyses that did not follow a common methodology. heater assessments were often little more than summations of data reporting combined with uncoordinated and inconsistent subordinate insights. he International Security Assistance Force (ISAF) in Afghanistan has recently attempted to incorporate analysis into assessment, but these eforts are limited and do not relect a comprehensive shift in approach to COIN campaign assessment. Assessments of the Afghanistan COIN campaign suffer from a lack of transparency and credibility. he various assessments of the Afghanistan campaign relect a mix of EBA-like measurement and pattern and trend analysis. Reports tend to emphasize measurement. his emphasis, in turn, relects U.S. efects-based doctrine that is heavily reliant on perception of efects, core metrics lists, and the presentation of aggregated or otherwise decontextualized data. Reports tend not to explain inconsistencies inherent in COIN assessment. Methods of data collection, data analysis, and assessment reporting vary from level to level, laterally from unit to unit, and over time. While the realities of COIN may require variation in assessment, there has been little efort to capture or explain these variations for consumers. Centralized military assessment in coalition and whole-of-government COIN warfare (e.g., ISAF in Afghanistan) does not and cannot relect consistent or controlled information or data collection. Producing even a nonscientiic centralized assessment based on pattern and trend analysis requires some degree of control over the collection of information and the inputs to the assessment process. But contemporary COIN demands a whole-of-government and, often, a coalition approach. With some exceptions, an external military force (e.g., the United States in Vietnam, Iraq, or Afghanistan) cannot explicitly task nonmilitary agencies—and can only indirectly task allies—to collect and report information or data. Similarly, host nation agencies may not respond to U.S. military assessment requirements or may respond to data requests selectively. Sets of data managed by U.S. or coalition forces often do not relect input from all areas of the battleield under coalition control or from all echelons of command. Summary xxiii Recommendations he study’s indings point to ive primary recommendations for those involved with COIN campaign analysis at both the tactical and policy levels. Conduct a thorough review of U.S. military assessment doctrine. Both U.S. joint and service stafs should conduct reviews with the purpose of rectifying contradictions and inconsistencies in doctrine. Train and educate military staff oficers to understand and incorporate assessment into COIN campaigns. he services should consider incorporating campaign assessment theory and practice into training and education programs that address COIN planning and operations. Conduct a review of interagency COIN assessment. An interagency panel should address the lack of coordination between DoD and nonmilitary government agencies that are likely to participate in COIN campaigns (e.g., the U.S. Department of State, U.S. Department of Agriculture, U.S. Agency for International Development). Incorporate all-source methodology as used by the intelligence community into the campaign assessment process. Analysis is generally absent from campaign assess- ment at least in part because there is no clearly deined assessment analysis process. Allsource intelligence methodology provides a good framework for assessment and could be used to both improve and structure campaign assessment reports. Implement a decentralized, objective-focused assessment process that incorporates sound analysis, or at least narrative context, at all levels. Any alternative to the current process must operate within the realities of the COIN environment and should aspire to transparency, credibility, and the other standards presented earlier. he data that it incorporates should be analyzed at the level at which they are collected, and this analysis should be balanced by in-depth analyses at higher levels. Assessments written to address campaign or theater-level objectives should be founded on layers of contextual analysis. Layered context will provide what might be described as “analysis in depth.” In addition, all data sources should be clearly cited. he U.S. military should adopt or at least attempt to implement some variation of contextual COIN assessment. A Note About Contextual Assessment Contextual assessment is a framework for COIN assessment that better aligns it with capstone military doctrine and COIN operational doctrine. his comprehensive, bottom-up assessment process builds layers of contextual narrative and data from the battalion to the theater level. he assessment process should be decentralized, but the reporting process should be standardized across the theater to ensure consistency and continu- xxiv Embracing the Fog of War: Assessment and Metrics in Counterinsurgency ity over time.6 Every report—from the battalion to the theater level—would be wholly captured by the inal theater report. In the best-case scenario, analysis would be conducted at each level. he creation of a long-form assessment would ensure that policymakers, the public, and senior leadership have access to detailed justiication for the indings in executive summary reports. More importantly, a comprehensive long-form assessment would provide a sound basis for theater- and policy-level assessments. hus, consumers will not have to read several hundred pages of assessment, analysis, and data, but the information will be available to provide depth, transparency, and credibility as needed. Commanders’ inputs are a critical element of holistic operational COIN assessment. A commander’s coup d’oeil, or eye for the battleield, is indispensable not only for sound decisionmaking in combat but also in helping to assess what has happened and what might happen in a complex combat environment. Military commanders are handpicked for command, and policymakers and the public necessarily rely on them to a signiicant degree in times of war. However, commanders’ assessments are often incorrectly deemed wholly subjective and thus suspect. In the absence of layered contextual analysis, it is diicult for commanders to argue that their personal analysis is more substantive than subjective opinion. Contextual assessment is designed to reinforce commanders’ assessments by giving weight to their analyses. All commanders, from the battalion to the theater level, would beneit considerably from a transparent, credible, and contextual approach to assessment. Figure S.2 shows how battalion reports are folded into regiment- or brigade-level reports and then into regional reports in the proposed contextual analysis framework. he theater report contains each of these component reports and may be several hundred pages in length, but it does contain built-in summaries at each level. By comparison, the unclassiied semiannual DoD assessment report on Afghanistan and Iraq is between 60 and 150 pages and includes only a single summary and almost no operational context or detailed citations. he theater-level report includes narrative justiications and superior analysis of subordinate reports at each level. hese analyses, or critiques, help address the inherent bias of narrative assessment. To further counter bias in the narrative reports, the theater-level analytic team provides a comprehensive analysis of the entire theater report. Finally, the commander provides an overall assessment. he theater-level assessment, the theater-level analysis, and the commander’s assessment contain summaries that can be used to build executive policy papers. he theater commander’s assessment is the top layer of the report, as shown in the igure. 6 A template for contextual assessment is presented in Appendix A. Appendix B provides a notional example of contextual assessment in practice. Summary xxv Figure S.2 The Contextual Assessment Reporting and Analysis Process Battalion report Brigade/ regiment report Regional report This assessment addresses the period This assessment addresses the period This assessment addresses the period This assessment addresses the period May. Over the between 01 April and 01 May past month a great deal of progress has been made towards achieving the mission. This is specifically shown in the reduction of violence that has been directly tied to a successful amnesty program and recent combat operations through intelligence reporting. The economic situation in the area has improved significantly over the past month, as shown in the growth of local markets that has been directly tied to increase in cross-border trade, as identified in civil affairs reporting. This information also shows increased growth Theater report This assessment addresses the period 01 April and 01 May. Over the between between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the been made towards achieving themade This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reductionpast of violence has been month athat great deal of progress has reduction of violence that has been directly tied to a successful amnesty been made towards achieving the the period This assessment addresses directly tied to a successful amnesty program mission. and recent combat operations This is specifically shown in the Over the May. between 01 April and 01 May through intelligence The hasand program recent combat operations reductionpast ofreporting. violence been month athat great deal of progress has the period This assessment addresses economicdirectly situation in to the area has through intelligence reporting. The tied a successful amnesty beenover made towards achieving the between 01 April and 01 May. Over the improvedprogram significantly the past and recent operations economic situation the hashas mission. Thiscombat isof specifically shown in the past month a greatin deal of area progress month, asthrough shown in the growth local intelligence The has been reduction ofreporting. violence that been towards achieving the past improved significantly over the markets that has been directly tied to made economic situation in to the has amnesty directly tied a area successful mission. This is specifically shown in the increase in cross-border trade, asover month, as shown in the growth of local improved significantly the past operations program and recent combat reduction of violence that has been affairs reporting. This identifiedmonth, in civil as af shown markets in the growth of has localbeen that tied to through intelligence reporting. The directly directly tied to a successful amnesty information also shows increased growth markets that has been directly to has economic situation in tied the area program and recent combat operations increase in cross-border trade, as increase in cross-border trade, asover the past improved significantly through reporting. The This identified inintelligence civil reporting. identifiedmonth, in civil as affairs reporting. This affairs shown in the growth ofin local economic situation the area has information also shows increased growth information also shows increased growth markets that has been directly tied toover the past improved significantly increase in cross-border month, astrade, shownasin the growth of local identified in civil affairsthat reporting. This markets has been directly tied to information also showsin increased growth increase cross-border trade, as identified in civil affairs reporting. This information also shows increased growth This assessment addresses the period 01 April and 01 May. Over the between between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the been made towards achieving themade This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reductionpast of violence has been month athat great deal of progress has reduction of violence that has been directly tied tomade a successful amnesty been towards achieving the the period This assessment addresses directly tied to a successful amnesty program mission. and recent combat operations This is specifically shown the Over the between 01 April and 01in May. through intelligence The has program and recent combat operations reductionpast ofreporting. violence that been month ahas great deal of progress has economicdirectly situation in to the through intelligence reporting. The tied a area successful amnesty been made towards achieving the improvedprogram significantly over the past and recent operations economic the area has mission. Thiscombat isofspecifically shown in in the month, asthrough shown in the growth localsituation intelligence reporting. The has been reduction of violence that improved over the past markets that has been directly tied to significantly economic situation in to the has amnesty directly tied a area successful increase in cross-border trade, asover month, as in the growth of local improved significantly theshown past operations program and recent identifiedmonth, in civil as affairs reporting. Thiscombat shown in the growth of has localbeen markets that through intelligence reporting. The directly tied to information also shows increased growth markets that has been directly to has economic situation in tied the area increase in cross-border trade, as increase in cross-border trade, as improved significantlyin over the past identified civil reporting. This identifiedmonth, in civil as affairs reporting. This affairs shown in the growth of local information also shows increased growth the period information also shows increased growth This assessment addresses markets that has been directly tied to increase in cross-border trade, as between 01 April and 01 May. Over the This assessment addresses identified in civil affairs reporting. This the period This assessment addresses the period past month great information also shows increased growth between 01 April and 01 a May. Overdeal the of progress has between 01 April and 01 May. Over the made towards the month a great deal of progressachieving has past month a great deal ofpast progress hasbeen been made towards achieving the been made towards achieving the mission. This is specifically shown in the This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reduction of violence that has been reductionpast of violence has been month athat great deal of progress has reduction of violence that has been directly tied to a successful amnesty been made towards achieving the the period directly tied toamnesty a successful amnesty This assessment addresses directly tied to a successful program mission. and recent combat operations This is specifically shown in the Over the May. between 01 April and 01 May through intelligence The hasand program and recent combat operations program recent combat operations reductionpast ofreporting. violence been month athat great deal of progress has economicdirectly situation in to the has through intelligence reporting. The tied a area successful amnesty through been made towards achieving the intelligence reporting. The improvedprogram significantly over the past operations and recent economic the area has mission. Thiscombat isofspecifically shown in in the month, asthrough shown in the growth localsituation economic intelligence The has been situation in the area has reduction ofreporting. violence that improved over the past markets that has been directly tied to significantly economic situation in to the has amnesty significantly over the past improved directly tied a area successful increase in cross-border trade, asover month, as shown in the growth of local improved significantly the past operations program and recent combat affairs reporting. This identifiedmonth, in civil as af shown markets in the growth month, of has localbeen as directly showntied in the that to growth of local through intelligence reporting. The information also shows increased growth markets that has been directly to has economic situation in tied the area increase in cross-border trade, that hasasbeen directly tied to increase in cross-border trade, markets as improved significantly the past identified inover civil reporting. This trade, as identifiedmonth, in civil as affairs reporting. This affairs increase in cross-border shown in the growth of local information also shows increased growth information also shows increased growth markets that has been directly tied to in civil affairs reporting. This identified increase in cross-border trade, as information also shows increased growth This assessment addresses identified in civil affairs reporting. This the period This assessment addresses the period information also shows growth between 01 increased April and 01 May. Over the between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the been made towards achieving themade This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reductionpast of violence has been month athat great deal of progress has reduction of violence that has been directly tied tomade a successful amnesty been towards achieving the the period This assessment addresses directly tied to a successful amnesty program mission. and recent combat operations This is specifically shown the Over the between 01 April and 01in May. through intelligence The has program and recent combat operations reductionpast ofreporting. violence been month athat great deal of progress has the period This assessment addresses economicdirectly situation in to the has through intelligence reporting. The tied a area successful amnesty been made towards achieving the between 01 April and 01 May. Over the improvedprogram significantly over the past and recent operations economic situation the hashas mission. Thiscombat isof specifically shown in the past month a greatin deal of area progress month, asthrough shown in the growth local intelligence reporting. The has been reduction of violence that been made towards achieving the past improved over the markets that has been directly tied to significantly economic situation in to the has amnesty directly tied a area successful mission. This is specifically shown in the increase in cross-border trade, asover month, as in the growth of local improved significantly theshown past operations program and recent reduction of violence that has been identifiedmonth, in civil as affairs reporting. Thiscombat shown in the growth of has localbeen markets that tied to through intelligence reporting. The directly directly tied to a successful amnesty information also shows increased growth markets that has been directly to has economic situation in tied the area program and recent combat operations increase in cross-border trade, as increase in cross-border trade, as improved significantly the pastreporting. through intelligence The This identified inover civil reporting. identifiedmonth, in civil as affairs reporting. This affairs shown in the growth ofin local economic situation the area has information also shows increased growth information also shows increased growth markets that has been directly tied toover the past improved significantly increase in cross-border month, astrade, shownasin the growth of local identified in civil affairsthat reporting. This markets has been directly tied to information also showsin increased growth increase cross-border trade, as identified in civil affairs reporting. This information also shows increased growth Theater-level analysis This assessment addresses the period 01 April and 01 May. Over the between between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the the period been made towards achieving themade This assessment addresses This assessment addresses the period mission. This is assessment specifically shown the This addresses the period This isthe specifically shown in theOver the between 01 Aprilmission. and 01in May. Over between 01 April and 01 May. reductionpast of violence that has been between 01 April and 01 the month a great deal ofMay. progress has reduction ofOver violence that has been month directly tied tomade a successful amnesty past month a great deal ofpast progress has a great deal of progress has been towards achieving the directly tied tomade aThis successful amnesty been towards achieving the the period program mission. and recent operations been made towards achieving the assessment addresses Thiscombat is specifically shown in the through intelligence reporting. The mission. This is specifically shown in the This assessment addresses the period program and recent combat operations mission. This is specifically shown in theOver the reduction of violence that has been between 01 April and 01 May. economicdirectly situation inof the has reduction violence that has been between 01 April and 01 May. Over the through intelligence reporting. Thehas tied to a area successful amnesty reduction of violence that been past month a great deal of progress has improvedprogram significantly over the past directly tied to month a successful amnesty past a great deal of progress has and recent combat operations economic situation intomade the area has amnesty directly tied a successful month, asthrough shown in the growth of local been towards achieving the program and recent combat operations been made towards achieving the intelligence reporting. The This assessment addresses the period improved significantly over the past operations markets that has been directly tied to through intelligence reporting. The shown mission. This isarea specifically in the program and combat This isthe specifically shown in the economic situation in the has between 01 Aprilmission. and 01recent May. Over increase in cross-border trade, as economic situation in the area has reduction of violence that has been month, as in the growth of local improved significantly over theshown past through intelligence reporting. Thehas been past month a great deal of progress has reduction of violence that identifiedmonth, in civil as affairs reporting. improved significantly over theof past directly toThis a successful amnesty shown in tied the growth local been made towards achieving the markets that has been directly tied to economic situation area has amnesty This assessment addresses the period directly tiedintothe a successful information also shows increased growth month, as program shown in and the growth of local recent combat operations markets that has been directly tied mission. This is to specifically shown the Over the in cross-border trade, as between 01significantly April and 01in May. improved over the past operations marketsinthat hasincrease been directly tied toprogram through intelligence The has and recent combat increase cross-border trade, as reduction ofreporting. violence been past month athat greatreporting. deal of progress has increase in cross-border trade, asthe situation in has in civil affairs This month, as shown in the growth of local The identified ineconomic civil identified affairs reporting. This through intelligence reporting. directly tied to a area successful amnesty been made towards achieving the identified improved inalso civilshows affairs reporting. Thisshows significantly over the past increased information also growth information increased growth program andThis recent combat operations markets that has beenshown directly tiedarea tothe assessment addresses economic in the hasperiod mission. This isof specifically in the information also shows increased month, asthrough shown in thegrowth growth localsituation intelligence The has inofreporting. cross-border trade, reduction violence that been improved significantly over past markets that hasincrease been directly tied to01 between April andas 01 the May. Over the economic situation in to the has directly tied a area successful amnesty increase in cross-border trade, identified inasover civil affairs reporting. This of local month, as shown in the growth significantly the past This assessment addresses the improved period past month a great deal of progress has program and recent combat operations identified in civilinformation affairs reporting. Thisshows increased growth as shown in thealso growth of has local markets that been directly tiedthe to period This assessment between 01 April andinformation 01 May.month, Over the through intelligence reporting. The addresses also shows increased growth been made towards markets that has been directly to has achieving the past month a great deal of progress has economic situation in tied the area increase in cross-border trade, as May. Over the between 01 April and 01 increase cross-border trade, asover the past been made towards achieving the thein improved significantly mission. This specifically shown This assessment addresses period identified in is civil affairs reporting. Thisin the identified in civil as affairs reporting. This a mission. This is specifically shown in the This assessment addresses the period past month great deal of progress has month, shown in the growth of local between 01 April andinformation 01 May. Over the information also shows increased growth also shows increased growth reduction of violence been the period This assessment addresses reductionpast of violence has been between April and 01 Over the markets that has been directly tied to that has month01athat great deal ofMay. progress has been made towards achieving the directly tied tomade a successful amnesty past month a great deal of progress has increase in cross-border trade, been towards achieving thedirectly tiedbetween to a as successful amnesty 01 April and 01 May. Over the program mission. and recent operations been made towards achieving the This assessment addresses the period identified in civil affairs reporting. mission. This is This specifically shown in the Thiscombat is specifically shown in the through intelligence reporting. The has mission. This is assessment specifically shown in the addresses the period program and recent combat operations past month great deal of progress has information also shows increased growth reduction ofThis violence that been between 01 April and 01 a May. May Over the been reduction of violence that has economicdirectly situation in the area has reduction of violence that has been between 01 April and 01 May May. Over the tied to a successful amnesty through intelligence reporting. The made towards the month a great deal of progressachieving has improvedprogram significantly over the past directly tied to month a successful amnesty past a great deal ofpast progress hasbeen and recent combat operations directly tied to a successful amnesty month, asthrough shown in the growth of localoperations been towards achieving the has program and recent combat been made towards achieving themade economic in the mission. This is area specifically shown in the intelligence reporting. The This assessment addresses thesituation period markets that has been directly tied to through intelligence reporting. The mission. mission. This isarea specifically shown the program recentshown combat operations This isand specifically in past the economic situation in the has between 01 April and 01in May May. Over the improved significantly the reduction ofover violence that has been increase in cross-border trade,in as economic situation the area has reduction of violence has been improved significantly over the past past month athat great deal of progress has reduction of violence that has been through intelligence reporting. The identifiedmonth, in civil as af affairs reporting. improved significantly over theof past directly toThis a successful month, astied shown intied the to growth of local shown in tied the growth localamnesty been made towards achieving the the period directly a successful amnesty This assessment addresses directly tosituation a successful amnesty information also shows increased growth month, as program shown in and the growth of local recent combat operations economic in the area has markets that has been directly tied to mission. This is specifically shown in the between 01 April and 01 has May. Over the markets been directly tied to operations marketsinthat has been directly tied toprogram through intelligence Thethat program and recent combat and recent combat operations increase cross-border trade, as reduction ofreporting. violence has been improved over the past past month athat great deal of significantly progress has increase in cross-border trade, asthe situation in has identified ineconomic civil affairs reporting. This through intelligence reporting. Theasreporting. The directly tied to a area successful amnesty increase in cross-border trade, through been made towards achieving the intelligence identified improved inalso civilshows affairs reporting. Thisthe significantly over past operations information increased month, asshown shown inarea thehas growth of local program andgrowth recent economic situation in the mission. Thiscombat isofspecifically in the information also shows increased growth month, asthrough shown in the growth local identified in civil affairs reporting. This economic situation in the area has intelligence reporting. The reduction of violence that has been improved thedirectly past markets that hasover been tied to markets that has been directly tied to significantly economic situation in to the has information also shows increased growth improved the past directly tied a area successful amnesty increase in cross-border trade, asover month, as shown in thesignificantly growth of localover improved significantly the past increase in cross-border trade, as This assessment addresses the period program and recent combat operations identifiedmonth, in civil as affairs reporting. This month, as directly showntied in the shown in the growth of has localbeen markets that to growth of local between 01 April andinformation 01 May. Over the through intelligence reporting. The also shows increased growth identified in has civil affairs reporting. This markets that has been directly to past month a great deal of progress has economic situation in tied the area increase in cross-border trade, that hasasbeen directly tied to increase cross-border trade, markets asover thealso been made towards achieving the thein information shows increased growth improved significantly past This assessment addresses period identified in civil reporting. This trade, identified in civil as affairs reporting. This affairs increase in cross-border as mission. This is assessment specifically shown the This the period month, shown in the growth of local between 01 April addresses andinformation 01in May. Over the information also shows increased growth also shows increased growth reductionpast of violence that has been between 01 April and 01 May. Over the markets that has beenidentified directly tied to in civil affairs reporting. This month a great deal of progress has directly tied tomade a successful amnesty past month a great deal of progress has increase in cross-border trade, as been towards achieving the program mission. and recent operations been made towards achieving the information shows increased growth This assessment addresses the period identified in civil affairs reporting. This also Thiscombat is specifically shown in the through intelligence reporting. The has mission. This is assessment specifically shown addresses the period information also shows growth reduction ofThis violence that beenin the between 01 increased April and 01 May. Over the This assessment addresses the period economicdirectly situation inof the has reduction violence that has been between 01 April and 01 May. Over the tied to a area successful amnesty past month a great deal of progress has between 01 April and 01 May. Over the improvedprogram significantly over the past directly tied to month a successful amnesty past a great deal of progress has and recent combat operations past month a great deal of progress has achieving the month, asthrough shown in the growth of local been made towards program and recent combat operations been made towards achieving the intelligence reporting. The This assessment addresses the period been made towards achieving the markets that has been directly tied to through intelligence reporting. The mission. This isarea specifically shown the mission. This isthe specifically shown in the economic situation in the has between 01 April and 01in May. Over mission. This is has specifically shown in the This assessment addresses the period increase in cross-border trade,in as economic situation the area has reduction of violence been improved significantly over the past past month athat great deal of progress has reduction of violence that has been reduction of violence has been between 01that April and 01 May. Over the identifiedmonth, in civil as affairs reporting. improved significantly over theof past directly toThis a successful shown in tied the growth localamnesty been made towards achieving thedeal directly tied to a operations successful amnesty past month a great progress hasamnesty This assessment addresses period directly tied tothe aofsuccessful information also shows increased growth month, as program shown in and the growth of local recent combat markets that has been directly tied to mission. This is specifically shown in the Overthe program and recent combat operations been made towards achieving between 01 April and 01 May. the markets that has been directly tied to through intelligence reporting. The program and recent combat operations increase in cross-border trade, as reduction ofintelligence violence that has been through reporting. The mission. is specifically shownhas in the past month aThis great deal of progress This assessment addresses the period increase in cross-border trade, asthe situation in has identified ineconomic civil affairs reporting. This through intelligence reporting. The directly tied to a area successful amnesty economic situation in the area has reduction of violence that has been been made towards achieving the between 01 April and 01 May. Over the identified improved inalso civilshows affairs reporting. Thisthe past significantly over information increased growth program andsignificantly recent economic situation in the hashas improved over the past directly tied to aoperations successful amnesty mission. Thiscombat isof specifically shown in the past month a great deal of area progress information also shows increased growth month, asthrough shown in the growth local intelligence reporting. The month, as program shown the growth of local and recent combat operations reduction ofintied violence that has been been made towards achieving the past improved significantly over the markets that has been directly to economic situation in the area hastied markets that has been through intelligence reporting. The shown in the directly tied to a directly successful amnesty mission. This is to specifically increase in cross-border trade, as month, as shown in the growth of local improved significantly over the past ineconomic cross-border trade, as situation in area has program and recent combat operations reduction ofthe violence that has been identifiedmonth, inincrease civil as affairs reporting. This shown in the growth of local markets that has been directly tied to identified inincreased civil affairs reporting. improved significantly the past amnesty through intelligence reporting. The directly tied over toThis a successful information also shows growth markets that has been tied togrowth information also shows increased month, asdirectly shown the growth of local economic situation in in the area has combat increase in cross-border trade, as program and recent operations increase in cross-border trade, as markets that has been directly to improved significantly theaffairs pasttied through intelligence reporting. The This identified inover civil reporting. identifiedmonth, in civil affairs reporting. This increase ineconomic cross-border trade, as as shown in the growth ofin local situation the area has information also shows increased information also shows increased growth identified in civil reporting. Thisthe past growth markets that has beenaffairs directly tied toover improved significantly information also shows increased increase in cross-border asin thegrowth month, astrade, shown growth of local identified in civil affairsthat reporting. This markets has been directly tied to information also showsin increased growth increase cross-border trade, as identified in civil affairs reporting. This information also shows increased growth Commander’s report NOTE: Theater-level analysis considers all reporting from battalion up to the theater report. The commander’s report addresses both the theater-level report and theater-level analysis. RAND MG1086-S.2 Acknowledgments I ofer my sincere thanks to Stephen Downes-Martin of the U.S. Naval War College and Jonathan Schroden of the Center for Naval Analyses for providing valuable insight into my research and indings. Hriar Cabayan of the Oice of the Secretary of Defense also provided support, access, and insight. he staf of the Information Dominance Center in ISAF Joint Command generously facilitated my research. LtCol Matthew Jones and LTC Vernon J. Bahm of ISAF Joint Command facilitated my research in Afghanistan, as did COL Robert Hume, who was director of the ISAF Afghan Assessment Group in early 2010. LtCol David LaRivee of the U.S. Air Force Academy and LTC Bret Van Poppel from the ISAF Afghan Assessment Group both facilitated my research and ofered very helpful guidance on the current assessment process. At RAND, Walter Perry provided mentoring on the subject of operations research and quantitative analysis and editorial input that greatly improved this monograph. Jerry M. Sollinger helped improve the organization and presentation of the study’s indings. Jason Campbell ofered helpful insight into the vagaries of COIN data sets. Col (ret.) T. X. Hammes, U.S. Marine Corps, and Michael Hix and Martin Libicki at RAND provided critical review and mentorship. RAND colleague Arturo Muñoz also ofered insight during the editing process. Christopher Paul at RAND provided a thorough critique that greatly improved the tone and approach of this monograph. Michael Baranick and Brian Eird at the National Defense University helped arrange forums for discussion and shared insight into modeling and simulations. James N. Bexield at the Oice of the Secretary of Defense helped facilitate discussions of my research and my participation in the NATO assessment review process. I also owe thanks to Philip J. Eles of the Government of Canada, Kyle Pizzey of NATO, MAJ Tad Hunter, MAJ Sang M. Sok at the Center for Army Analysis, COL (ret.) Richard A. Starkey of U.S. Central Command, Joseph Soeters at the Netherlands Defence Academy, COL homas Cioppa and LTC David Hudak at the Army’s Test and Evaluation Command Research and Analysis Center, MAJ Gary Kramlich of the 82d Airborne Division, and many others who helped provide insight into the U.S. military’s current assessment practice. Gail Fisher at RAND helped brainstorm some of the crossover points between social science research and assessment, and Rebecca homasson at the Oice of the Secretary of Defense served as a xxvii xxviii Embracing the Fog of War: Assessment and Metrics in Counterinsurgency sounding board and contributed valuable insight regarding the national assessment process. I would particularly like to thank the librarians and archivists at Texas Tech University’s Vietnam Center and Archive for the valuable service that they provide to researchers and historians of the Vietnam War. Abbreviations AAG Afghanistan Assessment Group ANSF Afghan National Security Forces ARVN Army of the Republic of Vietnam AWG assessment working group BLS U.S. Bureau of Labor Statistics CA contextual assessment CCIR commander’s critical information requirement CIA Central Intelligence Agency CIDNE Combined Information Data Network Exchange CJIATF-435 Combined Joint Interagency Task Force–435 COIN counterinsurgency CORDS Civil Operations and Revolutionary Development Support CUAT Commander’s Unit Assessment Tool DAM District Assessment Model DoD U.S. Department of Defense DoS U.S. Department of State EBA efects-based assessment EBO efects-based operations FM ield manual GST general system theory xxix xxx Embracing the Fog of War: Assessment and Metrics in Counterinsurgency GVN Government of (South) Vietnam HES Hamlet Evaluation System IBC Iraq Body Count IED improvised explosive device IJC ISAF Joint Command ISAF International Security Assistance Force JP joint publication MACV Military Assistance Command, Vietnam MCDP Marine Corps doctrine publication MOE measure of efectiveness MOP measure of performance MPICE Measuring Progress in Conlict Environments MtM mission-type metrics NATO North Atlantic Treaty Organization NC3A NATO Consultation, Command, and Control Agency NCO noncommissioned oicer NCW network-centric warfare NGO nongovernmental organization NVA North Vietnamese Army ONA operational net assessment OSA Oice of Systems Analysis OSD Oice of the Secretary of Defense PAAS Paciication Attitude Analysis System PACES Paciication Evaluation System PRT Provincial Reconstruction Team RC regional command RMA revolution in military afairs Abbreviations SIGACT signiicant activity SoSA system-of-systems analysis TCAPF Tactical Conlict Assessment and Planning Framework USAID U.S. Agency for International Development VC Viet Cong VCI Viet Cong infrastructure xxxi CHAPTER ONE Introduction his monograph examines how combatant stafs and policymakers assess and describe the progress of counterinsurgency (COIN) campaigns and provides an in-depth analysis of the military assessment process for COIN. It also ofers a critique of current doctrine and the performance of military campaign assessment across three case studies (Vietnam, Iraq, and Afghanistan). he purpose of this analysis and critique is twofold: (1) to identify the complexities and challenges of COIN assessment and (2) to frame standards and best practices to facilitate improvements to COIN assessment. he ultimate purpose is to help improve operational and strategic doctrine and practice for COIN assessment. Although the purpose of this research was derived from the sponsor’s objectives and builds on recommendations developed to support units operating in Afghanistan in 2010, the recommendations are designed to support commanders across the U.S. military in their eforts to produce credible COIN assessments for policymakers and the public. Assessment serves a dual purpose in that it tells the story of a campaign (the “up-and-out” assessment to policymakers and the public) while also helping to shape ongoing operations by informing operational and tactical commanders (the “downand-in” assessment for operational units). hese two goals are sometimes met through the production of separate assessment reports; in other cases, theater-level stafs attempt to address both audiences with a single type of report. his monograph focuses on the process used to produce theater-level “up-and-out” reports and recommends strategies to ensure that they provide credible and accurate input to national security strategy. Although it does not comprehensively address the “down-and-in” aspects of military assessment, the recommendations presented here should, if implemented, address both purposes. A thorough review of the assessment literature available through early 2011 revealed no efort to comprehensively address COIN assessment. While no study of this nature can be fully comprehensive, this monograph attempts to break new ground by ofering a holistic examination of COIN assessment. To this end, it describes and explains the theories that shape assessment, the environment in which assessments are developed, and the broad purposes of assessment while exploring the details of process and method. However, this monograph is not comprehensive in that it does not 1 2 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency address all aspects of COIN assessment (see the section “Issues Not Addressed,” later in this chapter). It does go into considerable detail on an array of issues pertaining to assessment. his detail is necessary to provide an in-depth understanding of a subject that has not received suicient attention in current literature. Although the indings presented here will be of interest to a broad audience, the research is speciically intended to inform the process used to assess ongoing COIN eforts in Afghanistan, and it uses Afghanistan as a central case study. Because COIN is prosecuted most often by the ground combat services (the Army and Marine Corps), the research, indings, and recommendations focus most directly on ground combat service or joint COIN assessment as it applies to the ground combat services. Background his section provides a brief description and explanation of assessment across the spectrum of warfare, outlines the diference between assessment and intelligence analysis, and examines policymaker and media dissatisfaction with the type of assessment reports issued during three relevant COIN campaigns: Vietnam, Iraq, and Afghanistan. An assessment is a report or series of reports written to inform operational and strategic decisionmaking.1 Military stafs build the assessments and, at least in theory, military commanders use them to make operational decisions in the ield. hese commanders add their comments to the assessment and then pass it along to policymakers. Assessments are typically delivered in the form of narratives, quantitative graphs, maps, or brieing slides.2 Senior military leaders and policy stafs use these reports to track the progress of military campaigns and to determine how to allocate resources, whether speciic operations or initiatives have met success or failure, and whether and when it may be necessary to alter strategy. Differences Between Campaign Assessment and Intelligence Analysis here are no clear standards for what a U.S. military campaign assessment should deliver or the form in which it should be presented. It is also not clear whether assessment should provide absolute clarity, reasonable understanding, an impression of events, or some combination thereof. A review of the assessment literature and interviews with staf and commanders indicated that assessments should give senior military leaders and policymakers a relevant understanding of what has happened on the 1 his monograph uses the terms strategy and policy interchangeably. Doctrinal or academic diferences between the two concepts are not always consistent, and strategy and policy are often conlated in practice. 2 A number of military assessments (of all types) from Vietnam, Iraq, and Afghanistan are cited throughout this monograph. he following are examples of military assessments: General Westmoreland’s Activities Reports, the Pentagon Papers, the ISAF Joint Command District Assessment map, the 1230-series reports on Iraq and Afghanistan, and General Stanley A. McChrystal’s COMISAF’s Initial Assessment (Unclassiied). Introduction 3 ground and why. In this way, a campaign assessment is very similar to intelligence analysis: It explains what has happened—and sometimes what is happening now— and why this information matters. However, it is diferent from intelligence analysis in that it is holistic and also addresses the performance of friendly organizations. he assessment explains how the entire campaign is progressing toward meeting the policymaker’s strategic objectives. Table 1.1 lists some of the diferences between campaign assessment and military intelligence analysis. hese are broad generalizations drawn from multiple sources and thus intended for general comparison only. he lines between assessment and intelligence analysis are sometimes blurred because military campaign assessment is poorly deined in doctrine and can involve the use of a wide array of methods, including many types of analysis. During the Vietnam War, at least one assessment staf clearly crossed the line into the realm of Table 1.1 Campaign Assessment Versus Intelligence Analysis Characteristic Campaign Assessment Intelligence Analysis Primary purpose Assess progress against operational and strategic objectives Explain behavior and events and predict future behavior and events Process Describes and explains progress, Describes and predicts behavior and recommends shifts in resources, strategy, actions in the environment, informs informs operations courses of action for operations and policy Method Any relevant and useful method All-source analysis using structured, doctrinal methods within prescribed intelligence oversight limits Sources Any available sources, including friendly operations reports and completed intelligence reports Limited to examination of enemy, foreign civilian, and environmental informationa Creators Representatives of all military staff sections and military commandersb Trained intelligence analysts Time frame Shows progress over long periods Timely, degrades in value over time Classification Can be mostly or entirely unclassified Almost always classified or restrictedc a Note that U.S. intelligence oversight regulations restrict the collection and analysis of information on U.S. citizens by designated intelligence organizations. While these policies do not fully preclude this kind of activity, they do regulate the activity to the point that participation is limited to a very few specific programs (e.g., those associated with international terrorism). There appears to be no clear regulation preventing U.S. military intelligence agencies from describing U.S. military activities in broad terms (i.e., without identifying individual U.S. service members), but in practice, military intelligence is confined to describing the ways in which U.S. military activities affect or shape the enemy, the environment, or foreign civilians. b Trained assessment analysts should ideally play a role in the process at some point. As of early 2011, there was no “assessment analyst” position in the U.S. military. Operations researchers and red-team analysts sometimes fill this role. c The exception to this general rule is open-source intelligence, but even many open-source products are restricted once analyzed. 4 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency intelligence analysis (a case described in greater detail in Chapter Six). As a result, this staf produced some reports that did not meet the standards or requirements for either assessment or intelligence analysis.3 While intelligence often feeds assessment, military campaign assessment should not attempt to replicate or replace intelligence analysis. Military commanders and policymakers are best served by receiving both types of reports. here is a inal yet important distinction between campaign assessment and intelligence analysis. While assessment is intended primarily to support policy decisionmaking, democracies also depend on assessment to inform the general public about the progress of a war. At some point, campaign assessment must be published in a transparent and unclassiied document. his unclassiied report must be clear and simple enough to be understood by a wide range of consumers. Intelligence analyses are rarely written with the express intent of public release, and in most cases they are not oicially declassiied until long after the war has ended. The Role and Purpose of Military Campaign Assessment With these distinctions drawn, this chapter turns to a question that serves as the starting point for considering the assessment process: Why is military campaign assessment important and how is it used to shape strategy? he literature review identiied three sources that ofer a general framework of military assessment theory as it is applied to both conventional and irregular warfare. First, Scott Sigmund Gartner describes the essential nature of conventional warfare assessments in Strategic Assessment in War: Leaders assess and, if necessary, alter their strategies based on information they gather from the battleield. . . . hese decisions can have enormous impact. Decisions on strategy play a signiicant role in determining a war’s nature, as well as its duration, intensity, and ultimately who wins and who loses.4 James Clancy and Chuck Crossett describe two broad approaches to conventional warfare assessment: Since World War II, the analysis of warfare has primarily been based upon two major concepts of efectiveness. In the grand movement of military forces, the gaining and control of territory is considered success. hose who control the land 3 his refers to a report on the situation along Highway 4 in South Vietnam by Oice of Systems Analysis Southeast Asia Intelligence Analysis and Force Efectiveness Division. As discussed in Chapter Six, the report was too detailed to serve as a stand-alone campaign assessment for policymakers and simultaneously insuiciently sourced, cited, or assigned appropriate caveats to meet contemporaneous or current standards for intelligence analysis. 4 Gartner, 1997, p. 2. Gartner takes the position that strategic assessment is achieved through the use of quantitative indicators. For additional discussion of the distinction between conventional and COIN assessment, see Clancy and Crossett, 2007. Introduction 5 control the resources, population, and legal structures within it. . . . Physical space is the battleield. he other traditional metric of success is the order of battle (OOB). . . .Such metrics assume large force-on-force battles in a Clausewitzianstyle engagement.5 Clancy and Crossett assert that this is not a useful paradigm in irregular warfare or in COIN, speciically. For COIN, they recommend a model that focuses on “sustainability, legitimacy, and environmental stability.”6 U.S. Army COL Bobby Clalin and panel co-chairs ofered the following deinition of military assessment at an April 2010 conference hosted by the Military Operations Research Society. He refers to COIN assessment, but his remarks could be applicable across the spectrum of conlict: Assessments are critical for an organization to understand how well the organization performs its mission; both what it produces as an output and what it achieves as an outcome. Assessments provide a rigorous analytic bridge between the functioning activities of the organization and those decisions necessary to make improvement; often articulated in policies and plans.7 U.S. military publications contain a range of deinitions for assessment. he United States Army Commander’s Appreciation and Campaign Design highlights ive “opportunities to learn” from an assessment that deine the purposes of the assessment process. Campaign assessments tell commanders, stafs, and policymakers 1. How to execute the planned course of action for a speciic operation; 2. Whether another course of action needs to be adopted; 3. Whether the operational design based on the problem frame is producing results; 4. Whether the problem framing needs adjusting; and 5. Whether the learning mechanisms of the organization are tuned to the particular operational problem.8 he pamphlet states that the Army has focused on the irst two functions but has put insuicient efort into developing functions 3 through 5. he U.S. Army ield manual (FM) 5-0, he Operations Process, reinforces the idea that assessment is a tool for inetuning current and prospective operations. It also states that assessment is a continuous process. he manual deines assessment fundamentals for the Army as follows: 5 Clancy and Crossett, 2007, p. 90. 6 Clancy and Crossett, 2007, p. 96. 7 Clalin, Sanders, and Boylan, 2010, p. 2. 8 U.S. Army Training and Doctrine Command, 2008, p. 18. 6 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Assessment is the continuous monitoring and evaluation of the current situation, particularly the enemy, and progress of an operation (FM 3-0). Assessment is both a continuous activity of the operations process and an activity of battle command. Commanders, assisted by their stafs and subordinate commanders, continuously assess the operational environment and the progress of the operation. Based on their assessment, commanders direct adjustments thus ensuring the operation remains focused on accomplishing the mission.9 his description seems to be focused on the down-and-in or operational purpose of assessment. Assessments are intended to feed operational analysis, but they are also tools for strategic decisionmaking and the primary means by which policymakers in the executive branch explain the progress of a campaign to lawmakers and the public. At least in theory, holistic theater-level assessments, such as the ones produced by the International Security Assistance Force (ISAF) in Afghanistan, are used by the United States and the North Atlantic Treaty Organization (NATO) to gauge campaign progress, determine COIN policy, dedicate resources, shape training and force transformation, and support strategic communication messaging. Dissatisfaction with Counterinsurgency Campaign Assessment Producing a truly holistic and efective COIN campaign assessment has proven to be an elusive goal. he Army points out in he Commander’s Appreciation and Campaign Design that the military has paid insuicient attention to certain aspects of campaign assessment. his conclusion is supported by the indings presented here: he research shows gaps in the process that afect not only strategic assessments but also operational, “down-and-in” assessments. After describing the intent of assessment, Clalin et al. describe some of the weaknesses in the assessment process as has been applied in Afghanistan: Despite the critical role that assessments play, organizations frequently treat assessments as an afterthought. Assessment capabilities are often recognized as lacking well after deployment and are subsequently generated out of the institutional force as a temporary loan. A lack of “operating force” assessment doctrine and analytic structure at echelons above corps may contribute to this assessment lag.10 Shortcomings in campaign assessments delivered to the public have not gone unnoticed in policymaking circles or in the public domain. Obvious gaps and inconsistencies in various reports in all three cases examined in this report (Vietnam, Iraq, and Afghanistan) have made consumers wary of military assessments. A few wellpublicized manipulations of the Vietnam-era COIN reports by military and civilian 9 HQDA, 2010, p. 6-1. 10 Clalin, Sanders, and Boylan, 2010, p. 2. Introduction 7 leaders sowed an underlying layer of distrust among some policymakers and members of the press and public, and this distrust has carried over to Iraq and Afghanistan assessments. H. R. McMaster (Dereliction of Duty), James William Gibson (he Perfect War), and Guenter Lewy (America in Vietnam) have documented the extraordinarily contentious and often disingenuous Vietnam-era assessment process. In his book, McMaster describes the foundation of these misgivings, particularly then–Secretary of Defense Robert S. McNamara’s predilection for shaping and (at least in McMaster’s view) whitewashing Vietnam assessments, beginning as early as 1963: Although he acknowledged that “statistics received over the past year or more from the GVN [Government of (South) Vietnam] oicials and reported by the U.S. mission . . . were grossly in error,” he irmly believed that tracking quantitative indices would give him a clear picture of how the war against the Viet Cong was going. Despite MACV’s [Military Assistance Command, Vietnam] protest that it was “impossible to measure progress in any meaningful way on a weekly basis,” McNamara insisted on “Weekly Headway Reports” that included “measurable criteria” to help chart the progress of the war. he very title of the report revealed his eagerness to demonstrate the South’s improvement under his program.11 In 1969, Congressman John V. Tunney of California delivered a scathing report on the Hamlet Evaluation System (HES)—one of the pillars of the Vietnam-era campaign assessment process—to the Committee on Foreign Afairs in which he used a quote from Lewis Carroll’s hrough the Looking-Glass to describe hamlet assessments. In his conclusion, Tunney stated, “It is diicult, after studying the HES, to understand how our oicials could have put so much uncritical faith in [that system] in the face of opposing facts.”12 Distrust of Vietnam-era military brieings became so endemic that members of the press corps referred to the daily military press brieings as the “ive o’clock follies.”13 Assessments performed during Operation Iraqi Freedom drew critical review from Congress and the press, at least until violence dropped precipitously in 2008, at which point interest in the Iraq War and Iraq policy plummet- 11 McMaster, 1997, p. 58. In 2010, U.S. Secretary of Defense Robert M. Gates told the Senate Appropriations Committee, “his is not something where we do ourselves any favors by tearing ourselves up by the roots every week to see if we’re growing,” in reference to the assessment of U.S. operations in Afghanistan (for written testimony, see Gates, 2010). McNamara explained his actions in the books In Retrospect: he Tragedy and Lessons of Vietnam (McNamara, 1995) and Argument Without End: In Search of Answers to the Vietnam Tragedy (McNamara, Blight, and Brigham, 1999). 12 Tunney, p. 9. HES is examined in some detail in Chapter Six of this monograph. A Brookings Institution study comparing public statements regarding Vietnam assessments and the assessments themselves showed remarkable congruence between the two. he report stated that concerns tended to revolve around the assessments’ optimistic emphasis and denial of speciic actions rather than the perception that there was a deliberate attempt to reshape information coming from the ield (see Gelb and Betts, 1979, p. 320). 13 Hammond, 1988, p. 239. 8 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency ed.14 More recently, some congressional staf members and members of Congress have expressed a deep and sustained dissatisfaction with the military assessments of the Afghanistan campaign.15 he indings presented in this monograph show a distinct linear association between Vietnam-era assessment processes and methods and the assessment processes used to gauge COIN campaign progress in Iraq and Afghanistan. For example, media criticism of contemporary assessment is very similar to that articulated during the Vietnam War. A widely promulgated 2009 commentary by Tom Engelhardt stated, he problem was that none of the oicial metrics managed to measure what mattered most in Vietnam. History may not simply repeat itself, but there’s good reason to look askance at whatever set of metrics the Obama Administration manages to devise [for Afghanistan]. . . . By the time they reach Washington, they are likely to have the best possible patina on them.16 he lack of conidence in military and political assessments of Vietnam, Iraq, and Afghanistan, at times, severely strained or (in the case of Vietnam) helped shatter national consensus and undermined a sustained focus on objectives and an honest debate on strategy. Such a loss of credibility can negate what FM 3-24, Counterinsurgency, refers to as a critical requirement in COIN: gaining and maintaining public support for a prolonged deployment.17 While assessments should not be written with the express intent of shaping public opinion, the failure to design and implement a transparent and credible assessment process can directly undermine national strategy. Literature Review his section discusses the sources used in the literature review portion of the study and also briely examines the state of literature on COIN assessment. he bibliography relects most (if not all) of the literature on COIN assessment published through early 2011. In addition, the literature review included a range of published, unpublished, oicial, and unoicial work across a number of ields. To build a comprehensive picture of assessment, it is necessary to capture lessons from a diverse range of ields, such as systems analysis, military operational art, time-series analysis, COIN doctrine, policy decisionmaking, intelligence analysis and policy, and military leadership. he initial 14 An Associated Press survey of press reporting showed a dramatic decline in press coverage of Iraq after 2007 (see Ricchiardi, 2008). 15 According to discussions with congressional staf members between November 2009 and February 2011, as well as discussions with analysts in Afghanistan and the United States who worked on Afghanistan assessment. 16 Engelhardt, 2009. 17 HQDA, 2006c, p. 1-24. Introduction 9 phase of the review considered a wide array of assessment and measurement publications, including business literature, professional journals on assessment and measurement, systems analysis literature, and publications on operations research.18 he second phase of the literature review had a narrower focus, on assessment theory and doctrine for military and stabilization operations, and included the few published books on military assessment and the somewhat more extensive corpus of journal articles on irregular warfare assessment.19 Relevant doctrinal publications included most joint, Army, and Marine Corps publications on operations and intelligence. A review of the literature on stabilization and development assessment tapped into this narrow ield and included lessons-learned publications from the U.S. Agency for International Development (USAID) and Canadian government sources.20 he examination of efects-based operations and assessment (EBO and EBA) relied extensively on oicial publications and doctrine, as well as journal articles and research reports.21 Analysis of the “wicked problem” was derived primarily from Rittel and Webber’s work and supplemented by the literature on complexity, chaos theory, and systemic operational design.22 he review of the policymaking and leadership literature focused on works associated with systems analysis, warfare, and COIN.23 he review of the intelligence analysis literature relied on oicial sources and (to a lesser extent) journal articles.24 18 For example, Measurement in the Social Sciences: heories and Strategies (Blalock, 1974), Handbook of Research Design and Social Measurement (Miller and Salkind, 2002), Measure heory (Halmos, 1950), and Selected Methods and Models in Military Operations Research (Zehna, 2005). he review also included selections from the Journal of Economic and Social Measurement, Measurement: Interdisciplinary Research and Perspectives, the Journal of Business Cycle Measurement and Analysis, Measurement Techniques, Applied Psychological Measurement, the Journal of Quantitative Analysis in Sports, Advances in Data Analysis and Classiication, Advances in Statistical Analysis, the Journal of the Royal Statistical Society, Statistical Science, and the Journal of the American Statistical Association. 19 he books included Strategic Assessment in War (Gartner, 1997) and Analysis for Military Decisions (Quade, 1964). Articles included “Measuring Efectiveness in Irregular Warfare” (Clancy and Crossett, 2007), “A Will to Measure” (Murray, 2001), and “How to Measure the War” (Campbell, O’Hanlon, and Shapiro, 2009b). 20 For example, “Military Operational Measures of Efectiveness for Peacekeeping Operations” (Anderson, 2001), Measuring Fragility: Indicators and Methods for Rating State Performance (USAID, 2005), and Measuring Efectiveness in Complex Operations: What Is Good Enough (Meharg, 2009). 21 Examples here include Army FM 5-0.1, he Operations Process (HQDA, 2006a); Lifting the Fog of War (Owens, 2000); Operational Assessment—he Achilles Heel of Efects-Based Operations? (Bowman, 2002); and Efects-Based Operations (EBO): A Grand Challenge for the Analytic Community (Davis, 2001). 22 hese texts included “Dilemmas in a General heory of Planning” (Rittel and Webber, 1973), “Chaos heory and Its Implications for Social Science Research” (Gregersen and Sailer, 1993), and “Systemic Operational Design: Learning and Adapting in Complex Missions” (Wass de Czege, 2009). 23 Speciically, A Question of Command: Counterinsurgency from the Civil War to Iraq (Moyar, 2009); How Much is Enough? Shaping the Defense Program, 1961–1969 (Enthoven and Smith, 1971/2005); and FM 3-24, Counterinsurgency (HQDA, 2006c). 24 Examples here include Joint Publication (JP) 2-0, Joint Intelligence (U.S. Joint Chiefs of Staf, 2007); FM 34-3, Intelligence Analysis (HQDA, 1990); A Compendium of Analytic Tradecraft Notes (CIA, 1997); and A Tradecraft 10 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency he approach to the case-study literature difered for the two primary cases considered in this monograph (Vietnam and Afghanistan). he literature review for the case of Vietnam necessarily focused on historical works and data, beginning with books that included both secondary source material and original information.25 It then narrowed to focus on work conducted by the Oice of Systems Analysis and on historical accounts of the HES, body counts, and policymaking.26 he subsequent detailed analysis of the Vietnam case relied heavily on historical documents from the Vietnam Center and Archive at the Texas Tech University, which provided access to a breadth of original source material, including obscure military documents, regulations, and reports.27 Unfortunately, Gregory A. Daddis’s examination of U.S. Army assessments in Vietnam, No Sure Victory: Measuring U.S. Army Efectiveness and Progress in the Vietnam War, was published after this study had concluded. he literature review for the Afghanistan case included oicial reports by NATO and the U.S. government, as well as published and unpublished journal articles and reports.28 he number of available publications on Afghanistan assessment paled in comparison to the ready availability of material on Vietnam assessment, but the literature review on Afghanistan was intended primarily to frame the observation and interview process. Primer: Structured Analytic Techniques for Improving Intelligence Analysis (U.S. Government, 2009). 25 For example, Vietnam: A History (Karnow, 1984), America in Vietnam (Lewy, 1978), and he 25-Year War: America’s Military Role in Vietnam (Palmer, 1984). 26 Accounts of Oice of Systems Analysis research include War Without Fronts: he American Experience in Vietnam (hayer, 1985) and A Systems Analysis View of the Vietnam War 1965–1972, Vols. 1, 9, and 10 (hayer, 1975a, 1975b, 1975c). Details about HES were found in he American Experience with Paciication in Vietnam, Vols. 1 and 2 (Cooper et al., 1972a, 1972b); Analysis of Vietnamization: Hamlet Evaluation System Revisions (Prince and Adkins, 1973); and Measuring Hamlet Security in Vietnam: Report of a Special Study Mission (Tunney, 1968). Sources for body counts include Report on the War in Vietnam (Sharp and Westmoreland, 1968) and selected debrieings. Accounts on policymaking include Dereliction of Duty: Lyndon Johnson, Robert McNamara, the Joint Chiefs of Staf, and the Lies hat Led to Vietnam (McMaster, 1997); In Retrospect: he Tragedy and Lessons of Vietnam (McNamara, 1995); and Lessons in Disaster: McGeorge Bundy and the Path to War in Vietnam (Goldstein, 2008). 27 For example, MACCORDS-OAD Fact Sheet: RD Cadre Evaluation System (MACV, 1968b), Commander’s Summary of the MACV Objectives Plan (MACV, 1969a), and “General Westmoreland’s Activities Report for September” (Westmoreland, 1967a). 28 Sources include selections from the series Report on Progress Toward Security and Stability in Afghanistan (DoD, 2009a, 2010a, 2010b), “Unclassiied Metrics” (ISAF Headquarters Strategic Advisory Group, 2009), “Transfer of Lead Security Responsibility Efect Scoring Model” (ISAF AAG, undated), “Measuring Progress in Afghanistan” (Kilcullen, 2009b), COMISAF’s Initial Assessment (ISAF Headquarters, 2009), and Afghanistan in 2009: A Survey of the Afghan People (Rennie, Sharma, and Sen, 2009). Introduction 11 he review of the Iraq assessment literature was limited by classiication restrictions, but it included journal articles, oicial documents and reports, and subjectmatter expert reports and data indexes.29 Compared with the detailed and exhaustive study of measurement or assessment that is characteristic of the hard and soft sciences, little efort has been made to develop a comprehensive assessment model for military operations.30 Much of the existing scholarship on the assessment of warfare is dedicated to historical review, strategic analysis of conventional war, or speciic technical assessment methods. he work on assessment methodology speciically for COIN also has been limited in scope and, in many cases, merely explores basic theory or narrow mathematical processes. homas C. hayer led the most comprehensive analyses of COIN metrics to date, published in the 12-part series A Systems Analysis View of the Vietnam War and in War Without Fronts, a retrospective examination of his work during the Vietnam War for the Oice of Systems Analysis, Southeast Asia Intelligence and Force Efectiveness Division. here are few thorough recommendations for practical application of assessment methods outside the scope of the U.S. experience in Vietnam, however, and only a handful of very recent academic eforts on Afghanistan. Standing U.S. military doctrine has not yet adequately addressed this gap, although various U.S. joint and service manuals attempt to explain assessment. But these manuals generally provide only a brief overview and, on occasion, ofer contradictory perspectives. his monograph attempts to address these gaps and inconsistencies in U.S. military assessment doctrine. Two themes emerged from the literature review. First, assessment is often conlated with measurement or some form of centralized quantitative analysis. Of the sources cited here that speciically address military campaign assessment, nearly all assume or accept that assessment is a process of centralized measurement or pattern analysis. None of the works reviewed for this study, including the most aggressive critiques of centralized efects-based assessment, ofer a comprehensive, decentralized alternative or a way to comprehensively incorporate nonquantitative data. Second, because the literature accepts the premise of centralized assessment, it tends to focus on detailed disagreement over centralized assessment issues, such as the selection of individual core metrics or the best method for aggregating data. While some sources do address the broader issues of assessment, few have done so comprehensively or in a way that might widen the aperture for the inclusion of other assessment theories. Professional debate over COIN campaign assessment is bounded in a way that seems to preclude the consideration of viable alternatives to doctrine or current practice. 29 See “Measures for Security in a Counterinsurgency” (Schroden, 2009), “Assessing Iraq’s Sunni Arab Insurgency” (Eisenstadt and White, 2006), selected Measuring Stability and Security in Iraq quarterly reports (DoD, 2009b, 2010c), the memorandum “Long-Term Plan for IRMO Metrics” (Sullivan, 2004), and Iraq Index: Tracking Reconstruction and Security in Post-Saddam Iraq (Brookings Institution, 2010). 30 his conclusion is derived from the research conducted for this study but also echoes Darilek et al., 2001, p. 98, and comments by senior U.S. military oicers made as recently as early 2010. 12 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency While this monograph considers only two existing approaches to COIN campaign assessment, there are three very general schools of thought on how to address the complexities and challenges of assessing a war (see Chapters hree and Four for a discussion of these approaches). he irst school of thought prioritizes EBO/EBA, while the second prioritizes pattern and trend analysis. he third school of thought is represented by traditionalists like General James N. Mattis and Milan Vego.31 Traditionalists, or advocates of capstone maneuver warfare theory, accept complexity and chaos as inevitable realities in warfare. hey believe that some of the structured methods designed and used to penetrate complexity are ill conceived and not applicable to COIN. However, a generally agreed-upon and tested alternative to EBA doctrine or pattern and trend analysis has yet to emerge from either the traditionalists or other advocates or experts. In the literature, debate among these three loosely deined groups is inconsistent, insuiciently documented, and as yet unresolved. It is distinctly possible that the analysis in this monograph has failed to capture other groups or schools of thought on the subject of COIN assessment, but it does attempt to add some clarity to the discussion. he irst step to understanding all three arguments is to clarify the nature of the debate: how best to address the fog of war to assess COIN campaigns. Research Methodology he methodology for this study consisted of a literature review, direct observation, interviews, and case-study research. he research was conducted in two phases, with the irst phase addressing COIN assessment in Afghanistan and the second addressing the broader issue of U.S. military COIN campaign assessment. he resulting monograph attempts to answer the following questions: • What challenges does the COIN environment present to assessment? • How does recent doctrine address these challenges, and what are the points of disagreement regarding doctrinal approaches to assessment? • What do the lessons of Vietnam hold for contemporary assessment? • What can recent COIN assessment contribute to an analysis of other assessment approaches? • What are the strengths and weaknesses of the current approach? • How could doctrine and practice be improved? 31 In 2008, Gen. James N. Mattis, U.S. Marine Corps, issued a memo to U.S. Joint Forces Command stating, “It is my view that EBO has been misapplied and overextended to the point that it actually hinders rather than helps joint operations” (Mattis, 2008). See also Vego, 2006. Introduction 13 he irst phase of this research aimed to develop alternative assessment methods for a combat element that was preparing to deploy to Afghanistan. his phase included a literature review, interviews, and observations derived from my participation in conferences and workshops on assessment. Preliminary indings provided a limited-scope analysis of assessment processes in Afghanistan and ofered limited recommendations. hese recommendations focused on the development of a process that would add context to assessment reporting and incorporate qualitative data. As the scope of the research expanded in the second phase of the study to include all U.S. COIN assessment, a framework for assessment criteria began to evolve. Although it would have been ideal to develop this framework at the outset of the project, it was necessary to dissect the existing theories of assessment irst. A number of conlicting and partial lists of standards have been proposed by various experts and in assessment manuals, as discussed in Chapter Five.32 With a few exceptions, nearly all of the criteria described in the literature applied only to centralized quantitative assessment. However, two generalized criteria for successful COIN assessment emerged from the initial phase of the research: transparency and credibility. hese two standards were introduced by ISAF in 2009 and are not necessarily consistent with assessment literature. Nonetheless, this study and prior RAND research showed that the two standards are well aligned with the theory, doctrine, and casestudy literature on COIN as it is currently practiced by Western democracies. hese standards may or may not be universally applicable (e.g., to COIN operations supported by dictatorships), but they are relevant for democracies that must sustain willing popular support for ongoing campaigns.33 Further examination of policymaker requirements, all-source analysis methodologies, COIN theory and doctrine, and scientiic method revealed ive additional criteria, or standards, that could be generalized for the production of holistic campaign assessment. Each was selected based on the examination of the literature, interviews with subject-matter experts, direct observation of assessment in Iraq and Afghanistan, and a close examination of the historical record on Vietnam. he irst two standards— transparency and credibility—are deemed requisite for successful COIN assessment in a democracy. All assessment reports should be relevant to policymakers. he other 32 hese lists tend to focus only on certain aspects of assessment, typically the selection of core metrics for centralized assessment. here are very few recommendations in the literature that might help the military establish standards for an overarching assessment methodology. 33 One could argue that transparency and credibility are more universally relevant since the advent of the Internet and a near-pervasive media environment. It has become more diicult for dictatorships, anocracies, and oligarchies to sustain unpopular external military operations when atrocities, costs, and opinions of military actions are widely promulgated on the Internet. For additional RAND research supporting the inclusion of transparency and credibility as assessment criteria, see the examination of 89 case studies in How Insurgencies End (Connable and Libicki, 2010). 14 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency four standards are practical requirements that should be met to establish an efective methodology. he seven standards are as follows: 1. Transparent: Transparent assessment reports are widely releasable to oicial consumers and, ideally, are unclassiied and without distribution restrictions; the inal iteration of the report is suitable for distribution on the Internet. Such reports reveal both the methods and data used at each level of assessment, from tactical to strategic, and allow for detailed and in-depth analysis without requiring additional research or requests for information. Any subjectivity or justiication for disagreements between layers of command is explained clearly and comprehensively. 2. Credible: Credible assessment reports are also transparent, because opacity devalues any report to the level of opinion: If the data and methods used to produce the report cannot be openly examined and debated, consumers have no reason to trust the reporter. In credible reports, all biases are explained, and laws in data collection, valuation, and analysis are clearly articulated. Data are explained in context. Such reports clearly explain the process used by commanders and stafs to select methods and are both accurate and precise to the greatest degree possible. 3. Relevant: Relevant reports efectively and eiciently inform policy consideration of COIN campaign progress. hey are suicient to help senior military leaders and policymakers determine resourcing and strategy, and they satisfy public demand for knowledge up to the point that they do not reveal sensitive or classiied information. 4. Balanced: Balanced reports relect information from all relevant sources available to military stafs and analysts, including both quantitative and qualitative data. Such reports relect input from military commanders at all levels of command and are broad enough in scope to incorporate nonmilitary information and open-source data. Balanced reports include countervailing opinions and analysis, as well as data that both agree with and contradict the overall indings. 5. Analyzed: Ideally, inished reports do not simply relect data input and the analysis of single data sets. hey also contain thorough analyses of all available data used to produce a uniied, holistic assessment. his analysis should be objective and predicated on at least a general methodological framework that can be modiied to it changing conditions and, if necessary, challenged by consumers. he requirement for holistic analysis is commonly voiced in the assessment literature. 6. Congruent: COIN assessment theory should be congruent with current U.S. joint and service understanding of warfare, the COIN environment, and the way in which COIN campaigns should be prosecuted. he standard for con- Introduction 7. 15 gruence was drawn from a combination of joint doctrine and service capstone doctrine, as discussed in greater detail in Chapter Two. Parsimonious: Assessment cannot show all aspects of a COIN campaign, nor should it seek or claim to deliver omniscience. Collection and reporting requirements for assessment should be carefully considered relative to the demands and risk they may leverage on subordinate units. Assessment should rely to the greatest extent possible on information that is generated through intelligence and operational activities, without requiring additional collection and reporting. Parsimony is a common theme in COIN assessment literature and is often identiied as a requirement by assessment experts.34 Notably absent from this list are the standards of scientiic method described in Chapter hree. hat chapter explains how some assessment analysts have attempted to apply scientiic rigor to assessment and suggests ways in which scientiic standards might be applied to the various phases of COIN assessment. However, this monograph maintains that assessment is not scientiic research. Because assessment is not research, none of these standards were deemed to be generally applicable to holistic campaign assessment. his does not negate their value in helping to shape assessment. Certainly, if a staf chooses to approach assessment as scientiic research, it should pursue methods that are reliable and produce valid indings. his study drew on several sources of information in addition to the literature review, including more than 20 interviews, both in Afghanistan and elsewhere. Questions for the interviews were crafted speciically for the position and experience of the interviewee; some of these interviews were planned in advance, while others were conducted in the ield as opportunities presented themselves. Interview subjects included military personnel at all echelons of command and civilians involved in the assessment process or who had related responsibilities. In addition, I participated in the assessment process during three tours in Iraq between 2003 and 2006 and used this knowledge to frame questions and identify research sources. I also participated in and helped lead several conferences and workshops on COIN assessment in Afghanistan between late 2009 and early 2011, such as the oicial NATO Systems Analysis and Studies conference series designed to help build and validate Afghanistan metrics. Finally, I was briely embedded with the ISAF Joint Command assessment team in May 2010. he in-depth case-study analysis of the Vietnam-era assessment processes was conducted over the course of a calendar year, and I obtained additional insight into the current assessment process while providing informal and oicial support to the ISAF Afghanistan Assessment Group (AAG) from mid-2010 through early 2011. 34 However, this call for parsimony tends to be focused only on the selection of core metrics for centralized assessment. he standard for parsimony in the framework developed for this study is intended to address the assessment process more broadly. 16 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Issues Not Addressed his research focused speciically on the military assessment process for COIN campaigns. Both in doctrine and in recognized best practices as detailed in the literature, COIN assessment should (in theory) incorporate input from civilian agencies, nongovernmental organizations (NGOs), and host-nation oicials and should therefore relect a holistic civilian-military approach. Joint doctrine on COIN recommends that civilian agencies take the lead in U.S. and coalition COIN operations, and it stresses unity of efort between military and nonmilitary activities.35 But doctrine has not translated neatly into assessment practice. For example, the U.S. Department of Defense (DoD) and U.S. Department of State (DoS) build separate theater-level assessments, and as of early 2011, they do not efectively communicate on assessment reporting or methodology. DoS and USAID use (or are in the process of testing) methods like MPICE (Measuring Progress in Conlict Environments) and TCAPF (Tactical Conlict Assessment and Planning Framework).36 While the military has tested TCAPF, it has not been fully incorporated into the theater assessment process. Although this monograph recognizes the gap between civilian and military assessment, a detailed examination was beyond the scope of the research efort.37 It does, however, address the incorporation of Provincial Reconstruction Team (PRT) information in the assessment process. Research on the Vietnam-era assessment process was extensive, but the analysis and recommendations presented here focus on contemporary COIN doctrine and practice because their purpose is to help inform current policy debate. It also bears mentioning that while a great deal of information on the Afghanistan assessment process is unclassiied, most of the oicial documentation of the Iraq war remains classiied. Although this monograph cites research on Iraq, it was as much by necessity as by design that it focuses on Vietnam and Afghanistan as its principal case studies. Of course, the United States has been involved in many other COIN operations over the past 60 years, but most of these cases have less relevance to current campaign assessment than Vietnam, Iraq, and Afghanistan because they were primarily shortterm, advisory, or covert-action missions and not what GEN David H. Petraeus calls “industrial-strength insurgencies.”38 One might use very diferent processes to assess the advisory mission to El Salvador and the COIN campaign in Iraq. Furthermore, while some historical cases prior to Vietnam informed this research (e.g., U.S. activities 35 JP 3-24, Counterinsurgency Operations, states, “It is always preferable for civilians to lead the overall COIN efort” (p. IV-11); see also pp. IV-1–IV-22 (U.S. Joint Chiefs of Staf, 2009c). 36 As of October 2010, TCAPF was being remodeled and expected to be replaced by the District Stability Framework tool. Statement by a U.S. USAID representative, October 14, 2010. 37 Armstrong and Chura-Beaver, 2010, address some of these other reports and processes. 38 Rubin, 2010b. Introduction 17 in the Philippines, Nicaragua, Greece), these cases were not suiciently germane to the current process to warrant in-depth examination and comparison. To ensure that the indings presented here were suiciently streamlined for use by policymakers and those involved in the assessment process, it was necessary to contain the initial scope of the research and focus on methodology and U.S. COIN operations since, and including, Vietnam. he disadvantage of this approach was that it excluded potentially valuable information on British, Canadian, Soviet/Russian, and other international eforts.39 Finally, NATO assessment doctrine is published at an unclassiied but restricted level and cannot be cited in public documents. herefore, this monograph does not refer to speciic NATO documents or doctrine. his monograph does not speciically address each and every aspect of campaign assessment. Most noticeable to assessment stafs and experts will be the absence of security force assessment and a detailed treatment of opinion polling. Both of these subsets of assessment are relevant to holistic campaign assessment. Security force assessment is shaped by many of the same concerns that shape holistic assessment, but it is usually more technical in nature than campaign assessment (in that it focuses on manpower, training, and logistics). Opinion polling is integral to campaign assessment, but it is a complex subset of holistic assessment that would require distinct and detailed treatment, beyond the time and resources available for this project. Organization of This Monograph his monograph provides a thorough examination of not only the practice of COIN assessment but also the theoretical roots underlying the various practical approaches to the assessment of contemporary COIN warfare. Its structure is intended to take the issue down to its roots and then build back up to a series of recommendations, with a thorough appreciation of the complex and contested issues in play. he narrative is designed to work from theory to practical application by delivering the broad themes underlying assessment, background on assessment theory and doctrine, a detailed historical case study (Vietnam), an examination of current practice and indings from that review, and recommendations that can be incorporated in the short and long terms. his progression is intended to put current practice in clear context to inform contemporary policy. Because COIN assessment is controversial and inadequately addressed in the literature, this monograph introduces and examines the seven framework concepts for assessment presented earlier in this chapter, in the section “Research Methodology.” 39 Ample information on international eforts is available online. he bibliography at the end of this monograph also includes many references examined for this study that ultimately did not igure prominently in the indings presented here. 18 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency he document is organized as follows. Chapter Two examines policymaker requirements for assessment, the need for transparency and credibility for policymaking, and the impact of the COIN environment on assessment. Chapter hree introduces the concepts behind centralized assessment, including systems analysis, scientiic rigor, and time-series analysis, and describes how pattern and trend analysis igures into the current assessment process. Chapter Four provides an overview of efectsbased theories and the doctrine of EBA, and Chapter Five explains the considerations for selecting core metrics, with examples from Afghanistan. It also describes some of the challenges of centralized assessment and the necessity for context in assessment analysis. Chapter Six presents the historical case study of COIN assessment during the Vietnam War, examining how assessment failed in Vietnam and why and placing pattern and trend analysis in a case context. Chapter Seven provides an overview of the Afghanistan assessment process as of early 2011 and identiies some concerns about the assessments being produced in that contingency. Chapter Eight explains how and why current assessment approaches have failed to deliver adequate support to policymakers, and Chapter Nine ofers several recommendations and options to improve the current process. Finally, Chapter Ten proposes an alternative to centralized assessment processes. his document also contains ive appendixes that provide resources and background information intended to supplement the discussions presented in the body of the document. Appendix A ofers a step-by-step framework for contextual assessment, including templates for assessment at the battalion, brigade/regiment, regional command, and theater levels. Appendix B provides a detailed hypothetical example of contextual assessment in practice, using notional data. Appendixes C and D present, as background to the historical case study outlined in Chapter Six, excerpts from a declassiied province-level narrative assessment report and a declassiied theater-level narrative assessment report from Vietnam, respectively. Appendix E concludes the monograph with a brief outline of the debate over efects-based operations and discusses its relevance to efects-based assessment. CHAPTER TWO Concepts That Shape Counterinsurgency Assessment he purpose of this chapter is to lay the groundwork for a focused examination of centralized assessment processes (EBA and pattern and trend analysis), as well as the Vietnam and Afghanistan case studies. To determine best practices for assessment, it is necessary to understand policymaker requirements. It is also important to know how the COIN environment will afect the ability to collect and analyze information. Ideally, this understanding should provide a common foundation for the development of a more efective assessment process that could be adapted to meet the unique challenges of any current or prospective campaign. his chapter describes the complex balance between policymaker requirements and the ability of military stafs to provide relevant assessment input to the decisionmaking process. his relationship foundered during the Vietnam War, and the patterns established during that period have posed ongoing challenges in both the Afghanistan and Iraq campaigns. he complexities and chaos of the COIN environment exacerbate the challenges that military stafs face in attempting to develop a relevant campaign assessment; these environmental impediments also complicate the process of strategic decisionmaking. Policy Requirements and Counterinsurgency Assessment National Public Radio: Is there a number that you track that makes you feel conident that you have made progress against the insurgents? hat you are winning militarily? Secretary of Defense Donald H. Rumsfeld: No one number is determinative, and the answer is no. We probably look at 50, 60, 70 diferent types of metrics, and come away with [sic] them with an impression. It’s impressionistic more than determinative.1 1 Rumsfeld, 2005. 19 20 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency For practical purposes, the end user of a military campaign assessment is the policymaker.2 A policymaker could be the President, a cabinet member or advisor, an executive oice staf member (e.g., on the National Security Council), a senior DoD oicial, or a legislator. What do policymakers require from COIN assessment?3 How will policy and policymaker concerns and behavior shape the U.S. military’s ability to provide useful assessments? How can the military shape assessment to best inform policy? Speciically, what should the relationship between COIN policy and assessment look like? he military is best positioned to produce a transparent, credible, and relevant campaign assessment when provided with clear strategic objectives or “strategic guidance” by policymakers. According to U.S. joint doctrine, this guidance should contain a “national strategic end state” and “termination criteria.” he former is deined as “the broadly expressed conditions that should exist at the end of a campaign or operation,” while the latter is “the speciied standards approved by the President or the SecDef [Secretary of Defense] that must be met before a joint operation can be concluded.”4 For the purposes of simplicity, this monograph refers to these criteria as national strategic “objectives.”5 Without clear or understandable objectives, military stafs are left scrambling to design operational objectives—and then assessments—that seem reasonable. his monograph addresses the confusion about assessment in Vietnam; some of that confusion can certainly be tied to the absence of clear objectives early in the war. A 1974 survey of U.S. general oicers showed that only 29 percent felt that pre1969 strategic objectives for the Vietnam War were clear and understandable, while 35 percent thought that they were “rather fuzzy.” A full 91 percent of these oicers listed “deining the objectives” as a key recommendation if they had to ight the war again.6 he commanding general of MACV in 1969 described the U.S. objectives as 2 Although policymakers are the “end users,” the public is the ultimate recipient of assessments. Public opinion cannot directly alter strategy, but it does inluence the course of events, the willingness of policymakers to commit forces and resources, and the decision to sustain or withdraw from a campaign. 3 his section focuses speciically on COIN policy requirements, but a broader understanding of policy decisionmaking would also inform this subject. he literature on policy decisionmaking requirements is voluminous and its focus ranges from technical aspects of intelligence production (Sherman Kent, e.g., CIA, 1966) to defense budget analysis (Charles J. Hitch and E. S. Quade) and strategic assessment (Scott Sigmund Gartner). 4 U.S. Joint Chiefs of Staf, 2006, p. III-5. 5 Military terminology regarding strategic guidance, objectives, and end states is complex and sometimes contradictory. JP 5-0, Joint Operation Planning, states that “strategic direction encompasses the processes and products by which the President, SecDef, and CJCS [Chairman of the Joint Chiefs of Staf] provide strategic guidance” to military forces (U.S. Joint Chiefs of Staf, 2006, p. II-1). In Army doctrine, strategic guidance is translated into military commanders’ intent, mission statements, end states, and intermediate objectives, which are, in turn, used to shape operations. FM 5-0, he Operations Process (HQDA, 2010), provides a more in-depth analysis of the operational-level terminology. his is somewhat contradictory to joint doctrine, which describes strategic and operational objectives, and these terms seem equivalent to Army mission statements and end states. 6 Kinnard, 1977, pp. 169, 176. Concepts That Shape Counterinsurgency Assessment 21 ill deined and misunderstood.7 A Brookings Institution report on the Vietnam War stated, “Administration leaders persistently failed to clarify U.S. objectives in concrete and speciic terms. Uncertainty and ambiguity in reports were therefore bound to emerge, for no one could be certain what he was measuring progress against or how victory could be deined.”8 Similar concerns dogged President George W. Bush’s strategy for both Iraq and Afghanistan, and criticism of both strategies has carried over into the current administration.9 With a clear and understandable policy in place, the military can develop an assessment process to answer questions that relate speciically to policymaker requirements. his envisions a best-case scenario: Assessments should be tied directly to clear and understandable national policy objectives. Policymakers can turn to such assessments with relative conidence. However, it seems to be more likely that policy on COIN campaigns will not be clear. U.S. Army doctrine describes COIN as an “illstructured” problem that is likely to create disagreements over the formulation of a clear end state.10 Of the three cases examined in this monograph—Vietnam, Iraq, and Afghanistan—it could be argued that none had a clear or generally agreed-upon end state, either throughout the campaign or at any one point in time. In cases without a clear end state, or “termination criteria,” the military assessment process is likely to be less clear, but it must still support policy decisionmaking. Whether policy provides clear or less clear strategic objectives, assessment will have to be lexible. By design, no assessment process is static or immutable, because the nature of war is luid. Just as policymakers will have to adjust policy to meet changing objectives, they will also have to anticipate and accept some changes in assessment methods and outputs. If the assessment process starts from the ideal point—a clear policy and a closely aligned assessment design that are agreed to prior to engaging in conlict—then these shifts will have only a marginally negative impact on the ability of policymakers to ascertain progress. If policy direction fails to stay abreast of the truth on the ground, then assessments will slide out of alignment as the military reports on events that seem to have little bearing on the original (and possibly outdated) policy. For example, in 2004, the multinational forces in Iraq were still focused on assessing post-invasion transition objectives as the country slid deeper into insurgency and civil violence. Fred C. Iklé describes another layer of complexity in the relationship between policy and military assessment. He states that the military is so wrapped up in day-to7 MACV, 1969a, p. 4. 8 Gelb and Betts, 1979, p. 307. 9 See, for example, Mark Schrecker’s 2010 article in Joint Force Quarterly, “U.S. Strategy in Afghanistan: Flawed Assumptions Will Lead to Ultimate Failure.” 10 HQDA, 2010, p. 2-4. In deining an ill-structured problem, the manual states, “At the root of this lack of consensus is the diiculty in agreeing on what is the problem.” 22 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency day operations that it has trouble seeing the big picture and often fails to produce adequate strategic assessment. Conversely, policymakers often do not have the background or context to understand the minutiae of military campaigns.11 here may be some undue generalization in this analysis, but it is practical: If the military fails to provide an adequate strategic assessment, it encourages micromanagement by policymakers, and policymakers waste valuable time by immersing themselves in arcane military detail with little to show for their eforts but frustration. It behooves the military to produce a report that is tuned to strategic decisionmaking, and it is incumbent upon policymakers to facilitate assessment by providing clear objectives and by remaining somewhat lexible as the military adjusts to meet changing ground conditions. his monograph describes many examples of military assessments that failed to adequately support policy decisionmaking. Figure 2.1 provides a visual account of extreme policymaker micromanagement. In the photo, President Lyndon B. Johnson (second from left) examines a detailed model of the Khe Sanh combat base in Vietnam. he battle for Khe Sanh preoccupied Johnson for much of early 1968; he had the model built in the White House Situation Room so he could track the location Figure 2.1 President Johnson and the Khe Sanh Model in the White House Situation Room SOURCE: February 15, 1968, photo via the National Archives, Archival Research Catalog, courtesy of the Lyndon Baines Johnson Library. RAND MG1086-2.1 11 Iklé, 2005, pp. 18–19. Concepts That Shape Counterinsurgency Assessment 23 and status of individual military units.12 Johnson was both reacting to a gap in useful assessment from the military and indulging his own proclivity to personally manage military operations. Chapter Six shows the efect that this kind of aggressive micromanagement had on the accuracy and integrity of military assessments during the Vietnam War. A relevant, or efective, assessment should preclude the need for this kind of micromanagement. But what does “efective” entail? What should assessment show and why? Previous RAND research on insurgency endings has showed that lasting victory in COIN comes not by military action alone but ultimately by addressing the root causes of the conlict to achieve naturally occurring stability.13 Most experts cited in How Insurgencies End concurred that addressing root causes is most often the key to victory, and FM 3-24 states, “Long-term success in COIN depends on the people taking charge of their own afairs and consenting to the government’s rule.”14 JP 3-24, Counterinsurgency Operations, describes an end state for COIN: COIN is successful when three general conditions are met. First, the [host-nation] government efectively controls legitimate social, political, economic, and security institutions that meet the population’s general expectations, including adequate mechanisms to address the grievances that may have fueled support of the insurgency. Second, the insurgency and its leaders are efectively co-opted, marginalized, or separated physically and psychologically from the population, with the voluntary assistance and consent of the population. hird, armed insurgent forces have been destroyed or demobilized and reintegrated into the political, economic, and social structures of the population.15 Assessing whether a COIN campaign is progressing toward long-lasting stability demands a clear assessment of human factors, or human terrain. However, this joint deinition of success does not apply universally to all COIN operations. COIN policy varies, and national objectives may be more limited in scope. Policy may require an outcome ranging from the short-term disruption of an insurgent cadre to long-lasting stability. Because there is such a range of possible strategic objectives in COIN, it is dificult to identify a singular approach to assessment. Some generalization is necessary. 12 Johnson requested and received direct cables on the situation in Khe Sanh on a near-daily basis from February 3 to March 30, 1968 (see, e.g., Wheeler, 1968). 13 See, especially, How Insurgencies End (Connable and Libicki, 2010) and Victory Has a housand Fathers: Sources of Success in Counterinsurgency (Paul, Clarke, and Grill, 2010). 14 HQDA, 2006c, p. 1-1. 15 U.S. Joint Chiefs of Staf, 2009c, p. III-5. 24 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency COIN objectives tend to be long-term objectives.16 herefore, it would be reasonable to assume that COIN assessment should show progress over time toward a clear (or at least deinable) end state.17 Policymakers who are accustomed to executive summaries and brieings tend to expect concise and often quantitative reports that show a near-mathematical path toward an end state. Unfortunately, as already discussed, deining a clear end state for a full-scale COIN campaign is diicult. “Addressing root causes” is not necessarily a quantiiable or even visible process over time or even at any one point in time. National strategy and the military operations conducted to address that strategy are likely to be complex, dynamic, and subject to regular review and sometimes drastic revision. herefore, while policymakers may desire a clear and easily digestible assessment report, a realistic assessment will necessarily be intricate, because it will relect complex and dynamic circumstances. Policymakers should be wary of simple and concise assessments of complex COIN operations. Both military oicers and policymakers will have to absorb some level of detail if they are to understand the arc of a speciic COIN campaign. Assessments of Vietnam, Iraq, and Afghanistan have produced reports that support what former Secretary of State Donald Rumsfeld referred to as either impressionistic or determinative decisionmaking. In one sense, these two terms describe a thought process: Rumsfeld asserts in the interview that he drew only impressions from assessment reports and did not take them as literal interpretations of events on the ground. In other words, he would absorb massive quantities of data and reporting from multiple sources and gain a broad impression of the war. Speciically, a policymaker using an impressionistic approach might see broad trends in violence going up, troop levels going up, and popular opinion generally rising and thus determine that a troop surge is working. his approach relies on one of two assumptions. On one hand, the policymaker may assume that it is unnecessary to know “ground truth” because the aggregated data provide adequate indication of broad patterns or trends; thus, impressions are not based on fact but are instead vague and indistinct.18 Chapter Six examines whether such patterns and trends are accurate enough to feed efective impressionistic decisionmaking for COIN and how the impressionistic approach fared in a speciic case during the Vietnam War. On the other hand, if the policymaker assumes or is assured that the assembled data are accurate, he or she must determine whether a determinative decision is possible. Typically, determinative decisions require highly accurate 16 According to U.S. Joint Chiefs of Staf, 2009c, p. III-16, “Insurgencies are protracted by nature, and history demonstrates that they often last for years or even decades.” 17 here was some debate over the value of having an end state versus an exit strategy in Afghanistan as of early 2011. hat debate may be useful to some readers, but because it is a speciic policy debate it was not clearly germane to this research efort. 18 In practice, however, trends are viewed as both precise and accurate, and policymakers tend to gravitate toward mathematical threshold–driven campaigns. Concepts That Shape Counterinsurgency Assessment 25 quantitative data, a sound theory of the problem at hand, and a clear understanding of the second- and third-order efects likely to emerge from the decision. With few exceptions, most policy decisions are impressionistic rather than determinative, and wartime decisionmaking tends to be especially so because the issues at stake are complex and often nebulous. All policymakers rely on impressions to make wartime decisions, but they must consider whether these impressions are derived from layered, transparent, and credible reporting or merely single-layer subjective analysis of aggregated and often inaccurate and incomplete data. Figure 2.2 is a simpliied depiction of the process of impressionistic decisionmaking using notional COIN data. It shows an array of data on aspects of the complex COIN environment, including economic performance, insurgent activity, and level of popular support. While these types of data might be presented in sequence for an executive assessment brieing, in practice, they are typically presented with little to no association or correlation. hese data might be used to determine force deployments and resourcing, or they might be used to describe campaign progress to Congress or the public. his is the process that Secretary Rumsfeld described in his interview. Policymakers taking a determinative approach might try to gather all this information, attempt to ind cause and efect between each data source (or have a miliFigure 2.2 Impressionistic Decisionmaking Using a Broad Array of Counterinsurgency Data 120,342 children attend school every day Only 923 new shops opened last month Employment of women fell 2.3% Spring offensive imminent Fear of U.S. troops reached all-time low 10,200 new army recruits signed on 62 civilians were killed this month 3 ministers were ired for corruption Type of attack January February March Direct fire 283 252 301 Indirect fire 110 93 92 Local security force casualties doubled Domestic spending up 3% Provincial elections were successful Conidence in peace low Aid spending increases 3% Highway construction is progressing slowly IED attacks are up 8% this month Violence is down 4.3% from last month NOTE: The data in the figure are notional. RAND MG1086-2.2 Impressionistic Decision 26 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency tary staf analyze the data), and produce inely tuned policy decisions based on what tends to be mostly quantitative analysis. Much of the Vietnam-era COIN reporting (e.g., HES) had the veneer of accuracy, and because it was presented in a heavily quantitative format, it fed determinative policymaking. But in the case of HES, signiicant data inaccuracies and gaps led to suboptimal determinative decisions. Policymakers can be pulled into making determinative decisions based on data that appear to be precise and accurate but that are in actuality imprecise and inaccurate, or precise but inaccurate. An overly aggressive drive to reduce uncertainty can be counterproductive when precise and accurate data are not available or cannot reasonably be collected. Determinative assessment does not have to be primarily quantitative. Detailed and layered contextual assessment can also support determinative decisionmaking if it is properly written and presented. Assessments can be used to create speciic, inely tuned policy, but military assessment in general is intended as only one input into policy decisionmaking. Policymakers should seek other inputs as well. Although military campaign assessments might recommend shifts in military strategy or national policy, these recommendations tend to be limited in scope or focused on resource allocation. Only rarely do they recommend major shifts in U.S. national security policy.19 Assessment can be useful in identifying the failure of the military, civilian agencies, or the host nation to execute a good policy, and also in restructuring a failing military strategy. But, ultimately, a poorly conceived COIN strategy will live or die on its merits. Inadequate or inaccurate assessment can help lose wars, but even the best assessment cannot by itself rescue a bad plan. Finally, some policymakers require assessment to be predictive. hey want to know what to expect in the months ahead as they determine how to allocate resources and shore up political support for the campaign. But because assessment is not intelligence analysis, COIN campaign assessments are not designed to provide well-analyzed predictive reporting. Some prediction is possible, of course, but prediction is an analytic quagmire and likely to produce inaccurate results. Compounding the inherent challenges of predictive analysis, most assessment stafs are not trained for the task. Policymakers must be cautious when requesting predictive analysis from a military assessment, at least until the gap in military analytic capability for assessment is remedied. Chapter Six examines these issues within the context of the Vietnam case. his section concludes with a summary of some of the aforementioned considerations, as well as other considerations drawn from the policy decision literature and primary-source research. No list of policy requirements can be comprehensive, but these are informative and provide a foundation for the discussion that follows. 19 See Commander’s Summary of the MACV Objectives Plan for an example of a top-level report that suggested changes in strategic approach (MACV, 1969a). Concepts That Shape Counterinsurgency Assessment 27 • What the policymaker requires from COIN assessment:20 – a close approximation of ground truth upon which decisions can be based – an appreciation for resource requirements (current and predictive) – transparent and credible reporting that can be used to inform Congress and the public – relevance to decisionmaking – predictive analysis (sometimes). • What the military requires from policymakers to facilitate sound assessment: – clear policy objectives – a description of what the policymaker wishes to learn from the assessment – an agreed-upon reporting format that is useful but also lexible and realistic. • Basic considerations for developing assessments for policymakers: – Ensure that the assessment reports on progress toward policy objectives. – Explain the assessment methodology in writing and in detail. – Communicate clearly with policymakers when assessment methods need to be changed. – he objectives should anchor the assessment, no matter what else changes; the link between current objectives and assessment should be immediate and unbreakable. – Because COIN campaigns are lengthy, assessment should help policymakers understand current and prospective timelines. – If policy objectives describe an end state, assessment should describe both how the campaign is progressing toward that end state and how events on the ground have informed changes to the described end state. – To establish credibility and, ultimately, deliver sound policy, it is important to understand the limitations that the COIN environment places on the ability to deliver precise and accurate assessments. – Policymakers must understand that, at best, assessment should be used to inform decisions. An assessment should not be relied upon as an independent means to provide clear answers to complex strategic challenges. Wicked Problems and Counterinsurgency A brief explanation of the complexities of the COIN environment is a necessary backdrop to the more focused examination of the case studies later in this monograph. he 20 his does not include responses to the question, “What do policymakers want from COIN assessment?” Want and need are not necessarily synonymous. For example, a policymaker might want explicit detail about a speciic military operation (e.g., President Johnson and the Khe Sanh model). More recently, policymakers tend to want quantiiable information. However, as discussed in greater detail later in this monograph, when such information is provided without context, it cannot help policymakers comprehend ground truth to any degree of accuracy. 28 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency issue of complexity is closely tied to policymaker requirements: If policymakers do not understand or appreciate the complexity of COIN they cannot extend realistic guidance on assessment to the military. If military stafs fail to account for complexity in an assessment or in the caveats of their reports, they do policymakers and the public a disservice. he term wicked problem is commonly used in both policy and military circles to describe COIN, so it is appropriate to use the concept to introduce the idea of complexity in COIN assessment. An understanding of wicked problems in assessment also highlights the daunting challenges facing counterinsurgents and COIN assessment stafs on the ground. In their landmark 1973 article in Policy Sciences, Horst Rittel and Melvin Webber use the term “wicked problem” to describe planning processes for governance, teaching, housing development, policing, and other real-world social challenges.21 One could read their article as a polemic or simply as an efort to draw a line between the hard and soft sciences. Perhaps the best interpretation is that it identiies a spectrum of complexity in research environments and shows that highly complex environments tend to complicate research and reduce the accuracy of indings. herefore, these complex, real-world problems are “wicked” because they do not ofer simple, inite, or necessarily replicable solutions. he concept is best captured in this excerpt: he problems that scientists and engineers have usually focused upon are mostly “tame” or “benign” ones. As an example, consider a problem of mathematics, such as solving an equation; or the task of an organic chemist in analyzing the structure of some unknown compound; or that of the chessplayer attempting to accomplish checkmate in ive moves. For each the mission is clear. It is clear, in turn, whether or not the problems have been solved. Wicked problems, in contrast, have neither of these clarifying traits; and they include nearly all public policy issues—whether the question concerns the location of a freeway, the adjustment of a tax rate, the modiication of school curricula, or the confrontation of crime.22 A scientist working on a cure for cancer might not agree. Indeed, there are some overstatements in Rittel and Webber’s argument. However, in their efort to describe a chaotic, hectic, unrewarding, and (in some cases) nearly opaque civil planning environment, they have also described COIN.23 Rittel (a systems analyst) and Webber (a city planner) lay the groundwork for several themes that run through this mono21 Rittel and Webber, 1973. As discussed earlier, Army doctrine refers to this as an “ill-structured” problem (HQDA, 2010, p. 2-4; U.S. Army Training and Doctrine Command, 2008, p. 9). 22 23 Rittel and Webber, 1973, p. 160. For a diferent and more in-depth exploration of the wicked problem construct as applied to military planning and operations, see Greenwood and Hammes, 2009, and U.S. Army Training and Doctrine Command, 2008. Concepts That Shape Counterinsurgency Assessment 29 graph. First, COIN is complex and unpredictable; second, COIN is both art and science, but mostly art; and third, context is critical to understanding the COIN environment. While “wicked problem” is now thrown about so freely that it has lost some meaning, COIN and therefore COIN assessment can still be described within the framework of Rittel and Webber’s concept: • • • • • • • War assessment does not have a deinitive formulation—there are many methods. COIN has no “stopping rule”—there is rarely a clear or concise ending.24 here are no black/white, true/false assessments, only better or worse. One cannot easily test COIN assessment for validity. here are no second chances; every assessment is consequential to strategy. No exhaustible or knowable set of COIN endings or outcomes exists. Every COIN campaign—and, consequently, assessment process—is a unique problem. • Tracing cause and efect and “weighting” various factors is a subjective process that makes it nearly impossible to scientiically determine solutions. • Anyone assessing COIN operations is subject to harsh and sometimes personal critique because assessors’ work is often politicized and has immediate, real-world impact in a high-stakes environment. Kenneth Menkhaus argues that state failure exacerbates the symptoms of wicked problems, and he presents varying degrees of failure in a series of typologies on a spectrum. He places Afghanistan within the typology of state failure, describing it as a case of “shorter term state collapse.”25 Iraq would also fall within this typology. One could infer from this analysis that these two active cases of U.S. COIN are not simply wicked problems but exacerbated wicked problems that make assessment particularly challenging. Wicked problems tend to emerge from or describe complex adaptive systems, a term used in mathematics, in the social sciences, and, more recently, by military practitioners to describe battleield environments. Wicked problem theory incorporates, or perhaps depends on, the idea that societies (or, in this case, populations in COIN), like groups on a battleield, form complex adaptive systems. In the social sciences, a complex adaptive system is deined by individual actors in a society—people or groups— changing behavior in response to an intricate and interdependent web of actions, reactions, and happenstance.26 he next section explains the complex adaptive system in the context of military assessment. 24 See Connable and Libicki, 2010. 25 Menkhaus, 2010, p. 89. 26 For a lay introduction to complex adaptive systems and mathematical approaches to understanding these systems, see Miller and Page, 2007. 30 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Complex Adaptive Systems, Nonlinearity, and Chaos Wicked problem is a general, nonscientiic term that is useful to frame the notion of complexity. To fully appreciate the degree to which complexity afects assessment, it is necessary to delve into a more structured discussion of the COIN environment. here is something about a battleield—or an operational environment—that confounds accurate assessment. A conventional battle between two opposing armies is certainly complex, but when one side is hidden among the population, the degree of complexity faced by the opposing side (the counterinsurgent) is magniied. he counterinsurgent must locate and separate the insurgents without damaging the population, all the while attempting to address a complex web of root causes. Eforts to assess progress against these confounding requirements often fall short, particularly since complexity is anathema to accuracy. his section introduces three broad concepts—complex adaptive systems, nonlinearity, and chaos—each of which should help build a more comprehensive understanding of the challenges posed to assessment stafs and policymakers. Assessment of COIN is commonly likened to the assessment of a system. Most literature refers to the COIN environment, the society within that environment, or all the groups and actors in the operational environment (e.g., friendly, insurgent, civilian) as parts of a larger and more complex type of system. Noted operations researcher and systems analyst Russell L. Ackof describes a system as follows: A whole consisting of two or more parts (1) each of which can afect the performance or properties of the whole, (2) none of which can have an independent efect on the whole, and (3) no subgroup of which can have an independent efect on the whole. In brief, then, a system is a whole that cannot be divided into independent parts or subgroups of parts.27 A complex system is a system that is typically self-contained but consists of a number of diferent moving and interacting sections or parts. A car is a complex system in that it is made up of thousands of intricate moving parts that interact with each other. It is possible to predict what will happen when someone turns the key in the ignition of a functioning car: It will start.28 It is also possible to know what each speciic part of the car will and will not do and what will happen when that part ceases to function. For instance, if the starter motor breaks, the car will not start. It is possible to isolate, deconstruct, and understand a complex system in great detail using a wide array of scientiic and nonscientiic methods. horough and persistent technical examination over many decades has led to steady and sometimes revolutionary improvements in the car, and millions of people have a good working understanding of the car’s complex system. 27 28 Ackof, 1994, p. 175. Commander’s Appreciation and Campaign Design (U.S. Army Training and Doctrine Command, 2008, p. 6) also uses the analogy of the car to explain the diference between a complex system and a complex adaptive system. Concepts That Shape Counterinsurgency Assessment 31 All cars have many predictable elements in common, and cars within one model are nearly identical in construct. People, on the other hand, are adaptive and not binary. No two people are alike; they can and often do react unpredictably. he reasons for these reactions are often unknown or cannot be easily discerned. People also adapt and change the way they think and behave based on an ininite number of possible inputs, which, in the absence of a controlled scientiic study, are unknowable to anyone but the individual. When people interact with each other, they form a broader complex adaptive system. It is possible to envision a society or population in a COIN environment as a complex adaptive system consisting of interrelated but individual complex adaptive systems. Indeed, this is how doctrine describes human terrain and how Ackof describes social systems.29 Because COIN depends on popular support and the behavior of individuals and small groups, some argue that success in COIN can be gauged only through an understanding of the complex adaptive social system that afects it.30 But complex adaptive systems are also self-adapting in that they change and react without the need for external stimuli. If a car were a complex adaptive system, it might be able to start and drive itself. Complex adaptive systems are “marked by selforganization and something called ‘emergence’—the capability to generate system changes without external input.”31 his adds a layer of complexity to the assessment problem: Measuring inputs and outputs may be insuicient to understand emergent behavior that is, for all intents and purposes, unpredictable. Another way of looking at the relative complexity of an environment is to assess its linearity. In general, simple, closed systems are linear, and open, complex systems, like COIN environments, are nonlinear. his implies that a complex adaptive social system is not a truly integrated system. In many cases, the parts, or people, are not interconnected in any clear or deinitive way. Small rural villages in Vietnam, Iraq, or Afghanistan might be completely unafected by violence or other activity hundreds of miles away. here are no physical boundaries to a nonlinear complex adaptive social system because people can move in and out, often at 29 See Ackof, 1994, p. 176. 30 here is some ongoing debate on this point and disagreement with the premise that COIN is a populationcentric endeavor. his monograph is primarily intended to inform U.S. government consumers and was written to address standing U.S. policy and doctrine on COIN. Both JP 3-24 and FM 3-24 contend that COIN is population-centric. his section cites several sources that link COIN and the concept of the complex adaptive system. 31 Wass de Czege, 2009, p. 7. Bousquet (2009, pp. 175, 181) uses similar terms to describe the same dynamic. While Bousquet argues that there may be patterns to self-organization, or emergence, he does not claim that these patterns are easily identiied in the kind of open complex adaptive system we see in COIN. he systems analysis literature on emergence blends concepts and approaches from a wide array of scientiic theories and methods and uses a similar range of methods to attempt to reduce uncertainty in such systems. 32 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency will.32 here are also no deinitive psychological boundaries to such a system: People can engage or disengage from the system while remaining physically in place. Barring physical necessity (e.g., threat of violence or requirement for essential services), they can also choose whether or not to act or react to any speciic input or output at any given point in time. he less linear a system is, the more complex and chaotic it tends to be. Chaos theory is closely related to the wicked problem proposition. In scientiic terms, chaos describes something akin to the complex adaptive system: A chaotic system is highly complex, interconnected, and dynamic. However, a chaotic system is also “wildly unpredictable” and susceptible to dramatic and (arguably) unpredictable changes that can alter the entire system.33 his is the concept of emergence taken to an extreme. Chaos theorists Hal Gregersen and Lee Sailer believe that “systems exhibiting chaotic behaviors can only be understood, whereas non-chaotic systems [like a car] can be understood, predicted, and perhaps controlled.”34 Understanding, of course, assumes that accurate and relatively complete data are available, but this is rarely the case in COIN. Kelly and Kilcullen claim not only that chaos (and therefore nonlinearity) is linked to complex adaptive systems but that chaos, in fact, “makes war a complex adaptive system, rather than a closed or equilibrium-based system.”35 his monograph argues that the COIN environment can be loosely analogized to a complex adaptive system but that the nonlinearity of the COIN environment, a lack of adequate data, and the chaos of war prevent analysts from accurately dissecting a COIN campaign as they would dissect a system. A complex adaptive social system marked by openness, emergence, nonlinearity, and some degree of chaos would be diicult enough to assess objectively. However, assessment stafs are part of a military organization that works within the COIN operating environment and has efectively become part of that environment. herefore, when analysts or assessment oicers refer to the COIN environment as a complex adaptive system, they are (or should be) referring to everyone and everything in the environment, including their own organization. Every U.S. infantry unit, PRT, advisor, and aid worker adds complexity to the assessment challenge. Analysts also need to consider any exogenous factors that might afect the environment. Exogenous factors that are not easily linked to behavior through systems analysis might include foreign 32 Ackof (1994, p. 176) describes social systems as “open systems that have a purpose of their own; at least some of whose essential parts have purposes of their own; and are parts of larger (containing) systems that have purposes of their own.” 33 Gregersen and Sailer, 1993, p. 779. he science behind chaos theory focuses on inding underlying patterns in what appears to be completely random behavior. 34 35 Gregersen and Sailer, 1993, p. 798. Kelly and Kilcullen, 2006, p. 66. hey also believe that “generating an analysis sophisticated enough to derive coherent and rational whole-of-government inputs that are required by EBO is probably unattainable.” Concepts That Shape Counterinsurgency Assessment 33 media broadcasts that shape public opinion, individual communiqués with external inluencers (e.g., emails or phone calls to expatriate organizers), hard-to-trace inancial transactions in both directions across a border (e.g., hawala transfers), the impact of insurgent or refugee groups operating from across a border (e.g., Afghan Taliban members and refugees in Pakistan), the impact of foreign sponsors of insurgents or counterinsurgents (e.g., China in Vietnam or NATO in Afghanistan), and so on. Assessments are most accurate and precise when describing environments that are simple, closed, linear, and predictable. Assessments are least accurate and precise when describing environments (or systems) marked by complexity, adaptation, emergence, and nonlinearity. he violence, the threat of violence, economic displacement, and social upheaval associated with COIN all contribute to chaos, making the COIN environment both complex and chaotic. When environments slide into chaos, they become increasingly diicult to understand and assess and, arguably, next to impossible to comprehend through systems analysis. Military Doctrine and the COIN Environment To understand the complex COIN environment, the military relies on doctrine, and all military COIN assessment is shaped to some extent by military doctrine. While doctrinal manuals are not necessarily intended to be literal “how-to” guides, doctrine plays a dominant role in training, education, and operations. It shapes the way military oicers think about assessment, and it can serve as an oicial arbiter to establish uniied policy and process in a joint environment. herefore, to understand past and current military assessment methods, some knowledge of doctrine is not only helpful but also necessary. And while the majority of the relevant literature on COIN was written in the 20th and early 21st centuries, the U.S. military has been studying complexity and the chaos generated by extreme violence for more than 200 years. Its amassed knowledge on warfare also beneits from more than two millennia of wisdom on the subject, from Julius Caesar and hucydides to Clausewitz and contemporary experts. Capstone military service doctrine relects the sum of this knowledge as well as cuttingedge philosophy. In their capstone literature, both the Army and Marine Corps use a pair of very simple terms to concisely describe wicked problems, complex adaptive systems, and chaos theory: friction and the fog of war. Both terms (as applied to warfare) are attributed to Carl von Clausewitz and describe the chaotic, confusing, and often unpredict- 34 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency able nature of battle.36 he seminal Marine Corps Doctrine Publication (MCDP) 1, Warighting, deines friction as the force that resists all action and saps energy. It makes the simple diicult and the diicult seemingly impossible. he very essence of war as a clash between opposed wills creates friction. In this dynamic environment of interacting forces, friction abounds.37 he concept of the fog of war relects the inherent uncertainty of battle as described by Clausewitz, B. H. Liddell Hart, and other noted experts. Friction and the fog of war are closely linked concepts in that uncertainty contributes to friction and allows it to lourish. MCDP 1 provides more detailed guidance: All actions in war take place in an atmosphere of uncertainty, or the “fog of war.” Uncertainty pervades battle in the form of unknowns about the enemy, about the environment, and even about the friendly situation. While we try to reduce these unknowns by gathering information, we must realize that we cannot eliminate them—or even come close. he very nature of war makes certainty impossible; all actions in war will be based on incomplete, inaccurate, or even contradictory information. War is intrinsically unpredictable. . . . At best, we can hope to determine possibilities and probabilities.38 hese inal propositions—that certainty is impossible; that information is incomplete, inaccurate, and even contradictory; and that war is intrinsically unpredictable— are also the most contentious in terms of applying the approaches found in any kind of science (including social science) to the wicked problem or to COIN assessment. MCDP 1 goes on to describe the battleield as nonlinear in that causes and efects are often disproportionate. he battleield is also complex in that the millions of moving pieces, or people in the environment, decide whether to act, react, or not to react to each other. It is also luid in that “each episode merges with those that precede and follow it—shaped by the former and shaping the conditions of the latter—creating a continuous, luctuating low of activity replete with leeting opportunities and unfore36 While fog of war is widely attributed to Clausewitz, Eugenia C. Kiesling (2001) argues that he never actually uses the phrase in a way that matches doctrinal interpretation. According to Kiesling, he refers to friction more than a dozen times but to fog only three. Nevertheless, she agrees that Clausewitz intended to describe the fog of war in the way that it is now used. Not all interpretations of On War agree with this analysis. For example, Alan Beyerchen (1992–1993, p. 77) states that, “his famous metaphor of the ‘fog’ of war is not so much about a dearth of information as how distortion and overload of information produce uncertainty as to the actual state of afairs.” 37 38 Headquarters, U.S. Marine Corps, 1997b, p. 5. Headquarters, U.S. Marine Corps, 1997b, p. 7 (emphasis added). MCDP 2, Intelligence, says, “We must continually remember that intelligence can reduce but never eliminate the uncertainty that is an inherent feature of war” (Headquarters, U.S. Marine Corps, 1997a, p. 19). Concepts That Shape Counterinsurgency Assessment 35 seen events.”39 Friction, fog of war, nonlinearity, complexity, and luidity all result in disorder and uncertainty, conditions in warfare that, according to the Marine Corps, can never be eliminated. he authors of MCDP 1 conclude the discussion on the nature of war by placing warfare somewhere between art and science. his statement describes the principal challenge of the wicked problem: Human beings interact with each other in ways that are fundamentally diferent from the way a scientist works with chemicals or formulas or the way an artist works with paints or musical notes. . . . We thus conclude that the conduct of war is fundamentally a dynamic process of human competition requiring both the knowledge of science and the creativity of art but driven ultimately by the power of human will.40 Whether or not one agrees with this statement, it is the philosophical foundation of one of two major land combat services in the U.S. military.41 All doctrine— including assessment doctrine—should stem from this understanding of war. U.S. Army doctrine is similarly direct and follows these same themes: he operational environment will become extremely luid, with continually changing coalitions, alliances, partnerships, and actors. . . . Finally, complex cultural, demographic, and physical environmental factors will be present, adding to the fog of war. Such factors include humanitarian crises, ethnic and religious diferences, and complex and urban terrain, which often become major centers of gravity and a haven for potential threats. he operational environment will be interconnected, dynamic, and extremely volatile.42 he complexity and chaos inherent in the COIN environment is also acknowledged in COIN doctrine. Two documents describe the military’s approach to COIN: JP 3-24, Counterinsurgency Operations, and the Army/Marine Corps publication FM 3-24/MCWP 3-33.5, Counterinsurgency. he latter prescribes the application of distributed operations and mission command: 39 Headquarters, U.S. Marine Corps, 1997b, p. 9. 40 Headquarters, U.S. Marine Corps, 1997b, p. 19 (emphasis in original). 41 In the foreword to MCDP-1, then–Commandant of the Marine Corps Charles C. Krulak describes warighting as a philosophy: Very simply, this publication describes the philosophy which distinguishes the U.S. Marine Corps. he thoughts contained here are not merely guidance for action in combat but a way of thinking. his publication provides the authoritative basis for how we ight and how we prepare to ight. his book contains no speciic techniques or procedures for conduct. Rather, it provides broad guidance in the form of concepts and values. It requires judgment in application. (Headquarters, U.S. Marine Corps, 1997b, Foreword) 42 FM 3-0, 2008, p. 1-3. Commander’s Appreciation and Campaign Design, an experimental pamphlet published by U.S. Army Training and Doctrine Command (2008), speciically describes war as a wicked problem. 36 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Mission command is the conduct of military operations through decentralized execution based upon mission orders for efective mission accomplishment. . . . Mission command is ideally suited to the mosaic nature of COIN operations. Local commanders have the best grasp of their situations. . . . hus, efective COIN operations are decentralized, and higher commanders owe it to their subordinates to push as many capabilities as possible down to their level.43 Mission command and distributed operations are intended to shape campaigns in a way that accepts and takes advantage of complexity and chaos. Perhaps the most salient line in the quote above is the one that describes the “mosaic nature of COIN operations.” his description not only acknowledges the existence of complexity and chaos but it also establishes the idea that local context is particularly relevant in COIN. According to FM 3-24, challenges and operational solutions are asynchronous from area to area across a COIN theater. Chapter Summary COIN is a wicked problem that presents unique challenges to assessment. Most literature on COIN supports the notion that the COIN environment can be loosely analogized to a complex adaptive system that is shaped in large part by unknown or unknowable events and actions. But complex adaptive system theory is not suicient to deine or explain the COIN environment for campaign assessment, because COIN is marked by chaos, friction, and the fog of war. Joint doctrine recognizes these challenges, as does the capstone doctrine of the U.S. Army and Marine Corps. Doctrine also recognizes that COIN is perhaps the most complex, opaque, and disorderly type of warfare. his monograph argues that the already daunting challenges faced by military stafs assessing conventional war are exacerbated many times over in COIN. In theory, military doctrine on COIN assessment should relect and account for the Army and Marine Corps’ positions on complexity and chaos. However, as discussed later, military doctrine instead proscribes a one-size-its-all approach to assessment, an approach originally crafted to address conventional air campaigns. 43 HQDA, 2006c, p. 1-26 (emphasis in original). he concept of mission command for COIN is deeply rooted in U.S. Army doctrine. FM 31-20, Operations Against Guerrilla Forces, published in 1951, was intentionally written to allow for maximum lexibility, initiative, and adaptation at the tactical level. See also Birtle, 2006, pp. 134–138. CHAPTER THREE Centralized Assessment Theory and Pattern and Trend Analysis he previous two chapters described the process and requirements of assessment, as well as the challenges that the complexity and chaos of COIN pose to assessment analysts. his chapter introduces the concepts and theories that have shaped the two existing approaches to assessment examined in this study: EBA and pattern and trend analysis. Both of these approaches are centralized and almost entirely quantitative. Each relies on the idea that the people, groups, and physical places and things in the COIN environment constitute a system that can be framed and assessed through systems analysis. EBA is explicitly linked to systems analysis theory, but it also incorporates some aspects of pattern and trend analysis. he salient diference between the two approaches is in the degree to which they assume the availability of complete and accurate data. As it is described in U.S. military doctrine, EBA requires the availability of plentiful and accurate data that can be used to develop a near-comprehensive picture of the battlespace. Pattern and trend analysts tend to accept that data on COIN campaigns will be incomplete and inaccurate to a rather signiicant degree but that these data are still suicient to produce relevant centralized, quantitative analyses. he irst part of this chapter examines the general propositions that shape both EBA and pattern and trend analysis. As mentioned, both approaches rely to varying degrees on systems analysis to help commanders and policymakers understand the COIN environment. hus, the irst part of this chapter introduces the concept of systems analysis. Subsequent sections address and dispel the notion that COIN assessment can be equated with scientiic research. his distinction—that military assessment is not scientiic research or analysis—has repercussions for all aspects of centralized assessment. his discussion provides a backdrop for an examination of the most commonly used centralized analytic method in COIN assessment: time-series analysis. Because time-series charts are so prevalent in assessment reporting, it behooves practitioners and consumers (policymakers) to understand the strengths and weaknesses of time-series analysis in the context of COIN assessment. he inal two sections of the chapter present the concepts of pattern analysis and trend analysis for COIN assessment. he more complex theories and concepts of EBO and EBA are addressed in Chapter Four. 37 38 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Centralized Quantitative Assessment As discussed, both EBA and pattern and trend analysis depend to varying degrees on the idea that the COIN environment is a system, and both approaches apply varying interpretations of systems analysis to the problem: EBA explicitly refers to the COIN environment as a “system of systems,” while pattern and trend analysts, such as hayer, have attempted to apply variations of systems analysis to campaign assessment.1 herefore, a basic appreciation for the theory and practice of systems analysis (in its various incarnations) is necessary to understand both EBA and other centralized assessment methods. here is no universally agreed-upon deinition of systems analysis; it is more of a general approach to problem-solving than a speciied set of tools or methods.2 Systems analysis is, in some ways, self-deining—it is an analysis of systems—but it encompasses a wide array of theories, methods, and applications. Business analysts use systems analysis to improve the bottom line of a company. he 1960s saw the irst dedicated application of systems analysis to military decisionmaking.3 hen–Secretary of Defense Robert S. McNamara created the Oice of Systems Analysis (OSA) to conduct cost-beneit studies of big-ticket weapon procurement projects and manpower issues. hese studies led to the development of the Planning-Programming-Budgeting System, which is still taught to students of national security and used to shape DoD policy.4 In the mid-1960s, OSA expanded its application of systems analysis to help assess the Vietnam War. his expansion was predicated on the notion that systems analysis was a useful approach not only for procurement but also for policy analysis. McNamara and the OSA team believed that although systems analysis could not necessarily provide a deinitive quantitative answer to the challenges of strategy, it could be used to present a reasonably accurate summary to support decisionmaking. Alain C. Enthoven and K. Wayne Smith made this argument in their defense of military systems analysis, How Much Is Enough? Shaping the Defense Program 1961– 1 Some EBO advocates argue that EBO and EBA are not reliant on system-of-systems analysis. See Appendix E for an overview of these arguments. 2 his assertion is based on a review of the systems analysis literature and discussions with systems analysts at RAND. here are several existing deinitions of systems analysis, but they all difer from each other to some extent. 3 Operations research was irst applied systematically to U.S. defense issues during World War II, and some aspects of systems analysis may have been applied prior to the 1960s. But it seems that the creation of the Oice of Systems Analysis was the irst institutionalized application of systems analysis to defense decisionmaking in the U.S. defense community. 4 See Hitch, 1965, for a general description of eforts by the Oice of the Secretary of Defense to apply systems analysis to budgeting. Hitch was the DoD comptroller under McNamara. Centralized Assessment Theory and Pattern and Trend Analysis 39 1969.5 Both Enthoven and Smith were leading members of OSA in the 1960s. hey argued, Even the best studies leave much to be desired. And no study can account for all the variables or quantify all the factors involved. But analysis can be an aid to judgment by deining issues and alternatives clearly: by providing responsible oicials with a full, accurate, and meaningful summary of as many of the relevant facts as possible, an agreed-upon list of disagreements and their underlying assumptions, and the probable cost of hedging against major uncertainties.6 Although the current study focused primarily on defense budget analyses and only peripherally on Vietnam, it is important to note Enthoven and Smith’s argument that more systematic analysis of the Vietnam War was necessary.7 hey may not have championed the use of speciic techniques or had any personal role in COIN assessment, but an OSA team conducted extensive systems analysis of the war. homas C. hayer was the point man on Vietnam assessment for the Oice of the Secretary of Defense. He served as the director of the Southeast Asia Intelligence and Force Efectiveness Division of the Southeast Asia Programs Oice under the Assistant Secretary of Defense for Systems Analysis from March 1966 until June 1972. He applied Enthoven and Smith’s description of systems analysis to COIN to produce a series of reports, A Systems Analysis View of the Vietnam War.8 McNamara, a leading proponent of systems analyst in the U.S. government, also became embroiled in systems analyses of the Vietnam COIN campaign (see Chapter Six for a more detailed discussion). Both McNamara and hayer also accepted the exacting challenges of COIN assessment, and like Enthoven and Smith, hayer believed that systems analysis could penetrate complexity to ind useful patterns even in the most challenging environments: he answer turned out to be inding the critical patterns of the war. . . . Any given action was seldom important by itself, and at irst glance no patterns were seen. Analysis, however, revealed them. From these we, in Washington, were able to monitor the war surprisingly well by examining trends and patterns in the forces, military activities, casualties, and population security.9 5 In the chapter entitled, “Some Problems in Wartime Defense Management,” Enthoven and Smith (1971/2005, pp. 307–308) argued that more systems analysis of the Vietnam problem would have better supported strategic decisionmaking. hey did not address hayer’s work at OSA directly or in detail, but Enthoven hired hayer, and hayer conducted a thorough systems analysis of the Vietnam War for OSA. 6 Enthoven and Smith, 1971/2005, p. 63 (emphasis in original). 7 According to Enthoven and Smith (1971/2005, p. 307), “he problem was not overmanagement of the war from Washington; it was undermanagement. he problem was not too much analysis; it was too little.” 8 See hayer, 1975a, 1975b, and 1975c, for the reports referenced for the current research efort. 9 hayer, 1985, p. 5. 40 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Some of the eforts to build a centralized assessment process have been channeled into developing speciic assessment tools, typically models of some sort. Figure 3.1 represents a widely publicized efort to capture and model the “system dynamic” of the COIN operating environment as described in FM 3-24, Counterinsurgency.10 his model was developed as a systems analysis tool for COIN assessment. It is telling that the diagram is highly complex and all but unreadable in the absence of a magnifying glass. he purpose of including it here is to highlight the complexity of COIN modeling as it informs assessment, not to show the speciic aspects of the model.11 Models like this are intended to help military stafs (not just assessment analysts) envision a COIN campaign or shape operations. A similar modeling study conducted, in part, to extend “the ield of quantitative system dynamics analysis to explore modern counterinsurgency theory” provides a more detailed prescription for operations.12 Other models have been speciically designed to apply systems analysis to COIN assessment. George T. Hodermarsky and Brian Kalamajka proposed applying a “systems approach Figure 3.1 System Dynamics Model of Field Manual 3-24, Counterinsurgency OUTSIDE SUPPORT TO INSURGENT FACTIONS ANSF TACTICAL ANSF INSTITUTIONAL INSURGENTS NARCOTICS COALITION CAPACITY & PRIORITIES CENTRAL GOV’T COALITION DOMESTIC SUPPORT OVERALL GOVERNMENT CAPACITY POPULATION CONDITIONS & BELIEFS POPULAR SUPPORT TRIBAL GOVERNANCE INFRASTRUCTURE SERVICES & ECONOMY SOURCE: Mowery, 2009. RAND MG1086-3.1 10 HQDA, 2006c. 11 he diagram originated in a brieing that was widely distributed within the U.S. military and, speciically, among ISAF staf (Mowery, 2009; see the bibliography for a link to a larger version of the model). 12 See Anderson, 2011, p. 112. Centralized Assessment Theory and Pattern and Trend Analysis 41 to assessments” to deal with complexity. heir model closely parallels the EBA model described in Chapter Four and Appendix E.13 EBA is a centralized assessment methodology derived from EBO theory. It is similar to other centralized assessment in that it relies on an inlux of reports from the ield to a central repository, where the information is aggregated and analyzed. Unlike pattern and trend analysis, doctrinal EBA relies on the availability of highly accurate data. It interprets models of the COIN environment not as merely esoteric and informative, but as literal maps depicting people, groups, and their relationships with each other. Chapter Seven examines EBA as practiced in the ield. Differences Between Scientiic Research and Campaign Assessment In some cases in Vietnam and Afghanistan, centralized assessment methods have been likened to scientiic research. hus, it is important to draw a clear distinction between scientiic research and campaign assessment and to explain why assessment indings are not comparable to research indings. It is irst necessary to draw a distinction between scientiic methods and scientiic research. Scientiic methods are simply analytic tools that can be applied to problemsolving, while scientiic research is a well-deined process with stringent guidelines and requirements. here have been many instances in which analysts have applied scientiic methods to centralized campaign assessment with the intent of adding rigor to the process and output. hese eforts include the application of social science methodology to population analysis, eforts to conduct regression analyses of multiple COIN data sets, and eforts to apply mathematical models to COIN assessment. While these analysts do not necessarily claim in explicit terms that their indings can be equated with research indings, nonanalysts in the military and policy community often take them as such. As a result, much greater faith is sometimes placed in these assessments than is warranted by the methods applied. For example, GEN William C. Westmoreland, the senior military oicer in Vietnam from 1964 to 1968, stated that the HES was “scientiically developed” with “certain very precise criteria.”14 he diference between “scientiically developed” and scientiic research can be unclear to nonscientists. While it is possible to apply scientiic methods to some aspects of assessment, the application of scientiic method alone does not constitute scientiic research. A number of scientiic ields have developed or reined methods to analyze complex adaptive systems. hese methods recognize the existence of data imperfections and gaps: hey are designed to reduce uncertainty rather than produce absolute accuracy. he more rigorous the study, the more bounded the scope of the study, and the 13 Hodermarsky and Kalamaja, 2008. 14 Sheehan, 1988, p. 697. 42 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency more accurate and complete the data, the more uncertainty can be reduced. he purpose of rigorous criteria is to ensure that scientiic indings can be judged according to generally agreed-upon standards. Indeed, this is the basis for peer review. Both research and review are highly structured and time-consuming processes that tend to produce moderate advances in knowledge in a focused area of interest. he Belmont Report, produced by the National Commission for the Protection of Human Subjects of Biomedical and Behavioral Research in 1978, provides a useful and widely recognized deinition of scientiic research: [T]he term “research” designates an activity designed to test a hypothesis, permit conclusions to be drawn, and thereby to develop or contribute to generalizable knowledge (expressed, for example, in theories, principles, and statements of relationships). Research is usually described in a formal protocol that sets forth an objective and a set of procedures designed to reach that objective.15 While it is possible to apply some scientiic standards to COIN assessment, or to conduct small, isolated scientiic research studies as a subset of COIN assessment, holistic military assessment of a COIN campaign is neither general nor applied scientiic research as it is deined in the scientiic literature.16 Military stafs might conduct analysis, but this analysis tends to be far more informally structured than scientiic analysis. For military purposes, assessment and analysis are often interposed, but these terms should not be confused with scientiic research (the process of shaping a study, collecting data, analyzing data, and reporting indings) or scientiic analysis (one aspect of research). Scientiic research or experimentation depends on the internal and external validity of both method and information. he purpose of research is to improve general knowledge or general scientiic knowledge, and research projects almost always have inite objectives. But holistic military campaign assessment is by necessity a loosely structured process that shifts according to conditions and the requirements of commanders and policymakers. It would be diicult to claim that a military campaign assessment produces generalizable scientiic knowledge. he overall process of assessment is ongoing rather than inite, and assessment can continue even after a military withdrawal. It would be nearly impossible to determine the scientiic validity of a military assessment that has been delivered to a policymaker and the public. Scientiic research also requires control of the data collection, data reporting, and data management process because these data are used to measure the attributes 15 National Commission for the Protection of Human Subjects of Biomedical and Behavioral Research, 1978, p. 3. 16 If one were to draw a comparison between military campaign assessment and scientiic research, the assessment might be loosely categorized as applied evaluation research, but the structured requirements for this kind of research obviate the comparison. Centralized Assessment Theory and Pattern and Trend Analysis 43 of selected variables.17 Data for research should be collected with a reliable measuring device (e.g., a structured survey or interview protocol) that has been created according to a set of strict standards. Lack of control over data collection and management throws a wrench in the scientiic process. In COIN, the term data is used to describe any kind of information available to assessment analysts, commanders, or policymakers. he following are examples of the kinds of data that might be used to feed assessment: • a one-time narrative report on a meeting with a key leader, iled by a civilian advisor • a quantitative report showing the number of people treated at a hospital, iled by an NGO • a radio report of an enemy attack, iled by a unit in theater • a human intelligence report containing the observations and opinions of both the source and the reporter. Each individual and unit reports information diferently, depending on training, personal predilection, available resources (e.g., radio or email), and ield pressures. Most initial ield reports are iled under fairly trying conditions and often during combat, and in coalition COIN warfare, they are sometimes submitted in a number of diferent languages. here is no way to control or determine how these reports are iled in their original formats because almost all reporting is edited, collated, and aggregated in some way for incorporation into a quantitative assessment. People who are not trained scientiic researchers are responsible for data management, so the way in which ield reports are inducted into databases is neither controlled nor scientiic (as discussed later in this monograph). Even data in a speciied category, produced to meet a directed collection requirement, are not necessarily suitable for scientiic comparison. As shown in the case studies presented here, this lack of control afects data validity. hese factors do not preclude the use of scientiic methods of analysis in the production of holistic COIN campaign assessment. hey simply draw a distinction between formal research and the less formal application of methods to feed what will eventually be subjective estimation. It is still possible for assessment stafs to conduct pattern or trend analysis using time-series data, but these analyses should be clearly deined as nonscientiic and derived from a loosely controlled data set. Any efort to apply correlation analysis to difering sets of aggregated quantitative data should proceed with the utmost caution, since the correlation of uncontrolled data sets in a nonscientiic environment tends to exaggerate or hide the impact of errors, gaps, and unknowns. Consumers of assessment reporting should not mistake a holistic COIN campaign 17 Validity itself is a complex and multifarious concept that has many diferent meanings and applications in science. It is commonly used to refer to the construct validity of a study, but this section refers speciically to data validity. 44 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency assessment for scientiic research, nor should they assume that results or individual elements of the report are scientiically valid without explicit substantiation. Applying Scientiic Rigor to Assessment he seven best-practice standards for COIN assessment introduced in Chapter One do not constitute an exclusive list that precludes the application of scientiic standards to assessment and analysis. If a staf determines that it will conduct some form of centralized campaign assessment, it might draw from scientiic standards to try to improve these nonscientiic assessment eforts. Noted cultural anthropologist H. Russell Bernard has stated that “nothing in research is as important as validity.”18 While validity and reliability are important criteria, there are other means of evaluating scientiic studies that are equally important to gauging the integrity of a scientiic study. Standards from many disciplines could be applicable to any one study or process:19 • Replicability: Can the measure produce commensurate indings when applied by diferent research groups (or, in this case, assessment stafs)? • Generalizability: Can indings from the measurement be generalized across a broad population? his is used to check validity. • Understandability: Can the intended audience understand not only the indings but also the process used to generate the indings? his feeds transparency. • Congruence: Are the research methods used and the subject to be studied compatible (e.g., using both polls and macroeconomic data to measure opinion)?20 • Credibility: Is the study objective, or is bias present? • Independence: Are the data and indings redundant or unique? • Conirmability: Is there a way to independently conirm the data or indings? • Relexivity: Have all possible shortcomings in the method or measure been assessed, and are they apparent to outsiders (transparent)? 18 Bernard, 2006, p. 53. Bernard goes on to say, Ultimately, the validity of any concept . . . depends on two things: (1) the utility of the device that measures it; and (2) the collective judgment of the scientiic community that a concept and its measure are valid. In the end, we are left to deal with the efects of our judgments, which is just as it should be. Valid measurement makes valid data, but validity itself depends on the collective opinion of researchers. (Bernard, 2006, p. 60) 19 Other criteria from a range of disciplines could be applied to metrics selection. his list is drawn from a survey of operations research, social science, political science, behavioral science, and other scientiic literature. ISAF identiies two of the criteria—transparency and credibility—and U.S. military doctrine adds in several others: measurable (can be gauged either quantitatively or qualitatively), collectable (data associated with the indicator can be obtained more easily), and relevant (provides insight into a supported measure of performance or efectiveness). 20 his difers from congruence of doctrine and assessment method. Centralized Assessment Theory and Pattern and Trend Analysis 45 • Precision: How precise is the metric? What does it portend to show in quantitative terms? his term might also be used to qualify data. • Accuracy: Is the measuring device properly calibrated, or, in assessment terms, is the collector capable of proper data collection and reporting? his term might also be applied to data. Some of these terms overlap or have similar meanings, but this list is far from exhaustive when considering all possible scientiic disciplines. Assessment stafs might consider these standards when developing their processes. he rigor of scientiic thinking is very useful when it comes to brainstorming the process of centralized assessment. However, it is not possible to prescribe the use of any one of these criteria for assessment because holistic assessment should not be held to scientiic standards. How can military assessment stafs make use of the concepts found in scientiic research to improve the transparency and credibility of their work? Skeptical consumers might take the application of scientiic terminology to military assessment as a veneer, so a good irst step would be to clearly deine holistic COIN campaign assessment as nonscientiic. he best time to start shaping assessment methods and processes would be prior to deployment. It would not be necessary or even possible to apply each of the aforementioned criteria to assessment, but the list could be used to gauge each step from direction to presentation. If analysts accept that COIN assessment is a nonscientiic process, thinking through these criteria would at least help deine caveats.21 Appropriate questions might include the following: • Direction: Are the commander’s requirements for information and subordinate assessment logical and reasonable? For example, if the unit is using centralized assessment, are the core metrics generalizable, reliable, replicable, precise, relevant, measurable, parsimonious, and understandable to subordinate commands and higher-level consumers? If not, how could these gaps and inaccuracies be reconciled or explained? • Collection: Are collectors trained to collect information consistently and in the right format? Is the information required actually collectable at the local level? If not, is it because it is simply not available (e.g., determining the cost of bananas when no bananas are sold) or because it would require signiicant risk to obtain? Additionally, is the collection method accurate? In other words, is it likely to produce valid and consistent results? • Analysis: Is it possible to determine whether there are nonrandom patterns that might show progress or setbacks? Is the information received conirmable, or are assessment analysts working with information that they cannot conirm as accu- 21 In this case, the step-by-step process relects the observations of actual assessment stafs and not the doctrinal process deined in FM 5-0 (HQDA, 2010). 46 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency rate? In this case, how would they add caveats to their analyses? Is there congruence among the methods used to collect information (e.g., polls), the original CCIRs (commander’s critical information requirements), and the actual information needed by the analysts? Are the sources of data independent or circular? In other words, is the same information from a single source being reported multiple times through separate channels and therefore being interpreted as corroboration rather than redundancy? • Production and presentation: Is the assessment clear and understandable to the consumer? Is it transparent and sourced? Can the consumer “see” through the summary to ind original reporting in order to check validity? Are all the decisions and methods behind the mathematical formulas, weights, values, thresholds, metric selection decisions, and other subjective inputs to the system visible or easily obtainable? It would be impossible to apply precisely structured scientiic research to ongoing combat decisionmaking because combat decisions require timely and necessarily imperfect input and output. While campaign assessment has more extended timelines, the pressures on assessment stafs are also constant and their control over their own processes is minimal. Access to data is always limited, and data—the essence of valid research—are often lawed and inconsistent to the point that they are not scientiically reliable. Some scientiic rigor can help improve informal, nonscientiic pattern and trend analysis, but these studies should contain appropriate caveats. Time-Series Analysis [T]he issue is not so much qualitative versus quantitative approaches to measurement as the degree of complexity one is willing to admit into the measurement theory. But complexity quickly leads to perplexity, vagueness, and defeatism unless systematic and rational grounds can be established for handling this complexity. —Hubert M. Blalock22 Centralized assessment often relies on the analysis of aggregated data or information over time. his section examines the basic concept of time-series analysis as applied to military assessments, as well as some of the speciic concerns that might apply to the use of time-series analysis for assessment. Unlike relatively high-tempo conventional campaigns, COIN operations tend to develop and end slowly, over a number of years. Progress in COIN tends to be incremental and vague rather than dramatic and obvious. he United States spent more 22 Blalock, 1974, p. 3. Centralized Assessment Theory and Pattern and Trend Analysis 47 than ten years in Vietnam, shifting from an advisory approach to direct combat and, eventually, back to advisors as the south gradually crumbled. U.S. forces in Afghanistan overthrew the Taliban in direct combat in a matter of months but then became embroiled in what has become the longest continuous war in American history. Similarly, in 2003, U.S. forces quickly decimated the Iraqi Army and then engaged in a ight against an amorphous insurgency that is still ongoing at low levels in 2011. When counterinsurgents win modern COIN campaigns, they tend to win slowly over time, with violence abating gradually.23 To track incremental progress toward strategic objectives in this kind of campaign, operational stafs try to develop a baseline of information and then track and compare data over time. Reliance on time-series graphs by military assessment stafs reinforces the notions that COIN campaigns tend to be lengthy and diicult to assess. Time-series analyses depend on data acquired over time, baseline data, and (in some cases) a threshold to assess progress. Stafs set a baseline by identifying the best available sources at the time and then tying information from those sources to a visual measurement standard that can be used on the horizontal and vertical axes of a graph.24 For example, attack reporting might show that there are typically 50 attacks per day across a country. In this case, the baseline might be set at 50. he staf then determines how the indicator might show progress toward the linked objective. To assess progress, the staf must decide how often data will be collected and reported to feed the analysis. he more often this cycle occurs, the easier it should be to show progress. However, frequent collection of volatile data sets can show variability that is simply “noise” or random. For example, if the indicator is the price of grain and the price luctuates wildly from day to day, frequent reporting may show tremendous variability without long-term efect. Stafs can address this problem by shifting reporting increments: hey can collect data frequently but report it only once per month or quarter. It is important to note, however, that assessment stafs usually do not control the reporting process or timetable. To show deinitive progress (at least for EBA or pattern and trend analysis), assessment stafs tend to select a time-series threshold or milestone.25 hresholds were proliic in COIN assessment in Vietnam (and remain so, to a lesser extent, in Afghani23 Connable and Libicki, 2010, p. 15. 24 A time-series analysis might involve determinate data or (typically quantitative) data that are clearly tied to a measure of efectiveness (MOE), measure of performance (MOP) or objective, or indeterminate data. Indeterminate data are informative but not necessarily deinitive. A determinate data set might be “number of children attending school,” while an indeterminate data set might contain public opinion polling. In practice, stafs do not ordinarily draw a distinction between determinate and indeterminate data, and indeterminate data are often used to track deinitive MOE targets (e.g., “x percentage of the population supports the coalition”). 25 he diference between a threshold and a milestone is that a threshold measures progress on the y-axis against a scale (e.g., number of attacks) while a milestone measures whether a threshold has been reached by a certain period. herefore, a milestone is a threshold on the y-axis coincident to a point on an x-axis timeline. 48 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency stan). Depending on the type of data reported, the threshold will be either higher or lower than the baseline. For example, if the indicator is violence and the goal (MOE) is a reduction in violence, then the threshold will be lower than the baseline. Conversely, if the MOE calls for increased electricity production, the threshold will be higher than the baseline. hreshold selection is almost always a wholly subjective process, so the use of thresholds adds a layer of subjectivity and opacity to the assessment.26 Time-series analysis is afected by both data collection and the dynamics of complex adaptive systems marked by chaos. For example, economists deal with leading, lagging, and coincident indicators to assess macroeconomic trends. hese kinds of indicators also appear in COIN time-series analyses: • Leading indicator: Information that helps predict future events. Intelligence professionals refer to these simply as “indicators” as part of the “indications and warning” process. For example, the massing of Iraqi troops on the border of Kuwait in 1990 was a leading indicator of possible invasion. • Lagging indicator: Information that surfaces after an event occurs but helps explain or conirm that event in retrospect. • Coincident indicator: Events or information that parallels other indicators to help analysts understand the meaning of either or both. For example, high winds coupled with dark skies may indicate a coming storm; the indicators are reinforcing. Leading indicators are used to support predictive analysis, but this analysis typically requires context of some sort. Lagging indicators can considerably reshape the assessment of an event after it occurs. An example of a lagging indicator in COIN would be a shift in popular perception in the days, weeks, and months after a U.S. operation or, perhaps, an election; “atmospheric” data like these come in from various sources over time. he efect on the population might emerge slowly over time, or reports of popular perception might not be available until well after the event. It might not be apparent whether the operation or election succeeded or failed until weeks or months later. In this case, any initial assessment might have to be retroactively adjusted. Time-Series Graphs in Military Assessment Military stafs tend to report time-series assessment through graphs to help strategic commanders and policymakers see progress, or lack of progress, over time. To build 26 It adds opacity because the consumer may have no way of determining how or why a threshold was set. herefore, achieving a threshold objective may have little meaning. If the policymaker sets the threshold, this is not a concern of the military assessment staf. However, it may afect the credibility and transparency of the report when it is issued to the general public. Centralized Assessment Theory and Pattern and Trend Analysis 49 one of these time-series graphs and use it in accordance with doctrine, the staf theoretically requires the following:27 • • • • clear strategic objectives and, in the case of EBO/EBA, efects and indicators a baseline of information provided by credible sources thresholds or milestones (typically for determinative assessments) systematic, consistent information over time that can be compared with the baseline • license to retroactively adjust data and timelines to relect new knowledge. Optimally, graphs help the consumer understand one idea or, at most, a very small set of interrelated ideas. For example, a time-series graph might show the number of friendly deaths compared to enemy deaths over time. Time-series graphs are intended to present a relatively narrow range of understanding: the meaning of the data becomes clear as the data change or (according to U.S. doctrine) show variance—“a diference between the actual situation during an operation and what the plan forecasted the situation would be at that time or event.”28 Overly complex graphs are distracting and ultimately of little use in explaining data. Time-series graphs can be used to represent small-scale events, but they are more often used to describe large-scale trends over long periods. Edward R. Tufte, the author of several popular manuals on the art of visual data display, states that time-series graphs are “at their best for big data sets with real variability.”29 DoD used time-series graphs extensively throughout the Vietnam era to show progress against both the Viet Cong (VC) and North Vietnamese Army (NVA) and to show progress in paciication eforts. Both military and policy stafs use time-series graphs to help describe and analyze the COIN campaigns in Afghanistan and Iraq. Based on a review of a broad sample of assessment products from these two theaters, it appears that time-series analysis is the most commonly used method of operational assessment in contemporary Western-led COIN campaigns. Figure 3.2 depicts attack incidents reported in Iraq from 2004 to 2007 in an aggregated time-series graph. Time-series graphs are so proliic in contemporary operations that policymakers at all levels expect to see them with every assessment and in some cases demand that the assessment consist solely of maps and time-series graphs.30 DoD reports to Congress 27 his list of requirements is based on an analysis of time-series methodology, basic requirements for COIN assessment according to RAND research, and U.S. military doctrine, most of which is examined in this monograph. 28 HQDA, 2010, p. 5-6. 29 Tufte, 2006, p. 30. 30 Author interview with an ISAF assessment analyst, Kabul, Afghanistan, May 6, 2010. his conclusion is corroborated by my personal observations during a May 2–8, 2010, visit to ISAF headquarters in Kabul and during three successive tours on division-level stafs in Iraq, as well as time spent with a service-level headquarters staf in Washington, D.C., during Operations Enduring Freedom and Iraqi Freedom. 50 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 3.2 Attack Incidents Reported in Iraq, 2004–2007 SOURCE: Multi-National Forces–Iraq, 2007, slide 2. RAND MG1086-3.2 on Iraq and Afghanistan convey many key points in time-series displays. For example, ISAF’s 2009 “Unclassiied Metrics” brieing consists almost entirely of graphic polling data, maps, and time-series graphs.31 Time-series COIN assessment is an unoicial standard in step 6 (delivering the report) in the assessment process as detailed in FM 5-0.32 Time-series charts are often used as centerpieces in oral brieings. hese brieings are intended to assist military oicers or civilian defense oicials in their eforts to characterize data for policymakers, the press, and the public. For example, GEN David H. Petraeus’s 2007 congressional testimony on Iraq was framed by time-series charts.33 Figure 3.3 shows the commanding general of U.S. Forces–Iraq, GEN Raymond G. Odierno, presenting a brieing at the Pentagon on February 22, 2010, on the thenupcoming Iraqi elections. he graph to his right shows attack incidents over time (i.e., signiicant activities, or SIGACTs), and the graph to his left shows attacks by explosives over time (a subset of SIGACTs). he photograph shows common practice for both brieings and more formal assessments in the Vietnam, Iraq, and Afghanistan cases. 31 ISAF Headquarters Strategic Advisory Group, 2009. 32 HQDA, 2010. 33 See Multi-National Forces–Iraq, 2007. Centralized Assessment Theory and Pattern and Trend Analysis 51 Figure 3.3 Iraq Brieing with Time-Series Graphs SOURCE: DoD photo by Robert D. Ward. RAND MG1086-3.3 Retroactively Adjusting Time-Series Analysis and Assessment Lagging indicators, data corrections, and late reports are commonplace in COIN. To be precise, then, time-series analyses require regular review and possibly retroactive adjustment in response to the discovery of errors or new information.34 Because operational COIN assessment involves the analysis of many interrelated types of data (e.g., attack data, popular opinion, information about friendly operations), any retroactive change would also change the holistic theater assessment. Command decisions can also afect the holistic theater assessment. If a commander or staf decides that the deinition of “attack” needs adjustment halfway through a campaign, all information recorded up to that point must be adjusted to meet the new deinition. For example, if IED “inds” are no longer counted as attacks, the oicial time-series analysis must be retroactively adjusted to remove reference to all IED inds as attacks. If the staf fails to adjust the graph or report, the assessment becomes inaccurate or suspect. It may appear that there was a sudden, unexplained reduction of violence when in fact a simple change in deinition altered the entire assessment. When operational assessment stafs do try to sustain accuracy and credibility by making retroactive adjustments, they should not expect to be rewarded for their eforts. Wicked problem theory again rears its head: Senior oicers and commanders 34 Stafs will have to deine “regular” and determine how often they should review time-series data based on reporting increments, available manpower, and consumer requirements. 52 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency who have been presenting and explaining the existing graphs to policymakers and the media for months, if not years, tend to be nonplussed when the accuracy of their entire data set is thrown into question by their own assessment stafs. he suspicions of civilian leadership and the international media regarding U.S. operational assessments may lead them to see any retroactive change in data as a potential whitewash. Enthoven and Smith describe this efect on a late 1960s efort to retroactively adjust inaccurate paciication data prior to the development of the HES: he reluctance to correct the paciication numbers is not surprising. he reason most widely given was that the public would never understand why past numbers being used now difered from those used in the past. Considering the “credibility gap” problems of oicials in Saigon and Washington, this was understandable, even though the gap was widening in part because the corrections were not made.35 Further, if time-series analyses are presented at face value and not carefully explained and contextualized over time, the negative reaction to retroactive adjustments can be exacerbated. he efort to simplify assessments through time-series graphs can backire precisely because the graphs are deceptively simple: Consumers will not expect the data or deinitions to change if military stafs have not helped them appreciate the complexities behind the colored lines and bars. Stafs also should assume that any time-series graph they publish will at some point be separated from the assessment brief or report and presented without context; this has occurred routinely in both Iraq and Afghanistan. In other words, any time-series graph should be “self-contextualizing”: It should be very diicult to misinterpret the value and meaning of the data even when viewing the graph as a stand-alone product. Because time-series graphs show only one narrow aspect of a larger data set and are hard to contextualize, this is a pressing challenge. Pattern and Trend Analysis Pattern and trend analysis is one of the two existing approaches to COIN campaign assessment discussed in this monograph. Modiied with some elements of EBA, it is currently the most commonly used assessment approach in Iraq and Afghanistan, and it was the most commonly used approach in Vietnam. Pattern and trend analysis is centralized assessment that leverages a number of statistical methods to show the direction of a campaign over time. While pattern and trend analysis can be presented through narrative, tables, or maps, the most common means of displaying such data are time-series graphs. he previous section explained many of the central methods and challenges in time-series analysis and thus, by default, pattern and trend analysis. 35 Enthoven and Smith, 1971/2005, pp. 302–303 (emphasis in original). Centralized Assessment Theory and Pattern and Trend Analysis 53 his section describes the overall methodology, delineates the diference between patterns and trends, and explains the general theory behind the application of pattern and trend analysis to COIN assessment. One of the Vietnam-era OSA’s unoicial mottos was, “It is better to be roughly right than precisely wrong.” Enthoven and Smith state that this motto was “a reminder to analysts to concentrate, not on pinpoint accuracy on a part of the problem, but on approximate accuracy over the total problem.”36 hey add, “Even when uncertainties are present, it is better to use numbers than to avoid them. Quantitative analysis is possible even if there are uncertainties.”37 his approach frames the efort to ind what might be described as reasonable patterns and trends in COIN information to support policy decisions. his implies that the precision required for EBA is not obtainable, and it calls into question the use of baselines and thresholds, both of which are associated with and depend on precise and accurate data. However, in many cases, both baselines and thresholds are used in pattern and trend analysis despite the lack of precise and accurate data. For the purposes of this research efort, I draw the following distinction between pattern and trend:38 • Pattern: a consistent arrangement or behavior, not necessarily over time • Trend: prevailing tendency over time. A pattern is an identiied consistency in any kind of data. Patterns can be found over time, but for the purposes of this monograph, they are not anchored in time. It would be possible to ind patterns of movement by looking at a single aerial photograph of a crowded city street; one could ind densities of people at crosswalks and on sidewalks. In another example, birds migrate during winter seasons; this pattern is observed and revealed only over a number of seasons. he pattern is in the consistency of the behavior over time (multiple winters) and not in one speciic winter season. However, birds may or may not migrate in any one speciic season; the pattern could change for some unpredicted reason. While patterns are often used for predictive analysis, applying pattern analysis to complex adaptive environments like COIN carries risks that will become evident in the discussion of the Vietnam case study in Chapter Six. Trends indicate a consistency or shift in behavior or activity over a speciic period of time. For example, a military assessment graph might show violence dropping steadily from month to month over a one-year period in Iraq (see Figures 3.2 and 3.3, 36 Enthoven and Smith, 1971/2005, p. 68. 37 Enthoven and Smith, 1971/2005, p. 70. 38 hese deinitions are aggregated from various dictionary deinitions and reference publications on time-series analysis. hey will not meet any one single standard, and many criticisms of these deinitions will probably be valid. However, these deinitions also generally relect the working deinitions of pattern and trend in the military and policy circles most interested in the information presented here. hey also loosely relect various assessment analysts’ interpretations of pattern and trend analysis. 54 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency for example). Such a graph might be used to show causation: Violence has been dropping over this period because, e.g., troop presence increased or troops were withdrawn. Assessment stafs might conduct centralized trend analysis to determine whether an input is having an efect.39 FM Interim 5-0.1, since superseded, described this process: A measure is a data point that depicts the degree to which an entity possesses an attribute. . . . Once two or more measures are taken, they can be plotted to determine patterns and trends. hese reveal whether an attribute is more or less prevalent at diferent times. Commanders and stafs also develop a standard or baseline against which they compare measures and trends. Once established, this baseline remains a ixed reference point. From this information and analysis of why a trend is up or down, stafs can identify trouble spots and plan operations to reverse negative trends. hey can also capitalize on positive trends by determining what is causing the positive increase and apply those tactics, techniques, and procedures more broadly.40 Trend analysis is part and parcel of centralized assessment.41 It is closely tied to the concepts of variability and correlation. his monograph argues that, in COIN, the identiication of a trend will rarely, if ever, show clear causation. Nevertheless, it is possible to ind patterns or trends in inaccurate and estimated data. For example, most U.S. economic reporting is based on data with inaccuracies and gaps. he U.S. Bureau of Labor Statistics (BLS) makes it clear that data are estimates based in large part on polling and do not relect accurate counts.42 Yet, few credible experts would dispute the idea that BLS unemployment statistics relect genuine trends in unemployment. Why are the BLS statistics diferent from statistics collected and analyzed in places like Vietnam, Iraq, or Afghanistan? he most obvious diference is the environment. COIN data are collected in a hostile, chaotic, and unpredictable environment. A mix of trained and untrained people, each of whom may be risking his or her life in the process, collect, and report these data. In contrast, BLS can generally rely on a well-structured government bureaucracy that operates at the submunicipal level across a very stable country. It uses a cadre of trained and experienced pollsters and statisticians to capture and analyze the data. Unlike people in COIN environments, who sometimes shape their answers to manipulate the questioner, Americans have no signiicant incentive to throw of BLS statistics with false responses to polling. And, if the 39 Perry, 2011. 40 HQDA, 2006a, p. 5-6. Although this interim manual is no longer in efect, it provided more detail on some aspects of assessment than the updated 2010 version, FM 5-0, he Operations Process. 41 In practice, the Afghanistan case shows that there is little structured analysis of time-series data for holistic assessment. 42 U.S. Bureau of Labor Statistics, 2009. BLS uses the Current Population Survey of approximately 60,000 households to help determine unemployment statistics. Centralized Assessment Theory and Pattern and Trend Analysis 55 BLS data inaccurately represent 15 percent of the U.S. population, for example, that error might not turn out to be signiicant in terms of national-level policy. In COIN, failing to address 15 percent of the population in a national-level assessment could be disastrous.43 Assessment analysts can try to identify trends through structured techniques, and they can also use these techniques to correlate trends with other inputs. To determine meaningful correlation with both partially known variables (e.g., number of violent incidents) and unknown variables (e.g., precise locations of insurgents), it is possible to use a range of approaches backed by any number of statistical or modeling techniques from scientiic studies. For example, an assessment staf could compare two or more trends, or a pattern and a trend, through correlation analysis. his analysis might show statistical correlation (e.g., variable A went up when variable B went down, consistently over time). But because there is no control over data collection and no means of ascertaining data validity, there is no way to prove that these correlations are meaningful. In other words, there is no way to prove causation through pattern and trend analysis using common COIN data sets available to assessment stafs. herefore, correlation analysis of time-series trend data does not meet the standards for efective campaign assessment as deined here. Pattern or trend analysis in COIN is possible, but these techniques are risky for both collectors and consumers of the analysis. hese approaches tend to produce a series of narrow windows into speciic issues without providing a means of holistic analysis. he Vietnam case shows how OSA produced these “windows” in the form of dissociated data sets and charts without producing an overarching analysis. his process would later be replicated in Afghanistan. his brief introduction to pattern and trend analysis is intended to frame a more detailed discussion of the process in Chapter Four. he Vietnam War provides a well-documented case of pattern and trend analyses, as discussed in greater detail in Chapter Six. Chapter Summary All analytic methods are designed to simplify complex problems to reduce uncertainty and foster better understanding. Centralized quantitative methods, such as systems analysis, EBA, and pattern and trend analysis, seem tailor-made to address a complex challenge like campaign assessment: hey provide commanders with ostensibly objective evidence to feed requirements from senior military leaders and policymakers. hese centralized analytic methods rely to varying extents on the idea that the COIN operational environment is a system, or it is made up of interconnected and interdepen43 In On Guerrilla War, Mao Tse-Tung states that in order to win, insurgents may need the support of only 15–25 percent of the population (Mao, 2000, p. 27). Data or “patterns” that do not address populations in hostile areas can easily miss a key element of the population. 56 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency dent systems. Analysts relying on efects-based theories contend that complexity can be overcome with technology and intuitive methodology. Some analysts associate their work with a more traditional understanding of complex problems; they believe that, at best, they can provide reasonable pattern and trend analysis of these systems. No matter which approach is taken, practitioners and consumers of centralized assessment should assume that military campaign assessment cannot relect the kind of control, precision, and accuracy ofered by scientiic research. And while some military commanders and stafs might prefer to deliver assessments that are grounded in strict scientiic method, they are prevented from doing so by the realities of both military operations and the COIN environment. Most of the operational reasons are practical. Policymakers can and should shift strategic objectives as needed. Commanders will always have control over assessment and they can—and should—be able to adjust the process to meet their needs or adapt to changing circumstances as they see it. No tactical commander will ever respond in lockstep to assessment information requirements; these requirements will be modiied and sometimes ignored due to the vagaries of combat. Even a team of highly trained scientists attempting to produce a valid centralized analysis of a COIN campaign would be prevented from doing so by the lack of control over data collection and data reporting, the decentralized application of policy, shifts in strategic objectives, and routine adjustments to core metrics. It might be possible to conduct limited scientiic studies in a COIN theater, but these studies would merely serve as inputs to the holistic campaign assessment process. A broadly accepted understanding that the military’s assessment of COIN campaigns is not scientiic research might actually improve the way in which assessments are built, reported, and consumed. Centralized assessment methods are heavily reliant on the capabilities of a single staf section (or small group of top-level stafs) to produce a holistic campaign assessment. hese stafs are, in turn, almost wholly dependent on the quality and low of data from the periphery to the center, but they have little actual control over the collection and reporting of these data. Military assessment stafs should not be burdened with the requirement to ind a scientiic method that might explain aggregated quantitative data; they could instead create a more reasonable set of analytic methods (preferably, all-source intelligence analysis methods) that would be applicable to the realities of their work. Policymakers should expect any methods applied to assessment to be somewhat ad hoc, but they can use this understanding of centralized assessment to decide whether to rely on determinative or impressionistic decisionmaking. All consumers of centralized assessment reports should be aware that quantitative time-series graphs and quantitative assessment indings are necessarily imprecise and potentially misleading. hey sometimes show consistency over time that is precise but not accurate. Without context, time-series analysis conducted according to OSA’s standard for “reasonable” accuracy will, by deinition, provide only reasonably accurate results, with “reasonable” lacking clear deinition or boundaries. With this standard so Centralized Assessment Theory and Pattern and Trend Analysis 57 loosely deined, it would be unreasonable to establish and rely on precise time-series thresholds and milestones to ascertain campaign progress over time. Finally, correlation does not infer causation, so correlation of loosely structured COIN data should not form the basis for holistic campaign assessment. CHAPTER FOUR The Effects-Based Approach to Assessment Chapter hree described the general approach to centralized assessment with a focus on one of the two assessment methods used by the U.S. military (pattern and trend analysis). his chapter describes the other approach, efects-based assessment, or EBA. EBA is the U.S. military’s oicial assessment process, according to doctrine as of early 2011. EBA is derived from EBO, a process rooted in conventional air operations but intended to be applied across the full spectrum of military operations, from humanitarian relief to conventional warfare to COIN. To understand EBA, it is irst necessary to understand some of the fundamental theories and language of EBO. Efects-based theories are spelled out in oicial U.S. military publications and are woven throughout standing doctrine. But any military process (e.g., logistics, operations, maintenance) is rarely spelled out in a single doctrinal publication or manual. Assessment is no exception. his chapter draws on what are often dissociated segments of various military documents to summarize and explain both EBO and EBA. EBA is a staf process designed to assess a military campaign by gauging progress toward objectives by examining efects (MOEs) or performance (MOPs) by tracking indicators.1 It is intended to help stafs envision the operating environment and to help commanders and stafs ind clear pathways to achieve campaign objectives. his chapter describes the centralized EBA process and the systems analysis theory that is central to this type of assessment model. It reveals some of the friction inherent in the application of precision measurement and analysis to complex and chaotic environments. It also shows some of the internal contradictions in U.S. doctrine that complicate the work of assessment analysts, commanders, and policymakers alike. 1 he term objectives is used to generalize several diferent concepts in joint and service doctrine, such as end state, task, mission, commander’s intent, or termination criteria. Joint, Army, and Marine Corps terminology vis-à-vis the steps and strata in the planning and operations process is inconsistent, and injecting these various terms into an examination of EBA would not serve to clarify the EBA process. Joint doctrine states, “Assessment uses measures of performance and measures of efectiveness to indicate progress towards achieving objectives” (U.S. Joint Chiefs of Staf, 2006, p. III-27). But Army assessment doctrine states, “A formal assessment plan has a hierarchical structure—known as the assessment framework—that begins with end state conditions, followed by [MOEs], and inally indicators.” (HQDA, 2010, p. H-2). Army doctrine describes objectives as intermediate goals that might be used to help shape end state, but objectives are not necessarily integral to the assessment process. 59 60 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Subsequent chapters argue that doctrinal EBA is not applicable to distributed COIN operations because EBA is a top-down, centralized assessment process, while COIN is a bottom-up, distributed type of operation. What Is Effects-Based Operations Theory and How Does It Shape Assessment? It is necessary to understand efects-based theories and EBA to understand the oicial U.S. military approach to assessment as published in doctrine. While assessment stafs do not necessarily apply EBA literally in real-world COIN campaigns, assessment processes in both Iraq and Afghanistan have been shaped by EBA doctrine. Because parts of U.S. doctrine (and nearly all assessment doctrine) are efects-based, the training and education on assessment that does exist in the U.S. military also tends to be efects-based. Efects-based operations, or EBO, describes an approach to military planning and operations that focuses on the precise and accurate production and assessment of efects with the intent of improving military performance.2 Arguably, it difers from traditional military theories such as maneuver warfare, which tend to focus explicitly (or more directly) on objectives or missions. Proponents of EBO contend that it is a means of helping stafs and commanders focus on those actions that will help them accomplish missions and achieve objectives. Doctrinal EBO appears to be systematic, precise, and focused on immediate and cyclical actions and reactions, while contemporary interpretations of EBO describe a less precise and prescriptive and more informative process. Critics of EBO tend to compare it to traditional approaches to warfare, or the more modern incarnation of maneuver warfare, which is nonsystematic, generally imprecise by design, and focused directly on a inal end state and tasks that lead to that end state. he two sides of this debate are explored in greater detail in Appendix E. Introduction to Effects-Based Operations No single document summarizes all interpretations of EBO—and they are many and varied. However, the U.S. Joint Forces Command’s Commander’s Handbook for an 2 EBO is derived from the broader concept of network-centric warfare (NCW) and the theory of revolution in military afairs (RMA). he literature on RMA is expansive, but Lifting the Fog of War by ADM (ret.) Bill Owens (2000) and Stephen Biddle’s Military Power: Explaining Victory and Defeat in Modern Battle (2004) might sum up the two sides of the debate. NCW proposes that technology ofers contemporary Western militaries the opportunity to dominate battleields by “lifting the fog of war.” his section also describes system-of-systems analysis (SoSA) and (more briely) operational net assessment (ONA), two components of NCW that are closely linked with efects-based theory. NCW, SoSA, ONA, and EBO shaped most of the military doctrine, tactics, techniques, and procedures promulgated in the late 1990s and early 2000s. As a result, at some point in their careers, most ield-grade U.S. and NATO military oicers have been trained and educated to study efects-based theory and to apply EBO to their plans, operations, and (to a lesser extent) assessments. The Effects-Based Approach to Assessment 61 Efects-Based Approach to Joint Operations and the U.S. Army FM 5-0, he Operations Process, are the most comprehensive and applicable to this discussion. Both describe not only the theory and process of EBO but also EBA.3 Figure 4.1 shows the low of EBO from action to cascading efect. In the irst step, an action is taken, resulting in an efect. his efect then ripples outward, creating new efects with each ripple. While the chart shows three levels of efects with arrows suggesting interconnected reactions between other efects, it gives a disarming appearance of simplicity. In COIN, this would be a much more complex, three-dimensional model made up of thousands of nodes and links, most of which would not be fully understood. In this model, the United States or its allies would take an action or inject an input into the system to achieve an efect or cascading series of efects. For example, in Vietnam, the United States attempted to force the North Vietnamese to the bargaining table by strangling their ground lines of communication with deep interdiction air strikes. Analysts calculated the amount of supplies needed to support operations in the south, how many trucks it would take to move these supplies, and how much disruption in the supply lines (the efect) could be achieved through a speciied amount of bombing.4 Air crews then dropped bombs on the targets, and intelligence analysts Figure 4.1 Effects-Based Operations and Cascading Effects Action A Second-order effect First-order effect Third-order effect Direct Indirect SOURCE: Mann, Endersby, and Searle, 2002, p. 33, Figure 3. RAND MG1086-4.1 3 See Joint Warighting Center, 2006, and HQDA, 2010. FM 5-0 incorporates some elements of EBO but clearly depicts EBA. he joint manual provides a more comprehensive discussion of EBO. 4 his approach failed and also proved very diicult to assess. A military intelligence analyst at MACV Headquarters in Vietnam stated, I have always, in my own mind, used my pi factor [failure rate adjustment factor] on any battle damage assessment. But even if you go down to 50 percent factor, it’s still staggering [the number of North Vietnamese trucks destroyed]. We’re getting 200 trucks a night. We’re going to reach 20,000 trucks. And even if you take it down to 50 percent, I still can’t igure out where the hell all the trucks are coming from. (quoted in Sorley, 2004, pp. 594–595) 62 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency attempted to gauge the efect of the bombing on supply lines and NVA and VC military operations through analyses of reports and data on the operation.5 In this case, efects-based theory was applied in a very practical sense: Force is used to achieve an efect that will lead to (or contribute to) victory. Modern efects-based theories originate from U.S. Air Force models designed to shape military campaigns through an analysis of battle damage assessment and the higher-order efects that might ensue from air-delivered attacks. Chapter Two showed that, in U.S. service doctrine, both strategic and operational objectives tend to be broadly prescriptive and vague to give leeway to subordinate commanders. Subordinate units interpret the commander’s intent and higher-order missions, deriving from them speciied and implied missions (what it is they need to achieve to accomplish their objective) and then tasks (what it is they need to do to accomplish their mission).6 Instead of supplanting traditional military theory, EBO is intended to enhance performance by bridging what some perceive to be a conceptual gap between objectives and tasks. Essentially, “Efects are derived from objectives. hey help bridge the gap between objectives and tasks by describing the conditions that need to be established or avoided within the [operating environment] to achieve the desired end state.”7 he next section describes the SoSA model of the interconnected and interdependent warfare environment, a central tenet of EBO. SoSA explains how efects are used to shift behavior within a complex system like a COIN operating environment. he subsequent section presents the debate over the applicability of EBO to modern warfare. Elements of Effects-Based Theory: Deinitions and System-of-Systems Analysis To some extent, all centralized assessment techniques incorporate elements of systems analysis, and EBO is a process designed to exploit systems. At least in some inter5 See Rehm, 1985, Part II, irst section. here are no page numbers in this section, but it shows the handwritten calculations that went into this efects-based process. Many of books on Vietnam strategy also describe this process. For example, Sorley (2004) documents how these operations were tracked at the theater level for a period of two years. he analysis of efects was conducted in part by intelligence stafs and in part by assessment stafs. It was generally acknowledged to have been a failed strategy, but this failure is not necessarily an indictment of the efects-based assessment process. 6 his is a very brief interpretation of a thoroughly documented process. For more detail on the command and control process, see HQDA, 2008. For example, Commanders analyze a mission in terms of speciied tasks, implied tasks, and the commander’s intent two echelons up. hey also consider the missions of adjacent units to understand their relative contributions to the decisive operation. Results of that analysis yield the essential tasks that—with the purpose of the operation— clearly specify the actions required. his analysis also produces the unit’s mission statement—a short description of the task and purpose that clearly indicates the action to be taken and the reason for doing so. (HQDA, 2008, p. 6-8) 7 Joint Warighting Center, 2006, p. III-5. The Effects-Based Approach to Assessment 63 pretations (e.g., the Commander’s Handbook), SoSA is a theoretical subset of EBO. It describes the operating environment as an interconnected set of systems, each of which is made up of interdependent nodes, key nodes, and links. Nodes are people or things that can be targeted, while links represent relationships that can be triggered or changed through efects. he Commander’s Handbook deines the elements of SoSA:8 • An efect is the physical and/or behavioral state of a system that results from an action, set of actions, or another efect. It is also a change to a condition, behavior, or degree of freedom. • A system is a functionally, physically, or behaviorally related group of regularly interacting or interdependent elements that form a uniied whole. Systems associated with national security include political, military, economic, social, infrastructure, information, and others. • A node is an element of a system that represents a person, place, or thing. • A key node is a node that is related to a strategic or operational efect or center of gravity.9 • A link is an element of a system that represents a behavioral, physical, or functional relationship between nodes. he SoSA concept is used to deine, shape, and explain EBO and EBA processes and methods, and these models can be found in a number of joint and service doctrinal publications. Figure 4.2 presents a graphical depiction of SoSA. his kind of graphical node-and-link depiction is common in modern U.S. COIN operations. It is intended to highlight “decisive points against which the joint force can act to render the terrorist [or other] system unable or unwilling to fulill its mission.”10 It shows smaller systems (e.g., the “military” system) and how they connect to other small systems (e.g., the “social system”) within a system-of-systems. SoSA drives operational concepts for both air and ground combat, and it is used to shape the ways in which analytic methods (typically statistical analyses) are applied to COIN assessment. It fosters a belief that not only can systems analysis be used to discern quantiiable progress but that it can also provide a clear understanding of the operating environment. Because nearly all the literature on warfare— including capstone U.S. military doctrine—rejects this possibility, the ways in which SoSA is applied to warighting and assessment are problematic. he authors of the Commander’s Handbook, which was published by U.S. Joint Forces Command, rejected 8 Joint Warighting Center, 2006, p. I-3. he list very closely relects the wording used in that document. 9 As used here, “center of gravity” is a military term referring to a center of power—military, political, or other—that is crucial to the targeted entity. 10 Joint Warighting Center, 2006, p. II-5. 64 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 4.2 System-of-Systems Analysis Perspective of the Operational Environment Joint Intelligence Preparation of the Battlespace SoSA Military Information Social Infrastructure Military System Node Link Node: People and things Link: Relationship between nodes Operational center of gravity Strategic center of gravity Political Economic SOURCE: Joint Warfighting Center, 2006, p. II-2. RAND MG1086-4.2 the notion that the battleield or COIN operating environment is a complex adaptive system: [T]he efects-based concept was founded on “General Systems heory,” not “Chaos heory” or “Complex Adaptive Systems” methods addressed in the mathematical sciences. In other words, [we] view the real world as a set of systems composed of tangible elements (nodes) and their relationships (links) to each other. he nodes represent discrete elements (people, material, facilities, and information) and the links portray the physical, functional and/or behavioral relations that can exist between and among nodes and systems. Both nodes and links are only symbolic. hey are “icons” meant to simplify the complexity of the real world.11 his description is self-contradictory: It presents SoSA as both a concrete and abstract depiction of the COIN environment or system.12 his excerpt also presents 11 Joint Warighting Center, 2006, p. V-1 (emphasis in original). While the handbook relies on general system theory (GST), other EBO literature clearly accepts the existence of chaos and the nature of complex adaptive systems as applied to COIN. Speciically, Edward Allen Smith (2002, pp. 231–352) makes an efort to both account for and exploit complexity. Complexity theory, which is tied to complex adaptive systems, is not entirely distinct from GST; the models used to explain complexity rely on the idea that there is a “universal pattern of life,” as Neil F. Johnson (2009, p. 56) states in his lay introduction to complexity. 12 See Ackof, 1971, p. 662, for deinitions of abstract and concrete systems in systems analysis literature. The Effects-Based Approach to Assessment 65 GST as an alternative to complex adaptive system theory. GST applies mathematical formulas and modeling to systems with the goal of clarifying and simplifying the complex. Indeed, it is one of the foundational theories behind systems analysis.13 GST assumes that all systems (e.g., mechanical, political, social) have a basic underlying structure that can be observed and deined.14 hrough the lens of GST, a SoSA schematic of the COIN environment portrays a system that is similar to other, familiar systems, such as a car or government bureaucracy. he scientiic intent behind both chaos theory and complex adaptive system theory also is to simplify, or reduce, the complexity of the real world, but the real world they describe does not necessarily possess an underlying system that is generally analogous.15 he more salient distinction between SoSA and complexity theories is that (arguably) SoSA and EBO claim to ofer a practical and readily achievable way to see through complexity to achieve desired efects, while chaos and complex adaptive system theories ofer only tentative formulas and limited results.16 he claim that “nodes and links are only symbolic” in the Commander’s Handbook is disingenuous; the handbook contradicts itself not only here but also on page II-2, where it states, “System nodes are the tangible elements within a system that can be ‘targeted’ for action, such as people, materiel, and facilities.”17 his depiction of EBO clearly tries to directly portray real people, places, and things on a schematic SoSA map. Indeed, some analytic software used by the U.S. military to help understand insurgencies is predicated on concrete node and link analysis. At least in this accounting, the system-of-systems approach relies on the belief that it is possible to see, understand, and predict all relevant (as ascertained by the military staf) nodes and links in the operating environment.18 As discussed later, at least in COIN, this is a faulty assumption. Appendix E presents 13 For a full explanation of GST, see Bertalanfy, 1974. Bertalanfy states, “A consequence of the existence of general system properties is the appearance of structural similarities or isomorphisms in diferent ields. here are correspondences in the principles that govern the behavior of entities that are, intrinsically, widely diferent” (p. 33). 14 Yehezkel Dror of Hebrew University of Jerusalem, working at RAND in 1969, stated that GST can “be used to better analyze and explain behavior and to provide a unifying and general theoretic framework for comprehending in common terms a large number of more heterogeneous phenomena” (Dror, 1969, p. 1). 15 Commander’s Appreciation and Campaign Design recognizes this disparity and cautions against reductionism when analyzing an interactively complex system (Joint Warighting Center, 2006, p. 6). 16 It is open to debate where Enthoven et al. would come out on a scale of certainty between these two positions, but the Commander’s Handbook is fairly deinitive. 17 18 Joint Warighting Center, 2006, p. II-2. he Commander’s Handbook makes brief mention of uncertainty and describes the role of intelligence in ferreting out unseen nodes and links using both quantitative and qualitative methods. It does not clearly ofer the possibility that the operating environment could obscure nodes and links to the point that EBO would be ineffective or inaccurate. his concern is relected in eforts to apply agent-based models to COIN problem sets. See Gilbert (2008) for a lay introduction to such models. JP 2-01.3, Joint Intelligence Preparation of the Operational Environment, states, 66 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency a counterargument to this analysis of EBO, contending that SoSA and its component elements are not to be taken literally and that warighters have misread and improperly applied EBO doctrine. Effects-Based Assessment he previous section described efects-based and systems analysis theories that shape some U.S. military doctrine on COIN operations. his section examines the oicial assessment method derived from efects-based theory. Like EBO, EBA depends on a system-of-systems understanding of the COIN environment. It also uses trend analysis and therefore incorporates some of the concepts described in Chapter hree. his discussion introduces distinct terminology and theories that form the framework for the EBA process. Measures and Indicators EBA is a system of quantitative measurement. It measures efects and performance by collecting and then sorting indicators. A measure of efect, or MOE, is a standard designed to gauge progress against a military end state; a measure of performance, or MOP, measures friendly performance; and indicators are the categories of data used in the measurement process. MOEs “assess changes in system behavior, capability, or operational environment. MOEs measure the attainment of an end state, achievement of an objective, or creation of an efect; they do not measure task performance.”19 FM 5-0 states, “MOEs help measure changes in conditions, both positive and negative. MOEs help to answer the question ‘Are we doing the right things?’ MOEs are commonly found and tracked in formal assessment plans.”20 On the other hand, MOPs measure task performance. MOPs are generally quantitative, but also can apply qualitative attributes to task accomplishment. hey are used in most aspects of combat assessment, since it typically seeks speciic, quantitative data or a direct Analyzing all possible nodes and links in the operational environment would be an insurmountable task. However, not all nodes and links are relevant to the [Joint Forces Command’s] mission. [Joint intelligence preparation of the operational environment] analysts should develop their understanding in suicient detail to identify relevant systems, subsystems, nodes, and potential key nodes. (U.S. Joint Chiefs of Staf, 2009b, p. II-44) his assumes that a staf can (1) know which nodes and links are relevant, (2) ind these nodes and links and describe them in suicient detail over time to create a realistic SoSA map, and (3) determine all the ways in which friendly actions, nonfriendly actions, and happenstance might shape or reshape nodes and links and make unseen nodes relevant while making other nodes less relevant. 19 U.S. Joint Chiefs of Staf, 2010, p. IV-33. 20 HQDA, 2010, p. 6-2. The Effects-Based Approach to Assessment 67 observation of an event to determine accomplishment of tactical tasks, but have relevance for noncombat operations as well.21 An indicator is “an item of information that provides insight into a measure of efectiveness or measure of performance.”22 Table 4.1 shows the relationships between MOEs, MOPs, and indicators.23 Table 4.1 Effects-Based Assessment Terminology in U.S. Army Doctrine MOE MOP Indicator Answers the question: Are we doing the right things? Answers the question: Are we doing things right? Answers the question: What is the status of this MOE or MOP? Measures purpose accomplishment Measures task completion Measures raw data inputs to inform MOEs and MOPs Measures why in the mission statement Measures what in the mission statement Information used to make measuring what or why possible No hierarchical relationship to MOPs No hierarchical relationship to MOEs Subordinate to MOEs and MOPs Often formally tracked in formal assessment plans Often formally tracked in execution matrixes Often formally tracked in formal assessment plans Typically challenging to choose the correct ones Typically simple to choose the correct ones Typically as challenging to select correctly as the supported MOE or MOP SOURCE: HQDA, 2010, p. 6-3, Table 6-1. NOTE: FM 6-0, Mission Command: Command and Control of Army Forces, offers up “criteria of success” instead of MOPs and the idea of a “running estimate” that can be used to supplement the standard assessment (HQDA, 2003, p. 6-17). Doctrinal Effects-Based Assessment Process: Overview Building on MOE, MOP, and indicators, FM 5-0 breaks the efects-based assessment process down into six steps: 1. Gather tools and assessment data. 2. Understand current and desired conditions. 3. Develop assessment measures and potential indicators. 4. Develop the collection plan. 21 U.S. Joint Chiefs of Staf, 2010, p. IV-33. 22 HQDA, 2010, p. 6-3. 23 HQDA, 2010, p. 6-3. 68 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency 5. Assign responsibilities for conducting analysis and generating recommendations. 6. Identify feedback mechanisms.24 Each of these steps can be taken prior to deployment. EBA starts to shape operations during this predeployment phase. he assessment staf section or, doctrinally, the Assessment Working Group, helps draft MOEs derived from the commander’s objectives.25 Depending on the approach this may take the form of a complex timeline with milestones and measurable efects, or it may simply be a more detailed restatement of the commander’s intent. he ultimate step in the doctrinal assessment cycle is producing the assessment report (or reports). While doctrine suggests ways of delivering assessments, there was no joint U.S. assessment report format as of early 2011; the instructions in FM 5-0 are generic and, as is typical of any military doctrine, suggestive rather than prescriptive. Assessment stafs develop their own reporting mechanisms or, more often, tailor reports to meet consumer (commander or policymaker) preferences. Joint intelligence doctrine provides a schematic of the efects-based assessment process. Figure 4.3 shows how efects (MOEs, MOPs) are integrated between military tasks and military objectives at each level of focus—national strategic (policy), theater strategic, operational, and tactical. he national strategic level might be equated with policymaking, the theater strategic level with a theater command like MACV or ISAF, the operational with a regional or division command (~15,000 people), and the tactical with a battalion command (~1,000 people). In this model, tactical assessment focuses on task performance reporting aligned with MOPs and also “combat assessment,” such as battle damage assessment from air or artillery attacks.26 Tactical units (e.g., battalions) are responsible for describing how they are afecting enemy forces and performing military tasks.27 his process is tailored for conventional military operations. 24 HQDA, 2010, p. H-1. 25 According to FM 5-0, he assessment working group is cross-functional by design and includes membership from across the staf, liaison personnel, and other partners outside the headquarters. Commanders direct the chief of staf, executive oicer, or a staf section leader to run the assessment working group. Typically, the operations oicer, plans oicer, or senior [operations research/systems analysis] staf section serves as the staf lead for the assessment working group. (HQDA, 2010, p. 6-9) 26 JP 2-0, Joint Intelligence, Tactical-level assessment typically uses MOPs to evaluate task accomplishment. he results of tactical tasks are often physical in nature, but also can relect the impact on speciic functions and systems. Tactical-level assessment may include assessing progress by phase lines; neutralization of enemy forces; control of key terrain, people, or resources; and security or reconstruction tasks. Combat assessment is an example of a tactical-level assessment and is a term that can encompass many tactical-level assessment actions. Combat assessment typically focuses on determining the results of weapons engagement. (U.S. Joint Chiefs of Staf, 2007, p. IV-22) 27 U.S. Joint Chiefs of Staf, 2007, IV-20. The Effects-Based Approach to Assessment 69 Figure 4.3 The Effects-Based Assessment Process Level Guidance National strategic Theater strategic Assessed using measures of effectiveness—MOEs (Are we doing the right things?) End state and objectives End state and mission Objectives MOE development is supported by Joint Intelligence Preparation of the Operational Environment Effects Tasks Operational Adversary courses of action Indicators Centers of gravity High-value targets Mission Objectives Effects Tasks Tactical Mission Objectives Combat tasks (particularly fires) use Combat Assessment Tasks Battle damage assessment Assessed using measures of performance—MOPs (Are we doing things right?) Munitions effectiveness assessment Reattack or future targeting SOURCE: U.S. Joint Chiefs of Staff, 2007, p. IV-20, Figure IV-6. RAND MG1086-4.3 Selecting Measures and Indicators he Commander’s Handbook describes how staf should gauge efects at the operational and strategic levels by tracking friendly MOPs and other indicators. It ofers the preferred EBA approach to developing MOEs and MOPs during the planning process: MOEs and MOPs can be qualitative or quantitative. Whenever possible, quantitative measurements are preferred because they are less susceptible to interpretation—subjective judgment. hey demand more rigor (or proof) and can be replicated over time even if the analysts and the users—the commanders—change.28 he following text box presents a set of recommended “progress indicators” for COIN. his example is drawn directly from FM 3-24. Table 4.2 is an example of an MOE spreadsheet used to track progress against an MOE using speciic indicators.29 his example is drawn from FM 3-05.40, Civil Afairs Operations. It shows an objective, then subordinate efects with an associated timeline on the right. 28 Joint Warighting Center, 2006, p. IV-15. 29 HQDA, 2006b. 70 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency he literature on COIN assessment tends to focus on which MOE, MOP, and indicators are appropriate for a speciic campaign or for COIN in general. Chapter Five extends this brief introduction, addressing the selection of EBA metrics and the complications associated with the application of core metrics lists, such as the one in Table 4.2. Box 4.1 Example Progress Indicators in Counterinsurgency • Acts of violence (numbers of attacks, friendly/host-nation casualties). • Dislocated civilians. The number, population, and demographics of dislocated civilian camps or the lack thereof are a resultant indicator of overall security and stability. A drop in the number of people in the camps indicates an increasing return to normalcy. People and families exciled from or fleeing their homes and property and people returning to them are measurable and revealing. • Human movement and religious attendance. In societies where the culture is dominated by religion, activities related to the predominant faith may indicate the ease of movement and confidence in security, people’s use of free will and volition, and the presence of freedom of religion. Possible indicators include the following: –Flow of religious pilgrims or lack thereof. –Development and active use of places of worship. –Number of temples and churches closed by a government. • Presence and activity of small- and medium-sized businesses. When danger or insecure conditions exist, these businesses close. Patrols can report on the number of businesses that are open and how many customers they have. Tax collections may indicate the overall amount of sales activity. • Level of agriculture activity. –Is a region or nation self-sustaining, or must life-support type foodstuffs be imported? –How many acres are in cultivation? Are the fields well maintained and watered? –Are agricultural goods getting to market? Has the annual need increased or decreased? • Presence or absence of associations. The formation and presence of multiple political parties indicates more involvement of the people in government. Meetings of independent professional associations demonstrate the viability of the middle class and professions. Trade union activity indicates worker involvement in the economy and politics. • Participation in elections, especially when insurgents publicly threaten violence against participants. • Government services available. Examples include the following: –Police stations operational and police officers present throughout the area. –Clinics and hospitals in full operation, and whether new facilities sponsored by the private sector are open and operational. –Schools and universities open and functioning. • Freedom of movement of people, goods, and communications. This is a classic measure to determine if an insurgency has denied areas in the physical, electronic, or print domains. • Tax revenue. If people are paying taxes, this can be an indicator of host-nation government influence and subsequent civil stability. • Industry exports. • Employment/unemployment rate. • Availability of electricity. • Speciic attacks on infrastructure. SOURCE: HQDA, 2006c, p. 5-28, Table 5-7. The Effects-Based Approach to Assessment 71 Table 4.2 Example of Measures of Effectiveness Spreadsheet with Indicators MOE Spreadsheet Objective 1: Gain public support for U.S./coalition military forces and interim Iraqi government. Effect A: General populace supports U.S./coalition efforts. Measures Number of offensive gestures directed at U.S./coalition patrols by Iraqi civilians Number of instances involving anti-U.S./coalition graffiti Number of anti-U.S./coalition demonstrations Number of pure Iraqi events U.S./coalition respresentatives are invited to attend October November December 10 12 9 9 11 8 12 11 5 4 3 5 Effect B: Civil leadership at district and local levels supports U.S./coalition efforts Measures Number of civil or religious leaders actively supporting U.S./coalition initiatives Number of civil or religious activities U.S./coalition representatives are invited to attend October November December 20 20 25 8 10 12 Baseline Neutral Positive Negative SOURCE: HQDA, 2006b, p. 4-9, Table 4-3. Weighting the Assessment Both the Commander’s Handbook and FM 5-0, he Operations Process, encourage the use of quantitative indicators because EBO theory generally sees quantitative data as objective and qualitative reports as subjective opinion.30 hus, data are used to populate weighted mathematical models that theater-level stafs then use to determine overall progress.31 Figure 4.4 shows the mathematical assessment model suggested by FM 5-0. It is broken down into levels from end state to “condition” (another term that is not thoroughly explored in doctrine but is equivalent in some ways to end state), MOE, and subordinate indicators. 30 While FM 5-0 recognizes the need for both types of indicators, it states, “Quantitative indicators prove less biased than qualitative indicators. In general, numbers based on observations are impartial” (HQDA, 2010, p. 6-8). 31 See HQDA, 2010, Appendix H. 72 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 4.4 Field Manual 5-0 Mathematical Assessment Model End state Value: 5.5 Condition 1 Value: 4.8 Weight: 1.0 MOE 1 Value: 5.5 Weight: 1.0 Condition 2 Value: 6.2 Weight: 1.0 MOE 2 Value: 4.3 Weight: 1.5 MOE 1 Value: 7.8 Weight: 1.0 MOE 2 Value: 5.2 Weight: 1.5 Indicator 1 Value: 6.8 Weight: 1.0 Indicator 1 Value: 5.8 Weight: 1.0 Indicator 1 Value: 8.1 Weight: 1.0 Indicator 1 Value: 6.9 Weight: 1.0 Indicator 2 Value: 8.1 Weight: 1.0 Indicator 2 Value: 2.1 Weight: 2.0 Indicator 2 Value: 6.2 Weight: 1.0 Indicator 2 Value: 3.4 Weight: 1.0 Indicator 3 Value: 1.5 Weight: 1.0 Indicator 3 Value: 7.3 Weight: 1.0 Indicator 3 Value: 9.0 Weight: 1.0 SOURCE: HQDA, 2010, p. H-6, Figure H-3. NOTE: “Value” is also a subjectively assigned number in the weighting process. RAND MG1086-4.4 By weighting, the staf gives each metric a mathematical percentage value in relation to the other metrics. FM 5-0 both describes and recommends the implementation of a comprehensive weighting scheme to establish an overall single measurement of progress: A weight is a number that expresses relative signiicance. Some indicators may be more signiicant than others for informing a given MOE. hey count for more in the real world and should literally count for more in the mathematical assessment framework. Weights are used as multipliers for MOEs and indicators. he standard weight of 1.0 implies equal signiicance. A weight of 2.0 for an MOE (or indicator) implies that MOE carries twice the signiicance.32 For example, in a simpliied case, a staf would assign a percentage weight out of 100 percent to each indicator and MOE. First, they would have to determine a weighting scheme and record the justiication for the scheme to ensure transparency (this 32 HQDA, 2010, p. H-5. The Effects-Based Approach to Assessment 73 latter step rarely happens in practice). Next, the staf would have to assign a weight to each MOE: • • • • Violence: 50 percent Public opinion: 20 percent Government staing: 15 percent33 New construction: 15 percent. Once the weights were assigned, the staf could then develop a process to track percentage changes over time, precisely compare those changes, and show overall progress toward an MOE, or even an overall campaign objective, using weighted indicators. However, the doctrinal model displayed in Figure 4.4 also incorporates “value,” another subjective term tied to a 1–10 rating scale.34 If weights show the importance of an indicator to the consumer (commander or policymaker), value purports to show whether a reported indicator is “good” or “bad.” he notional example presented in FM 5-0 shows that more overall ransom money paid for hostages in a month is bad, while less would be good. his scheme interlaces two subjective rating scales and then combines them in a mathematical formula. he inherent subjectivity of this mathematical model is both compounded and opaque. Determining Thresholds for Indicators Not every assessment incorporates weighting, but some efects-based products derived from core metrics lists require the identiication of quantitative time-series thresholds. To select a threshold, the commander or staf selects a point on a graph that will identify when an efect has been achieved. he Commander’s Handbook states that quantitative measurement is not complete until the metric is compared against a speciic criteria [sic] (standard or threshold). hese thresholds can be minimums, maximums, or both. Comparison of measures against established criteria gives commanders a sense of whether they are making progress in accomplishing their objectives, efects, or tasks.35 33 his is referred to in Afghanistan as tashkil ills. 34 he FM ofers the following explanation for value: Standardization means that each component is expressed as a number on a common scale such as 1 to 5 or 1 to 10. Setting a common scale aids understanding and comparing as well as running the mathematical model. For example, Indicator 1 for MOE 1 for Condition 1 in igure H-3 [reprinted as Figure 4.4 here] could be monthly reported dollars in ransom paid as a result of kidnapping operations. For the month of June, that number is $250,000. hat number is normalized to a scale of 1 to 10, with 1 being bad and 10 being good. he value of that indicator within the framework is 6.8. (HQDA, 2010, p. H-5) 35 Joint Warighting Center, 2006, p. IV-15. 74 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 4.5 depicts this process, showing the number of attacks over time in a notional location. It shows a measure as a single point on the scale, a metric as the diference between two points (or the change), and the threshold as the acceptable number of hostile actions (about 12). FM 5-0 ofers a similar deinition and ties thresholds to the use of color-coding for assessment reporting: A threshold is a value above which one category is in efect and below which another category is in efect. hresholds answer the question for a given indicator or MOE of what good and bad is. he categories can be whatever the commander inds useful, such as colors or numbers.36 he Commander’s Handbook does not suggest any systematic method (or any method) for setting thresholds, but FM 5-0 states, “A signiicant amount of human judgment goes into designing an assessment framework. Choosing MOEs and indicators that accurately measure progress toward each desired condition is an art.”37 If this is the case, how should one choose the number 12 as an acceptable number of attacks, as depicted in Figure 4.5? Would this number be selected through a careful modeling Figure 4.5 Example of Thresholds in Quantitative Assessment 50 Number of hostile actions Measure 40 30 Metric 20 Threshold 10 0 84 91 98 105 112 Days since start of operations SOURCE: Joint Warfighting Center, 2006, p. IV-15, Figure IV-7. RAND MG1086-4.5 36 HQDA, 2010, p. H-6. 37 HQDA, 2010, p. H-5. 119 126 The Effects-Based Approach to Assessment 75 process that would show how many attacks host-nation forces could handle per reporting period, or would it simply be chosen because it “seemed about right” to a staf or commander? If the latter method is used—it seems to be the most common choice— then the selection of thresholds is a highly subjective process that exposes the entire mathematical assessment scheme to legitimate critique. Examples of Color-Coded Reports Stafs at all levels tend to deliver EBA in both graphic and narrative format. Because graphics are easier to digest, such reports are more commonly produced than narrative reports. he Commander’s Handbook ofers an example of a graphic color-coded assessment report. Figure 4.6 shows color-coded ratings for a number of desired efects. Color codes relect input from indicators: red is bad, yellow middling, and green good. Figure 4.7 depicts a modiication of the color-coded assessment as developed by stafs in Iraq and Afghanistan in the 2000s.38 his version also relies on a sliding color scale from red to green, with red depicting poor performance or results and green depicting positive performance or results. In the igure, the irst example (speciic assessment) attempts to display pinpoint accuracy, the second (ranged assessment) Figure 4.6 Example of a Color-Coded Effects-Based Assessment Report Effects Days since start of operations Assessment trend 84 1 Foreign terrorists assist an insurgency (undesired) 2 Brown regular forces stop fighting 3 Coalition forces control lines of communications 4 Populace votes 5 Media coverage in region is balanced 6 Coalition gains regional partners 7 Utility output exceeds prewar levels 85 86 87 88 89 90 91 92 93 94 95 96 SOURCE: Joint Warfighting Center, 2006, p. IV-14, Figure IV-6. RAND MG1086-4.6 38 he examples in Figure 4.7 were compiled based on observation of and participation in assessment report development groups. Data in the igure are notional. 76 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 4.7 Example of Color-Coded Assessment Reports on a Sliding Color Scale Governance 1—Specific assessment Governance 2—Ranged assessment Insurgent shadow governance Friendly governance 3—Comparative assessment RAND MG1086-4.7 is intended to account for the relative inaccuracy of COIN data, and the third (comparative assessment) shows a comparative assessment of progress and setback for both insurgents and friendly forces. hese notional examples use “governance” as the MOE (Is governance better or worse?), but this technique could be applied to any MOE or line of operation. Chapter Summary he original design of EBO has been expanded over the past two decades and is now central to joint U.S. military theory, planning, operations, and assessment. While there is ongoing debate over the relevance and eicacy of EBO (see Appendix E), efectsbased and related systems analysis concepts continue to shape standing doctrine and operations. Because EBA is derived immediately from EBO, it is also a process originally intended to address conventional warfare. It is predicated on the need to interpret a cohesive enemy force or national threat through the application of systems analysis; identify centers of gravity, critical capabilities, and critical vulnerabilities of that threat; and subsequently predict and assess second- and third-order efects of friendly actions designed to shape the threat. To be successful, EBA depends on the availability of a system-of-systems understanding of the environment and the threat, a steady low of The Effects-Based Approach to Assessment 77 complete and accurate data, and the ability to manage and analyze these data with mathematical precision. Doctrine describes some ways in which EBA might be interpreted for COIN operations, but it does not clearly describe how to modify EBA so that it can better account for the complexity, chaos, and mosaic nature of the COIN environment. here does not seem to be a way to reconcile the centralized and precise nature of EBA with the distributed and imprecise nature of COIN. Eforts to explain how EBA might contend with complexity, chaos, and the absence of clear and comprehensive data seem to highlight rather than alleviate these stark inconsistencies. Some of the literature on EBO is more articulate in attempting to address these inconsistencies, but no document reviewed for this study seems to resolve them. While this very brief introduction to the inner mechanics of EBA is not comprehensive, it does address all the essential elements that might go into an EBA product. Issues only hinted at here—subjectivity, complexity, opacity—come to life in the case studies (Chapters Six and Seven). he underlying inconsistency between complexity and the near-perfect accuracy needed to produce an EBA assessment requires the use of ield modiications designed to make EBA work in the COIN environment. hese modiications tend to result in a scattershot and sometimes inconsistent assessment process that lacks transparency and credibility. he following chapter goes into greater detail on how assessments are actually developed and presented. CHAPTER FIVE Choosing Core Metrics In today’s increasingly dynamic operating environment, [commanders] can only gain suicient situational awareness and adapt their current operations, future operations, and future plans if the staf is assessing “the right things” [the right metrics] in the operational assessment. —Commander’s Handbook for an Efects-Based Approach to Joint Operations1 Our metrics suck. —Director, ISAF Afghan Assessment Group, 2009 2 All centralized assessments, including both EBA pattern and trend analysis, use a set of “core metrics” to assess COIN campaigns. Core metrics serve a dual purpose: hey direct the collection of what is deemed to be relevant information for the assessment, and they shape the analysis of information and data once they are acquired. Selecting core metrics for centralized quantitative assessment is tricky. What matters most and why? Is information available, or will the commander have to induce risk to get new information? How should a core metric be deined for military units in the ield and for policymakers? Selecting and using a set of core metrics is diicult in practice. he indicators described in this chapter have been used in Afghanistan, and some were used in Vietnam. he purpose of providing these examples is not only to present real-world metrics but also to show, in detail, the friction points between theory and reality. hese examples also help illuminate and inform other considerations: Can one consider speciic metrics to be intrinsically good or bad for campaign assessment? Which set of metrics should be incorporated into doctrine and used in the ield? What can one learn from looking at metrics like irst-to-ire ratios? he irst-to-ire ratio is a 1 Joint Warighting Center, 2006, p. IX. 2 Quoted in Soeters, 2009, p. 11. 79 80 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency particularly timely example because this metric is in vogue among both experts and operational staf oicers in Iraq and Afghanistan.3 Other considerations shape the analysis of the core metric selection process. Is the current process of selecting core metrics adequate to provide good centralized efectsbased assessments, pattern analysis, or trend analysis? How do commanders and stafs usually select metrics, and is this process adequately rigorous? his chapter answers these questions and raises others in an efort to plumb the depths of complexity faced by policymakers, commanders, and military assessment stafs in COIN. It shows that there are no inherently good or bad metrics because no core metric is equally applicable to all situations. It also shows that the process of selecting metrics is highly subjective and often poorly understood. his analysis reveals the hidden “what-if” questions that erode the underlying assumptions behind centralized assessment and preclude the ability to apply core metrics to distributed operations overlaid on a mosaic environment. Choosing “Metrics” and Determining Values his section describes both the process of selecting individual criteria for centralized assessment and the challenges in doing so. It presents subject-matter expert positions on the selection of core metrics and examples of speciic core metrics that highlight selection challenges. he purpose is not to recommend any one process over another, but instead to explore the topic that tends to dominate contemporary assessment discussion: inding the right things to count. Many experts working to help improve centralized assessment focus not only on what to count (“good metrics”) but also on what not to count (“bad metrics”). hey tend to refer to metrics in the vernacular; they use the term in the same way FM 5-0 would describe MOE, MOP, or, more commonly, indicators. his section uses the term metrics to describe EBA-related concepts, such as MOEs, MOPs, and indicators, as appropriate, to relect various expert positions. However, core metrics are also used in pattern and trend analysis and could probably be used for any other kind of centralized assessment not identiied here. Policymakers, commanders, and stafs face considerable diiculty in selecting metrics for centralized assessment. Military stafs are tasked by doctrine with inding the “right things to count” but typically cannot settle on what those things might be for a variety of reasons. Each of the following considerations shapes the selection and analysis of core metrics to some degree: 3 I have personally observed contentious debate over the deinition of “irst to ire” in both theaters. he irst-toire ratio was often highlighted as a key metric at several levels of assessment. Choosing Core Metrics 81 • Any single core metric must be generally applicable everywhere at all times, because core metrics are intended to show progress across the entire theater.4 However, inding a metric that meets this requirement has proven to be very dificult, if not impossible. • Core metrics demand the generalized and recurring collection of information, but this is not possible in distributed COIN operations, because the information varies greatly in consistency, quality, and availability from village to village (or hamlet to hamlet). • here is no agreed-upon set of necessary qualities for a core metric, even in U.S. doctrine. Experts tend to agree that metrics should show variability (their words) and that metrics lists should be parsimonious, but they otherwise disagree or fail to provide clear guidance. • here are no agreed-upon deinitions for individual metrics, including the ones listed as examples in this monograph. herefore, it is unclear to both stafs and reporting units what they should be looking for and why. • Deinitions that do exist not only are contested but also change with policy or commander preference. When this happens, the metric and associated collection requirements change, and any eforts to maintain reliability or track variability fall apart. It is very diicult to predict how deinitions might change when selecting metrics to avoid this outcome. • It is not clear how individual core metrics should be tied to end states or how they might show progress toward those end states (or, to use a more common term, objectives). Doctrine provides insuicient guidance as to how metrics should show progress in assessment reports. Intense debates over “good” and “bad” metrics are commonplace. he Commander’s Handbook clearly states that stafs should be able to ind a list of inherently good metrics, or “the right things to assess.”5 David Kilcullen lists what he describes as good and bad core metrics in a December 2009 paper on Afghanistan metrics. First, he addresses metrics (or MOE, MOP, and indicators) that he believes can be deceptive and should be avoided. hese include counts of enemy dead, reported number of enemy attacks (reported as SIGACTs in both Iraq and Afghanistan), any metric that produces a single number out of context, and any metric that measures inputs, such as the number of schools built. He builds a case describing why each of these should be avoided. He then provides an extensive list of things that should be considered metrics. his is a sample from Kilcullen’s list: progress of NGO construction projects, 4 his consideration applies even when core metric indicators are broken into provincial or district-level blocks of data. Indicators can show diferent levels of activity or efect from area to area, but core metrics are designed to provide reliable measurement using a like standard across all areas. 5 Joint Warighting Center, 2006, p. IV-18. 82 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency transportation prices, the inluence of Taliban versus government courts, kill ratios, kill-versus-capture ratios, civilian casualties, irst-to-ire ratios, and several others along ive primary lines of operation.6 Kill ratios measure the number of casualties inlicted versus the number of casualties sufered, while the irst-to-ire ratio measures which side ired irst to determine who has the tactical initiative in combat.7 Considerations for Selecting Core Metrics Kilcullen’s intent is for operational stafs to select appropriate metrics from his list and elsewhere, shape them to it speciic needs, and adjust them over time as necessary. He argues that it is better to develop a common set of core metrics at the theater level (a point at which the operational and strategic meet) that can be readily tracked and adjusted, rather than creating long lists of metrics that may be cumbersome and narrowly applicable; he recommends parsimony. Core metrics lists are typically designed and maintained at the theater level, but in both Iraq and Afghanistan, several such lists (each somewhat diferent from the next) existed or continue to exist simultaneously at various levels. In theory, subordinate lists would be derived from a centralized list, but in practice, core metrics lists relect a mix of speciied, implied, and ad hoc items. Since the term metrics is nondoctrinal and rather vague, a core metrics list for EBA can contain MOEs, MOPs, output indicators, and input indicators. Some stafs have found these lists diicult to interpret, particularly since indicators should be subordinate to MOEs and MOPs in accordance with FM 5-0.8 In their various works on Afghanistan assessments between 2008 and 2010, both Kilcullen and Anthony H. Cordesman of the Center for Strategic and International Studies argue that selecting the right criteria is essential to the overall assessment process. Cordesman also argues for a holistic approach to objectives-based assessments and calls for the careful selection of core metrics.9 Jonathan J. Schroden of the Center for Naval Analyses ofers a detailed examination of security metrics in the Journal of Strategic Studies. He accepts the wicked problem construct and has produced a body of work on adapting various methodologies to try to compensate for data laws. In his article, he tries to steer the counterinsurgent away from traditional metrics, like casu6 Security, development, governance, rule of law, and essential services (Kilcullen, 2009b). 7 Kilcullen, 2009b. 8 Anthony H. Cordesman of the Center for Strategic and International Studies concurs with Kilcullen on core metrics but also believes that there is a lack of coordination in developing core metrics lists: Various elements of the US government—civil and military—often generate metrics and analytic models without adequate eforts to ensure comparability. Net assessment and fusion are still the exception and not the rule. he end result in war after war has been that diferent value systems are applied to some degree to US, allied, host country, and threat forces; and kinetic measures of tactical progress have more emphasis than the equally important data on ideology, politics, and governance, economics and the perceptions of the population. (Cordesman, 2008, p. 3) 9 Cordesman, 2010. Choosing Core Metrics 83 alty counts and body counts, and toward a complex analysis of attack reporting to analyze security. He states that these more complex metrics have a direct relation to several key aspects of counterinsurgency operations: removing the insurgent’s advantage of initiative; developing local forces to fight and generate intelligence; and gaining the support of and securing the local population. he same cannot be said about many other metrics currently in use.10 Parsimony is a central theme in many professional discussions of COIN assessment.11 here is general agreement that a core metrics list should be kept brief and broadly focused to ensure that subordinate commands are aforded maximum lexibility. A detailed and extensive list of metrics might not be equally applicable across an entire country (an idea that will be explored later), and the micromanagement of the assessment process even at the operational level of command—in this case, the theater operational level—would deter full participation from units in the ield. his, in turn, would prevent the acquisition of the information necessary to feed the centralized assessment process. U.S. Army assessment doctrine states, “Excessive reporting requirements can render an otherwise valid assessment plan onerous and untenable.”12 Campbell, O’Hanlon, and Shapiro discuss the idea of parsimony in metrics lists but argue that it is sometimes useful to have large quantities of data available to various stafs and researchers. heir Brookings Index on Iraq consists of about 50 indicators, while the index on Afghanistan is slightly more condensed due to the lack of available data. In their report, Assessing Counterinsurgency and Stabilization Missions, they state, If we were conident about which 10 or 15 or 20 metrics could best tell the story of the eforts in Iraq or Afghanistan, and had reliable data for those categories, we might have focused more narrowly on them. he truth is that the wars in both Afghanistan and Iraq have usually demonstrated an ability to confound short lists of metrics.13 10 Schroden, 2009, p. 741. 11 Hriar Cabayan of the Oice of the Secretary of Defense produced a list of MOEs and associated metrics for the ISAF Joint Command staf in Afghanistan. His working group consisted of approximately 70 assessment, modeling, and COIN experts from government, academia, and the military. Most of the MOEs that the group proposed were shaped to measure popular perception. he group generally concurred that if the population is the center of gravity in COIN, and winning popular support is critical to winning the campaign, then measuring popular support is the most important method in campaign assessment (Cabayan, 2010a, 2010b). 12 13 HQDA, 2010, p. H-3. Campbell et al., 2009a, p. 4. hey add, “In retrospect, some key metrics seem to emerge with greater clarity.” his trend away from parsimony and toward providing a detailed, or at least speciic, list of metrics is partially attributable to an efort to meet speciic requirements from the ield, but it also may relect a natural inclination among both academic and military experts to seek and present detail in assessments. 84 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Parsimony was one of the key axioms identiied by the participants of an assessment workshop held by the Military Operations Research Society. he results of the workshop relected the combined viewpoints of many of the leading proponents and practitioners of operations research in the U.S. government and in academia. heir list of axioms included the following: • Parsimony: Use the simplest tool that will get the job done. • Sensitivity: A good measure should support change detection within the review time frame. • Validity: he indicator must measure what we need to measure.14 • Reliability: If collected again by a diferent person or a short time later, it should yield the same answer. • Not everything that can be counted counts, and not everything counts that can be counted (citing Albert Einstein).15 It may be possible to identify desirable traits for metrics, but deciding on individual metrics is challenging. he varying subjective views of policymakers, commanders, and staf oicers can generate debate that muddies deinitions and reduces reporting incentive. If the description is too vague, each tactical unit will deine the metric differently. For example, a battalion commander might count IED attacks as enemy ire and found IEDs (i.e., discovered before they are detonated against a friendly target) as a friendly success, while another might not consider an IED to be a type of ire at all. Since IEDs make up a large portion of incidents in modern COIN, this kind of diferentiation would skew the metric to the point that it would be misleading when aggregated only one level above that at which the attacks occurred. Schroden conducted a detailed analysis of the “ratio of incident initiation” between May 2005 and August 2007 in Anbar Province, Iraq.16 While he views this as a useful metric when properly analyzed, he elaborates on the problem of deinitions: [W]e realized it was necessary to have a clear understanding of what it meant to “initiate” an incident. . . . However, in Iraq the situation is not always this clear. 14 here are several types of validity, including construct, internal, and conclusion validity. For the sake of simplifying a complex subject, this report takes a broad approach to the concept of validity, drawn primarily from social science literature. Also, occasionally the term valid will be used to describe the relative value of data. 15 16 Grier et al., 2010, p. 9. Some of these bullets are paraphrased. Schroden diferentiates between the more simplistic irst-to-ire ratio and the ratio of incident initiation. He states that “the latter allows one to deine an actual categorization scheme that makes operational sense for incidents like IEDs.” Schroden also believes that while it is not possible to obtain total accuracy with any indicator, it is possible to analyze large data sets like this one and produce operationally useful conclusions (Jonathan J. Schroden, email exchange with the author, September 29, 2010b). Choosing Core Metrics 85 . . . hus, a more rigorous deinition was needed for what it meant to initiate an incident.17 Each staf from the tactical to strategic level must agree on deinitions, or the values of the data will change (possibly more than once) as they are assimilated up the chain of command. If too much detail is provided in the description of the indicator, however, tactical commanders might feel that they are being micromanaged or might report too narrowly. Finding a middle ground is tricky, and the more contested the deinition the more likely it is to be changed over time. Each change in deinition sets of a ripple of retroactive changes in assessment or, worse, distorts current assessments. he examples in the remaining sections of this chapter are intended to demonstrate the complexity encountered by military oicers as they attempt to select core metrics by which to measure a COIN campaign. Example of a Core Metric: First-to-Fire Indicator his indicator is designed to show which side—counterinsurgent or insurgent—ired irst in a speciic engagement to determine which side generally has the initiative in combat. he idea behind irst-to-ire is that relative initiative shows either momentum or lack of momentum. Kilcullen recommends this as a useful metric because it shows which side controls the casualty rate. Further, he irst-to-ire ratio is a key indicator of which side controls the initiation of ireights, and is a useful surrogate metric to determine which side possesses the tactical initiative. If our side ires irst in most ireights, this likely indicates that we are ambushing the enemy (or mounting pre-planned attacks) more frequently than we are being ambushed. his in turn may indicate that our side has better situational awareness and access to intelligence on enemy movements than the insurgents, and it certainly indicates that we have the initiative and the enemy may be reacting to us.18 Because it is presented as a ratio, it is a quantitative metric (or indicator) requiring hard, quantitative information from the ield. Combat units would have to report which side ired irst for each speciic combat incident. Once the metric is deined and reporting requirements (i.e., how often and in what format) are established, the data start to low in from subordinate units. he theater assessment stafs will then have a data set that, in theory, shows friendly to enemy ire ratios over time. he total number of incidents would then be added up to show a comparison between friendly 17 Schroden, 2009, pp. 724–725. 18 Kilcullen, 2009b, p. 17. 86 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency and enemy-initiated attacks. First, though, the staf has to deine and incentivize the capture of data to ensure the successful use of the indicator: • What is an “incident”? his may be rather simple to deine for a single ambush. However, if there is a running, interconnected ireight between multiple units over the course of a day, should the whole battle be considered one incident or should each ight be considered an incident? his would have to be precisely deined for tactical commanders. • What is “ire”? Does this metric track direct-ire attacks only, or does it also include IED attacks? What about indirect ire? Mine strikes? A single sniper shot from an unseen foe? Does an assassination of a key government or military oicial count, or would that fall under another category? Does a friendly airstrike or artillery strike not called in by infantry units—perhaps a missile strike based on intelligence—count? • Do these deinitions match the deinitions of other metrics? In other words, is a irst-to-ire incident in the SIGACT database exactly the same as a irst-to-ire incident in other databases, or in the original deinition of the metric? If not, how will this afect cross-comparison of information or data? • Why does this metric matter? he staf should provide a detailed explanation to subordinate commands to ensure that they are incentivized to report accurately. If they do not believe that the metric is important, they may fail to report the data or simply ill in forms with little thought. How should an assessment staf interpret this ratio? Is it true that, as Kilcullen states, “If our side ires irst in most ireights, this likely indicates that we are ambushing the enemy,” or might other complexities be hiding under the surface? Assessment stafs would need to seek parity on the following considerations, among others: • What if the rules of engagement require positive identiication before a friendly unit can initiate ire? If this is the case, how often do counterinsurgents identify insurgents—i.e., friendly spots enemy irst—but have to wait to be ired upon before they are comfortable engaging? What if they are engaged while waiting for approval to ire on an enemy that they observed irst? In these cases, the incident report might be misleading. • here is little incentive for tactical units to report that they are routinely caught of-guard by insurgents. here may be great personal motivation for the combat leader to report that his side had the initiative when the matter is in doubt. his is particularly true if the command has emphasized iring irst by creating a irst-to-ire reporting requirement. • When opposing units chance upon each other (in conventional terms, a meeting engagement), the issue of who ired the irst shot may be irrelevant or misleading. Choosing Core Metrics 87 In instances like this, in which both sides are surprised, the irst shots are often wildly inaccurate or ill-considered; it may be more tactically sound to hold ire until a clear shot is ofered. • One side may initiate most of the ighting but also cause fewer casualties and, overall, be less efective.19 Undisciplined units—particularly poorly trained hostnation security forces on which counterinsurgents might have to rely in the early stages of a campaign—often discharge their weapons at the slightest hint of danger. U.S. advisors in Iraq referred to this reaction as the “death blossom.” Behavior like this is a sign of incompetence. hese units are also the ones most likely to make false claims of initiative when delivering unsupervised combat reports. In Vietnam, even U.S. combat activity reports were often conlicting and misleading for this reason.20 • Conversely, insurgents may be incompetent or relatively less competent than the security forces. hey may get of the irst shot because they can blend in with the population, but they may lose—badly—once the tactical combat is under way. Sometimes, the act of attacking is suicient to achieve a propaganda victory, but what about situations in which insurgents tend to ire irst and are then defeated, weakening their structure and possibly undermining the psychological value of their attacks? Who controls the loss rate in these situations? Is iring irst always “good,” or can it sometimes backire? While it may be possible to balance this indicator against other indicators, doing so with aggregated data would make accurate analysis diicult. • Is it possible to retain analytical context as irst-to-ire information or data are aggregated up to the policy level? Schroden does not speciically address all of these complexities, but he believes that trained analysts can break through some of the chaos of combat reporting to provide useful analysis. He argues that a shift from enemy- to friendly-initiated attacks in mid-2007 in Anbar Province was a clear indicator of success for the coalition: “Since insurgents rely on initiative to compensate for being overmatched in personnel, technology, and irepower, the shift of the incident ratio to favor friendly forces was a clear and quantitative indication that signiicant progress was being made against the insur- 19 Analysis of core metrics might try to compare these data with another data set to show whether iring irst also produced more control or more casualties. However, it is not clear how this might work in the absence of a clearer deinition of control or with body count reporting (see Chapter Six for additional discussion of this point). At any rate, comparison would have to be conducted at a relatively low level of aggregation to ensure accuracy; this kind of low-level assessment might require extensive efort at the tactical level and also obviate the need for core metric assessment. 20 hayer, 1985, p. 56. hayer describes the problems with “operational days of contact” statistics. 88 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency gency in Al Anbar.”21 his explanation closely matches Kilcullen’s description of the irst-to-ire ratio. So, assuming that good and bad metrics or indicators do, in fact, exist, is irstto-ire inherently good? To what degree do unaddressed complexities shape the data? Can one description of irst-to-ire be used in every case, or even within one campaign? Diferences in the meaning of data from place to place call into question not only the reliability and generalizability but also the validity of the irst-to-ire metric. A measuring device (e.g., an MOE or indicator) is valid if “it does what it was intended to do.”22 First-to-ire might not show what it portends to show. Taken at face value, the device might be considered valid if it showed which side ired irst. However, for the purposes of operational assessment, this may not be suicient. If the irst-to-ire metric claims to show an advantage in initiative for one side or the other, it must show actual advantage. For example, direct-ire attacks in a notional “Sector A” might show that the insurgents have found a way to initiate attacks against friendly forces while taking few casualties in return. his might show both a tactical and psychological advantage for the insurgents in Sector A. In a notional Sector B, however, the use of sniper ire (a type of direct ire that allows insurgents to ire irst in most cases) might show that the insurgents have been pushed from operating in strong, semiconventional units to relying on asymmetric attacks. his might be especially relevant if the attacks in Sector B have been inefective in that they have not caused many casualties. In this case, friendly casualty rates would be only one of many coincidence indicators or confounding variables that are not incorporated into the irst-to-ire assessment.23 A confounding variable is a variable that questions the assumption drawn from a dependent variable (indicator) or places the dependent variable in context. For example, if insurgents ired irst more often than counterinsurgents but also caused very few casualties, would the irst-to-ire data have the same meaning as it might when viewed independently? What would it mean for the overall progress of the campaign if, at the same time, 30 stores opened in the market, insurgents bombed ive schools, ten government jobs were illed, eight new roads were built, and so on? his kind of complexity can quickly overwhelm an analyst attempting to correlate incomplete and aggregated quantitative data sets. For the irst-to-ire indicator, confounding variables could include both friendly casualty rates and civilian casualty rates. What if coalition (friendly) forces in a speciic area are getting the jump on the enemy but then attack so aggressively that they cause increasing numbers of civilian casualties? his could result in tactical improvement but also a strategic setback. Considering these two variables in conjunction would require 21 Schroden, 2009, p. 726. 22 Carmines and Zeller, 1979, p. 12. 23 his metric and the SIGACT metric are also particularly vulnerable to the kinds of data paradoxes that are described later in this chapter. Choosing Core Metrics 89 the kind of context that can be found only in disaggregated, primary sources. Schroden notes that another admitted drawback of the irst-to-ire metric is “that we have counted all incidents equally, so for example a suicide bomber who kills 100 people in a single explosion is counted the same as a single bullet from a sniper. However, thus far we have been unable to identify a weighting scheme that is both clear and non– ad hoc.”24 Lack of centralized control over the collection and reporting of data in distributed operations like COIN degrades the reliability of measurement and the validity of indings. Many questions about such individual metrics remain unanswered in the peerreviewed literature, doctrine, and (particularly) in the ield in Iraq and Afghanistan. he debate over the deinition of irst-to-ire has been subjective and generally unrelated to reliability or validity, although Schroden has done some empirical analysis on this speciic metric for ISAF.25 he value of irst-to-ire as a stand-alone metric, or even as part of a broader operational assessment, appears to be uncertain but Schroden concurs with Kilcullen: “hese drawbacks aside, we maintain that the incident ratio is a better measure of progress for a counterinsurgency than many others in common use.”26 Example of Input Measures of Performance or Indicators Some indicators in use in Afghanistan and Iraq track inputs rather than outputs. In other words, they count or describe things that coalition forces or government agencies deliver (input) into the environment with the goal of achieving a certain result. For example, a government agency might input money into the economy to achieve the efect of improved economic opportunity, and then attempt to determine the success of this efort by measuring the amount of money input into the economy and not the efect of the action (or the degree to which it furthered objectives). Kilcullen places input indicators in the “bad” column. He states that they “are indicators based on our own level of efort, as distinct from the efects of our eforts. . . . hese indicators tell us what we are doing, but not the efect we are having.”27 A typical input metric in Afghanistan is tashkil ills. A tashkil is a government staf in a particular oice or station. his indicator (or MOP) shows how many government jobs in a particular district were illed by civil servants. In theory, this could show how successful the government has been in placing its civil servants and, in turn, how responsive the government was to the population. In practice, there may be laws: 24 Schroden, 2009, p. 731. 25 Jonathan Schroden, email exchange with the author, October 27, 2010d. 26 Schroden, 2009, p. 731. 27 Kilcullen, 2009b, p. 7. 90 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency • Are these civil servants actually doing any work of value, or are they adding to the anger and frustration of the population by being visibly inefective? • Have the civil servants been through any kind of training, or are they untrained and therefore inadequately prepared to do their jobs? • his input does not track corruption. If the civil servants are corrupt by Afghan standards and perpetuate the perception of the Afghan government as corrupt, their presence is counterproductive. • Are they showing up for work or staying at home? Are they actual civil servants or “ghost” employees whose paychecks are sent directly to a corrupt oicial in Kabul? his input does not measure physical presence, which, itself, is also not necessarily valuable. • Are the civil servants local or from out of the area, and how might this bear on their ability to survive and operate in their oicial positions? A counter to the criticisms of this input metric might claim that the data would not be analyzed in isolation but would be compared to other data, such as corruption levels, popular opinion of the government in that area, whether or not the oicials live near their places of work, and other factors. If a systematic and thorough analysis of this kind were being carried out with reasonably sound data and kept in context, then this might be a valid defense of the tashkil ill metric. However, this would require an individual analyst to match corruption reporting, polling data, or some like-opinion data to the performance of an individual Afghan government oicial. his is not the kind of all-source analysis that can be easily aggregated for assessments. Nonetheless, even aggregated and out-of-context input metrics like this are used to decide resource allocation at top levels of command: A commander might assume that if a district has “100-percent tashkil ills” then it must not need any more civil servants. Stephen Downes-Martin describes how even the basic arithmetic used to describe tashkil ills in a police unit—before the data were aggregated—can be misleading: Another observed example of junk arithmetic (this time leading to an overly optimistic claim) was an Afghan National Police (ANP) assessment that claimed nearly 100% tashkil illed. he underlying data, however, was that the patrolmen were overilled and NCOs [noncommissioned oicers] and oicers under-illed by signiicant amounts.28 Tashkil ills is a useful input indicator only if it is used to measure input and not as a campaign assessment tool. In other words, it can only describe how many people were assigned to a speciic oice. his kind of number is useful to budgeting oicers 28 Downes-Martin, 2010b, p. 4. Downes-Martin delivered his draft report after being embedded with the Regional Command Southwest Assessment Section both prior to its deployment and then for three months in Helmand Province, Afghanistan. Choosing Core Metrics 91 and government planners, but only for very limited purposes. It cannot describe how many of those people showed up for work, how many were competent, or how many were corrupt, efective, responsive, or tied to the insurgency. In the absence of additional information, such as performance reports, it would seem unwise to draw any operational (as opposed to logistical or bureaucratic) conclusions from this indicator. Example of Output Measures of Effectiveness or Indicators A popular output MOE or indicator counts the number of new businesses opened in a district or village.29 he metric report would state, “District A—10 new businesses opened—increase of 20 percent from last reporting period.” At the operational level (battalion, regiment, brigade, or perhaps region), the assessment staf would see that not only had ten new businesses opened but that these data relect a positive trend: More new businesses are opening in this area than was the case in the previous period. At irst glance, this would seem to show that security is improving, business is returning, people are inding ways to get their goods to market, and so on. However, if “new businesses” are reported as a core metric, the following issues would have to be taken into consideration: • Like the tashkil ill problem, there is no method to show direct correlation between opening markets and improvement in security outside of individual intelligence analysis, which is not suitable for data aggregation in operational assessment. • Further, only intelligence reporting or the local commander’s analysis could show whether these new stores relected civil stability or whether insurgents owned them, whether they were fronts for criminal gangs, or whether these businesses otherwise represented an increase in corruption or hostile control. • Is it common for businesses to open and then close quickly in this speciic area, perhaps according to the seasonal harvests? What are the prospects for longevity? If new businesses open and then fail in rapid succession, the population might lose faith (perhaps unjustiiably) in the local or national economy. • In the absence of detailed contextual narrative or very clear measuring criteria, both the increase in businesses and the percentage increase over time have little meaning. What is the optimal number (or threshold) of businesses in this district or village?30 Is there any way to show that this “optimal” number, if it can be identiied, could be tied to improved security or an improved perception of government legitimacy? his is unlikely, and there appears to be no evidence that 29 Business and commercial availability metrics have been used at various times in both Afghanistan and Iraq and were used to track performance in both post–World War II Germany and Bosnia-Herzegovina. Sources for this information are the author’s observations, interviews conducted for this study, and Grier et al., 2010, p. 8. 30 See the discussion of thresholds in Chapter Four. 92 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency this level of analysis has occurred in practice—at least not successfully—at the operational assessment level in Vietnam, Iraq, or Afghanistan. Every core metric used in Vietnam, Iraq, and Afghanistan contains internal contradictions like these; they are unavoidable because context shapes information in complex and chaotic environments. Eforts to compensate for these problems tend to focus on inding ways to compare metrics to each other to determine nonscientiic correlation (e.g., Does it seem like one thing is making another thing happen?). his kind of correlation assessment then leads the analyst to bring in more sources and focus at lower and lower levels in an efort to obtain context, which, in turn, makes centralized and aggregated quantitative assessment all but impossible. Chapter Six describes OSA cycled through this process; concerns with centralized core metrics in Afghanistan led to the development of a contextual analytic report called District Deep Dive.31 Example of Measures of Effectiveness and Indicators from Doctrine he following text box presents an example list of MOEs and indicators for stability operations as provided in FM 5-0. Although it is merely an example using notional data, the list shows many of the inconsistencies inherent in the selection of core metrics. None of the MOEs are worded similarly; most identify general, undeined standards (e.g., “improved”), while MOE 2 for condition 2 is simply a report and not an efect at all. Indicators for some MOEs are quantitative data categories that are ostensibly tied to MOEs (e.g., monthly number of reported kidnappings), while others are polling questions and are not strictly indicators. here is no explanation for why any of these MOEs are important, and the “efects” are vague. he list does not explain what “disrupted” means for condition 1, MOE 1, and it does not deine “improved” for condition 1, MOE 2. Nothing explains how these measures might be related to each other, how indicators for one MOE might show a change in another MOE, or how any of these criteria should be analyzed holistically. While this doctrinal example may not have been intended to show such detail, it does set a standard for assessment. his is typically how MOEs and indicators are presented to operational units in U.S. COIN operations. Any subordinate staf (e.g., a brigade headquarters) attempting to put such a list into efect would be forced into developing highly subjective interpretations of each MOE and indicator. It would be possible to interpret ield reporting for nearly all of these indicators as either positive or negative in the absence of clear and contextual criteria for evaluation. he MOEs and indicators suggested by FM 5-0 are shown in the following text box. 31 Because the District Deep Dive is a relatively new process, an examination of its eicacy or applicability to the overall assessment process was outside the bounds of the current research efort. Choosing Core Metrics 93 Box 5.1 Measures of Effectiveness and Indicators from Field Manual 5-0 Condition 1: Enemy defeated in the brigade area of operations. MOE 1: Enemy kidnapping activity in the brigade area of operations disrupted. • Indicator 1: Monthly reported dollars in ransom paid as a result of kidnapping operations. • Indicator 2: Monthly number of reported attempted kidnappings. • Indicator 3: Monthly poll question #23: “Have any kidnappings occurred in your neighborhood in the past 30 days?” Results for provinces ABC only. MOE 2: Public perception of security in the brigade area of operations improved. • Indicator 1: Monthly poll question #34: “Have you changed your normal activities in the past month because of concerns about your safety and that of your family?” Results for provinces ABC only. • Indicator 2: Montly K–12 school attendance in provinces ABC as reported by the host-nation ministry of education. • Indicator 3: Monthly number of tips from local nationals reported to the brigade terrorism tips hotline. MOE 3: Sniper events in the brigade area of operations disrupted. • Indicator 1: Monthly decrease in reported sniper events in the brigade area of operations. (Note: It is acceptable to have only one indicator that directly answers a given MOE. Avoid complicating the assessment needlessly when a simple construct suffices.) Condition 2: Role 1 medical care available to the population in city X. MOE 1: Public perception of medical care availability improved in city X. • Indicator 1: Monthly poll question #42: “Are you and your family able to visit the hospital when you need to?” Results for provinces ABC only. • Indicator 2: Monthly poll question #8: “Do you and your family have important health needs that are not being met?” Results for provinces ABC only. • Indicator 3: Monthly decrease in the number of requests for medical care availability from local nationals by the brigade. MOE 2: Battalion commander estimated monthly host-nation medical care availability in battalion area of operations. • Indicator 1: Monthly average of reported battalion commander’s estimates (scale of 1–5) of hostnation medical care availability in the battalion area of operations. SOURCE: HQDA, 2010, p. H-4, Figure H-2. Chapter Summary his chapter asked the question, “How hard could it be to select a practical core metrics list?” Policymakers and military stafs have discovered over the past ten years or so that the answer to this question is, “Quite hard if not impossible.” hat there was no agreed-upon single list of metrics for Afghanistan as of early 2011 drives this point home. here is a clear lack of consensus among experts and practitioners, and between experts and practitioners, over how to tackle a seemingly simple task like selecting core metrics. No assessment group applies a clear set of broadly agreed-upon standards for academic, scientiic, or even military analytic rigor in the selection and valuation of metrics. Doctrine provides little clarity and, in most cases, fails to address the realities of COIN assessment. Ultimately, this failure to settle on a core metrics list stems from 94 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency the failure of centralized COIN assessment theory to account for and address the reality of COIN. In other words, policymakers, commanders, and stafs cannot settle on a uniied core metrics list because core metrics are inherently impracticable for COIN campaign assessment. CHAPTER SIX Vietnam-Era Assessment his chapter describes key elements of the Vietnam-era assessment process, examines the use and misuse of data in the resulting assessments and in policy decisionmaking, and ties this historical example to contemporary centralized assessment. It begins with an explanation of the assessment processes used during the Vietnam War era and then explores the details of assessment requirements, assessment methods, ield collection, and reporting. he intention is to show how assessments and assessment data were used in the military and civilian strategic decisionmaking processes of the period and why lessons from Vietnam are relevant to Iraq and Afghanistan. Assessment eforts in Iraq and Afghanistan bear an irrefutable resemblance to Vietnam-era assessments, and there is a clear linear connection between U.S. assessment practices in these two periods.1 he Vietnam War ofers a breadth of resources on assessment that are unavailable from either the Iraq or Afghanistan campaigns. Arguably, the assessment processes used during the Vietnam War were the most complex and comprehensive in the history of modern COIN. Assessment analysts from Saigon to Washington, D.C., employed what were then cutting-edge computer programs to tabulate millions of reports of all kinds, including attack data, hamlet paciication data, and operational data on U.S. forces; the sheer amount of data collected in Vietnam is probably unparalleled in the history of warfare.2 Many if not most of these reports are now available to the public online or in hard copy—often measured in thousands of linear feet of paper—at the U.S. National Archives or in any one of the other Vietnam databases and archives across the country (e.g., Texas Tech University, Cornell University). Not only are the 1 his does not imply that there is a direct linear connection between the Vietnam, Iraq, and Afghanistan COIN cases. Each COIN case is necessarily unique, and there are distinct diferences from one case to another. his chapter focuses on the similarities in the assessment process employed by the United States and its allies and the similarities in the COIN environment. 2 Part of the reason for this volume of data was the sheer number of military personnel deployed in Vietnam across the peak of operations (between 1967 and 1969)—approximately ive times the number of U.S. forces in Afghanistan as of late 2010. he National Archives alone house more than 9 million Vietnam War reports of various types (including oicial cables and memos), and these collections represent only a fraction of the overall data reported. 95 96 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency raw data available and relatively easy to ind, but the scholarship on Vietnam-era assessment is also extensive, detailed, and typically of very high quality. Research for this study uncovered not only detailed analyses of the assessment process by key participants (e.g., hayer, Brigham, McNamara, Komer, Race) and contemporary observers (e.g., Karnow, Lewy) but also a series of lesser known but equally insightful analyses commissioned by the U.S. government both during and after the war (e.g., Cooper et al., U.S. Defense Logistics Agency). he Vietnam case is rich with examples and revealing analyses, but it is also a complex analytic mineield. Any research on the Vietnam War runs the risk of being either too narrow (thereby missing critical strategic context) or too shallow (missing details that explain critical decisions or complex combat environments). It is impossible to ind a perfect balance or to produce a comprehensive analysis of Vietnam-era assessment in a single chapter. his chapter is therefore admittedly incomplete. It does not address all the reasons behind key decisions at any individual level, nor does it explain all the challenges—analytic, physical, or bureaucratic—faced by those tasked with reporting and assessing the war. It also does not directly address one of the most contentious assessment issues of the war: estimates of enemy strength. his issue is too complex and contested to be adequately examined here.3 What this chapter does show is the depth of the COIN assessment challenge. It proiles the well-meaning, at times intellectually courageous, and exhaustive eforts made to collect and analyze data and why it was so diicult to collect and report those data accurately. It describes an efort to apply mathematical processes and (to some extent) scientiic rigor to centralized assessment. Indeed, there is a two-sided struggle in the centralized assessment cycle: On one side, analysts ight to obtain, collate, and understand vast reams of decontextualized data while under intense pressure from policymakers and senior military leaders to show progress; on the other side, troops in the ield are tasked with reporting data that often do not exist, in formats that make little sense, for objectives they do not understand or believe in, while also under intense pressure to show progress. his chapter provides clear evidence that combat data were often erroneous or fabricated during that period and shows how eforts to analyze these data with pattern and trend analysis fell short of expectations and were ultimately not efective in supporting policy. It also examines homas hayer’s assertion that the data quality was “reasonable” and therefore suicient to ind “deinite patterns” with which to support decisionmaking. Finally, because the Vietnam system is so similar to the Iraq and Afghanistan systems, this chapter presents a case study analysis on what was efectively modiied EBA. 3 Instead, I refer to speciic elements of data and strength estimates as part of broader examinations. For example, this chapter addresses estimates of VC infrastructure (VCI, the intelligence and political wing of the VC). Vietnam-Era Assessment 97 Orientation to Vietnam, Circa 1967–1973 A July 26, 1970, U.S. military intelligence brieing listed 44 provinces, 257 districts, 2,464 villages, 11,729 hamlets, and 1,500 miles of coastline from the demilitarized zone with North Vietnam to the border with Cambodia in the Gulf of Siam.4 Figure 6.1 is an undated map of Vietnam showing the demilitarized zone between North and South Vietnam, surrounding geography, and the U.S. military corps sectors under MACV. Figure 6.2 depicts most of the key igures in Vietnam policy and assessment that are cited or referred to in this chapter. Figure 6.1 Map of Vietnam, Including Military Sector Delineation CHINA NORTH VIETNAM Gulf of Tonkin LAOS Demilitarized zone Con Thien THAILAND I Corps II Corps SOUTH VIETNAM CAMBODIA III Corps Gulf of Siam IV Corps South China Sea NOTE: MACV consisted of four corps-level sectors. RAND MG1086-6.1 4 Sorley, 2004, p. 454. he number of villages and hamlets should have been presented as an estimated number, since it was in lux throughout the war, depending on the information available. 98 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 6.2 Photographs of Key Figures Cited or Referenced in This Chapter SOURCE: U.S. government archival photos. NOTE: From left to right, top row: William E. Colby, Robert W. Komer, Creighton W. Abrams, McGeorge Bundy; bottom row: Robert S. McNamara, William C. Westmoreland, Julian J. Ewell, Paul D. Harkins. (Thayer is not pictured.) RAND MG1086-6.2 Overview of Vietnam-Era Assessment Processes [T]he more statistics you assemble the greater your appetite becomes for even more statistics. Somehow—somehow you feel that, in the end, you can solve the whole goddamn problem if you just had enough statistics. —GEN Creighton W. Abrams5 his section briely describes the various assessment processes in use during the Vietnam War; subsequent sections explore some of these processes in detail. Assessments of the Vietnam War varied in type, purpose, and intended consumer. he entire process changed and grew between the early 1960s and the early 1970s; there is no single “Vietnam War assessment.” As in Iraq and Afghanistan, various Vietnam War assessments were published by a number of diferent organizations, all in diferent formats and based on diferent sources. While no single organization or entity was entirely responsible for providing campaign assessments to the President 5 Quoted in Sorley, 2004, p. 195 (emphasis in original). Abrams appears to have made this comment with some intended irony at a May 24, 1969, intelligence brieing in Saigon. However, he went on to state that the statistics were helpful. Vietnam-Era Assessment 99 and his advisors, two groups played key roles in the process. he Military Assistance Command, Vietnam, or MACV, was the theater-level military command in Saigon, South Vietnam. MACV was the neck of the funnel for nearly all ield reports on operations, intelligence, paciication, and other data categories. MACV commanding generals including GEN Paul D. Harkins, GEN William C. Westmoreland, and GEN Creighton W. Abrams wrote formal assessments for the President based on the work of their stafs, accumulated data, and their personal perspectives on the war.6 MACV both collated and iltered data for submission to the Oice of the Secretary of Defense (OSD) and wrote its own reports, including detailed accounts on the development of the Army of the Republic of Vietnam (ARVN).7 he other key assessment group outside the intelligence analysis community was the Oice of Systems Analysis (OSA) in the Pentagon. homas C. hayer, whose work is examined later in this chapter, was hired by Alain C. Enthoven to work on Vietnam issues in OSA.8 In an efort to improve insight into force deployments, hayer took it upon himself to put together a comprehensive scientiic review of the available data and publish reports based on his analyses.9 His oice became an assessment center that applied systems analysis methodology to COIN campaign assessment. hayer points out that he never had the opportunity to task data collection to the ield, but others in OSD, including Secretary of Defense Robert McNamara, had that authority and exercised it often. hayer’s work was published separately in what were labeled “unoicial” reports, while the data he amassed were used to feed other reports in various sections of OSD. Some of the MACV reports and nearly all of the OSD reports relied heavily on aggregated quantitative data and pattern and trend analysis. In some cases, data collection requirements were developed to meet perceived operational or strategic needs; in other cases, they were speciically designed to provide data that would show some kind of progress without context. For example, in 1968, MACV reported the number of cakes of soap it had issued to Vietnamese villagers in 1967 (572,121), an irrelevant input metric.10 Data lowed up from the hundreds of thousands of troops on the ground, province advisors, military advisors to Republic of Vietnam units, U.S. civilian oicials, and U.S. intelligence oicers, as well as Republic of Vietnam military units, government agencies, and civilian development teams. hese data were then fed into catalogs and computer databases, including the Hamlet Evaluation System (HES), the 6 See, for example, Westmoreland, 1967a. he Measurement of Progress report was the most consistently published MACV theater-level holistic assessment. 7 MACV, 1969c. 8 Rehm, 1985, p. 4-3. 9 Rehm, 1985, p. 4-3. 10 Lewy, 1978, p. 93. It was irrelevant to campaign assessment, though it might have been relevant to the oicer in charge of the budget. 100 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Terrorist Incident Reporting System, the Territorial Forces Efectiveness System, the Paciication Attitude Analysis System (PAAS), the Situation Reports Army File, and many others.11 “Data” could mean anything from a simple number (e.g., rounds ired in a single artillery attack) to a more complex set of ostensibly correlated survey data (e.g., hamlet evaluation data). Often overlooked are the thousands of detailed narrative assessments written by military and civilian oicers working in the ield.12 hese assessment reports, containing quantitative and qualitative data as well as analysis, are probably the best sources of contextual reporting on the Vietnam War. hey describe in depth the successes and failures of the Americans and the Vietnamese (both North and South), potentially provide rich context, and serve as de facto histories of the war at the district and province levels. hese reports sufered from the same kind of subjectivity as any ield report, and most were necessarily focused in scope; thus, they do not provide comprehensive, theater-wide assessments of the entire U.S. COIN campaign in Vietnam. Despite the fact that MACV relied on narrative province reports as one of two “primary documents upon which the periodic theater-wide assessment of overall progress in the paciication efort is made” (a claim that is not relected in the campaign assessment), a 1969 MACV order limited province reports to four pages for easier reading by senior leaders.13 he challenges of assessment in Vietnam are the same challenges faced in any COIN campaign, including Iraq and Afghanistan: How should policymakers determine progress and decide strategy if all they have to choose from is inaccurate, decontextualized, and aggregated numbers or thousands of pages of lengthy narrative? In Vietnam, theater-level reports produced by military and military intelligence oicers tried to bridge the gap between these two options, as did those produced by various government and intelligence agencies in Washington, D.C. While some of these 11 Some of these are reporting categories, others are databases, and not all were necessarily fully computerized. hayer describes some of these databases and systems, but many others were available to him during his service with OSA. He did not necessarily control all of these databases, and some may have been compiled after his departure. he Cornell University Library and the U.S. National Archives now hold many of these reports. Introduction to the Paciication Data Bank, a government pamphlet released in 1960, lists and describes the automated paciication databases that existed in November 1969. hey included HES, the Territorial Forces Efectiveness System, the Terrorist Incident Reporting System, the Revolutionary Development Cadre System, the Assistance in Kind System, the Refugee System, the Village and Hamlet Radio System, the People’s Self Defense Force System, and the VCI Neutralization System, the last of which was crossed out in the document. he Paciication Data Bank itself was designed to provide “the ability to construct reports to meet the management and analysis requirements of individual users” (Introduction to the Paciication Data Bank, 1969, p. 1). he list is not exhaustive; there were a number of other automated databases in use that were not related to paciication but also fed campaign analysis at OSA. he Southeast Asia Province File, Project Corona Harvest, the System for Evaluating the Efectiveness of [South Vietnamese] Subsystems, and the Air Summary Data Base are all available through the Cornell University Library in Collection Number 4406. 12 See, for example, MACV, 1968a. 13 MACV, 1969b, Annex A, p. 7. Vietnam-Era Assessment 101 reports were useful and many were well written, they often contained the laws by virtue of their overaggregated data analyses or lengthy narratives. Douglas Kinnard’s 1974 survey of 110 U.S. general oicers who had been involved with the Vietnam War showed that only 2 percent felt that the system used to measure the war was valid.14 his is a resounding dismissal from those who used and fed the assessment system throughout the war. heir viewpoint is by no means an empirical refutation of the MACV or OSD assessments, but it shows that Vietnam-era COIN assessment was not adequate. As the remainder of this chapter shows, the laws in the system were pervasive. he application of this lawed system had dire consequences for the personnel in the ield and for U.S. national security. Thomas C. Thayer and the Southeast Asia Analysis Reports An examination of the Vietnam-era assessment process should ideally begin with an overview of the assessment process at OSA. As director of OSA’s Southeast Asia Intelligence and Force Efectiveness Division, homas C. hayer led the most comprehensive analyses of COIN metrics to date. his section explains hayer’s role in the Vietnam-era assessment process, the problems hayer faced in terms of data accuracy, and hayer’s proposition that “reasonably accurate” data are suicient to ind deinite patterns and trends in a COIN campaign. hayer’s reports, and his analysis of his own work in War Without Fronts, are perhaps the best and most accessible insights into the nuts and bolts of COIN assessment. Upon joining OSA, hayer already had credibility as an analyst, having served three years in the ield in Vietnam as a civilian (albeit on speciic technical projects and operations research initiatives).15 During the war, he published his analyses in more than 50 editions of the Southeast Asia Analysis Report, later republished in a 12-volume series titled A Systems Analysis View of the Vietnam War 1965–1972.16 Although his reports were ostensibly unoicial, there is little doubt that they were taken by most readers as oicial reports issued by OSD.17 hayer himself lists the accolades he received from consumers of his reports, including the White House, DoS, and MACV. Because hayer’s reports were then the only serious efort to aggregate all the available and hard-to-obtain war reporting, and because hayer’s analyses were well written and interesting, the Southeast Asia Analysis Reports may have been as inluential as those issued by MACV or the Central Intelligence Agency (CIA)—and perhaps even 14 See Kinnard, 1977, p. 172. 15 Rehm, 1985, p. 4-1. 16 See hayer, 1975a, 1975b, and 1975c, for volumes referenced for this study. 17 hayer (1985, p. 261) acknowledges this concern. 102 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency more so.18 With their COIN assessments, hayer and his analysts shaped decisionmaking on the Vietnam War, and hayer himself could easily be considered the most inluential operational and strategic assessment analyst in the history of U.S. COIN. Limitations of Thayer’s Data To understand how hayer used the data he collected, it is important to irst understand the quality of the data in question. hroughout War Without Fronts and in his other publications and public discussions, he reairms that the data were often of very poor quality. While he has some biases as the senior analyst and he can never know the full degree of data inaccuracies or gaps, he is remarkably forthright in showing the errors that he knew existed. It is possible that the actual inaccuracies in these data were much more or less problematic than hayer’s reports. he following quotes from War Without Fronts address the quality of speciic data types encountered by hayer and his team of analysts. Counting Insurgents. “he estimators [of enemy force data] were aware of the great uncertainties in their estimates and did their best to furnish a reasonably accurate picture of the communist forces” (p. 30). he postulated combined estimate of total forces ranged between 395,025 and 482,452 (p. 31). Palmer concurs that estimates of combined enemy force data—NVA and VC—were generally “best guesses” (Palmer, 1984, p. 79) Communist Attacks. “he most serious problem in dealing with the oicial U.S. igures is that they do not include communist actions reported by the Vietnamese National Police and other civilian authorities” (p. 44). Regarding ground assaults: “he igures probably understate the actual rate at which ground attacks took place because communist attacks in reaction to allied operations were seldom, if ever, included in the data” (p. 46). Anti-Aircraft Incidents. “[W]hile the number of air sorties tripled the number of anti-aircraft incidents fell from 6,800 in 1971 to 800 in 1972. Such a drastic change in the face of intense combat suggests that the pilots simply stopped reporting antiaircraft ire, not that the ire itself had stopped” (p. 51). hayer’s speculation may be accurate, but in the absence of ield investigation, there is no way to determine why these numbers dropped or whether they were accurate in the irst place. ARVN Leadership. “American advisors’ assessments gave the impression that all of the ARVN division commanders were capable. Experienced observers disagreed” (p. 62). If this was commonly the case, as hayer asserts, then these speciic data as used to determine host-nation military readiness for the period in question are lawed. 18 hayer is quoted describing how he managed to obtain the data for his reports in Rehm, 1985, pp. 4-2–4-3. He essentially found the data with the assistance of his analysts; there was no systematic process to funnel data to hayer’s oice, no order from on high granting him access to hidden sources, and no comprehensive data collection plan. Vietnam-Era Assessment 103 Civilian Casualties. “[N]o oicial U.S. estimate of civilian casualties exists for the Vietnam War. [A]n estimate is developed here and compared with another made by the U.S. Senate Subcommittee on Refugees but both estimates are unoicial guesses” (p. 125). Hamlet Security and Development Data (pre-HES) 1964–1967. “he data support the notion of an optimistic bias” (p. 138). “he method of reporting also suggests that the population in communist controlled areas was understated because the pre-1967 joint system counted only the hamlets planned for paciication. It ignored the ones not in the paciication plan, most of which were probably communist hamlets” (p. 143). herefore, the 1964–1967 data described only hamlets in friendly control. his statement also assumes a great deal about those hamlets not covered by this process. Hamlet Security and Development Data (HES). “In terms of percentages, the gain in [GVN control over the population] is from 42 percent of the total population in 1964 to 93 percent in 1972” (p. 141). his shows that in the year in which the United States essentially declared defeat and withdrew from Vietnam, the notoriously inefective and corrupt GVN controlled 93 percent of the South Vietnamese population to some degree. his unlikely statistic is based on what are questionable data (as discussed later in this chapter), and, as Race points out, control is irrelevant in the absence of willing support. hayer also quotes a hamlet-level poll showing that only “54% of the respondents generally agreed with the HES description of their hamlet” (p. 151). On January 16, 1971, a senior oicer at a MACV intelligence brieing referred to this disparity between HES and the polling data, stating that this “means our HES rating is optimistic—in the view of the people” (quoted in Sorley, 2004, p. 523). By the end of 1972, there were only about 35,000 U.S. troops in Vietnam, so most of this information must have come from GVN and ARVN reporting, which was typically less reliable than U.S. reporting.19 The People’s Self-Defense Forces Rosters. “he igures are notoriously unreliable” (p. 170). he data purported to show how many members of the People’s SelfDefense Forces, a local civil defense organization, were on the active roll on a yearly basis. PAAS (a survey of popular sentiment). “he PAAS attempted to portray urban and rural South Vietnamese attitudes toward security, politics and economic development. . . . Any systematic efort to portray attitudes and beliefs is subject to error . . . and the conditions in South Vietnam further limited the accuracy.” Because it relied on semistructured interviews and not standard polling methods, “the results must be viewed as being much less precise” than those of opinion polls conducted in the United States. “Quota rather than probability sampling techniques were used to select the hamlets and the individual respondents, so the sample from which interviews were drawn was not necessarily an accurate representation of the South Vietnamese Popu19 he purpose of these quotes is to identify data laws only. 104 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency lation” (pp. 174–176). hayer uses the remainder of this chapter of his book to draw fairly comprehensive conclusions from the PAAS data. In March 1970, MACV’s Civil Operations and Revolutionary Development Support (CORDS) director William Colby stated, “here is no fully reliable system for assessing overall public attitudes in Vietnam so most judgments are only estimates” (U.S. Senate Foreign Relations Committee, 1970, p. 417). Estimating VC Infrastructure (clandestine agents). American oicials “attempted to estimate the numbers and types of clandestine communists by adopting the techniques used by police everywhere to compile lists of persons wanted for crimes . . . [and] communist subversive presence.” his information “can be obtained from questions found in the HES and the PAAS, although these estimates are also tenuous” (pp. 205–206). Estimates of VCI were tenuous and used methods that were not clearly articulated. Furthermore, it is not clear that every estimate of VCI from all U.S. sources relied on standard police techniques; poorly informed guesswork probably fed these statistics to a signiicant degree. hese estimates, however, were used to measure success based on a body count formula or, according to hayer, “VCI neutralization.” A September 12, 1970, estimate by MACV put the number of VCI personnel at 65,000, of which 37,000 had been “identiied by name and position.”20 his number did not include “guerrilla or low-level supporters.” So, assuming that analysts were actually able to identify 37,000 individuals as clandestine agents—in and of itself a staggering intelligence coup—the total number was still a gross estimate of only part of the VCI. his did not stop MACV from setting VCI “neutralization” objectives; in one case, the objective was to kill or capture 1,800 VCI personnel per month. It is not clear how this number was agreed upon.21 Refugee Statistics. he statistics were designed to “identify numbers of individuals to whom payments were due, not to count all refugees and war victims in South Vietnam.” However, “outside observers believed they represented the total number of refugees. . . . If [the refugees] didn’t register [with the government] they didn’t show up in the statistics” (pp. 221, 225). “Despite their lack of precision, the refugee igures are fairly reliable in indicating the magnitude of the problem and they did luctuate with the tempo of combat. When combat increased, so did the number of refugees. When it decreased the number of refugees did too” (p. 222). hayer’s caveats are straightforward and helpful in understanding the statistics. However, it may or may not be true that the actual number of refugees rose and fell with combat. While this statistic may relect a broad trend, it might instead relect increased eforts by GVN and U.S. refugee assistance program oicers to travel to areas that had recently experienced ighting 20 21 Sorley, 2004, p. 478. his estimate is taken from the MACV Weekly Intelligence Estimate Update. Sorley, 2004, p. 303. Colby states that the primary objective was to capture, not kill, VCI personnel (Sorley, 2004, p. 381). Vietnam-Era Assessment 105 to register new refugees.22 his would relect an artiicial bump in the data created by proactive registration and would not relect the existence of refugees in areas that had no registration oicials. Because refugees could not register where there were no GVN oicials, NGOs (e.g., Red Cross), or U.S. oicials, no refugees in communist-held or in hotly contested areas could be counted.23 Refugees also may lee when faced with threats or intimidation. How many refugees were created when U.S. forces ceded areas to communist control, as often happened? How many refugees did the communists displace in their own areas, thereby losing popular support? How many people registered for refugee payments because they needed money and were not, in fact, refugees?24 Were GVN statistics—the primary source of refugee data, according to former Ambassador Robert W. Komer— accurate? Were the United States and GVN deining refugees the same way? hese relevant inputs are unknown, but Komer doubted the accuracy of the GVN data.25 he CIA described a ghost population of unregistered refugees who did not register “to circumvent local government policies which exclude them from certain areas.”26 William Colby described the magnitude of the data gap at a brieing with the MACV senior staf: “We estimate something between three and four million people have been in refugee status in this country sometime in the past three years.”27 Further, it seems incongruous to state that the statistic was not designed to identify all refugees in the war zone and then to state that the level of combat had a clear impact on the rise and fall of the “number of refugees.” It would be safe to say that more ighting probably causes more refugees in a speciic area for a speciic period of time, but the data do not support any broader conclusions. South Vietnamese Military Casualties After 1972. According to hayer, the low number of battle deaths reported by the South Vietnamese was used by Congress “as 22 Field reporting shows that this is exactly what transpired in at least some cases. For example, in Phong Dinh Province in early 1968, the GVN surged refugee services to meet an increase in refugees after intense ighting in the provincial capital. he oicial numbers did not include an estimate of an additional 4,000 refugees who had not registered (MACV, 1968a, pp. 12–13). 23 While there may be exceptions to this assumption, they are probably few and statistically insigniicant. 24 Refugee and damage claims could be lucrative. A Vietnamese citizen could claim solatia, or compensation payments of thousands of dollars, and could pursue claims against both the United States and GVN simultaneously for the same incident. he U.S. government paid more than $4.7 million in claims as of March 1970 (U.S. Senate Foreign Relations Committee, 1970, p. 420). 25 Komer, 1967, p. 12. 26 CIA, 1966, p. VII-2. his declassiied report also states that there may have been between 30,000 and 40,000 refugees living in only one of nine Saigon precincts, while GVN statistics listed only 1,518 in the entire city. 27 Sorley, 2004, p. 459. Colby presented this 3–4 million igure on July 26, 1970. But in his congressional testimony in March of that year, he stated, “About 3,500,000 people have been uprooted during the past six years” (U.S. Senate Foreign Relations Committee, 1970, p. 420). he latter estimate seems much more accurate than the former by virtue of its presentation as a more precise number. his disparity exempliies the casual way in which statistics are misrepresented—even unintentionally—in oicial brieings. 106 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency part of the rationale for slashing aid to South Vietnamese forces during the summer of 1974. he problem was that the oicial South Vietnamese igures for battle deaths [for 1974] turned out to be twice as high as the igures reported to Washington in the operational messages” (p. 256). his comment highlights the danger of depending on host-nation reporting to determine policy. The Concept of “Reasonable” Accuracy hayer believed that ield reporting in Vietnam was signiicantly inaccurate, and that many of the aggregated data sets that he worked with were also signiicantly inaccurate. Yet, like some of his contemporaries and many modern assessment analysts, he believed that the data were “reasonably” accurate and could be used to show useful patterns and trends. He presented these patterns and trends in his reports, along with the data he collected from MACV and other sources. he idea that useful patterns and trends can be gleaned from aggregated data that are known to be both signiicantly inaccurate and incomplete drives to the heart of the debate over centralized assessment. It is helpful to understand the depth of the chasm between “accurate” and “reasonable” before examining the “good-enough” proposition, or the idea that useful patterns and trends can be identiied through centralized assessment. he following two sections of this chapter build upon hayer’s descriptions of data quality to provide in-depth analyses of two speciic data sets used by hayer and other Vietnam analysts: body counts and hamlet paciication evaluations. he subsequent section, “Pattern and Trend Analysis of Vietnam War Data,” addresses the idea of pattern and trend analysis in holistic assessment. he chapter concludes with a discussion of how trend analysis was used to guide strategy and to try to shape popular opinion on the war and whether lessons from Vietnam are relevant to modern COIN campaign assessment in Iraq and Afghanistan. Body Counts [hey were the] bane of my existence and just about got me ired as a division commander. hey were gross exaggerations by many units primarily because of the incredible interest shown by people like McNamara and Westmoreland. I shudder to think how many of our soldiers were killed on a body-counting mission—what a waste. he immensity of the false reporting is a blot on the honor of the Army. A fake—totally worthless. —U.S. general oicers’ opinions of body-count data28 28 Quotes from Kinnard, 1977, p. 75. he irst comment was in response to Kinnard’s survey of U.S. general oicers in 1974. Vietnam-Era Assessment 107 his section addresses the body count metric in the context of the Vietnam War. he purpose of this section is to (1) reveal laws in this set of aggregated data, (2) provide a detailed example of data collection and reporting problems in COIN, and (3) build a foundation for the later discussion of data analysis and strategic decisionmaking. Body count was a term for the oicially mandated tallying of the bodies of enemy combatants killed by U.S. and GVN military forces. his assessment method had been used in conventional contests, including World War II and Korea; in Vietnam, its implementation is attributed to Secretary of Defense McNamara, General Westmoreland, and Westmoreland’s predecessor, GEN Paul D. Harkins.29 he previous section described homas C. hayer’s caveats of Vietnam-era aggregated data. In War Without Fronts, hayer also expressed doubt regarding the veracity of estimates of enemy dead coming from the ield: “It is doubtful whether anybody, including Hanoi [Democratic Republic of Vietnam], really knows how many communist troops died.”30 hayer goes on to detail why body count estimates from the ield were so inaccurate: the communists’ emphasis on reclaiming bodies, diicult terrain, duplicate reporting, and the temptation to exaggerate, among other reasons. He then describes in detail the unsuccessful methods that his assessment staf used to try to compensate for the problems with body count data and time-series analysis. “Considerable efort was made to check the validity of the communist loss estimates but the results were not conclusive.”31 Just how accurate were the body-count statistics? How were these data used? here is both anecdotal and oicial evidence from the tactical to the strategic level that body-count reporting was taken seriously and that it shaped both strategy and assessment, for better or worse. LTC Douglas S. Smith, a mechanized infantry commander in Vietnam, had his reservations about body counts and weapon seizure statistics but thought them accurate enough to be a useful indicator: [Before I arrived, one of our units’] body count for VC . . . was 103 in a six-month period. he U.S. losses during this time frame was 36 which was about a 3:1 ratio. . . . [In] the past almost six months now that I’ve had the battalion, we see almost a reversal. We see our body count in the same period of time nearing 750 and our losses down in the vicinity of 25 killed. So our casualties have gone down and the number of VC eliminated has gone up tremendously.32 It is impossible to know whether or not Smith’s information was accurate, as he claims, down to single digits; he does not describe how he checked the accuracy of information from his subordinate commanders, and there is no record to show whether 29 he process started with Harkins (Sheehan, 1988, p. 287). 30 hayer, 1985, p. 101. 31 hayer, 1985, p. 102. 32 Smith, 1969. 108 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency his data were checked by a higher-level unit. Body counts igured prominently in oicial unit reporting and were typically used as the primary gauge of success in both combat operations promotions.33 Most anecdotal and much of the oicial reporting, however, seems to show that body count data were inaccurate or created from whole cloth. Oicial Army history shows that the MACV intelligence staf iled a series of formal complaints to MACV that claimed military stafs were underestimating enemy strength and overstating enemy losses.34 It is not clear that senior staf and oicers really had any idea how accurate the numbers were from any source at any one time.35 A transcript of an intelligence brieing delivered to GEN Creighton W. Abrams (Westmoreland’s replacement as the senior oicer in Vietnam) on March 5, 1971, reveals this uncertainty at the top of the military intelligence and operations hierarchy in Vietnam. In response to Abrams’s question regarding the veracity of ARVN body count reporting, the briefer states that “it is probably a physical impossibility for them to tabulate the bodies, especially after B-52 strikes. . . . But at the same time, we feel that the number from air and artillery and so forth that are never uncovered probably make up the diference.”36 When asked whether the ARVN estimates were accurate to within an order of magnitude, he responded, “I feel so.”37 An oicial CIA analysis of the body count provided by U.S. and GVN military forces in the wake of the Tet Ofensive called the estimate “exceedingly diicult to accept.”38 Guenter Lewy describes a survey of former ground commanders at the Army War College in 1968. He states that 60 percent of the oicers thought that body counts were based on a “combination of body counting and estimates” and that “body counts were usually ‘upped,’ sometimes honestly and sometimes with great license.” Some of these oicers reported pressure to turn in higher counts and that “this pressure was 33 See, for example, Sykes, undated. 34 Cosmos, 2006, p. 86. As stated earlier in this chapter, estimates of enemy strength were the most contested elements of campaign and intelligence assessments throughout the war. Westmoreland sued CBS News for broadcasting a report that accused him and MACV staf of falsifying estimates to sustain the impression that the United States was succeeding in its attrition warfare strategy. Testimony and evidence from the case are revealing and provide good insight into the issue of estimates. he RITZ and CORRAL programs, the testimony of former Marine and then-Congressman Paul N. McCloskey, and the letters of MACV intelligence oicer James A. Meacham are particularly informative. See Westmoreland v. CBS, 1984. 35 here is considerable evidence that the numbers were sometimes intentionally falsiied, or known errors were papered over. Sorley (2004, pp. 221–222) includes an example of a military intelligence briefer preparing to deliver a brief on enemy strength who talks about how to “explain away” a gap of 72,000 enemy forces between varying estimates. Another briefer states that the number is “too big to cry, and it hurts too much to laugh.” 36 Sorley, 2004, p. 555 (emphasis added). 37 Sorley, 2004, p. 555. 38 CIA, 1968, p. 4. he report presented a detailed analysis of available reporting and compared the body count to estimates of enemy forces, which were also suspect. he more salient points in this report address the ratio of civilian casualties to those of VC irregular or cadre forces. It asserts that many of the bodies that were physically counted may have been civilians. Vietnam-Era Assessment 109 especially pronounced when the actual score was in the enemy’s favor.”39 One division commander, Julian J. Ewell, put pressure on his subordinate units to report high body counts. He allegedly set up a quota system and threatened commanders with relief of command if they did not meet their quotas.40 his pressure to rack up high body counts may have resulted in the deaths of civilians or at least encouraged the counting of dead civilians to achieve higher “scores” during Operation Speedy Express in the Mekong Delta in early 1969.41 Ewell described his irm belief in the body count system in a discussion in Saigon with General Abrams, commanding general of MACV, on June 21, 1969. In the quote below, he is responding to Abrams’ statement that “going out there and killing a few of them is not, in my opinion, going to have the efect [sic]. I think they’re [the VC combatants] all sort of fatalists.”42 Well, I don’t agree with you, General. You can get a sapper [combat engineer] unit mining the road, and you kill two or three and they’ll knock it of. It may be that a month later they’ll come back. hese people can count. And, boy, when you line them up [bodies] and they count one, two, three, four, their enthusiasm is highly reduced. hat’s the way we opened up Highway 4—just killing them. It doesn’t take many.43 Other anecdotal evidence calls into question the accuracy of Vietnam-era body count statistics. A U.S. Army advisor to the ARVN in Phong Dinh Province in 1967 observed an ARVN unit to which he was assigned fabricating a body count and probably a weapon count after engaging with a VC unit in rough terrain. While this single anecdote is not relective of all body count reporting in Vietnam, it is similar to most other anecdotal evidence on the veracity of body counts: We got into a ight, boxed in this group of VC and they tried to duck out through the perimeter. [We] killed a few of them. he senior advisor said, “Verify the body count.” No way! he ight was all over the place in rough terrain; we just couldn’t count them. he best gauge [of success] was the weapons that we had recovered, but not all VC carried weapons. Some were specialists that carried ammunition or 39 Lewy, 1978, p. 81. 40 Lewy, 1978, p. 81. 41 hese accusations remain contested, and Ewell denied them until he died in 2009. Evidence supporting the accusations is anecdotal and circumstantial. For example, the unit reported 10,899 enemy dead and 748 weapons seized during the same period. It is possible that inlated body counts provided during this operation backired in that they exaggerated the discrepancy between the body count and recovered weapons. 42 Sorley, 2004, p. 213. Abrams responded, “One last thing you ought to consider when you handle this is how it’s going to look when the New York Times and Newsweek and some of those describe it to the American people. I’m not quite ready to take that one on. hey might lavor it. hey might lose their objectivity.” 43 Sorley, 2004, p. 213 (emphasis in original). 110 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency other materiel. And the ARVN had no interest in body counts. hey were paid to recover weapons, so they were motivated by diferent reporting criteria.44 Gibson relates several anecdotes that describe the falsiication and extreme inlation (ten times higher in one case) of body counts by combat units.45 For example, one oicer claimed, “If you come across dead bodies, you count the dead bodies. You re-sweep the area, recount the numbers, double it, and call it on in.”46 Alain Enthoven and K. Wayne Smith eviscerate the body count process: “Errors could and did frequently creep in through double counting, counting civilians . . . or counting graves, or through ignoring the rules because of the pressures to exaggerate enemy losses or the hazards of trying to count bodies while the enemy was still in the area.”47 hey explain how the body count system encouraged inlation, exacerbating an already dificult problem: he incentives for ield commanders clearly lay in the direction of claiming a high body count. Padded claims kept everyone happy; there were no penalties for overstating enemy losses, but an understatement could lead to sharp questions as to why U.S. casualties were so high compared with the results achieved. Few commanders were bold enough to volunteer the information that they had lost as many men in an engagement as the enemy—or more. he net result of all this was that statistics regarding body counts were notoriously unreliable.48 Enthoven and Smith go on to state that General Westmoreland (the senior oicer in Vietnam at the time) claimed to know the body count to within 1.8-percent accuracy based on captured enemy documents but that OSA thought the same documents were 30-percent understated.49 Contemporary COIN doctrine has resolved this question somewhat, uncovering the critical law in the use of aggregated body count 44 Former U.S. Army psychological operations advisor to Phong Dinh Province, Vietnam, interview with the author, Washington, D.C., April 19, 2010. Note that he states that the ARVN unit still provided a body count. 45 Gibson (2000, pp. 125, 127) quotes irst-hand accounts from the ield: “I know of one unit that lost 18 men killed in an ambush and reported 131 enemy body count. I was on the ground at the tail end of the battle and I saw ive enemy bodies. I doubt if there were many more.” And “in counting, a weapon captured is counted as ive bodies. In other words, if you shoot a guy who’s got a gun and you get that gun, you’ve shot six people.” 46 Gibson, 2000, p. 127 (emphasis in original). 47 Enthoven and Smith, 1971/2005, pp. 295–296. hey add, Of-the-record interviews with oicers who had been a part of the process revealed a consistent, almost universal pattern: in a representative case, battalions raised the igures coming from the companies, and brigades raised the igures coming in from the battalions. In addition, something had to be (and was) put in for all the artillery and air support, which the men on the ground could not check out, to give the supporting arms their share of the “kill.” (p. 298) 48 Enthoven and Smith, 1971/2005, p. 295 (emphasis added). 49 Enthoven and Smith, 1971/2005, p. 297. Vietnam-Era Assessment 111 metrics. FM 3-24 irst states, “Body count can be a partial, efective indicator only when adversaries and their identities can be veriied.” his might be true if the data were not aggregated or if they were used to track the decimation of a speciic, known unit (usually unlikely in COIN). More importantly, it points out that if body counts are used to measure attrition of an enemy, one must know the total enemy strength for the numbers to have any meaning.50 hayer makes it clear that estimates of total enemy strength were estimates and that estimates of irregular forces were not accurate. Without accurate data in both data sets, the entire premise of employing an aggregated body count metric to determine attrition unravels. his approach would also assume that the North Vietnamese had a casualty threshold, an idea that is explored later in this chapter. GEN Bruce Palmer, Jr., had extensive experience in Vietnam dating back to the 1950s and served as General Westmoreland’s deputy for part of the war. He believed that counting enemy dead encouraged exaggerated claims but that “higher headquarters, nonetheless, have ways to judge the validity of unit claims.”51 While some reports were undoubtedly questioned from above, there is no indication in Palmer’s book, in hayer’s work, or in any of the oicial records reviewed for this study that there was any comprehensive or systematic efort to correct inlated body counts. As a retired general with extensive Vietnam experience, Palmer’s voice is relevant. However, 61 percent of the generals who responded to Kinnard’s 1974 survey believed that body count reports were “often inlated.”52 And, as stated earlier, hayer admitted that his analytic eforts to rectify body counts were inconclusive. he body count metric shaped behavior and put the lives of U.S. personnel at risk.53 Body counts in several U.S. military units, including Ewell’s, were informally but implicitly tied to promotions and assignments through the use of published scoring systems. For example, Ewell described a “highly skilled” U.S. military unit as having a 50:1 ratio of enemy to friendly casualties and an “average” unit having a 10:1 ratio. 50 HQDA, 2006c, p. 5-27. My professional experience, both in Iraq as a senior intelligence analyst and working with senior analytic stafs in the United States tasked with estimating the number of insurgents in Iraq, suggests that these numbers either were pure guesses or relected inappropriate, haphazard, or debunked methodology. In one case, a staf attempted to count insurgents using a method that U.S. park rangers employ to count wild deer. here is no method that can be employed to accurately count something that is almost impossible to deine at the individual level (in the absence of uniformed cadre) and usually only reveals itself by choice. Intelligence activities can uncover speciic insurgents and elements of an insurgent cadre, but intelligence estimates of overall insurgent strength are rarely (if ever) derived from comprehensive and accurate data. his inding also stems from previous RAND research conducted for How Insurgencies End (Connable and Libicki, 2010). Order-of-battle assessments of conventional units are a separate matter, and certainly a complicating factor in a case like Vietnam. For an example of a declassiied Vietnam-era intelligence order-of-battle summary, see MACV, 1972. 51 Palmer, 1984, p. 165. 52 Kinnard, 1977, p. 172. 53 In some cases, the use of published metrics lists may have encouraged risk aversion. According to Gibson, points were deducted for friendly casualties in the 25th Infantry Division’s “Best of the Pack” contest. 112 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency he 25th Infantry Division held a running “Best of the Pack” contest that awarded points for “possible body counts,” and the 503rd Infantry Regiment had a military performance indicator chart that awarded points for body counts and captured prisoners, and deducted points for various problems (e.g., cases of malaria).54 Because body counts were informally but clearly tied to promotions, the desire to raise the body count may have led some oicers to conduct unnecessary missions. here is no clear way to tell whether any speciic mission was conducted to attrite the enemy— a valid if misguided objective in an attrition warfare strategy—or to simply “up the score.” his is a subtle yet critical diference between these two types of missions: he former might be conducted thoughtfully and within the scope of larger strategic objectives, while the latter are often conducted haphazardly and carry risk that far outweighs any potential military beneit. Gibson presents a number of anecdotes describing how troops were wounded and killed on missions designed to raise body counts or simply to count existing bodies.55 In a self-critical evaluation of its tactics, the U.S. Army published a primer on lessons learned written by veteran war historian and retired Army BG S. L. A. Marshall. In it, Marshall and his co-author, David Hackworth, describe several incidents in which lives were needlessly endangered or lost in the hunt for enemy bodies. For example, A U.S. rile company in a good defensive position atop a ridge is taking steady toll of an NVA force attacking up hill. he skipper [company commander] sends a four-man patrol to police weapons and count bodies. hree men return bearing the fourth, who was wounded before the job was well started. Another patrol is sent. he same thing happens. he skipper says, “Oh, to hell with it!”56 It probably was not General Westmoreland’s, Secretary McNamara’s, nor President Johnson’s intention to cause additional or unnecessary risk to U.S. troops to count enemy dead. But their undue emphasis on this assessment information requirement— like many other requirements for information in COIN—became a military objective unto itself. In the end, it failed to provide accurate data, failed to accurately or usefully inform the decisionmaking process, and cost the lives of an untold number of troops. he Vietnam-era body count metric is a cautionary tale not only for those who might apply attrition warfare to COIN but also to those who might apply centralized metrics without consideration for the risk and the unwanted behavior they might induce in the ield. Putting aside the ultimate sacriice that some men may have paid to feed the body count system, it certainly is not conducive to military good order and discipline— 54 Gibson, 2000, pp. 113–115. Gibson’s chart on p. 115 indicates the 503rd Infantry Division, but it should probably read 503rd Infantry Regiment. 55 Gibson, 2000, pp. 112–122. 56 Marshall, 1966, p. 33. Vietnam-Era Assessment 113 or to civil-military relations—to establish an assessment system that encourages institutionalized lying. his section called into question the use of directed collection requirements for data that are diicult or dangerous to collect and likely to be tied to performance ratings. he following section presents an examination of hamlet paciication metrics. The Hamlet Evaluation System his section considers the Vietnam-era HES and places it in context with modern assessment methods. his examination is important for a number of reasons. he HES is arguably the de facto gold standard for assessment, at least in the community of experts interested in assessment. It is referenced as the only truly comprehensive and structured efort to assess population security in a COIN campaign.57 It was also intended as a structured endeavor that employed polling techniques, computer-assisted analysis, time-series analysis, and correlation analysis. herefore, it ofers excellent insight into the application of scientiic process in COIN assessment. Hamlet Evaluation System Concept and Execution Perhaps the best single measure of paciication is the extent to which the population has been brought under Government control and protection. To provide a more valid standard of measurement, we have developed a device called the hamlet evaluation system. —Secretary of Defense Robert S. McNamara58 We’ve been using [HES] and defending it, over the years. We’ve emphasized that we don’t think it’s a precise thermometer for the situation, but it’s been a very handy tool. It’s given us an idea of diferences over time and it’s given us an idea of differences over space. —William E. Colby, CORDS director59 While the HES may be the gold standard for quantitative COIN assessment, it is also emblematic of the failures of centralized assessment in COIN (despite Colby’s arguments to the contrary). he CIA developed the HES in 1966 and DoD implemented it in 1967 as part of the Paciication Evaluation System (PACES), a “fully automated 57 See, for example, Gayvert, 2010. 58 McNamara, 1968, p. 266. 59 Sorley, 2004, p. 367 (emphasis in original). Colby made these comments at a February 11, 1970, brieing to then–Secretary of Defense Melvin R. Laird and then–Chairman of the Joint Chiefs of Staf Earl G. Wheeler. 114 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency system” designed to determine who controlled the population of Vietnam.60 he CIA and DoD created PACES to measure paciication programs under MACV’s CORDS program. he establishment of the HES mandated the collection of standardized security and development survey data on approximately 11,000–13,000 Vietnamese hamlets. A hamlet is a community that is smaller than a village; many hamlets made up one village in Vietnam, a district contained many villages, and every province contained several districts. Hamlets varied in size dramatically, from approximately 50 people to as many as 20,000.61 he HES aggregated the survey data describing these hamlets into centralized computer databases and then presented them as quantitative outputs. hese outputs were then used to generate a number of diferent types of analyses and analytic reports.62 As mentioned earlier, the system is perhaps the single most structured and thoroughly implemented assessment collection system in the history of COIN, but it is also remarkable in that it closely relects contemporary efects-based theory and EBA. Indeed, the HES has been suggested as a possible model for Afghanistan, 63 and the Hamlet Evaluation Worksheet (survey sheet) guidelines are remarkably similar to the rating guidelines used to develop the ISAF Joint Command (IJC) district assessments. Furthermore, HES metrics are very similar to metrics used in both Iraq and Afghanistan. Other systems used in Afghanistan and suggested for incorporation into contemporary COIN doctrine, including TCAPF (to be replaced by the District Stability Framework) and MPICE, are similar to the HES or incorporate some elements of HES methodology. he HES was a controversial program and was the subject of intense congressional and media scrutiny from its inception in 1967. Archival information on the system is expansive and includes reams of data, maps, reports, scorecards, press conference transcripts, handbooks, brieings, and oicial testimony. While this section provides a review of the published and archival data, it cannot and does not ofer a thorough examination of the entire issue of hamlet paciication.64 Instead, the purpose of this section is to use the HES as a window into the complexities of centralized analysis for COIN assessment. Concluding his chapter on the HES in War Without Fronts, hayer states, “In the absence of an absolute criterion of truth, the data can be interpreted in many ways 60 Kinnard, 1977, p. 107; Brigham, 1968, p. 22. 61 CORDS director Robert Komer (1968a, p. 5) described this size disparity as comparing “apples to grapefruit.” 62 he HES produced six published reports in both graphic and tabular format. Brigham (1968, pp. 12–13) provides a list of these reports. 63 64 See Gayvert, 2010. See Komer, 1970, Race, 1972, and Cooper et al., 1972a, 1972b, for a more in-depth examination of the paciication program. his monograph does not address the inherent validity of the paciication approach in the context of Vietnam or COIN strategy. Vietnam-Era Assessment 115 and at various levels of aggregation. Indeed, this chapter has done so and has shown the futility of assuming that the data represent a completely accurate statement at any point in time.” But, he adds, “he trends seem reasonable.”65 In 2009, noted civil violence expert Stathis N. Kalyvas and Matthew Adam Kocher wrote, “he HES is likely to remain for some time the only systematic micro-level database of civil war dynamics covering such a large territory and time period.” In their analysis of the HES, they address criticisms of the data but ind the data set to be a “unique resource for the study of the dynamics of civil wars in terms of its scope, detail, and level of analysis.”66 Anders Sweetland conducted a statistical item analysis of December 1967 HES data and found the system to be meaningful and mathematically stable. 67 HES survey data were drawn from a series of worksheets illed out by oicials at the district and province levels (by provincial advisory teams, which may have had as many as six people).68 he HES worksheets are too large to reprint here, but they included a range of questions in a format similar to a modern bubble sheet. Some of the forms had spaces for narrative input.69 he questionnaire rated six major “factors”—perhaps the equivalent of MOEs or MOPs—each of which had three associated indicators (analogous to the indicators in FM 5-0). District or province advisors 65 hayer, 1985, pp. 151–152. 66 Kalyvas and Kocher, 2009, pp. 343 and 342, respectively. 67 Sweetland, 1968, p. ii. Sweetland’s study is discussed in greater detail later in this chapter. 68 here are disparities in the description of the HES process even among MACV staf members. he MACCORDS Field Reporting System order of July 1, 1969, stated that province advisors were responsible for the PACES reports but that district advisors were responsible for HES reporting (Mission Coordinator, U.S. Embassy, Saigon, Vietnam, Annex C). Brigham (1968) indicates that data collection occurred primarily at the province level, while Komer (1970), MACV (1971a), Roush (1969), and Cooper et al. (1972a, 1972b) state that the data were collected and recorded by more than 200 district advisors. Anecdotal ield reports are also conlicting; for instance, the U.S. Army psychological operations advisor to Phong Dinh Province illed out the report himself at the province level, while Donovan describes a district-level process (Former U.S. Army psychological operations advisor to Phong Dinh Province, Vietnam, interview with the author, Washington, D.C., April 19, 2010; Donovan, 1985). Nearly all sources show that GVN and ARVN oicials continued to have input in even the HES 70 and 71 reports. Clark and Wyman (1967, p. 26) summarize the original OSD plan for the HES, published in late 1966: he objective was to obtain from each U.S. subsector (district) advisor a detailed evaluation of each hamlet with some degree of GVN control in the district to which he was assigned. here are on the average about 50 hamlets per district, the range extending from about 25 to a maximum of more than 100. he basic element of the proposed system was a questionnaire containing 191 separate evaluations to be illed out for each hamlet with some vestige of GVN control (about 70 percent of all the hamlets in the country). Responses to the questions were to be recorded on a one-page checklist as a yes/no response. After completing the evaluation checklist the advisor was to be asked to compute from his responses an overall evaluation index for the hamlet. hese forms were to be submitted to sector (province) headquarters where the number of hamlets in ive evaluation levels were to be tabulated and the summary data from them forwarded through division and corps headquarters to RDSD [Revolutionary Development Support Directorate] and to the Oice of the Deputy Ambassador. 69 HES statistics can be found in hayer (1975b) and Sweetland (1968); for worksheets, see hayer (1975b) and Brigham (1968). 116 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency tasked with illing out the forms would mark each of the categories with grades ranging from A to E. his is no diferent than a color-code scheme: like colors, the letters are directly equated to numbers, 1–5. he numerical responses to these questions were then averaged. In later versions of the system (HES 70 and 71, for the years 1970 and 1971), respondents illed out slightly more qualitative questionnaires, and analysts at some level above them would assign codes to the data.70 he HES exempliies most of the criteria used to deine centralized and heavily quantitative assessment. Every version of HES, including the 70 and 71 versions, aggregated subjective input from a range of oicials, each of whom had diferent priorities, faced diferent challenges, and had very diferent degrees of access to the hamlets assessed; their individual experiences relected the mosaic nature of COIN. hese oicials were then channeled into providing information that they may or may not have had, within narrow categorizations that may or may not have relected the range of possible conditions from hamlet to hamlet. hese subjective inputs were then transformed into narrow numerical codes that were devoid of context once aggregated and processed. (It is impossible to retain context when aggregating more than 8,000 reports.) hese aggregated data were used to produce a countrywide perspective on hamlet security and development. hayer’s data on GVN control of the countryside is derived primarily from the HES (and also from PAAS and the Territorial Forces Efectiveness System). Narrow, standardized guidelines for the rating system forced the respondents to provide answers for each indicator in one of ive categories. For example, the 1968 version of the HES rated VC military activities, focusing on the activity of “VC guerrilla units” to help the respondent rate the category.71 he precoded options were as follows: A. Guerrillas (Gs) remain driven out. No threat of harassment or intimidation from Gs in adjacent villages. B. G control reduced to 1–2 hamlets on village periphery or 2–3 hours travel to hamlet: could make a desperation raid. Activities of Gs from adjacent villages limited by no havens or by friendly defenses. C. Military control of village broken, most Gs identiied, 50 percent losses, havens destroyed, activity below platoon level; can harass but not prevent GVN activities in hamlet. 70 Revision recommendations can be found in the Simulmatics report, described later in this chapter, and in a 1973 report by the Department of Applied Science and Technology at the Bendix Aerospace Systems Division, Analysis of Vietnamization: Hamlet Evaluation System Revisions (Prince and Adkins, 1973). he latter provides a more in-depth analysis of the questions and data than the former. A December 27, 1969, MACV brieing in Saigon reported that when HES was modiied to HES 70, 42 percent of hamlets changed category (e.g., from B to C). See Sorley, 2004, p. 331. 71 Each indicator on the worksheet is broken into ive rating categories. his example is indicative of issues with the other indicators. Vietnam-Era Assessment 117 D. Village Gs reduced somewhat in men and defenses; can attack in platoon strength from within [village] or 1–2 hours travel to hamlet. E. Village Gs combat efective though some identiied or eliminated; VC village defenses largely intact.72 his is very similar to the IJC district assessment rating deinition levels in Afghanistan, which also use in a pre-coded 1–5 scale and provide standards for each level.73 For example, a notional IJC district assessment metric used to determine the progress of local Afghan government leadership might have rating deinition levels that span from 1, “Leader is efective and not corrupt,” to 5, “Leader inefective and corrupt.”74 But it is possible for a leader to be efective and also corrupt (especially in Afghanistan); therefore, it is feasible for a leader’s behavior to rate a 1 on the rating scale and simultaneously rate a 5. Would this equal a 3? here is already a distinct deinition for a 3 (in this example, “Leader is moderately efective and moderately corrupt”), so it could not. herefore, in this notional Afghanistan case, the district assessment does not include an option that permits accuracy. he section of the HES on VC activity (which is representative of the other sections of the HES) was similarly restrictive. It might have been possible to have no threat of attack or harassment from adjacent villages (an “A” rating, equivalent to a 1) but also face intact VC defenses in the village in question (an “E” rating, or 5). his situation cannot equal “C” because that rating has a separate deinition. In these cases, respondents probably would have chosen to average the two options together, despite the fact that averaging is illogical in that the resulting report would not relect reality.75 he process of forcing respondents—each of whom is facing very diferent circumstances—into narrow metrics and deinitions results in inaccurate data. It also appears that the authors of the deinitions in the HES were attempting to elicit positive input from the respondents. In the best-case scenario, or “A” rating, the village (an area larger than a hamlet) and all surrounding villages are not only free of guerrillas but also free of harassment. his is, in essence, a peacetime situation and a laudable goal for counterinsurgents. But in the HES example presented here, the irst two rating options both depict optimistic situations, and even the third (C) option shows a rather impressive degree of success: he guerrillas cannot prevent GVN activities in the hamlet. he contemporary equivalent might be the ability of Afghan government oicials to operate freely and without harassment in an Afghan village; this would be a dramatic success in Afghanistan and probably even in Vietnam. In fact, 72 hayer, 1975b, pp. 49–50. 73 his relects the process as of late 2010. 74 his is similar but not exactly representative of the actual rating scale used in IJC documents. 75 Interviews with tactical commanders and assessment stafs in Afghanistan support this assumption. Quantitative inputs are routinely averaged for aggregation, regardless of context. 118 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency in Vietnam, a “C” rating showed the hamlet as being “secure” in aggregated reports and on maps.76 he last two ratings (D and E) both describe the guerrillas in terms of how they have been degraded, and in the “E” rating—the worst possible rating for a GVNcontrolled hamlet—some of the guerrillas have been identiied and killed (“eliminated,” in the parlance of the HES). herefore, even in the worst-case situation, there are signs of success.77 Might a retooling of the deinitions and rating levels make this system, or any similar system (e.g., the District Assessment Model used in Afghanistan) more realistic? he right deinition would give suicient leeway to account for any and all situations faced by province advisors. But in practice, any deinition is channelizing and will create incongruities among the respondents; these are core metrics. he only way to allow for any and all types of input and to eliminate bias from the formula is to clearly deine the data requirement (if this is possible) and then use a rating scale to “rate the hamlet in this category of assessment from 1 to 5 based on your understanding of the situation.” his kind of dissociated quantitative response would make the aggregated data all but meaningless because there would be no standard associated with the numbers. Subjective bias would guide the selection process absolutely, and the consumers of the data would have no way to determine whether a 3 from one rater was equivalent to a 3 (potentially an average of 1 and 5) from another rater. he kinds of quantitative and graphic reports produced from the data would be meaningless. Once this was realized, the process of deining rating levels would begin anew and would almost certainly result in the same muddled and unsatisfactory deinitions already found in the HES and the District Assessment Model. here was so much disagreement over the HES deinitions that as late as 1973—six years after the implementation of the program—external auditors were still recommending revisions to the pre-coded ratings.78 Because each hamlet in the HES was rated on an A–E or 1–5 scorecard on several indicators (rather than just one), it was necessary to average all the scores to come 76 his was true as of 1967, according to Komer, and late 1968, according to Tunney. his may have changed with later modiications to the program. See Komer, 1967, p. 4, and Tunney, 1968, p. 5. 77 Komer states that e-rated hamlets did not include those hamlets controlled by insurgents or VC and that there should have been an “f ” rating to show VC-controlled hamlets. here was no such rating in the HES at that time. Brigham (who attended the Komer news conference) stated that there should also be another category for “not rated,” but that category did not exist either. hese categories were later added (at least by 1969) as “VC” for VC-controlled, or “other” (see Komer, 1967, p. 14; Brigham, 1968, p. 12). For another oicial interpretation of the rating system and its relation to paciication objectives (i.e., the MACV Strategic Objectives Plan), see Sorley, 2004, p. 203. 78 See, for example, Prince and Adkins, 1973. Vietnam-Era Assessment 119 up with a single overall rating.79 Colonel Erwin R. Brigham, then an HES manager, explained how this worked: “A hamlet rated in security with 3 As, 4 Bs, and 2 Cs; and in development with 2 Cs, 4 Ds and 3 Es would be given an overall [unweighted] rating of C. Similarly, A, B, D, and E, hamlets may have individual ratings of A, B, C, D, and E.”80 here are at least two problems with this system. It can be misleading because it is impossible to know what is most important in this particular area at any one point in time. For example, is development inconsequential because security is insuicient? What do the people of the village actually think about their own security? he system begs for weighting, but weighting would be subjective and reduce transparency. his kind of system would make the data murky at best and opaque at worst. Ambassador Robert W. Komer helped develop the HES and established and ran the CORDS program in the late 1960s. He told reporters at a 1967 news conference that the HES was so complicated that it would take much longer than a 20-minute brieing and a 30- or 40-minute question-and-answer session to explain how it worked and how it compensated for the complex realities of South Vietnam.81 his is not a transparent process. Once the CORDS staf collected the data and averaged the responses to produce a rating for each hamlet, they would then average all of the hamlets together to develop a progress number that would represent overall paciication in the country.82 According to Komer, Now it’s pretty hard to measure paciication results. . . . [Y]ou’ve got 12,700 hamlets in this country that we’ve identiied so far, and no doubt there are some others too. So you’ve got to look at the averages.83 It was this averaged, overall paciication number that was typically delivered to senior decisionmakers. herefore, the overall score used to shape policy was composed of averages within an average. Averaging the total score further degraded the transparency and accuracy of the HES data to the point that the overall number was relatively meaningless.84 William E. Colby, who was the CORDS director from 1968 to 1971 79 Various descriptions of the HES also include some mention of weighting the numbers, but these descriptions are not consistent across all sources. It is not entirely clear how averaging and weighting igured into the inal product from version to version. 80 Brigham, 1968, pp. 9–10. 81 Komer, 1967, p. 16. 82 hayer, 1975b, pp. 53–58. Note that there are sometimes two mismatched page numbers on some of the pages in this report. hese pages are the ones on the lower right of the page. 83 84 Komer, 1968a, p. 4. Further complicating this calculation was the inclusion of a “conidence rating” on the HES worksheet (at least in the 1969 version). he advisor would use a 1–5 scale to rate conidence in the information provided on the worksheet. If the reports were not aggregated, this rating might have been a valuable tool to help analysts assess 120 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency and responsible for the HES during that period, said that “a mathematical complication of the responses from all the hamlets suggested a false average instead of the wide individual variations that lay beneath it.”85 It would be possible for the single, top-level average to hide a major security imbalance. For example, the heavily populated areas around Saigon could be improving, while the less densely populated areas in contested spaces along the demilitarized zone in the north or in the Mekong Delta in the south—the populations that probably mattered most to the COIN efort—might be declining rapidly; but in this case, the average would show only improvement.86 OSA analysts could address this problem by delving into deeper layers of data, but by producing a single, quantitative inding, they made their own analysis irrelevant to busy policymakers or interested members of the public who did not have more than an hour (by Komer’s reckoning) to igure out the HES system. Every month, the HES produced approximately 90,000 pages of data and reports.87 his means that over the course of just four of the years in which the system was fully functional, it produced more than 4.3 million pages of information, and each page may have contained ten, 20, or more discrete elements of data—perhaps 40 million pieces of data, as a round estimate.88 his is a remarkably dense and complex reporting process. Such a vast quantity of information presented extraordinary logistical and analytical challenges for Komer’s, Colby’s, and hayer’s analysts in the 1960s and early 1970s. Computer-aided analysis has improved exponentially since 1971, of course, but even now it would be diicult to account for the possible reading, coding, aggregating, or transmission errors that could creep into these pages as they are written by advisors in the ield, passed up to the province level for collation and retransmission, passed to the division level for another round of collation and retransmission, and then provided to MACV for a inal round of collation and retransmission to OSD. Despite obvious concerns about data accuracy as voiced by the program director, HES reports as delivered were said to be accurate to within a tenth of a percentthe validity of the reporting for an individual hamlet. However, it is not clear that conidence ratings were incorporated into the total HES number or, if they were, how the variations in conidence ratings could be accounted for in an aggregated total. An averaged conidence rating could be misleading in the absence of disaggregated context. See MACV, 1969d. 85 Colby and Forbath, 1978, p. 259. Colby, who would later become director of the CIA, believed that the HES was most valuable as an organizing tool at the local level; he thought that it helped motivate GVN oicials to improve hamlet security and development. 86 his is a notional example. In reality, the 1967 version of the HES did not even address the population in Saigon. herefore, the 1967 data did not include approximately 3.6 million people out of a population that may have been about 17 million (hayer, 1975b, p. 12). 87 88 Brigham, 1970, p. 48. Some of these pages may have included reports that repeated or re-reported raw data from the ield, but for the purposes of information management and analysis, the overall quantity is what is relevant. Vietnam-Era Assessment 121 age point. A January 11, 1969, HES report to military and civilian leaders in Saigon reported a “record gain” of 3.5 percent in secured hamlets and a decrease of 1.1 percent in the VC-controlled population. hese igures were presented without caveat for accuracy or a margin of error. he way in which the igures were presented implied a degree of accuracy that would not be achievable even under ideal conditions. None of the assembled senior leaders, including the commanding general of MACV and the U.S. ambassador, questioned the report or the accuracy of the data.89 his unquestioning acceptance of HES data may have been due to the pressure on senior leaders and HES program oicers to meet quantitative paciication thresholds. For example, in a July 17, 1969, brieing, CORDS director Colby stated that the national objective was to achieve at least a C-rating for 90 percent of the total number of hamlets by the end of that year, and at least B-ratings (if not A ratings) for 50 percent of the total number of hamlets. According to Colby, “he President very clearly says that the purpose of that is to get it so the government can absolutely count on at least 50 percent of the population to vote for them in a challenge with the Communists.”90 Since the HES did not relect anywhere near 100 percent of the South Vietnamese population (even when taken at face value), there is no way that the President’s 50-percent objective for A- and B-rated hamlets could have guaranteed that 50 percent of the population would do anything.91 his statement shows how policymakers can be misled by or misunderstand complex quantitative COIN data. Both hayer and Brigham cite two independent studies of the HES conducted in the ield in Vietnam. he purpose of both studies was to substantiate the reliability of the HES model and to improve the system.92 One of these studies was managed by the Simulmatics Corporation but conducted primarily by U.S. military and Vietnamese ield collectors and analysts. It examined only the reliability of the HES system and not (according to the report) the validity of the data. It describes the system as generally reliable and the information produced by the HES as reasonably accurate.93 his study is important because it ostensibly validates the HES methodology for campaign assessment. he security conditions and lack of access undermine the study’s reliability, however. According to the report, it covered 18 percent of the districts (40 out of 244 total districts) in which advisors were “present,” and it is not clear how many hamlets were 89 Sorley, 2004, p. 96, transcript of audio recording of a commander’s conference in Saigon, South Vietnam. 90 Sorley, 2004, p. 220. It is not clear whether he is referring to a benchmark established by Lyndon Johnson prior to the end of his term or by Richard M. Nixon during the irst part of his irst term. 91 Colby stated that, as of 1970, 40 percent of the South Vietnamese population lived in urban centers, and the majority were not covered by HES reporting (U.S. Senate Foreign Relations Committee, 1970, p. 420). he propriety of the President’s efort to ix the South Vietnamese election is immaterial to this discussion. 92 he Simulmatics Corporation and Paciic Technical Analysis, Inc., conducted these studies in 1967 and 1968. 93 Pool et al., 1968, p. 1. 122 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency located in these districts.94 It is unclear what is meant by “present,” but the sample could not then be extrapolated to cover the hundreds or perhaps thousands of hamlets that were rated by advisors but not visited. he samples primarily included heavily populated areas, areas that tended to receive more attention from MACV than the more contested rural districts and that would be safer and easier to rate. U.S. military oicers who outranked the district and province advisors managed and coordinated the study, and two special forces oicers of unknown rank (but who may have outranked the district advisors) helped conduct the ield interviews. here is at least some impression that the military was “grading its own work,” and there may have been some pressure on the relatively junior advisors to respond positively to interview questions. Regardless of the reliability of the Simulmatics sampling process, at the end of the day, the study relied on district and province advisors to tell the interviewers whether or not their own reporting was accurate.95 While this reliability report provided some excellent recommendations for program improvement, it is insuicient to prove the reliability of the HES; ultimately, the reliability of the overall system was never clearly established.96 he next section addresses the validity of the HES data and the ways in which those data were used for paciication and strategic analysis. Hamlet Evaluation System Data and Analysis [T]he greatest possible restraint should be exercised by senior oicials and policymakers in using highly aggregated reporting results in which vital qualiications have been “summarized out.” —Report on Vietnam hamlet paciication assessment, 1972 97 94 Pool et al., 1968, p. 14; Brigham, 1968, p. 3. Brigham reported that there were 244 districts in Vietnam. he Simulmatics Corporation stated that it covered “approximately” 40 districts. 95 he study did include some observations by the research staf, but the staf was small and consisted of either oicers from out of country or Vietnamese oicials of unknown ailiation; the visits were structured and of insuicient length to determine actual survey reliability or data validity. he study also included interviews with hamlet chiefs in 22 of the 40 districts covered. While there was coincidence between the responses of the hamlet chiefs and the ratings for the hamlets, this is a nonscientiic sample. It also does not show whether district advisors relied heavily on the hamlet chiefs to provide them with their original ratings; this was a common tactic. In these cases, there would be natural coincidence between the original ratings and the ield observations that would not necessarily speak to the reliability of the HES. 96 RAND conducted a statistical item analysis of HES in 1968 that showed the system to be “meaningful and statistically stable and can be used as the basis for developing extensions to the construct of paciication” (Sweetland, 1968, p. ii). hat study did not address validity. he Simulmatics study found problems with the development scores, and a survey of all information relating to the HES available in 2011 showed that the item analysis may have implied meaningful relations and statistical stability based on a steady stream of precise but inaccurate reporting. hat said, any analysis of the overall reliability of the HES should include Sweetland’s indings. 97 Cooper et al., 1972b, p. 234. he source is a comprehensive report on the Vietnam civil paciication programs published by the Institute of Defense Analyses for the Advanced Research Project Agency. Vietnam-Era Assessment 123 [HES will become] the body count of paciication. —Attributed to Vincent Davis, friend of John Paul Vann98 Some district and province advisors made every efort to ill out the HES worksheets accurately, and many HES reports probably contained some data that were very accurate. One district advisor described the HES as “a valid system which is correct in my district.”99 But the preponderance of available evidence seems to show that much of the information in the HES reports—it is impossible to know how much—was incomplete and inaccurate. Komer (the head of the program from 1967 to 1968) stated that HES data were “not glamorous.”100 Anecdotal evidence that a sizable amount of HES data were inaccurate or simply fabricated is similar to anecdotal evidence showing oicers fabricated data for body counts in Vietnam. Environmental and bureaucratic complexities also eroded the quality of HES information. Instructions for illing out the four worksheets were long and complicated: he simplest instructions were about 25 pages in length.101 he collection of HES data was necessarily ad hoc and difered from location to location. hayer stated that it was infeasible to deine HES data as completely accurate at any one point in time, and he described the factors that “combine to make HES reporting diicult”: 1. Only a few of the 18 indicators can be rated on the basis of direct observation of a clear-cut condition. 2. Much of the HES information can be obtained only from Vietnamese [sources]. Surveys to date [late 1967] indicate that advisors rely on their Vietnamese counterparts for at least half of the raw data they use in answering the HES questions.102 98 Sheehan, 1988, p. 697. 99 Arthur, 1970, p. 193. 100 Komer, 1967, p. 21. 101 Mission Coordinator, U.S. Embassy, Saigon, South Vietnam, 1969, Annex C and appendixes. See also the Hamlet Evaluation System Advisors Handbook (MACV, 1971c), at 40 pages, and the Hamlet Evaluation System District Advisors’ Handbook (MACV, 1971a), at 50 pages, including a helpful Vietnamese translation. Komer and others state that advisors were trained to ill out the HES worksheets prior to deployment, and mobile training teams that provided ield instruction visited them occasionally. here was also an HES “newsletter” written by an oicer at MACV CORDS to address ongoing issues with worksheets. A sampling of these handbooks, worksheets, and newsletters can be found in the Texas Tech University Vietnam Center and Archive. 102 Despite this, the system prominently reported indings in statistical analyses from data that were gathered from single-source counterpart reports . For example, the HES showed that, in 1970, hamlet chiefs governing 80 percent of the population were regularly present both day and night in the hamlets. Since the vast majority of hamlet visits were often brief and infrequent (a few hours per month or once every few months) and did not occur at night (and district advisors could live in only one hamlet at a time), this information had to have come primarily from Vietnamese sources. Yet, the data are presented without caveat. Further, hayer compares these data to perception data from PAAS showing that only 59 percent of the population believed (and could not conirm) that 124 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency 3. Most advisors cannot visit all of their hamlets during one month. In [one area], advisors reported that US personnel visited about two-thirds of the accessible hamlets in their district in any given month, but managed to cover all but 6% over a four-month interval.103 One district advisor claimed to have visited two diferent hamlets per day, approximately every day of every month (there were 60 in his district) for the duration of his tour. his is an incredible pace that, if true, probably made him one of the top HES reporters in Vietnam. Still, sometimes these visits lasted 15 minutes and sometimes they lasted “a couple of hours,” and they rarely took place at night when the VC were most active.104 his means that if an advisor were able to visit every hamlet in a district every month (an anomaly, according to hayer), the observation period would consist of (at best, according to Arthur) approximately three hours out of a potential ~720 hours in a monthly reporting cycle, or 0.41 percent of the month for a given hamlet.105 It would be unreasonable to claim that such a limited observation period could prepare an advisor to write a comprehensive assessment of a hamlet, or that this assessment could be deemed either precise or accurate. In a monograph on paciication published by the U.S. Army, South Vietnamese Brigadier General Tran Dinh ho questioned the validity and accuracy of HES data and addressed the ability of advisors to consistently visit and rate all the hamlets in their respective districts: he data recorded by district senior advisers were obtained partly from village and hamlet oicials’ reports, partly from information provided by friendly forces or through the advisory channel, and partly from actual visits to villages and hamlets by the senior adviser and the district chief. A question arose, however, as to the validity and reliability of the reports thus obtained; doubts about accuracy and timeliness. he most reliable way to have accurate data was to make visits to the villages and hamlets and see for oneself. But the truth was that even if all the time available were devoted to visits, and even if road communication and transportation facilities and helicopters were available all the time, no one could possibly cover all the villages and hamlets of a district in a single month.106 their local oicials remained in or near the hamlets overnight (hayer, 1975b, p. 168). Popular polling often failed to substantiate HES indings. 103 hayer, 1975b, p. 61. 104 See, e.g., Arthur, 1970, p. 3. he transcript of Arthur’s testimony showed that he was reluctant to state that he did not visit the hamlets at night. He eventually admitted that he received reports on nighttime activity from South Vietnamese militia units working for the GVN. Some advisors slept overnight in speciic hamlets from time to time or lived in one speciic hamlet, but the HES study was concerned with the typical coverage of all the hamlets in a district during any given month. 105 his calculation is meant only to describe an average overview of HES data collection, not the overall efort of the advisors or other oicials working at the hamlet level. In most cases, advisors lived in hamlets or spent the majority of their time working in hamlets. 106 ho, 1980, pp. 107–108. Vietnam-Era Assessment 125 Observation visits may have been structured and sometimes announced in advance, so it is not clear whether the advisors would get an accurate picture of life in the hamlet. (It would be easy to put on a show for a visiting advisor.) Other advisors lew over hamlets in helicopters. Sometimes, they counted huts from above and then multiplied the number of huts by the average number of residents per hut (eight) to build population estimates. his process might work in peaceful areas but is suspect in war zones with high rates of unpredictable civilian movement and millions of displaced persons. With few exceptions, none of the advisors spoke luent Vietnamese. hey could communicate with the population only through an interpreter, each of whom would vary in terms of capability, motivation, and availability.107 A psychological operations advisor in Phong Dinh province was tasked with illing out the HES reports for the province with no input from the district level.108 His HES training consisted of a few minutes of overview during a 15-minute turnover from his predecessor.109 here were between 250 and 300 hamlets in his area of operations, and he had traveled to only a handful of them. In the absence of accurate input, he sat down with his ARVN counterpart in front of a map of the province and, pointing to each hamlet in turn, asked, “Can you go here?” If the answer was yes, he coded it as friendly; “maybe” rated a neutral code; “no” rated an enemy code. He had no way of verifying this oicer’s subjective opinions and had no other input to work with. He sent the data to the division headquarters and was castigated for altering the report and changing the reporting pattern set by his predecessor; an oicer on the staf told him to change the results back to avoid any controversy that might arise from a signiicant shift.110 If he rated approximately 250 hamlets (at the low end), and fewer than 107 Komer (1967, p. 12) stated that, in 1967, somewhere between 20 and 50 percent of the advisors spoke some Vietnamese with some (undeined) level of proiciency. Gibson (2000, p. 308) cites an informal study showing that most Vietnamese interpreters spoke fewer than 500 English words. Oicial statistics recorded during a March 1970 series of congressional hearings on CORDS (the “Fulbright Hearings”) showed that 36.4 percent of district senior advisors had achieved level 1 on the Interagency Language Roundtable for reading and speaking, that 6.6 percent had achieved level 2, and that none had achieved levels 3, 4, or 5 (with 5 being the highest level of proiciency). Only 18.7 percent of assistant district senior advisors achieved level 1, 5.1 percent achieved level 2, and none achieved higher. A level 1 speaker has only elementary proiciency and cannot hold an extended conversation on a complex subject. Level 2 speakers are more capable of holding complex conversations on limited subjects but may still miss critical nuance and vocabulary. By these standards, no advisors were luent in Vietnamese, and only about 6 percent of all advisors could be considered to have linguistic abilities that would allow them to function with only minimal assistance from an interpreter. On the other hand, 28 percent of civilian advisors could communicate at level 3 or better (U.S. Senate Foreign Relations Committee, 1970, p. 418). See also Interagency Language Roundtable, undated. he U.S. government implemented the program in stages from the late 1950s through the 1980s, and the system remains in use as of 2011. 108 he fact that he received no input from the district level may have been indicative of structural problems in the reporting system only in this province and only during this period. 109 Training 110 Former improved over time as MACV and CORDS implemented training courses for advisors. U.S. Army psychological operations advisor to Phong Dinh Province, Vietnam, interview with the author, Washington, D.C., April 19, 2010. 126 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency 9,000 of the known 12,000 or so hamlets were evaluated through the HES in 1967,111 this single anecdote shows that nearly 3 percent of the overall HES data during this period were falsiied (because his changes were not accepted), inaccurate, or based on extraordinarily poor collection methods.112 he story of another advisor reinforces the idea that it was very hard to downgrade an HES rating. In a conversation with Congressman John V. Tunney in 1968, this advisor stated that he downgraded four hamlets after the Tet Ofensive in February of that year; these downgraded reports would have generated an automated PACES message to the MACV headquarters. He was “immediately hit with a barrage of cables from Saigon demanding a full explanation for downgrading them.” He claims that he spent the next couple of weeks justifying his report, a time during which he says he neglected some of the other pressing demands of his job. he cables from his higher headquarters may have been sent in an earnest efort to get more context from the advisor, but they had a chilling efect: “I believe I am an honest man, and although I hate to admit it, it may be a long time in hell before I downgrade another hamlet.”113 Tunney presented this anecdote in support of his argument that, no matter how the HES is programmed, it will tend to elicit biased responses. David Donovan’s experience with HES was similar to the advisor’s in many ways. Donovan was a district-level military advisory team leader in Vietnam in the late 1960s.114 In his memoir, he provides his observations of the HES system. While Donovan thought the intent of the reports was good, “like so many good bureaucratic intentions, the idea was weakest at the point of practical application.” He stated that he saw district senior advisors delegate the HES worksheet to “less informed and less experienced subordinates,” sometimes with instructions to “just ill in the blanks with anything that seemed reasonable.” Donovan identiied timeliness of submission as having precedence over reporting accuracy, and many reports on hamlets were illed in “when the hamlet had never been seen by the [district senior advisor] or any of his team members.” In many cases, overworked advisors would simply ask their Vietnamese counterparts for their opinion so they could ill in the HES (like the Phong Dinh Province advisor did), though the Vietnamese had good reason to be biased.115 Lewy cites one case of outright cover-up by Vietnamese oicials who were feeding data into the HES 70 report, a case that reinforces Donovan’s suspicions regarding host-nation inputs to the HES. he senior advisor for Phu Yen Province discovered that local oicials in the district of Hieu Xuong were hiding incidents and abductions to 111 Tunney, 1968, p. 1. 112 He reported that he made only incremental changes after this point, but the accuracy of the report never changed much during his tour in late 1967. 113 Tunney, 1968, p. 8. 114 Donovan, 1985, pp. 157–158. David Donovan is the pen name of Terry T. Turner. 115 Donovan, 1985, p. 158. Vietnam-Era Assessment 127 retain the district’s “paciied” rating. Lewy also points out that it was still possible for even U.S. advisors to “game” the HES 70 and 71 reports. He states that while advisors no longer assigned their own ratings and that scores were added to the reports in Saigon, “it probably was not very diicult to igure out what kinds of conditions in the hamlets would result in a high rating, and advisors, in efect, were still rating their own success in achieving paciication.”116 Douglas Kinnard relates a conversation with an oicial from the U.S. Embassy in Vietnam, who told him that “the gaps between the HES numbers and actual conditions in the provinces had reached ludicrous proportions.” Kinnard and the oicial determined that, out of the ive A-ranked hamlets in their area of operations, they could safely drive through only two. U.S. Marine Corps Col. William A. Corson served as a paciication advisor in Vietnam and as a systems analyst in OSD’s Southeast Asia Programs Division in 1968. He states that the C-rated hamlets did not pass the paciication advisors’ “sleep test”—they were not safe to sleep in—despite the fact that they were labeled secure.117 He also stated that the GVN gamed the HES to protect its own people by putting its development cadre in A- or B-ranked hamlets, undermining one of the key purposes of the HES survey program; in theory, the HES should have been used to shift support to less secure or less developed hamlets an efort to bring them up to the A or B level.118 John Paul Vann, perhaps the most successful and well-known paciication oicer in CORDS and the subject of Neil Sheehan’s Bright Shining Lie: John Paul Vann and America in Vietnam, discovered that the HES could also overstate the degree to which the VC controlled certain hamlets. Sheehan describes Vann’s discovery upon arriving at a new posting in Can ho, the capital of Phong Dinh Province in the MACV IV Corps military region: He discovered that many of the 2,100 hamlets the HES listed as under Viet Cong ownership in IV Corps in February 1969 (another 2,000 hamlets were listed in varying degrees of Saigon control) were actually held by half a dozen guerrillas.119 Jefrey Race was a civilian province advisor in Long An Province in 1968 and 1969. In his classic War Comes to Long An, Race describes the HES as useful in that it showed changes in some indicators over time. However, he also called the system’s data a “misleading indicator of progress.” He points out that, in many of the C-rated hamlets in his province—those considered paciied by HES standards—there was actu- 116 Lewy, 1978, p. 192. 117 Corson, 1968, p. 236. 118 Corson, 1968, pp. 236–237. 119 Sheehan, 1988, p. 732. 128 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency ally more revolutionary activity than government activity.120 He describes in detail the eforts to pacify a hamlet called Hoa huan II. He states that local GVN oicials could not travel in and around the hamlet, but clandestine VCI party oicials had the freedom to move about of the main roads. Despite these conditions, the U.S. advisory team “upgraded the status of Hoa huan II to the C category because of the ‘progress’ it had shown.”121 Race considered the Hoa huan II assessment a cautionary tale. Because the HES was rigidly structured (even with the HES 70 and 71 modiications), it did not allow the systematic reporting and analysis of hidden problems like those described by Race.122 he lack of obvious VC activity could (and sometimes did) hide a very sizable insurgent presence in the hamlets; in some cases, the hamlet chiefs and GVN oicials simply came to an arrangement with the VC, allowing them to operate with impunity. For example, in Vinh Binh Province (in 1971, so probably using HES 71 or at least the HES 70), there was a B-rated hamlet that also contained the 312 Main Force VC Battalion. he VC held “veto power over life and death issues” in the hamlet, but since there had been no overt incidents and the people had responded positively to the GVN in public, “the HES marks this a B hamlet.”123 Only detailed and contextual all-source intelligence could have identiied and reported this situation in a format that was meaningful to operational decisionmakers. Did any of this ultimately matter? Did policymakers take the HES data seriously? he evidence shows that the HES was not only a central pillar of the Vietnamera assessment process, along with body counts, but that it also shaped both strategy and combat operations. A review of oicial message traic from the U.S. Embassy in Saigon to the President and DoS showed that senior oicials like Komer and Ambassador Ellsworth Bunker based their strategic recommendations at least in part on hamlet security data.124 Brigham describes the HES as a “management tool” and explains how HES reports were used to direct combat operations and, alarmingly, also used by artillery units to determine population density in what may have been an efort to avoid civilian casualties.125 At least some policymakers took the data seriously, relying on them to show progress to Congress and the public. Secretary of Defense Robert S. McNamara reported paciication data to the press and the public in news conferences, highlighting them as a major indicator of progress.126 George Allen had 17 years of 120 Race, 1972, p. 214. 121 Race, 1972, p. 260. 122 Later versions of HES allowed for subjective input, but it is not clear that this input was seen or analyzed at the MACV or OSD level due to data aggregation. 123 Lewy, 124 See 1978, p. 194. BACM Research, Disk 2, as well as Bunker, 1968a and 1968b, and Komer, 1968b. 125 Brigham, 126 See, 1968, p. 21; Brigham, 1970, p. 53. for example, McNamara, 1967. Vietnam-Era Assessment 129 experience studying Vietnam for various intelligence agencies. His account provides some insight into the way the system was created and how HES results were misunderstood and misused. Here, he describes pressure from the executive level to show progress through using the aggregated and twice-averaged HES total score: [W]e got pressure from the White House, the way they used the Hamlet Evaluation System. hey were the ones that decided this A-B-C-D-E, or 5-4-3-2-1 scale, you could come up with an aggregate average score for all the hamlets in Vietnam, monthly, and it was 2.163 maybe this month. And they would sit there with [bated] breath waiting to see what next month’s score would be. And if it were 2.165, two one-thousandths higher than last month, “Great, we’re progressing, we’re progressing!”127 his anecdotal insight into the strategic use of HES data by senior policymakers is reinforced by other sources. For example, Vietnam-era CIA oicer and hamlet paciication expert Rufus Phillips stated that, in 1965 (pre-HES), there was a “demand from Washington for instant or nearly instant paciication results.”128 John Tunney was both a critic of the HES and, as a sitting U.S. congressman, a consumer of HES data. He conducted a special study mission to Vietnam to investigate the HES in May of 1968. Reinforcing Allen, Tunney states that on numerous occasions, “Presidential advisors have briefed Members of the Congress and told them that a certain percentage, usually carried to one or two decimal places, of the South Vietnamese people were living in relatively secure areas under the control of the Saigon government.” Tunney went on to state that “this information was being used at the highest levels of the U.S. government as a guide to operational planning and as a means of informing the American people of the progress of the war.”129 hat Sharp and Westmoreland highlight a full-page HES chart showing near-steady improvement in HES reporting from 1965 through 1968 in the conclusion of their Report on the War in Vietnam reinforces this supposition.130 his critique of the HES is not intended as an indictment of CORDS or the overall paciication strategy; both programs were probably successful to some degree, and the theory behind bottom-up COIN has generally proved to be sound. Nor is this section intended to criticize the personal eforts of any of the key players in the HES assessment chain.131 It is quite clear that the HES was a best try in the absence of 127 Rehm, 1985, pp. 6-14–6-15. 128 Phillips, 2008, p. 267. 129 Tunney, 1968, p. 1. 130 Sharp 131 To and Westmoreland, 1968, p. 199. what degree paciication succeeded will never be known because the hard work by CORDS teams literally went up in smoke in 1975. Colby claims that paciication was a clear success, citing the lack of VC activity during the 1972 and 1975 NVA invasions (Colby and Forbath, 1978, pp. 285–286). However, this shows only that the VC may have been largely defeated by 1972. his is coincident with half of the objectives of paciication accord- 130 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency better options. No similar or better options had been articulated by 1967.132 he MACV CORDS staf seems to have made every efort to ensure the accuracy of the HES data and to improve the program over time. Komer, the man best positioned to comment on the system from the inside, stated, “All that we think we can say is that it seems to be much better than anything we have had before. It’s more objective, it’s more accurate, it’s more systematic [than previous eforts].”133 Komer points out that there was insuicient analysis of the war (and he had only ten analysts in the CORDS Operations and Analysis Division to pore over the HES data),134 but it is not clear that more analysis of aggregated data like that from the HES would have better served policymakers than some kind of contextual analysis.135 Kinnard’s survey showed that 75 percent of general oicers believed that the HES “had weaknesses but was about as good as could have been devised.” his far-from-ringing endorsement seems to support the “good-enough” proposition for COIN assessment. However, only 2 percent of these senior leaders found the HES to be “a good way to measure progress in paciication.” Because the HES required narrow and centralized quantitative inputs and aggregation, it relected the most misleading elements of contemporary centralized assessment and EBA. In fact, it was an EBA system in all but name: It used a set of core metrics to gather and aggregate standardized data to measure the efects that U.S. and GVN paciication programs were having on the population of South Vietnam, down to two or three decimal places.136 EBA promises and depends on accuracy, though neither hayer nor Komer thought that HES data were ever truly accurate. But the structured collection of HES data, the statistical analysis, and the defense of the aggregated HES indings are collectively equivalent to the process of producing an EBA report. As Allen and others showed, policymakers took the indings quite literally, even when they were warned not to. hayer, Komer, and Colby cannot be wholly to blame for this overreliance on the data; the policymakers who seized on the reports as a quantiication of the war are also somewhat to blame. But no one involved fully appreciated that the act of aggregating and publishing statistical HES data was enough to give those numbers a life of their own. Caveats should have extended from data quality to the validity of the overall process and indings. ing to MACV orders and the HES; it does not prove the success of paciication programs. Both the Simulmatics Corporation (Pool et al., 1968) and Bendix Aerospace (Prince and Adkins, 1973) reports hint at a split between the relative validity of HES security statistics (more valid) and development statistics (less valid). 132 Ahern (2010, p. 252) describes the Saigon manager’s statement that the HES was “far and away superior” to any other method, a “modest claim given that there was no other method.” 133 Komer, 1967, p. 9. 134 MACV, 1968b, p. 16. 135 Komer, 1970, p. 70. 136 Even a quick survey of oicial publications and transcripts of oicial statements on the HES from 1967 to 1972 will show that the program was intended as a quantitative tool to measure efects. Vietnam-Era Assessment 131 By the end of the war, it was clear that the HES had not successfully informed policy (or, at the very least, had not informed a successful policy). Because they were presented as scientiically accurate, the quantitative results, with their false precision, misled the executive branch, Congress, and the American public as to how the United States was actually performing in Vietnam. he lessons from the HES alone cannot be empirically generalized to all similar eforts, but they are similar to those generated by an examination of body counts, operational days of contact, hostile incidents, and other categories of assessment reporting. Weighted by these other examples, the HES appears to show that the most extensive and structured efort to obtain ground truth in a centralized quantitative system likely produced fabricated or irrelevant data that had no real utility at the tactical, operational, or strategic levels of operation and decisionmaking. It also shows that these eforts potentially distracted tactical-level counterinsurgents from focusing on what should have been their overarching purpose: mission accomplishment. Pattern and Trend Analysis of Vietnam War Data Perhaps only a single thread separates us from the truth, or perhaps an entire ream, but we will know for certain only when we look at the whole weave. —Hassan, in he Storyteller of Marrakesh137 If you want [the statistics] to go up, they will go up. If you want [the statistics] to go down, they will go down. —A South Vietnamese general oicer138 his section considers the value of aggregated Vietnam COIN data to campaign assessment and the possibility that useful patterns and trends could be found in what tended to be inaccurate, incomplete, and aggregated quantitative data. Eforts to apply pattern and trend analysis by MACV and OSA provide good insight into the ways in which assessment stafs use scientiic (in this case, statistical) analysis of data collected and reported through nonscientiic means to measure shifting objectives. hroughout War Without Fronts and in his 1985 presentation to analysts on assessing the Soviet war in Afghanistan (included in Rehm, 1985), hayer asserts that, despite obvious laws, Vietnam-era COIN data were reasonably accurate and suicient to ind useful patterns and trends. Speciically, “the patterns we found were very 137 Roy-Bhattacharya, 2011, p. 222. In the novel, the storyteller Hassan retells the story of the disappearance of a pair of tourists, each time embellishing the story with additional details and contradicting the memories of his listeners, who contribute their own (subjective) narratives, further muddying the waters. 138 Gibson, 2000, p. 314. 132 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency deinite.”139 He mentions patterns (consistent arrangements or behaviors, not necessarily over time) far more frequently than trends (prevailing tendencies over time) and, at times, he conlates these two concepts.140 Were the aggregated quantitative data from the Vietnam War reasonably accurate and therefore useful for assessment? Was it possible to ind deinite and useful patterns and trends in these data? Which elements of hayer’s analyses might show actual patterns or trends, and how might they show progress in a COIN campaign? Which elements probably do not show actual trends, and why not? What are the dangers of reporting trends that probably do not exist? While “reasonable” and “accurate” are loosely deined terms, it makes sense to deine them as distinct assessment objectives. Absolute accuracy is, of course, more diicult to achieve than reasonable accuracy. Accurate pattern and trend recognition requires accurate data. But hayer uses terms like “unoicial guess,” “notoriously unreliable,” and “tenuous estimate” to rate the quality of key categories of Vietnam War data. He describes the PAAS polling as “much less precise” than polls taken in the United States.141 he evidence clearly shows that the COIN environment played havoc with data collection and reporting in Vietnam. Both aggregated body count and HES data were signiicantly inaccurate and incomplete, but all the data sets contained substantial gaps and inaccuracies. Except in the most exacting and controlled of circumstances, no aggregated data set is ever 100-percent accurate or complete; as McKnight et al. show, even most scientiic studies may have data gaps that exceed 30 percent.142 But the degree of inaccuracy in Vietnam-era assessment data—as described by hayer, who supervised the analysis of those data for six years—strongly suggests that no policymaker should have placed strategic value on a single chart or graph. Does reasonably accurate pattern or trend recognition for COIN require accurate or only reasonably accurate data? And if it is possible to produce patterns from COIN data, are they useful to COIN assessment let alone operational planning? Can behavioral patterns drawn from reasonably accurate data—or any data for that matter—be equated with progress or lack of progress in a COIN campaign? Chapter hree showed that patterns and trends can be found in data that are somewhat inaccurate but consistently so. All hayer was looking for was consistency in reporting, with the idea that consistency shows through inaccuracy. A consistent pattern might or might not be revealing by itself, but it might be useful when compared to other data, or it might serve to highlight an area of concern for further analysis. For example, hayer shows 139 Rehm, 1985, p. 4-4. 140 A pattern can also be a trend if it is a pattern over time, but a pattern does not necessarily need to be identiied in time series. For example, the drop in violence in Afghanistan over the winter, year after year for many years, is a pattern over time; the coincidence of decreased civilian activity in the presence of an impending insurgent ambush is a general pattern; the steady increase in oil revenue over time is a trend but not necessarily a pattern. 141 PAAS data and related analysis can be found in hayer, 1975b. 142 McKnight et al., 2007, p. 3. Vietnam-Era Assessment 133 that friendly casualty levels always went down in the rainy season and always rose in the dry season—a pattern.143 he one aggregated metric that is almost always highly accurate in U.S. COIN assessment is U.S. casualties.144 Historical weather reports from Vietnam were reliable (it is always possible to know that it rained after it has rained). If hayer’s analysis uncovered a genuine pattern, was this identiied pattern useful to policymakers trying to determine success or failure in the Vietnam War? By itself, could it show the American public that the United States was achieving its objectives? It may have been useful to operational planners in the same way that the pattern of lower violence during Afghan winters is helpful to ISAF planners. But this inding is of questionable strategic assessment value, since, by itself, it reveals nothing about the progress of the campaign. his pattern is not reliably predictive because historical analysis of human behavior in warfare (even when matched to weather patterns) is not predictive. In an example that parallels the weather-casualty pattern, a July 26, 1970, MACV intelligence report stated, In the Laotian panhandle, the rainy season runs from the middle of May through the middle of October. During that period the enemy logistical activity along the [Ho Chi Minh] trail is almost at a stop. his year he is making a very deliberate efort to continue his logistical efort into the rainy season.145 Recognizing that there are traditionally fewer casualties and less ighting during heavy rains might help a policymaker understand historical aggregated casualty data by showing why casualties may have risen or fallen during speciic periods. But this kind of historical analysis, while it can have clear statistical correlations, cannot show clear cause-and-efect relationships: While it may be true that casualties rise and fall with the rains, casualty levels over any one speciic period will rise or fall for a complex web of reasons. At best, it would be possible to say the following about hayer’s weather pattern analysis: • • • • • here typically seems to be less ighting during heavy rain than during dry periods. Historically, friendly casualties have dropped during periods of heavy rain. It seems likely that this drop in casualties is due to reduced ighting. However, we cannot show clear cause and efect in all cases. his analysis is not reliably predictive. 143 Rehm, 1985, p. 4-5. 144 Although hayer shows that even these data were often lagging and inaccurate at any one point in time. 145 Sorley, 2004, p. 445 (emphasis in original). In this case, the pattern could be useful in showing how the North Vietnamese were applying additional efort to logistics. his information could be used to support an all-source intelligence analytic inding that could be incorporated into campaign assessment, but it would be one of many inputs in the assessment. 134 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Even if this pattern could be predictive, predicting insurgent behavior based on patterns (or trends) might be a recipe for disaster if insurgents are savvy: Insurgents are also capable of identifying the same pattern and can take advantage of pattern analysis. he North Vietnamese achieved strategic surprise against the United States and GVN by launching a major countrywide ofensive during the traditional Tet holiday in 1968. In the years prior to this ofensive, North and South had observed a cease-ire during Tet. he historical pattern showed that there was much less ighting during Tet than immediately before and after the holiday. his pattern analysis was so convincing that it likely precluded the United States and GVN from capitalizing on intelligence reporting that indicated that an attack was likely.146 It would not be much of a stretch for the North Vietnamese to take advantage of predictable U.S. behavior (e.g., a maintenance stand-down when VC units are assumed to be less active) during the rainy season.147 Could the inclement weather and casualty pattern have been useful as a coincidence indicator to support a broader assessment inding? Perhaps, but because it was a nonspeciic pattern and not a time-series trend based on information accumulated over a speciic period, it could not be safely compared to the kind of time-series trends that are used for centralized assessment (e.g., attack data or casualty data over time, updated to the date of the assessment). For example, one could compare the weather and casualty pattern to current time-series casualty data to show that the rainy season might have led to a reduction in casualties. But this analysis would not stand up to scrutiny because the nonspeciic pattern could not show clear causation. From an analytic standpoint, it would be dangerous and potentially misleading to claim that casualties were dropping because “they always drop this time of year due to the rain.” In this same case, then, could the pattern be used as a “lashlight” to point analysts in the direction of more explicit information? It could be used as a tool to help discern patterns from trends, but this kind of examination is better suited to intelligence analysis. In practice, hayer presented these patterns in his Systems Analysis series of reports as 146 Sharp and Westmoreland, 1968, p. 41. See also Wirtz, 1994, p. 66. Former CIA analyst John T. Moore testiied before Congress in the Westmoreland v. CBS case that the Defense Intelligence Agency had suppressed pre-Tet warnings based on a disagreement over troop strength estimates between the two agencies and MACV (Westmoreland v. CBS, 1984; see, speciically, Westmoreland’s memorandum in opposition to CBS’s motion to dismiss, October 19, 1982). 147 See Gayvert, 2010, p. 7. In a thought-provoking article on the HES, Gayvert points out that HES data showed a decline in GVN control across the countryside in the six months prior to the Tet Ofensive and that the HES might have clued analysts in to impending trouble. he HES data may have shown an increase in insurgent activity, but the gross HES graphics presented to the public showed an almost steady increase in GVN control across the population and only lat levels of VC control from January 1967 to January 1968 (Sharp and Westmoreland, 1968, p. 199). OSA analysts described the pre-Tet decline in support as a “slight net downgrading” and not a clear indicator of trouble (hayer, 1975b, p. 78). HES data also showed a very brief decline in GVN control after Tet. his decline may have relected the efects of the Tet Ofensive, or it may have been an indication that the preHES data were inaccurate. here is no way to tell whether the drop in control as reported by the HES in the six months prior to Tet was in fact accurate (and therefore indicative of problems) or whether it relected a shift in reporting criteria, a change in personnel, or any number of other factors. Vietnam-Era Assessment 135 key components of his assessments, and in his own words, he based his entire assessment process on the ability to ind useful patterns in reasonably accurate COIN data. What about trends identiied by OSA analysts? One of the standard methods of analysis was to compare U.S. operations to either enemy operations or environmental changes. his kind of analysis lines up well with contemporary trend analysis. Yet, two of the most confounding and inaccurate aggregated metrics used to measure progress in the Vietnam War were “battalion days of operation” and “operational days of contact.” he battalion days of operation metric purported to show how frequently U.S. and GVN military units were operating and at what scale (large or small). Operational days of contact purported to show only those days that battalions actually saw combat.148 In War Without Fronts, hayer inds that “the igures are not useful for even simple analysis of trends and comparisons.”149 He also calls any attempt to use these data for statistical analysis futile.150 However, he uses them for statistical trend analysis on the following page to produce what he calls tentative conclusions. Later in the book, he directly correlates statistical data on operational days of contact with casualty data in a table titled “Correlation Analysis: Friendly Battalion Operations with Contact Against Combat Deaths.”151 his statistical correlation analysis is presented by hayer without caveat. Anyone reading this table without the context of hayer’s earlier comments (pp. 55 and 56; the table appears on p. 93) would have every reason to believe that these data were accurate and that the correlation analysis was derived through a sound statistical method using valid data. he method may or may not have been sound, but the fact that OSA would use data described as unsuitable for comparisons to produce a correlation analysis shows how compelling aggregated quantitative data can be to both analysts and policymakers. he trends shown in the table are not useful for any analysis by OSA standards, so they should not be used for campaign assessment. he table on casualty comparison in War Without Fronts also compares bad operational data not just with accurate U.S. casualty data but also with bad enemy body count data.152 When data are used in correlation analyses (like the comparison of friendly operations to combat deaths), the most inaccurate and gapped data shape the overall analysis. For example, if it were possible to know the absolute accuracy and consistency 148 hayer points out that a battalion “day” was actually an aggregation of company days and that one battalion could have 1.33 days of potential activity. he entire system was nonsensical (hayer, 1985, p. 56). Further, Sheehan (1988, p. 636) describes how Marine Corps units were forbidden from counting paciication missions as battalion days of operation, despite the fact that paciication was part of their mission. 149 hayer, 1985, p. 55 (emphasis added). 150 hayer, 1985, p. 56. 151 hayer, 1985, p. 93. 152 It also calculates indings using assumed troop strength for a Republic of Vietnam Armed Forces infantry bat- talion and a U.S. infantry battalion. hese numbers do not relect reality; very few Vietnamese or U.S. infantry battalions were ever close to full combat strength. 136 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency of a data set, it might be possible to determine the overall inaccuracy of analysis based on the correlations among the data elements in two or more sets. If the operational days of contact data set was 40-percent accurate and the U.S. casualty data set was 95-percent accurate, the correlation analysis based on the two data sets would still be, at best, only about 40-percent accurate. his means that any efort to analyze inaccurate aggregated data by correlating them with other data—the most common way to place aggregated data in context or to identify cause-and-efect linkage—would be counterproductive. Since there is no way to determine absolute data accuracy in COIN, there is no way to determine the accuracy of any correlation-based analysis for COIN assessment. It is also possible to show patterns and trends in the quantitative analysis of Vietnam War data that are precise and consistent but misleading. he HES may have produced some consistent data, but those consistencies probably relected a good deal of consistent inaccuracy, or what Downes-Martin refers to as “precision without accuracy.”153 For example, the advisor in Phong Dinh Province consistently reported most of the same erroneous data for almost his entire tour. hayer contends that HES data correlated well with non-HES data to show deinite patterns. For example, he shows that anti-aircraft incidents “tend to occur over VC controlled hamlets, and incidents around A or B hamlets tend to feature terror.”154 his may have been true, but those A- and B-rated hamlets may also have been hiding large VC units or VCI elements that were savvy enough not to draw attention to their presence by conducting anti-aircraft or large ground attacks nearby (e.g., Vinh Binh Province in 1971). he VC could easily igure out that advisors rarely visited hamlets and then typically did so in daylight; it would not be diicult to deceive them by simply terrifying the population into putting on a good face and hiding for an hour or two once a month. It is very possible that many A- and B-rated hamlets were, in fact, sanctuaries for insurgents.155 he steady rise in population security recorded in the HES data could have been due to an actual improvement in paciication, but it could also have been the result of consistent but inaccurate reporting from advisors who were unoicially deterred from downgrading reports and simultaneously rewarded for showing at least incremental improvement. his kind of pressure could lead to a slow but steady improvement in the percentages, and this trend could have been sustained as U.S. advisors transferred reporting responsibility to GVN partners who may have been more likely to report 153 Downes-Martin, 154 hayer, 2010b, p. 7. 1975b, p. 61. 155 Both Corson (1968) and Ahern (2010) substantiate this assumption. Ahern states that the VCI may have been present when main-force VC units were not. He describes an operation in a hamlet that sat only a few hundred yards from a standing Marine Corps infantry battalion’s headquarters. he Marines and accompanying GVN civil and military forces discovered “booby traps, concrete-lined bunkers, camoulaged pits concealing sharpened bamboo stakes . . . and caches of documents and weapons” (Ahern, 2010, pp. 249–250). Vietnam-Era Assessment 137 improvement as a matter of course.156 Trends that emerge from consistently poor and misleading data are poor and misleading. he assertion that deinite and useful patterns and trends for COIN campaign assessment could be gleaned from aggregated quantitative data of notably poor quality is not proven. Leslie H. Gelb and Richard K. Betts, authors of he Irony of Vietnam: the System Worked, believe that these analyses were often mistaken and misleading: [S]tatistics by their very nature could not go deeply enough. Much of the most important information about Vietnam was essentially unquantiiable, and even when it could be quantiied the dangers of misinterpretation and overinterpretation were ever present. Comparison with years past was an illusory measure when it was not coupled with judgments about how far there still was to go and how likely it was that the goal could ever be reached. It was all too easy to confuse shortterm breathing spells with long-term trends and to confuse “things getting better” with “winning.”157 he last sentence in this quote is particularly relevant for pattern and trend analysis. Because pattern and trend analysis in Vietnam was necessarily imprecise, it was not possible to establish meaningful time-series thresholds for operational or strategic patterns and trends. Consequently, analysts could not plot an accurate, deinable path to victory along time-series graphs because there was no reasonably deined quantitative objective. his means that all pattern and trend analyses required interpretation that was almost entirely subjective. he degree of subjectivity inherent in these analyses undermined their value in the absence of extensive narrative explanation, but narrative explanations (e.g., province advisor reports) became superluous when quantitative data (e.g., HES data) were made available. Policy debate then boiled down to the subjective interpretation of inaccurate data by people far removed from the battleield. In the few cases in which quantiiable objectives or thresholds were established during the Vietnam War, they proved to be both meaningless and misleading. Population security objectives established at various times by Komer and Westmoreland did not produce victory even when they were surpassed, and, as discussed later in this chapter, enemy casualty thresholds failed to deliver a quantiiable milestone for success.158 hayer recognized the inherent subjectivity in pattern and trend analysis, and he always couched his analyses with narrative. He had experience in Vietnam and continued to draw on viewpoints from the ield to support his analysis at OSA.159 Even a 156 Details 157 Gelb of this transition to GVN control can be found in Cushman, 1972, p. 27. and Betts, 1979, p. 303. 158 For example, Komer described a paciication threshold objective of 1,800 hamlets in 1966. his particular threshold was not met, but others were (Komer, 1968a, p. 2). 159 See, e.g., Sheehan, 1988, p. 683. Sheehan describes hayer’s debrieing of Army LTC Hal Moore, commander of a U.S. Army infantry battalion that had recently been involved in heavy ighting in the Ia Drang Valley. 138 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency quick review of the Southeast Asia Analysis Reports shows that hayer and his analysts sometimes had a much more in-depth understanding of the issues than the aggregated data alone could relect.160 hey had obviously done extensive research on speciic subjects and relied on multiple sources for some elements of their work, although these other sources were rarely cited. While some of the assessments in the Southeast Asia Analysis Reports relected shallow quantitative interpretation, others were nearly (for lack of sourcing) on par with high-quality all-source intelligence analyses. For example, hayer’s analysis of the security situation along Route 4 between the Mekong Delta and Saigon in 1967 contained what appears to be useful insight that might have supported operational decisionmaking. he core of this analysis hinges on two captured enemy documents that, if accurate, showed clear intent to disrupt traic on Route 4. he HES and security force data presented in charts and tables add almost nothing to the contextual narrative; they could have easily been left out of the report. he important quantitative data are those that show a twofold increase in VC attacks along Route 4 in two speciic provinces. In this case, hayer showed a useful trend— one that helped verify the validity of the captured documents.161 he important distinction between this kind of analysis and his analysis of broader statistics in the Southeast Asia Analysis Reports and War Without Fronts is that (at least in this example) the quantitative data set was small, locally focused, and contextualized with other sources and narrative. Even this rather useful analysis would have beneitted signiicantly from citation and added subordinate context. OSA pattern and trend analysis was least useful and potentially most misleading when it was highly aggregated and presented without suicient context. Komer, who was working out of Saigon, took hayer’s analysts to task for failing to conduct a more detailed analysis of the paciication program. He questioned why OSA was not incorporating reams of other available data into the analysis published in the Southeast Asia Analysis Reports and asked the analysts not to “over-read” the HES data. From his vantage point in Saigon, Komer recognized that assessment requires context. Yet, the kind of contextual analysis he was demanding from OSA was diicult to develop for a small, centralized staf. he OSA analysts replied that they did not have access to detailed data streams and then fell back on their quantitative processes to defend their indings.162 160 he reports were later republished in the 12-volume series A Systems Analysis View of the Vietnam War 1965–1972. 161 hayer, 1975c, pp. 9–22. his analysis would have been much more credible if it contained source notes. In the absence of sourcing, it is unclear how much of this analysis was hayer’s opinion and how much was grounded in speciic reporting. It is also unclear to what degree the reporting he did use was accurate, although he does caveat some of his information. 162 hayer, 1975b, pp. 75–83. his analytic exchange ofers fascinating insight into the inner workings of both CORDS and OSA. It reveals Komer’s deep dissatisfaction with elements of OSA’s analysis. (At one point, he calls the analysts’ assumptions “silly.”) He directly and angrily confronts the analysts, identifying them as “desk types Vietnam-Era Assessment 139 OSA analysis was most useful and least likely to be misleading when minimally aggregated and presented in context; the Route 4 study is a case in point. But these detailed narrative assessments crossed the line from strategic campaign assessment into the realm of operational intelligence analysis.163 he study of VC activity along Route 4 could have been useful for operational planning, and it certainly informed aspects of COIN assessment. But the report was no more a holistic assessment than any other intelligence analysis report. By itself, it did not show strategic progress, and it did not present a suicient stand-alone case for increasing or decreasing troop deployments. he report, and others like it, would have been more appropriately written by an operational intelligence staf and then incorporated into a layered contextual analysis of the war. here is clear indication in hayer’s writing and postwar statements that the more he became immersed in the data, the more he and his analysts slid away from comprehensive assessment and into the realm of intelligence analysis. he desire for meaning and context in the aggregated data drove them to seek more and more contextual detail even as they continued to try to ind broad meaning in gross quantitative patterns. hayer analogized the process to cooking.164 hayer and OSA made a valiant efort to pull together data and analysis from a cold start. hey had no COIN assessment doctrine to work with, and they had no historical case studies to rely on (at least not one that compared to the complexity of the COIN campaign in Vietnam). hey were not always fully assisted in their eforts by MACV and other military organizations, and they often had to hunt down their own data: In April 1968, they stated that they were “hampered by an almost total in Washington,” and challenges them to visit Vietnam to check their work. he analysts’ reply was detailed. his back-and-forth debate has been replicated in both form and (with necessary changes in terminology) context in both Iraq and Afghanistan. Some might argue that debate of this sort is healthy, and from a strictly analytic standpoint they would be right: It is healthy for analysts to have iterative engagement with people in the ield to gain ground-truth perspective on their reporting. However, as the head of the paciication program, Komer was necessarily an interested party in the results of the analysis. his two-way debate would have been well served by other input. 163 his may have technically fallen under the purview of OSA’s Southeast Asia Intelligence Analysis and Force Efectiveness Division, but it is not clear from the literature whether the oice was speciically conducting intelligence or assessment. However, hayer’s interpretation of his division’s work seems to make it clear that it was focused on assessment and systems analysis for strategic decisionmaking. 164 Rehm, 1985, p. 9-11. In another example of the shift to intelligence analysis, hayer stated, If we had known before the Tet Ofensive . . . that Westmoreland thought there were 25,000 iniltrators a month coming down we would have lipped to McNamara or whoever was there at the time. We would have said, “there is something coming. We don’t know what it is or where or when, but we suggest that you want to ask the Command or the Joint Chiefs of Staf the following questions.” (Rehm, 1985, p. 9-13) his is clearly an intelligence indications and warning analysis by all deinitions. hayer may not have been precluded from making these analyses (his division had the word “intelligence” in its title), but if this was the focus of OSA assessment then the oice should have been resituated within an intelligence analysis organization, such as the CIA or Defense Intelligence Agency, or intelligence analyses should have somehow been segregated from assessments. 140 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency absence of meaningful data.”165 he data they did receive were signiicantly inaccurate and incomplete. Strategic objectives were often unclear. he systems analysts and operations researchers at OSA reverted to centralized quantitative analysis by default. As a result, OSA analyses were presented in slices that did not form a comprehensive picture. he patterns and trends that were identiied could not be defended as accurate and did not build an overarching understanding of progress in comparison to strategic objectives. Policymakers and the public were left to contend with vague patterns and misleadingly precise trends drawn from quantitative systems analyses that were so complex as to be opaque to the average American. his was a best efort in the absence of other clear options, but it was not transparent, credible, relevant, balanced, or parsimonious. One could argue that the data were analyzed, but analysis was conducted without an overarching method designed to produce a holistic impression of the campaign. Even during this period, ample COIN doctrine and literature were available to OSA.166 It would have been possible to create an analytic process that was congruent with a more realistic appreciation for the COIN environment, but probably not from a top-level staf. Casualty Data and Impressionistic Assessment of the Vietnam War Every quantitative measurement we have shows that we are winning the war. —Secretary of Defense Robert S. McNamara, 1962167 Building on the discussions of Vietnam-era assessment presented in this chapter, this section develops a case example showing how senior leaders used aggregated data (speciically casualty data) to shape strategy and support decisionmaking during the Vietnam War. he example revolves around the strategic assessment of aggregated body count data by both military commanders (e.g., GEN William C. Westmoreland) and civilian leaders (e.g., Secretary of Defense Robert S. McNamara and National Security Advisor McGeorge Bundy). It refers to General Westmoreland’s analysis of a casualty data chart in his Report on the War in Vietnam as of 30 June 1968, which he (and his staf) coauthored with ADM U. S. G. Sharp, commander of U.S. Paciic Command.168 165 hayer, 1975b, p. 77. 166 his includes the U.S. Marine Corps’ Small Wars Manual, the Army’s FM 31-series publications dating back to the 1950s, Mao Tse-Tung’s On Guerrilla Warfare, and Che Guevara’s Guerrilla Warfare. 167 Sheehan, 168 As 1988, p. 290. commander of the U.S. Paciic Command, Sharp was efectively the next level of command above Westmoreland. he speciic analysis of casualty data was probably authored by Westmoreland and MACV staf, not by Sharp and his staf. he layout of the book seems to show that Sharp addressed “area of responsibility” issues outside of Vietnam, while Westmoreland authored or edited the portions on the war within Vietnam. he discus- Vietnam-Era Assessment 141 At more than 300 pages, the report is replete with statistics, charts, and photographs. It represents the consolidated military data on the war up to mid-1968. It also contains contextual narrative of the war efort and what amounts to the holistic assessment of the U.S. COIN efort in Vietnam. Westmoreland conveys most of his key assessment points through annotated charts. Westmoreland’s assessment is predicated on a theory of attrition warfare and not population-centric warfare as described in the 1940 Marine Corps Small Wars Manual or in the 2006 COIN manual, FM 3-24.169 After Harkins, he was the leading proponent of the attrition warfare strategy and believed that when the enemy “loses one man it’s equivalent to our loss of more than ten.”170 he entire attrition strategy was based on unsustainable quantitative assumptions like this. he purpose of this section, however, is not to debate the relative merits of attrition versus population-centric COIN strategies. Instead, it is intended to explain how centralized quantitative assessment can present and sustain a misleading understanding of a COIN campaign. It shows how timeseries analysis, trend analysis derived from time-series data, and quantitative thresholds can both confuse policymakers and mask gaps in knowledge and understanding. Figure 6.3 presents a comparative casualty ratio. It shows the ratio of enemy casualties (body count statistics) compared to allied casualties (U.S. casualty statistics) between January 1965 and December 1968. hroughout their report, Sharp and Westmoreland use these types of charts to support their narrative comments, and the narrative tends to rely heavily on the charts.171 Westmoreland annotated this graph with an assessment on the opposite page (p. 190). He does not describe how he reached his conclusions, what other inputs he might have used to analyze these data, or how he drew a connection between cause and efect: he chart on the facing page shows the general upward trend in the ratio of friendly to enemy casualties. . . . he trend is clearly up, and at the present time it has reached a ratio in which six of the enemy are killed for each allied soldier lost. From a purely military standpoint this trend shows the impact of the introduction sion here does not present a comprehensive analysis of the Report on the War in Vietnam; such an examination was beyond the scope of this study. 169 See Headquarters, U.S. Marine Corps, 1940, and HQDA, 2006c, respectively. 170 Westmoreland, 1967b, p. 5. On page 31, he explains that this ratio was derived from the diference between North Vietnamese and U.S. population igures. He should have stated in his earlier remarks that his ratio was purely mathematical. In this press brieing, Westmoreland clearly deines and explains the attrition strategy and his theories of counterrevolutionary warfare. Westmoreland reads from an operational update that he received that morning. It described a captured NVA oicer who had oicial military orders on his person to ight to the last man. Westmoreland also states on page 37 that the Battle of Dak To was “the beginning of a great defeat of the enemy.” 171 Sharp and Westmoreland, 1968, p. 191. 142 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency of U.S. troops, the steady improvement in performance by all allied forces, and the steady decline in battleield performance by the enemy.172 When parsed, this paragraph reveals some of the limitations of operational timeseries assessment as a holistic assessment tool. Because one of the data sets (enemy body count) was inaccurate to a signiicant degree, the entire chart is inaccurate to a signiicant degree. If the reporting had been accurate it might have helped Westmoreland support the supposition in the inal sentence: Allied forces were steadily improving while the enemy was steadily declining in performance. However, even a reasonably accurate trend could be the result of other factors. A higher VC and NVA casualty rate might show increasing boldness among the enemy or an increase in active manpower (more troops in the ield leads to more casualties). his would be a negative and not a Figure 6.3 Ratio of Enemy to Allied Casualties, Vietnam, 1965–1968 TET 8.5 8.0 7.5 Annual ratio 7.0 6.5 6.0 YR 60 61 62 63 64 65 66 67 68 5.5 5.0 2.5 2.0 1.5 1.0 3rd offensive 3.0 2nd offensive 3.5 Air strikes vs. North Vietnam 4.0 U.S. begins large-scale build-up 4.5 0.5 0 J F M A M J J A S O N D J F M A M J J A S O N D J F M A M J J A S O N D J F M A M J J A S O N D 1965 1966 1967 1968 SOURCE: Sharp and Westmoreland, 1968, p. 191. RAND MG1086-6.3 172 Sharp and Westmoreland, 1968, p. 190. It is concerning that Westmoreland misidentiies his own chart as friendly-to-enemy casualties rather than enemy-to-friendly in this assessment. his shows either a simple editing error or a deeper failure to understand the data. Vietnam-Era Assessment 143 positive trend for the United States and GVN. Myriad other factors that play into a ratio like this are not—and could not be—depicted in such a graph. Westmoreland caveats his inding by adding the phrase, “from a purely military standpoint.” He seems to suggest that one should not draw any holistic (i.e., national strategic) conclusions about the COIN campaign from the graph.173 However, despite his caveat, Westmoreland clearly implies that this improved ratio is a key indicator of success over time. His next time-series graph shows the ratio of enemy to allied weapon losses over the same period. In the text accompanying this graph, he states, “Surely the combination of these irst two charts relects a situation which must be of major concern to the enemy.” his supposition, ofered without any clear analysis or corroborating evidence, was mistaken.174 Neither President Ho Chi Minh nor the commander of the NVA, General Vo Nguyen Giap, was overly concerned with casualties or weapon losses.175 hey made this clear in statements that would sound boastful coming from anyone other than 1960s-era North Vietnamese communists. In a seemingly direct refutation of Westmoreland’s casualty ratio assumption, Ho famously stated, “You can kill ten of my men for every one I kill of yours, but even at those odds, you will lose and I will win,” while Giap stated, “Every minute, hundreds of thousands of people die on this earth. he life or death of a hundred, a thousand, tens of thousands of human beings, even our compatriots, means little. . . . Westmoreland was wrong to count on his superior irepower to grind us down.”176 While the VC and NVA may have been losing more weapons in 1968 than in 1965, China was pouring materiel into the North, replenishing supplies perhaps as fast as they were consumed.177 By 1965, the communists had settled in for a long war, and Ho Chi Minh was prepared to ight for “ive, ten, twenty-ive years, or even longer.”178 Westmoreland looked at the graph, saw what appeared to be a positive trend, and then wrongly extrapolated what should 173 Even taking into account the potential contextualization provided by the rest of the nearly 300-page report would likely leave the consumer with little more than a vague impression of progress and no comprehensive understanding as to how this progress it into a strategic framework. 174 To be fair to Westmoreland, as late as 1970, the intelligence community continued to assert that the leadership in Hanoi could not sustain manpower losses indeinitely (“he Outlook from Hanoi: Factors Afecting North Vietnam’s Policy on the War in Vietnam,” Special National Intelligence Estimate 14.3-70, February 5, 1970, in National Intelligence Council, 2005, pp. 506–526). his reporting may not have been accurate, but it probably inluenced Westmoreland’s thinking. 175 It is diicult to analyze Westmoreland’s comment because the phrase “of major concern” is a vague and mean- ingless qualiier to a policymaker attempting to determine courses of action. here may have been times that both leaders were concerned about the logistics of manpower and materiel, but these concerns did not deter them from believing in eventual victory against the South and the United States. 176 Karnow, 1984, p. 20. 177 Zhai, 2000, pp. 179–180. 178 CIA, 1966, p. VIII-9. 144 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency have been a very narrow set of data—or a trend—into a broad conclusion about the overall success of his COIN eforts. He had reams of other analyses, raw reports, and advice at his ingertips, but his strategic bottom line was derived immediately and incorrectly from centralized quantitative charts. his episode shows the degree of inluence that aggregated and decontextualized charts can have on even seasoned and well-informed decisionmakers. McNamara also used body count statistics to assess the war and to inform policymakers and the public. As Secretary of Defense during the period that Westmoreland was in command in Vietnam, he is ultimately responsible for the Report on the War in Vietnam and for the analysis coming out of OSA. He defended his role in counting VC and NVA casualties, stating, “We tried to use body counts as a measurement to help us igure out what we should be doing in Vietnam to win the war while putting our troops at the least risk.”179 He generally attributed the details of the attritionbased assessments to Westmoreland. He described Westmoreland’s conviction that the North would hit a “crossover point” at which it could no longer sustain the level of casualties inlicted by U.S. and South Vietnamese forces. Westmoreland thought that the point was reached in the spring of 1967, a year before the strategically disastrous Tet Ofensive.180 It is obvious in retrospect that they were wrong, but how were Westmoreland and McNamara determining that this crossover point existed or had been reached? McNamara stated that “we needed to have some idea what they could sustain and what their losses were,” but he failed to explain how they determined the irst part of the equation.181 his crossover point is simply a time-series threshold or milestone.182 In the case of Westmoreland’s casualty ratio graph, the determined threshold would appear as a number along the y-axis of the graph. At that speciic quantitative point, the North Vietnamese would lose the will to ight or would be physically defeated. Identifying that point would require knowing that (1) this method could have the desired efect (killing wins the war); (2) the North Vietnamese leadership had a speciic casualty limit in mind, and when that number was reached they would quit; or (3) the United States knew more about the willpower and intent of the North Vietnamese leaders than those leaders knew about themselves, down to the speciic ratio of casualties they were willing to accept. Any shift in the Vietnam casualty ratio would result in a change to the assessment of both the efect and the overall strategic outlook. his change would 179 McNamara, 1995, p. 238. his statement is sadly ironic, considering that the body count metric put troops at risk. 180 McNamara, 1995, p. 238. 181 McNamara, 1995, p. 238. 182 If Westmoreland had predicted that the threshold would have been achieved at a certain point in time, it would also have been a milestone. His “tipping point” became a rather dubious threshold and milestone only in retrospect. Vietnam-Era Assessment 145 have to be justiied in some detail to retain transparency and credibility, assuming that it was possible to justify the quantitative threshold in the irst place. National Security Advisor McGeorge Bundy played a key role in shaping the Vietnam War strategy and particularly in estimating enemy strength and resiliency. He is generally recognized as one of the most inluential advisors in the White House in the early to mid-1960s. In a self-critical statement, he admitted that there were insuficient analyses at all levels of the national security apparatus and that all analytic stafs failed to grasp “how the enemy would take it and come back for more.”183 According to Bundy biographer and conidant Gordon M. Goldstein, “here was no analysis or evidence to validate Bundy’s expectation that Ho Chi Minh and his fervent followers would capitulate.”184 Not only did Bundy fail to adequately examine the policy of attrition, but he also failed to analyze or shape the assessment of the attrition strategy. While it is not clear what role Bundy had in shaping or approving Westmoreland’s ill-conceived body count threshold, Goldstein takes him to task for failing to follow through on the mechanics of the attrition assessment process: How many casualties would be required to compel them to quit? How many years would it take? Where, exactly, was the tipping point—the threshold of pain and loss that would extract a fundamental reversal in Vietnamese nationalist ambitions?185 Goldstein’s questions are rhetorical, even sardonic. But to Westmoreland, McNamara, and Bundy, each struggling in his own way to win the war, these questions might have seemed quite reasonable early on. Vietnam arguably was the most complex irregular war in which the United States had engaged to that point: he military was battling against both a conventional army supported by powerful external sponsors and a full-strength insurgency consisting of tens (if not hundreds) of thousands of experienced and aggressive combatants. he U.S. military was simultaneously trying to hold together a faltering ally while its own morale ebbed. here was extraordinary pressure on the top advisors and decisionmakers in the military, the Pentagon, and the White House to bring the war to a successful conclusion. Each of these three men, in his own way, bought into the idea that a inite solution might exist, that this solution could be achieved by applying graduated pressure and speciic amounts of resources against the threat, and that the solution could be measured through systems analysis 183 Goldstein, 2008, pp. 178–179 (emphasis in original). 184 Goldstein, 2008, p. 179. 185 Goldstein, 2008, p. 179. 146 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency and displayed on a time-series graph. In the end, they were all proven wrong, and both McNamara and Bundy agreed in retrospect that theirs was a failed approach.186 his example shows the dangers of relying on a narrow set of aggregated and decontextualized data to inform strategic decisionmaking in COIN. It is not a precise allegory for either Iraq or Afghanistan because neither of these wars has been prosecuted according to attrition warfare theory and neither relects the kind of hybrid conventional/COIN war fought in Vietnam. However, it does show how even “the best and the brightest” can fail to understand the complexities of COIN and COIN assessment. It also shows that despite the availability of massive amounts of aggregated quantitative data and inished quantitative analyses from MACV and OSA, both senior military leaders and policymakers relied on supposition and overly optimistic readings of quantitative charts to reach at least some of their major decisions. Some of the criticisms aimed at these senior leaders may be valid, but it is also apparent that there was something missing from the assessment process. Is Analysis of Vietnam Assessment Germane to Iraq and Afghanistan Assessment? Comparisons between Vietnam and contemporary COIN campaigns suggest that it is reasonable to equate the Vietnam-era assessment processes and those used in Iraq and Afghanistan for the purposes of research: • here were signiicant advances in technological capability between 1972 and 2001–2011, particularly in the ields of communication and surveillance. While technology can certainly account for some improvement in data reporting, particularly in the speed of reporting, units in both Iraq and Afghanistan continue to pass reports in written form (even if it is electronic) and via tactical radio. • COIN is still conducted by infantry units patrolling on foot through underdeveloped urban and rural terrain, and these units face the same challenges that their counterparts faced in Vietnam. heir riles, radios, rockets, hand grenades, boots, and helmets typically are improved versions of Vietnam-era equipment. here has been no “revolution in COIN afairs.” • Centralized core metrics have not changed signiicantly, and in most cases, Iraq and Afghanistan metrics appear very similar or identical to each other and to Vietnam-era metrics.187 Units in contemporary campaigns are generally trying 186 McNamara tended to equivocate on analytic methods and in several statements defended the application of systems analysis to war assessment. However, he also stated that the attrition and graduated pressure strategies were poorly conceived. 187 For example, many of the questions in the HES are similar to both the security and development metrics in use in Afghanistan today. Vietnam-Era Assessment • • • • 147 to count the same things that their counterparts were attempting to count in Vietnam—including bodies in some cases. Polls are still taken using face-to-face engagement and human intelligence is still captured in face-to-face discussion; all collection and reporting continues to depend on people for precision and accuracy. Assessments in both Iraq and Afghanistan rely, in part, on host-nation reporting, which is often unreliable. Most host-nation reports are delivered in writing (typically on paper and not in English), in face-to-face meetings, or via radio.188 Time has not changed the geometry of war. Accurate population-centric data cannot be collected in insurgent-held territory or in places where there is no friendly presence.189 his was true in Vietnam and it is true in Iraq and Afghanistan. Vietnam-era assessment tools like the HES were efects-based despite the fact that the language of efects-based theory did not permeate U.S. doctrine until well after the end of the war. It would not be far-fetched to state that the current process is immediately derived from, and closely mirrors, the Vietnam-era process. Similarities between some of these systems are general, while others are speciic. his section concludes with a striking comparison of the 1964 paciication assessment process in Vietnam and the 2010 district assessment process in Afghanistan: • In mid-1964, MACV implemented the Paciication Reporting System. he system purported to show the degree of control that the GVN had over key areas of the countryside. here were ive rating codes in this system: secured, undergoing securing, undergoing clearing, uncontested, and VC-controlled. hese ive codes were represented by colors (respectively, dark blue, light blue, green, white, and red) on a map that was maintained by the Revolutionary Development and Support Directorate in Saigon.190 • In early 2010, IJC implemented the District Assessment Model (DAM), a system that purported to show the degree of control that the government of Afghani188 Iraqi reporting might be more accurate than Afghan reporting because the Iraqis have a tradition of structured and centralized administration from the provinces to the center and the Afghans do not; the South Vietnamese had a highly structured bureaucracy designed by French colonial administrators. While host-nation reports during war should be considered suspect for a variety of reasons, there is no clear reason to value Iraqi or Afghan reporting over Vietnamese reporting. Cooper et al. (1972b, p. 235) advise, “Never again should the United States, while lending advice and assistance to a host nation ighting an insurgency, allow itself to become . . . dependent, for the vital information it requires for policymaking, on distorted and otherwise unsatisfactory reporting and evaluation by host nation agencies.” 189 It is always possible to collect intelligence from anywhere on earth, but the quality and completeness of the data drawn from intelligence collection is highly dependent on collection capabilities, access, and resource availability. Collection of accurate population data requires some kind of immersion in the population. 190 Cooper et al., 1972b, p. 210. 148 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency stan had over key districts in the countryside. here were ive rating codes in this system: population supporting the government, population sympathetic, population neutral or on the fence, population sympathizing with insurgents, and population supporting the insurgency. hese ive codes were represented by colors (respectively, green, blue, yellow, olive, and red, with white indicating “not assessed”) on a map that was maintained by the IJC Information Dominance Center in Kabul, Afghanistan. Chapter Summary Assessment of the Vietnam War was haphazard before MACV became a theater combat command in the early 1960s. Between the mid-1960s and the early 1970s, MACV and OSD (with CIA assistance) created the single largest and most comprehensive military COIN assessment apparatus in the history of warfare. It involved the eforts of hundreds of thousands of military personnel, civilians, Vietnamese nationals, intelligence experts, and analysts over the course of several years. hese contributors produced hundreds of millions of data items (in DoD parlance), tens of thousands of tactical and operational analytic reports, and hundreds of comprehensive assessments that addressed nearly every aspect of the war. While it is not possible to state that poor assessment led to the loss of the Vietnam War, the war was distinguished by internal confusion, poor decisionmaking, and, ultimately, strategic defeat. It is apparent from analyzing the way in which assessments were presented and used that they contributed to many of the poor decisions that led to this defeat. Why did operational COIN assessment of the Vietnam War fail? his chapter explored the following issues and reasons:191 • No adequate precedent or doctrine on assessment existed prior to the war. • he assessment data were intrinsically lawed due to the complex and chaotic environment of Vietnam. • Centralized pattern and trend analysis was often inaccurate or misleading. • Top-down assessment requirements resulted in widespread data fabrication. • Senior leaders and policymakers misunderstood the value and meaning of available quantitative data. • Senior leaders, policymakers, and analysts overrelied on centralized, quantitative analyses of aggregated data that were devoid of useful context. 191 his is not to imply that defeat resulted directly from poor campaign assessment. Policymakers, military oicers, the North Vietnamese, the South Vietnamese, and many other players all had a role in shaping the end of the war. Vietnam-Era Assessment 149 Overreliance on centralized, quantitative analysis was commonplace. here seems to have been no systematic efort to integrate qualitative reporting (e.g., human intelligence reports, engagement reports) into the OSA assessment process. Similarly, the MACV assessments were heavily quantitative.192 his is an unfortunate oversight. In contrast to both Iraq and Afghanistan, Vietnam was a holistic U.S. government campaign from the beginning of the war (and certainly by the mid- to late 1960s). Hundreds of civilian advisors worked at the district and province levels across South Vietnam.193 hese advisors wrote hundreds if not thousands of highly detailed narratives that placed both qualitative and quantitative reporting within (often) insightful and structured context. he MACV CORDS report on Phong Dinh Province for February 1968 is an excellent example of the kind of information that was available from the ield.194 Military advisors and intelligence oicers also produced reams of detailed narrative analysis that contained useful ield context. While analysts and policymakers certainly read some of these reports, neither the U.S. military nor the civilian establishment systematized the collation and analysis of these reports for campaign assessment. Some of the narratives written by senior leaders were relatively useful in presenting a step-by-step history of the war (e.g., Sharp and Westmoreland’s Report on the War in Vietnam) or in presenting senior policy positions. hese senior-level reports did not deliver consistent, transparent, or credible assessments, however. Published reports were often suspect in the minds of consumers because they contained obvious biases or were not clearly tied to contextual ield reporting.195 his research shows that the data produced by MACV and OSA obviated the requirement for comprehensive qualitative insight or narrative context, at least in the minds of some policymakers.196 he section “Pattern and Trend Analysis of Vietnam War Data,” earlier in this chapter, compared the assessment methodology employed by OSA to the standards for assessment framed in the introduction to this monograph. It showed that the kind of pattern and trend analyses applied to the Vietnam campaign did not stack up well against these standards. Table 6.1 shows how assessment of the Vietnam War compared to the standards. Subsequent chapters show a clear connection between COIN assessment processes in Vietnam, Iraq, and Afghanistan. here are certainly some diferences among the cases, and each campaign has unique characteristics: Vietnam is not Iraq is not 192 See, for example, Sharp and Westmoreland, 1968. his refers to campaign assessment and not necessarily to MACV intelligence analysis. 193 Some civilian advisors work closely with military units in remote areas of Afghanistan and some did so in Iraq. However, neither of these eforts compared well with Vietnam-era eforts. 194 MACV, 195 See, 196 he 1968a; see Appendix C of this monograph. again, Sharp and Westmoreland, 1968, p. 9. notable exceptions to this inding are the comprehensive intelligence analyses written by the CIA and other intelligence agencies, as well as the Pentagon Papers. 150 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Table 6.1 Performance of Vietnam-Era Assessment Against Assessment Standards Standard Vietnam-Era Assessment Analysis Transparent No Data aggregation and methods opaque Credible No No clear method, data questionable, oversold to policymakers and public Relevant No Did not effectively support decisionmaking Balanced No Did not systematically incorporate qualitative and mixed data or subordinate reports Analyzed No Analysis was not consistent or holistic Congruent No Did not reflect understanding of COIN Parsimonious No Data requirements were overbearing Afghanistan. But the challenges of assessment in Iraq and Afghanistan are consistent with most of the challenges faced in Vietnam. While Vietnam provided a precedent for COIN assessment, the following points are consistent for Iraq and Afghanistan assessment: • No adequate doctrine on assessment existed prior to the war. • he assessment data were and are intrinsically lawed due to the complex adaptive environments of Iraq and Afghanistan. • Centralized pattern and trend analysis was and is often inaccurate or misleading. • Top-down assessment requirements have resulted in data fabrication. • Senior leaders and policymakers misunderstand the quantitative data. • Senior leaders, policymakers, and analysts overrely on centralized, quantitative analyses of aggregated data that are devoid of useful context. he U.S. military and the policy community have not adequately incorporated the lessons of the Vietnam COIN assessment case into military doctrine or into the U.S. government’s approach to warighting. Or, quite possibly, they have incorporated the wrong lessons. As a result, military stafs and policymakers must continue to rely on assessment processes that are “better than nothing.” It remains to be seen whether this prescient post–Vietnam War quote from Don Oberdorfer, Washington Post reporter and author of Tet! he Turning Point in the Vietnam War, will also apply to future U.S. COIN campaigns as it has to Iraq and Afghanistan: Originally intended to place the conduct of the war on a “scientiic” and thus manageable basis and to enhance public conidence, the practice [of aggregated quantiication of war data] ultimately consumed vast amounts of time and energy, Vietnam-Era Assessment 151 led to misconceptions and erroneous conclusions about the war and was a major factor in the erosion of public conidence. Nevertheless, the “numbers game” may be now so deeply imbedded in military, press and public thinking that it will persist in future military conlicts.197 197 Oberdorfer, 2001, p. 262; also quoted in U.S. Defense Logistics Agency, 1980, pp. 15-26–15-27. Oberdorfer’s account of the Tet Ofensive was originally published in 1971. CHAPTER SEVEN Assessment in Afghanistan Going forward [in Afghanistan], we will not blindly stay the course. Instead, we will set clear metrics to measure progress and hold ourselves accountable. We’ll consistently assess our eforts to train Afghan security forces and our progress in combating insurgents. We will measure the growth of Afghanistan’s economy, and its illicit narcotics production. And we will review whether we are using the right tools and tactics to make progress towards accomplishing our goals. —President Barack Obama, in a speech on the U.S. strategy in Afghanistan, March 27, 2009 he Vietnam case study ofers a comprehensive view of the COIN assessment process from data collection in the ield to policy support. It is also a static historical case and is therefore relatively easy to dissect and explain compared to ongoing cases. he deinitive scholarly work on assessment of the Iraq and Afghanistan campaigns has yet to be written. And due to classiication and the ongoing nature of operations, neither of these contemporary cases can provide the kind of detail and perspective ofered by the Vietnam case. Nevertheless, Iraq seems particularly suited to examination because the war there is nearing termination, with the U.S. withdrawal in progress as of this writing. Unfortunately, the tendency to classify Iraq assessment reporting precludes a useful examination of Iraq assessment here, even as an ongoing case study; this chapter and the next refer to some of the available unclassiied Iraq publications. he Afghanistan case is far more accessible to researchers due to eforts by senior military oicers to keep at least some assessment material unclassiied. Because the war there is ongoing, examination of the Afghanistan case can contribute to improvements in current U.S. and NATO assessment eforts. Indeed, this was the original intent of this research efort. hus, this chapter focuses on how ISAF and its subordinate stafs are conducting assessment in Afghanistan. Assessment of the Afghanistan campaign did not begin in earnest until 2003. Still, there was no comprehensive efort by the military to gather information and assess the war until at least 2008, so this chapter does not address the assessment pro- 153 154 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency cess prior to that period.1 From late 2008 through 2011, commanders, policymakers, and assessment stafs looked at assessment processes and methods with greater interest. hey had very little doctrine to fall back on, and they had only the Vietnam and Iraq cases to guide them through an “industrial-strength insurgency.” By default, they tended to borrow from Iraq because many had experience in that conlict; these processes were, in turn, duplicative of many Vietnam-era eforts. But commanders and staf oicers also did exactly what the Vietnam-era OSA did in the mid-1960s: hey made due with whatever tools they had on hand. hey borrowed from a range of disciplines and tried to come up with assessments that made sense through trial and error. hey fell back on heavily quantitative processes at irst and then struggled to igure out how to capture contextual narrative and popular sentiment. he number of assessments grew uncontrollably as various oices and commands within NATO, ISAF, the U.S. military, U.S. civilian agencies, and policy circles struggled to understand the war. At one point in 2010, all the interested parties were collectively producing nearly 50 assessments of all kinds, delivered at various intervals by stafs from the battalion to the NATO headquarters level.2 he number of assessments produced relects bureaucratic momentum to some degree, but it also relects a general dissatisfaction with assessment reporting. As of early 2011, eforts to improve and reine the holistic campaign assessments of Afghanistan were ongoing. his chapter irst describes the assessment process as it developed between late 2008 and 2010 and then provides an original analysis of this process, supplemented by the perspectives of two independent researchers. he assessment process in Afghanistan is in a near-continuous state of lux: he descriptions in this chapter are current only through late 2010. By the time this monograph is published, an entirely diferent process may be in place. However, an examination of the late 2008–2010 assessment process is useful not only to support continuing improvements in Afghanistan assessment but also to contribute to future doctrine and practices. he theater-level Afghanistan assessment process is derived from doctrinal EBA, but it does not strictly follow EBA guidelines that have proven diicult to implement. Instead, it borrows from EBA, pattern and trend analysis, and other scientiic and statistical methods to respond to centralized requests for information. he formal assessment process does not, in and of itself, provide an overarching holistic understanding of the campaign; instead, these insights are contributed by senior commanders in narrative and subjective reports. he lack of a comprehensive methodology, the inability to account for inconsistencies in centralized COIN assessment, and the onerous demands 1 During this period, OSD continued to try to assess the war through various oices and programs. Most of the brieings from these oices are restricted and cannot be included in this discussion. 2 his number was identiied through a collaborative efort by military oicers and researchers working on a NATO project in support of ISAF assessments from early 2010 to early 2011. Assessment in Afghanistan 155 of a coalition bureaucracy have resulted in an assessment process that arguably does not meet most of the standards for quality assessment as deined in this monograph. Overview of the Assessment Process in Afghanistan as of Late 2010 Between early 2009 and early 2011, ISAF and IJC attempted to implement various iterations of centralized assessment. GEN Stanley McChrystal, ISAF commander in 2009 and early 2010, stated that assessments should be both transparent and credible.3 he irst efort to create a transparent and credible system was initiated in 2009 under the newly formed ISAF AAG. hese assessment documents described an intricately detailed and highly quantitative efects-based process that would rely on the collection and analysis of approximately 100 discrete ield indicators. One of the most senior and experienced operations research oicers in the U.S. Army led this efort. he ISAF commander rejected the standard operating procedure and its requirements on the basis that it was overly complex and opaque.4 Subsequently, in late 2009 and early 2010, the IJC developed and implemented the DAM. he DAM, discussed later, is an efects-based process that incorporates the commander’s qualitative input but relies heavily on a list of data-driven indicators to provide what FM 5-0 refers to as quantitative objectivity.5 District assessments are translated into a color-coded map product similar to the one maintained by the Revolutionary Development and Support Directorate in Saigon, South Vietnam, as discussed in Chapter Six.6 Both the AAG and the IJC Information Dominance Center have experimented with several versions of narrative assessment reports, as well as some 3 he director of the AAG stated, “Credibility and transparency are the watchwords for our process. COMISAF [commander ISAF] and COMIJC [commander IJC] often ask, ‘Is this credible?’” (AAG director, interview with the author, Kabul, Afghanistan, May 2010). 4 Assessment analyst, interview with the author, Afghanistan, May 6, 2010. ISAF currently produces quarterly and monthly reports derived from IJC reports and reports from other functional commands in theater. ISAF reports include a monthly assessment report that is submitted to NATO Joint Forces Command Brunssum, as well as a quarterly report, a monthly report on Afghan National Security Forces (ANSF), and documentation to support semiannual 1230 and 1231 reports. (hese reports are required to comply with sections 1230 and 1231, respectively, of the National Defense Authorization Act for each iscal year.) 5 See HQDA, 2010. I spent a week with the IJC Assessment Section in early May 2010 and continue to review IJC assessment publications. Contrasting this inding, a senior U.S. military oicer stated at a recent conference on assessment, “If the commander disagrees with the way the data is being interpreted at higher headquarters, his judgment is always given precedence” (statement by a senior U.S. military oicer, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010). 6 Cooper et al., 1972b, p. 210; former U.S. Army psychological operations advisor to Phong Dinh Province, Vietnam, interview with the author, Washington, D.C., April 19, 2010; Sharp and Westmoreland, 1968; Donovan, 1985. 156 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency highly quantitative eforts to assess ANSF and transition.7 An example of these eforts is the Transfer of Lead Security Responsibility Efect Scoring Model, which relies on a system of scores and weights to build a quantitative transition map.8 In addition to IJC and AAG, a number of other groups produce formal assessments of the campaign in Afghanistan, including the NATO Joint Forces Command Brunssum; the NATO Consultation, Command, and Control Agency (NC3A); U.S. Central Command Afghanistan-Pakistan Center of Excellence; several U.S. and foreign intelligence agencies; and various policy groups within DoD and the U.S. executive branch. As of September 2010, it was unclear to a group of 150 experts, including many NATO and ISAF oicials, exactly how many or even what types of assessment existed.9 Few of these reports are thoroughly coordinated outside their agencies, and their comparative value is minimized by broad inconsistencies in methodology, format, data content, and varying requirements for security access. For example, while the IJC DAM relects some qualitative input from subordinate commanders, the NC3A assessment produced in 2010 was built from an almost purely quantitative model that used automated color-coding. he semiannual DoD reports on Afghanistan—the de facto U.S. government assessment of the COIN campaign—difer to some degree in terms of method and means of coordination from report to report.10 Few of these reports relect a methodical or comprehensive process that might tie requirements to objectives, or collection to requirements. here has been little to no efort to untangle the speciied and implied data collection and production requirements that these reports place on subordinate stafs. Aside from a few narrative products from AAG and some of the intelligence analysis reports, theater-level assessment reports on Afghanistan are efects-based. All refer in some way to efects and indicators, and most require subordinate commanders to provide a range of quantitative data to senior stafs. his approach relects U.S. Army and joint Army–Marine Corps assessment doctrine as spelled out in FM 5-0 and FM 3-24.11 Assessments conducted by both ISAF and IJC relect the following dynamics:12 7 Commanders have signiicant input into the principal tool to rate ANSF efectiveness, the Commander’s Unit Assessment Tool (CUAT) for the Afghan National Army and the Afghan National Police CUAT (P-CUAT). 8 ISAF AAG, undated. 9 To address this problem and improve coordination in the assessment process, ISAF and IJC established the Assessment Synchronization Board. 10 his discussion of the 1230-series reports is based on extensive conversations with assessment personnel in both Afghanistan and the United States who are familiar with the development of these reports. One person writes the 1230 and 1231 reports. 11 12 HQDA, 2010, and HQDA, 2006c, respectively. hese observations are based on ongoing interviews and interactions with ISAF and IJC assessment staf members, ten months of reviewing assessment products and methodologies, and a dissection of the doctrinal and professional literature used to develop current models. Assessment in Afghanistan 157 • • • • COIN assessment is efects-based and is used to gauge efects-based operations. Data low up to feed a centralized assessment process.13 Commanders’ informed judgment is valued but remains almost wholly subjective. Quantitative data are inherently objective and useful in evaluating subjective judgment. • Qualitative data (e.g., engagement reports) are subjective and not suitable for systematized, centralized analysis. Regional command (RC)–level reports (those conducted just below the theater level but above the brigade or regimental level) are designed and written in accordance with the RC commander’s preference, staf capabilities, and ISAF and IJC guidance. While all RCs must ile centralized reports to IJC, as of late 2010, each RC used diferent methods, reported on diferent timelines for some products, and relied on diferent data sources. IJC had intentionally allowed each RC to develop its own methodology and reporting process to give commanders leeway in assessments, an approach recommended in FM 5-0. Table 7.1 compares how each RC (e.g., Capital, East, North) conducts its assessments. It shows the commands’ reporting cycles, primary sources (“input elements”), and level at which they focus their assessments (“geographic level”). Each RC uses diferent sources, reports on diferent frequencies and in diferent formats, has varying focus within its respective region, and has varying assessment capabilities. hese RC reports are processed at the IJC Information Dominance Center and are used to feed various assessment reports, including the district assessment report. hese reports are also captured by ISAF AAG, where they are compared to reams of quantitative data, including SIGACTs, economic data, reporting on the development of ANSF capabilities, and polling data. AAG develops a range of reports for ISAF, NATO, and U.S. customers at the military theater, combatant command, and alliance (NATO) levels and for policymakers at the National Security Council and in Congress. AAG analysts spend a great deal of time managing data, rectifying errors in databases, and producing time-series charts (sometimes annotated, sometimes not) to answer one-of requests for information from senior oicers, congressional stafs, and policymakers. As of early 2011, AAG staf, including LTC Bret Van Poppel, Ryan McAlinden, Capt. Eric S. Gons, and Marcus Gaul, were attempting to tackle the challenges associated with centralized assessment, data aggregation, and the lack of holistic analysis methodology. hese eforts may have signiicant impact on ISAF’s approach to assessment in the coming year, but the work was not yet available as of this writing. 13 ISAF and IJC intent, as published in the various brieings cited throughout this monograph, is to build assessments from the bottom up and not to conduct centralized analysis. his does not appear to be current practice. 158 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Table 7.1 Regional Command Assessment Processes as of Late 2010 Regional Command Frequency and Format Input Elements Remarks Geographic Level Capital No assessment reported for RC-Capital. IJC receives threat assessment directly from the district. Also contributes to IJC CUAT. East Quarterly brief Perception surveys, line-of-operations ratings, commander comments Robust assessment; recent District assessment focused on perception of the population based on survey responses; also contributes to ICJ CUAT North Weekly slides SIGACTs Qualitative determination of the security situation; improving, stable, or deteriorating; also contributes to IJC CUAT South Quarterly brief Qualitative indicators, Robust assessment; perception surveys assessment can be projected along effects or by geographical level; also contributes to IJC CUAT District West Weekly slides Staff and battlespace District-level assessment owner assessment utilizes IJC definitions; also contributes to IJC CUAT District Southwest Quarterly brief Perception surveys, line-of-operations ratings, commander comments Province Robust assessment; recent District assessment focused on perception of the population based on survey responses; also contributes to IJC CUAT SOURCE: Sok, 2010, p. 5. Data Management Challenges: The Case of Afghanistan We’re running the world’s greatest OPSEC [operational security] campaign in the world—against ourselves. —Senior ISAF information analyst, 201014 Eforts at the IJC Information Dominance Center and ISAF AAG to build a transparent, credible, and relevant assessment have been hampered by the lack of accurate and complete data. he spread of assessment requirements from the United States, NATO, ISAF, and IJC have created high demand for both raw and aggregated data from subordinate military units and civilian reconstruction units; there may be a requirement for 14 Cited in a brieing delivered by Jonathan Schroden at the U.S. Naval War College on October 19, 2010c. he quote is derived from an interview that Schroden conducted in February 2010. Assessment in Afghanistan 159 RCs to address as many as 2,000 metrics to feed these demands.15 Data for these assessments are submitted through many diferent channels, and, because some NATO stafs (including some U.S. stafs) report directly to their national commands, data pipelines are often isolated.16 Coalition civil-military operations also incorporate non-DoD civilians who report through nonmilitary channels. his stovepiped relationship prevents military commanders and assessment stafs from directly tasking some groups, such as coalition partner-led PRTs, with assessment information requirements. For some coalition members, these data management rules are cemented in diplomatic memorandums at the ministerial level. he PRTs sometimes control battlespace in Afghanistan, particularly in more stable areas. his means that there are areas under ISAF theater control that are essentially black holes for many types of assessment data.17 Until recently, there were 39 (known) separate databases in use just to store information on attack data, or SIGACTs.18 As of April 2011, no oicials interviewed for this study had a clear grasp of all existing data sources, let alone a method for metadata tagging, coding, and sharing of those data, but there has been a concerted efort to consolidate all information into a single system. Graphic data reports, typically in time-series format, are developed at multiple levels of command and are published and distributed by oicial email outside the theater, separately from formal assessments. Often, senior staf oicers and policymakers use these time-series graphics as standalone campaign assessment devices. Much of the data—quantitative and qualitative—required for comprehensive central analysis cannot be collected for a variety of reasons. In some cases, the data exist but never make it to the combined (U.S. and NATO) theater headquarters data networks. Problems with collection go well beyond any backlash against excessive data demands. Subordinate units may attempt to collect against centralized requirements but ind data to be unavailable, intermittently available, or, when available, of poor quality. Collecting data in a chaotic environment shaped by high threat levels, broken lines of communication, and mountainous geography is necessarily dificult and often unfruitful. As a result, there are entire sets of data that cannot reasonably be collected, and commonly referenced data sets can be chronically incomplete and are probably not representative of reality. 15 Midgrade U.S. military oicer, email correspondence with the author, June 27, 2011. 16 An August 2010 analysis of the data reporting processes in Afghanistan conducted by the National Defense University produced an unpublished data-sharing map that resembled the plumbing schematic on an aircraft carrier. 17 Statement by a subject-matter expert on Afghanistan assessments and SIGACT reporting, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. 18 Information analyst, interview with the author, Afghanistan, May 6, 2010. 160 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Despite leaps forward in information technology since the Vietnam War, there has been little progress toward creating an efective data collection, management, and sharing system for U.S. COIN operations. Afghanistan ofers salient examples. Data collection and management problems in Afghanistan are persistent, and corrections to known capability gaps have been slow and are often marginally efective. Ongoing concerns were relected in mid-2010 discussions of theater data management. For example, a brieing by the Center for Army Analysis concluded, “OEF [Operation Enduring Freedom] assessments . . . within the DoD and in theater have been hindered by a lack of necessary data: [here has been] no standardization over time for many data types [and there is] little to no readily usable historical data available.”19 he proceedings from a Military Operations Research Society workshop in April 2010 identiied the following problems with data knowledge and management in Afghanistan: • Mismatch between data required by analysts and data collected by operators • Disconnect of data needs among operators—analysts—and commanders • Lack of standard data architecture, resulting in information excesses and gaps • Lack of centralized repository or management approach, leaving existing data inaccessible or unknown • Lack of adequate collection program standards to address irregularities, inconsistencies, and incompleteness of data Additionally, data mining is inhibited by: • Not knowing what databases are out there • Lack of standardized database formats • Multiple levels of security among databases • Lack of access to networks housing databases.20 Communication infrastructure in Afghanistan is immature. According to a 2010 ISAF Headquarters brieing, “he main communication challenge is the IT [information technology] Infrastructure in Afghanistan. Currently, the communications system in Afghanistan works via satellite (V-SAT) which is not powerful and not fast enough to fulill all the communication needs of the country and also it is very expensive.”21 he ISAF Knowledge Management Oice has also stated that “Afghanistan and Coalition forces cannot communicate efectively across Afghanistan.”22 19 Center for Army Analysis, 2010. 20 Baranick et al., 2010, p. 5. 21 ISAF Headquarters, 2010. 22 ISAF Knowledge Management Oice, 2010. Assessment in Afghanistan 161 Data that are centrally assessed at the IJC or ISAF level are often inaccurate or incomplete due, at least in part, to database and data sharing issues. According to interviews with key assessment personnel, combat-experienced oicers, and others with inside knowledge of the Afghanistan assessment processes, the SIGACT databases, including those found in the centralized Combined Information Data Network Exchange (CIDNE) database, are severely lawed, as discussed later in this chapter. Despite ten years of efort and considerable expense, the U.S. military and ISAF cannot efectively capture, manage, and share the data necessary to conduct efective centralized assessment. Both operations and intelligence stafs place considerable data requirements on subordinate stafs, and these stafs generate thousands of reports each week; not all of these data feed assessment reports. hus, it is worth considering whether additional requirements to collect, store, and share data for assessment are worth the output that centralized assessment delivers. Problems with the Signiicant Activities Data Sets SIGACTs are reported violent incidents ranging from threatening letters to key leaders to major assaults on coalition outposts. SIGACTs are captured at various levels of command in both Iraq and Afghanistan and then consolidated in a central database (CIDNE). Because SIGACTs are commonly used as key indicators of progress in Afghanistan, it is worth noting some expert opinions on the value of these data. Most of the conclusions regarding the Afghanistan SIGACT databases also apply to Iraq databases.23 he following insights are drawn from interviews with subject-matter experts, comments from workshop participants, and oicial documents. Some of these points address database problems and not laws in the data, but they are listed here to provide a holistic view of the issues pertaining to SIGACT data in Afghanistan (as well as Iraq). • An expert on Afghanistan assessments and, speciically, SIGACT data, has stated, “SIGACTs are never accurate. he database isn’t even complete. If you think that every incident that happens is entered into CIDNE you’re kidding yourself.” • Furthermore, according to this analyst, “A lot of the ields in the SIGACT reports are listed as ‘unknown,’ so each report has varying levels of information.”24 • A senior U.S. military oicer with intimate knowledge of Afghanistan databases stated, “I’d say 10 to 20 percent of attack reports in CIDNE had misinterpretations of category. When you go through and read the narrative of the attack, 23 his is not to imply that Iraq SIGACTs were suiciently accurate to support accurate or meaningful centralized assessment. Just as there is no way to prove or disprove the accuracy of Afghanistan SIGACTs, there is no way to prove or disprove the accuracy of Iraq SIGACTs. 24 hese statements were made at the Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. 162 • • • • • • Embracing the Fog of War: Assessment and Metrics in Counterinsurgency you ind that it’s very clear the wrong codes were plugged in to label the type of attack.” his same oicer noted, “Sometimes, reports are changed or updated eight or ten times after they are initially recorded.” he same oicer also noted, “here’s stuf that’s in the civilian casualty database [a subset of SIGACTs] that’s not in CIDNE. he data that is in CIDNE is in text format in very large ields and is very hard to analyze.”25 A U.S. government oicial stated, “When [a DoS security contractor] is attacked, they don’t report through the military; they report through the embassy. And when NGOs get attacked, they often don’t report to anyone, but they keep their own databases.”26 In other words, there are entire sets of attack data that are never recorded in SIGACTs. A government-sponsored academic review of the assessment process in Afghanistan concluded, “As currently available in CONUS, CIDNE data provide limited support for empirical analyses owing to geographical or temporal gaps in coverage, and missing, incomplete and erroneous data. SIGACTs are limited in support of COIN-type security assessments; for example, those bearing on Afghan population security conditions outside areas of NATO force presence, non-kinetic threats, or attacks that do not involve forces with access to CIDNE reporting.”27 A U.S. military oicer with two combat tours in Afghanistan stated, “We were pretty good about reporting SIGACTs. What bothered me was what happened to the reports after we submitted them. We would submit long narrative reports with lots of detail, and the CIDNE report would have almost nothing. Some clerk along the way must have cut out information he thought was unimportant.”28 Finally, according to a senior information manager in Afghanistan, “SIGACTs data reports above the ISAF level are of by about 30 percent.”29 hese anecdotal comments are not suicient to call into question the validity of the entire SIGACT data set. Are SIGACTs generally accurate, and, if so, how can that accuracy be proven? Disproving the accuracy of a database that may contain more than 100,000 individual reports collected from thousands of sources over a ten-year period 25 Statements by a senior U.S. military oicer, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. 26 Statement by a U.S. government oicial, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. 27 Cabayan, 2010b. 28 Statement by a U.S. military oicer, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 2, 2010. 29 Senior information manager, interview with the author, Kabul, Afghanistan, May 2010. Assessment in Afghanistan 163 presents many challenges.30 Empirically disproving accuracy would require a separate and very costly study that might ultimately prove to be inconclusive. But if it is diicult to conclusively disprove the accuracy of the SIGACT data, it is also diicult to prove their accuracy. To enable such an analysis, it seems that the following would have to be made widely available: • a single, complete database of incidents that could be examined for accuracy • proof that consistent and accurate retroactive changes were made not only to baseline data but also to correct for lagging errors and changes in metric deinitions (e.g., IED inds counted and then not counted as attacks) • systematic means of backtracking a methodologically sound, representative sample of incidents to the originators of the reports • with a system in place to check ield reporting, a methodologically sound, peerreviewed study showing the degree of accuracy of information on incidents in the centralized database. If any part of NATO has made a systematic efort to validate SIGACT data by conducting ield interviews, the report from this efort has not been suiciently publicized. his process would require a team of trained ield researchers to select an appropriate sample size from the CIDNE SIGACT database, identify the personnel involved in the incidents in question, locate these individuals, and then interview them to determine whether their initial tactical reports had been accurately relected through multiple layers of re-reporting and aggregation. his would be a timeconsuming and very costly efort. Just inding the persons involved in each incident would prove to be nearly impossible due to rotations, retirements, and outprocessing from the services. Since SIGACT data also include information from Afghan security forces, this research would have to study the accuracy of Afghan reporting as well: he team would require translators and interpreters to study these reports and conduct ield interviews, assuming that the Afghans in question could be identiied. here seems to be no reasonable way to prove the accuracy of the entire SIGACT database. his is problematic: SIGACT data not only are used in stand-alone assessment graphs, they are also igure prominently in holistic theater assessments, operational analyses, and research reports.31 here are many other problems with the SIGACT data, and there are some conceptual problems with the approach to tracking violent incidents in COIN that have 30 31 his is an estimate; the actual number of reports is not speciied in publicly available sources. See, e.g., Howard, 2007, or Ramjeet, 2008. Both sources present examples of detailed operations analysis research using SIGACT data. hese kinds of analyses are common and often identify helpful patterns for tactical and operational analysis. here is no question that some of these reports have tactical and operational utility as long as the data are not taken at face value. However, they have far less utility when aggregated for theater campaign assessment. 164 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency yet to be adequately addressed in practice. Some of these problems were identiied in Chapter Five, but other issues have plagued the validity and utility of SIGACT databases. For example, there is no sound method for diferentiating between insurgencyrelated violence and other acts, like common criminal activity or the kind of intratribal, intertribal, and interethnic violence that is common in some areas of Afghanistan even during times of relative peace.32 his means that there is also no accurate method for tracking trend shifts in non–insurgency-related violence, so it is not possible to clearly state that insurgency-related violence is rising or falling across the entire country at any one time. It would be possible to argue that this diferentiation is unnecessary if the objective is to simply reduce overall violence, but SIGACT data are used to help stafs determine which type of forces to deploy in various areas (e.g., the army or police) and also to help determine whether speciic areas are ready for a transition from military and paramilitary forces to normal police forces. hey are also used to justify the continued need for U.S. military forces—forces that might not be needed to deal with a surge in criminal or intratribal violence. With few exceptions, SIGACT data only relect oicial reporting from U.S., coalition, or host-nation sources.33 he government oicial quoted earlier stated that there are entire categories of reports from coalition oicials (e.g., embassy employees, NGO staf) that are not counted in SIGACTs. here are other insurgent-related violent acts that that occur “of the grid,” or out of sight of oicials, and thus are not counted. If an insurgent attempts to intimidate a government oicial, that intimidation might be reported as an incident through oicial channels and eventually incorporated into SIGACTs and CIDNE. However, if insurgents intimidate local villagers in a remote area devoid of military or government oicials, that incident most likely will never be reported. For the assessment staf, this unobserved incident—and the thousands like it—simply does not exist; these incidents are not relected in time-series SIGACT reports. he same is true for IEDs that have been planted by insurgents but never discovered or detonated, direct-ire rounds shot at friendly vehicles that miss and are not recorded, insurgent hijackings of civilian vehicles at unobserved checkpoints, deliveries of night letters that go unreported, or attacks in which IEDs, rockets, mortars, riles, machine guns, or other weapons misire and the attack is canceled, to name a few examples. While these incidents are not reported, they do represent insurgent 32 here may be ields in some versions of SIGACT report templates that allow the reporter to make these kinds of diferentiations. However, very often, the reporter does not know the reason behind the incident. In both Afghanistan and Iraq, it is often impossible to diferentiate between insurgents, criminals, and tribal members, because it is possible to be an insurgent criminal tribal member. Divining motivations for individual attacks would be diicult even with near-perfect intelligence. Further, data ields are often inaccurate or incomplete. Separating insurgent activity from criminal or intertribal activity is all but impossible through aggregated data analysis. 33 Exceptions include eforts by data management teams to capture incidents from press reports that might not have been coded through the oicial SIGACT reporting process. However, press coverage may be even more scattered and inconsistent than military and government coverage. Assessment in Afghanistan 165 capability and intent, and in many cases, they afect the population and conditions on the ground. hus, SIGACT reporting is not synchronized with reality. here is no method for determining how many insurgency-related incidents go unreported across the country on a daily basis, so the signiicance of this additional gap between reporting and reality is unknown. his gap undermines the validity of theater assessment that depends on timeseries SIGACT data. Problems caused by this gap can be exacerbated when friendly forces move into previously uncovered areas and begin reporting a theretoforeunreported stream of insurgent-related incidents. Trend lines can jump, making it appear that there is a greater insurgent presence in a particular district or province when, in fact, coalition forces merely revealed the presence of insurgents who had been operating unnoticed. hese kinds of jumps can also occur when military or other government forces instigate attacks by moving into an area where insurgents, criminals, or simply hostile tribal members were living or operating peacefully. And the number of actual events can drop in one area when government forces move in, only to increase in another area that is not under observation as insurgents set up shop away from prying eyes. As Kilcullen points out, “Violence tends to be high in contested areas and low in government-controlled areas. But it is also low in enemy-controlled areas, so that a low level of violence indicates someone is fully in control of a district, but does not tell us who.”34 he Vietnam case showed that insurgents often purposefully choose to refrain from violence in an efort to avoid detection or maintain freedom of movement, so the absence of violence can be and probably often is characteristic of a location where insurgents have a signiicant presence. In general, the presence or absence of violence, by itself, is not always meaningful for assessment. Comparing SIGACT data to other variables is no less problematic than attempting to analyze the data on their own merits. A number of assessment staf oicers and scholars have applied a variety of statistical methods to Afghanistan SIGACT data in an efort to ind cause and efect between two variables. But at some point, statistical analysis tends to require (1) some quantity of known, accurate data, (2) a data set with generally known inaccuracies, or (3) strong evidence of correlation between two variables across a number of locations (e.g., violence always drops when soccer balls are distributed). However, this monograph argues that none of these requirements can be met through centralized assessment. It shows that analyses can be precise but inaccurate and that, in general, the completeness and quality of the data do not correlate with meaningful results. Any large-scale SIGACT study that claims to ind a causal link between SIGACT data and any other variable should be carefully scrutinized. As long as accuracy is not assumed, violent incident reporting may have value in context when local commanders can compare speciic incidents to intelligence reporting and to their own (known) operations. hey may still not be able to explicitly con34 Kilcullen, 2009b, p. 5 (emphasis added). 166 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency clude cause and efect between their operations and a rise or fall in violence, but local commanders are best placed to interpret these incidents in a way that is meaningful for assessment. For example, Figure 7.1 shows operations analysis of SIGACT data at the RC level conducted for a UK military unit. It shows the number and types of reported attacks over a given period and at speciic times of the day in two-hour increments. As long as this chart was used only to help analyze the time of day during which insurgents were more or less likely to attack coalition forces in RC South, then it could be useful. Any efort to extrapolate these data, or this chart, to show cause and efect for assessment would be inadvisable. heater-level aggregated SIGACT data should be used to assess the Afghanistan campaign only with the utmost caution, and then only when accompanied by stringent caveats. Figure 7.1 Time-of-Day Attack Pattern Analysis for Regional Command South, Afghanistan 30 IED Indirect fire Direct fire Other 25 20 15 10 5 0 2 4 6 8 10 12 14 16 18 20 22 24 Hour of day SOURCE: Howard, 2007, p. 9. RAND MG1086-7.1 Analysis of the Afghanistan Assessment Process he joint/combined force in Afghanistan has created a modiied EBA process that generally mirrors the one applied in Iraq and is clearly evocative of the model applied in Vietnam.35 As of early 2011, ISAF (which oicially relies on NATO’s Efects-Based 35 See hayer, 1985, 1975a, and 1975b. Assessment in Afghanistan 167 Approach to Operations doctrine) framed both its operations and assessments in terms of efects.36 here may be intent by senior leadership to emphasize commanders’ analysis, but in practice, the various assessments of Afghanistan are derived primarily from quantitative and centralized assessment of individual data streams. Many millions of dollars have been spent improving data management systems, and no fewer than ive interagency working groups are assigned to improving data collection and sharing in theater, across U.S. agencies, and among NATO partners. Eforts to improve assessment and further develop MOEs, MOPs, and indicators have incorporated work by mathematical modelers, social scientists, and EBA experts from the U.S. Army’s operations research community. Both the ISAF and IJC stafs have aggressively pursued mathematical modeling processes in an efort to improve assessment design. Far from placing the planning and assessment process in the hands of the commander, strategic review and transition planning relies on input from several agencies, hundreds of staf oicers, and hundreds of external experts.37 he Afghanistan assessment model that Gregor accurately condemns as “inconsistent” is the U.S. military’s and NATO’s best efort at EBA. Research conducted by the U.S. Naval War College and the Netherlands Defence Academy reinforces the notion that the modiied version of EBA in use in Afghanistan has not met military or policymaker requirements for COIN. his research speciically addresses the failure of assessment to provide transparency and credibility. According to Stephen Downes-Martin of the U.S. Naval War College, “While no assessments process can be perfect or free of any criticism, the laws in the currently used approaches are suiciently egregious that professional military judgment on assessments is, rightfully, distrusted.”38 He inds that, “in the absence of a credible numbers-based theory for COIN, there is no objective numbers-based assessment. At best one obtains precision without accuracy.”39 Downes-Martin also recorded complaints from the ield: here is serious evidence that the collection capacity of regional commands’ partner civilian organizations and major subordinate commands are overwhelmed by the number of metrics demanded, and that they rightfully do not trust the value of collecting on those metrics or the value of the assessments done using them. his evidence is in the form of verbal admissions from senior oicers that they make 36 To emphasize this point, most stafs at the RC level and below have “efects cells” designed to shape efectsbased operational eforts. 37 For example, from August 30 to September 3, 2010, at NATO Joint Forces Command, Brunssum, Netherlands, NATO held a conference of 150 oicers and experts to examine data collection, data sharing, and metrics in Afghanistan. 38 Downes-Martin, 2010b, preface. Downes-Martin was embedded with an operational assessment section both during its predeployment training and then for three months in Helmand Province in 2010. He has irst-hand knowledge of the assessment process at both the RC and theater levels. 39 Downes-Martin, 2010b, p. 7 (emphasis in original). 168 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency up what they do not have, and do not check the quality of what they do have, in their submissions. An additional reason given for not taking the metrics seriously was that they received no feedback from the requesting organization.40 Downes-Martin’s observation that centralized quantitative assessment requirements generated falsiied reports closely parallels the evidence of falsiication described in the Vietnam case. Interviews and discussions conducted for the current study between November 2009 and February 2011 uncovered similar admissions from some oicers who had served in Afghanistan. An operations researcher who had worked on a PRT in Afghanistan stated, “When we were asked to produce data we didn’t have we just made it up.”41 I personally observed similar actions during two of three tours in Iraq.42 Direct observation of the assessment process supports the conclusion that military oicers often attempt to push back against requirements that they cannot logically or practicably meet, but some do make up data to avoid irreconcilable friction and to allow them to focus on what they see as more important tasks. In Afghanistan, this behavior—seemingly common, judging by accounts from assessment staf, and observations by other researchers—is a direct result of what Downes-Martin refers to as “promiscuous metrics collection” driven by a core metrics list that cannot be relevant or collectible in all places across the country at the same time.43 Joseph Soeters of the Netherlands Defence Academy provides some insight into the EBA approaches attempted in Afghanistan in early 2009.44 Soeters found that “any system that tries to cover all information needs—both operational and strategic— 40 Downes-Martin, 2010b, p. 8 (emphasis added). 41 Statement by an operations research analyst, conference working group, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. Two military oicers who overheard this statement concurred and added that they, too, made up data when mandated to report on indicators that did not exist. 42 Commentary based on observations in Iraq should be taken as a professional, irst-person perspective on such incidents. hese observations do not qualify as empirical evidence of widespread falsiication. 43 Sometimes, these core metrics lists are included in CCIR lists and therefore address both assessment and intelligence requirements. Campbell, O’Hanlon, and Shapiro of the Brookings Institution have worked extensively with core metrics lists. hey state, If we were conident about which 10 or 15 or 20 metrics could best tell the story of the eforts in Iraq or Afghanistan, and had reliable data for those categories, we might have focused more narrowly on them. he truth is that the wars in both Afghanistan and Iraq have usually demonstrated an ability to confound short lists of metrics. However, they go on to state, “In retrospect, some key metrics emerge with greater clarity” (Campbell, O’Hanlon, and Shapiro, 2009a, p. 4). 44 In January 2009, Soeters led a team of researchers in Kandahar and Kabul, Afghanistan, on a study of the operational assessment process. he team conducted 63 semistructured interviews with the commander of RC South and the deputy chief of stability in Kabul, among others. hey also attended assessment staf meetings and reviewed assessment documents. Soeters speciically focused on the use of EBO and EBA in COIN assessment, as opposed to the application of civilian business models in military operations. Assessment in Afghanistan 169 becomes overly complex. . . . At the aggregate level of outcomes it is diicult, if not impossible, to attribute changes and results to certain inputs and activities.” Forming standards of achievement is diicult when “results have multiple or even contradictory readings [or] when the environment is highly changeable.”45 Soeters’s interviews revealed some of the underlying and rarely publicized frustrations with EBA from within the military. In addition, military oicers and other oicials have called the centralized EBA process “an illusion” that cannot react to a complex system and that assumes an unachievable level of predictability. One interviewee stated that EBA fails to deliver useful direction to subordinate stafs.46 Soeters found that assessment stafs had poor tools to work with, because the “measures of efect and performance are unclear, change frequently and are diicult to measure. . . . he deinitions these measurements should be based upon are often unclear, vague, and diferent.” Chapter Summary Campaign assessment in Afghanistan is derived to a great extent from U.S. joint doctrine on EBA. On the surface, it looks a great deal like EBA in that it uses indicators to assess efects and to measure progress toward MOEs and MOPs. However, assessment in Afghanistan has typically been ad hoc and loosely coordinated, incorporating both EBA and pattern and trend analysis. In many cases through late 2010, assessment reports consisted of little more than republished time-series charts or statistical data that were devoid of meaningful context or explanation. he best analysts spend most of their time managing the vast quantities of centralized data necessary to feed EBA and pattern and trend reporting. To date, there has been no comprehensive efort to capture context from the ield to shape time-series assessment reports, no common assessment methodology across the theater or over time, no method for systematically incorporating qualitative data into theater assessment, no common framework for pattern and trend analysis, and little or no comprehensive efort to incorporate caveats regarding data quality or the degree of subjectivity in assessment reporting. As of late 2010, the point at which the focused research on the internal assessment process on Afghanistan ended, assessment of the Afghanistan campaign did not fare well against the assessment standards identiied in this monograph. Table 7.2 applies the assessment standards described in Chapter One to the assessments produced by assessment stafs in Afghanistan, though not necessarily to the individual reports written by senior oicers. 45 Soeters, 2009, pp. 7–8. 46 Soeters, 2009, p. 11. 170 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Table 7.2 Performance of Afghanistan Assessment Against Assessment Standards Standard Afghanistan Assessment Analysis Transparent No Data aggregation and methods opaque, insufficient declassification of assessment reports Credible No No holistic methodology, multiple and often contradictory reports, large data gaps Relevant No While senior officer analysis is respected, assessments do not sufficiently support this analysis; this makes the process seem subjective Balanced No Does not systematically incorporate qualitative or mixed data or subordinate reports Analyzed No Analysis is not consistent across the theater, nor does it deliver a methodologically sound, overarching framework or report Congruent No Does not reflect COIN joint/capstone doctrine Parsimonious No Data requirements are extensive, poorly defined, and typically inconsistent from level to level CHAPTER EIGHT Why Does Centralized Assessment Fail in Counterinsurgency? Our whole notion [is] that we can somehow develop a mathematical model that includes concrete achievements, factor in a time frame and voilà. Iraq doesn’t work that way and Afghanistan doesn’t work that way. —Ryan C. Crocker, U.S. ambassador to Afghanistan and former ambassador to Iraq1 his chapter builds on the Vietnam case but speciically relects the indings on assessment in Afghanistan and (to a lesser extent) Iraq. Before answering the question posed in the title of this chapter (Why does centralized assessment fail in COIN?), it is worth reiterating a point made in the introduction to Chapter Seven: A poorly conceived process—not the individuals executing the process or receiving its output—is the primary reason for the inherent weaknesses of U.S. military COIN assessment. More than a year of research, several years of personal involvement with the COIN assessment process, and numerous interactions with military and civilian assessment stafs from the battalion to the joint staf level revealed teams of assessment analysts that were both understafed and overworked. Leaders and policymakers cannot adequately communicate requirements to collectors and analysts because COIN assessment is poorly deined and rarely addressed in the literature, professional education, or staf training. In many cases, operational assessment was a secondary or tertiary responsibility for oicers and noncommissioned oicers in combat theaters. here is little or no all-source analytic capability for campaign assessment at any level, and there is no doctrinal guidance on how one might conduct a comprehensive analysis of a COIN campaign. Doctrine and other available oicial guidance tend to be inherently contradictory, inadequate, and confusing. Almost all interactions with assessment stafs revealed hard-working people struggling to make the best of a bad situation. his observation of the current process seems to echo conditions during the Vietnam War. 1 As quoted in Shadid, 2010. 171 172 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Interviews with staf involved with assessment, subject-matter experts, and policymakers, along with a survey of the civilian and military literature, uncovered widespread dissatisfaction with the military assessment process as it is applied to Afghanistan. his dissatisfaction is evident from the company-grade oicer level to the highest levels of policy (e.g., congressional and national security staf). In 16 months of research, interviews, conferences, workshops, and engagement with most of the producers and some of the oicial consumers of Afghanistan assessments (in total, well over 300 oicers, civilian analysts, policymakers, and experts), no one was willing to characterize the current system as well-designed, transparent, or credible. his chapter presents detailed indings matched to discussions of the seven standards framed in the beginning of this report and used to assess the Vietnam and Afghanistan cases: transparency, credibility, relevance, balance, analysis, congruence, and parsimony. It shows that centralized assessment does not meet any of these criteria and that the doctrinal and de facto approaches to U.S. COIN assessment are inefective. his chapter also addresses some additional issues and provides examples of data challenges from the Iraq case. Neither Transparent nor Credible his section examines how EBA doctrine, pattern and trend analysis, and the modiied versions of these approaches in Afghanistan have failed to deliver transparent or credible assessments to senior military oicers and policymakers. he concepts of transparency and credibility are closely linked, and there is some necessary overlap between the two. Transparency is a broad concept, but it is a helpful standard for reports that will be used to shape operations and support public review of the overall regional strategy of the United States. If democracies are to prosecute COIN, they must be able to openly justify to their populations the inherent risks and extended timelines associated with the typical COIN campaign.2 Transparency allows policy stafs to dig into speciic data to address any pressing concerns they might face, but without forcing an assessment staf to scramble to produce new reports. Transparent assessments help policymakers achieve depth of understanding when attempting to grasp the complexities of COIN. A classiied but transparent assessment can also be fairly compared to an intelligence analysis product (in terms of sourcing), so transparency gives policy stafs a means to develop contrasting viewpoints to support decisionmaking. After some necessary declassiication steps, including the removal of sensitive material, a transparent assessment can be used to help interested members of the public understand the complexities and challenges of the campaign. In this ultimate format, a transparent 2 See the summary of indings in Connable and Libicki, 2010, for insight into COIN campaign dynamics. Why Does Centralized Assessment Fail in Counterinsurgency? 173 campaign assessment can prevent creeping disillusionment with government policy, combat the impression that military assessment is overly optimistic, and reduce the impression that reports relect unanchored subjective interpretations rather than indepth reporting and analysis.3 Ultimately, transparent assessment reports are widely releasable to oicial consumers and, ideally, unclassiied and without caveat; the inal iteration should be suitable for distribution on the Internet. Such reports reveal both the methods and data used at each level of assessment, from tactical to strategic, and allow for detailed and contextual analysis. Any and all subjectivity or justiications for disagreements between layers of command should be comprehensive, available, and understandable. It is diicult to see how a centralized process can deliver transparency for COIN assessment. Centralized assessment is almost entirely reliant on quantitative data aggregation to produce assessment reports. At the highest levels of aggregation, these data are opaque to government and civilian consumers. hey provide little opportunity for skeptical consumers to “see” into the assessment report to determine data or methodological validity. It would be possible to imagine opaque aggregation and transparent disaggregation on opposite ends of a continuum. Figure 8.1 shows transparency and disaggregated context on one end of a continuum and aggregation and opacity on another. No matter how reports are presented or declassiied, data aggregation is likely to undermine eforts to present assessments transparently. Because aggregated data tend not to have been analyzed in context at the level at which they are collected, accurate retrospective analysis of these data may not even be possible: It is diicult to ind other sources of information that are suicient to place such data in context well after Figure 8.1 Relative Transparency of Aggregated and Disaggregated Data Opaque Transparent Disaggregated Aggregated RAND MG1086-8.1 3 his assumes, of course, that the transparent report meets other standards (e.g., it is credible, balanced, analyzed). 174 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency the situation on the ground has changed.4 Aggregation essentially precludes efective analysis by anyone outside the combat theater. his analytic roadblock has likely led the press and subject-matter experts to be more critical of COIN campaigns than may be warranted otherwise. Lack of transparency undermines credibility, but the credibility of assessment depends on a number of factors. Credible assessment reports must be transparent because opacity devalues any report to the level of opinion: If the data and methods used to produce the report cannot be openly examined and debated, consumers have no reason to trust the reporter. In a credible report, to the greatest extent possible, biases are revealed through iteration between commanders at various levels and toplevel analytic stafs. Flaws in data collection, valuation, and data analysis are clearly explained. Methods are consistent and clear. Data are explained in context. he process used by commanders and stafs to select methods is also clearly explained. he report is both accurate and precise to the greatest degree possible. It is unlikely that centralized assessment can meet these criteria. It is not possible to accurately or suiciently account for data laws and gaps because the degree to which aggregated data are inaccurate or incomplete cannot be known: COIN data are not collected using reliable scientiic methods. he lack of control and consistency across data collection and management eforts prevents an assessment from showing even a reasonable estimate of bias and laws in data collection. It is also not possible to determine how subordinate units valued any speciic type of data over another once those data have been aggregated, so data valuation is typically unknown and often inaccurate. It is possible to rectify the existing lack of overarching analytic methodology in the campaign assessment process. But, at best, structured analysis would provide dressing for an unappetizing array of dissociated and opaque sets of often inaccurate and gapped data. No core metrics list from any campaign has been widely accepted as credible and objective, so even expert analysis of core metrics is unlikely to produce a credible report. he following two sections address additional considerations regarding eforts to develop credible assessment reporting. Lack of Data Caveats and Citations Undermines Transparency and Credibility Assessment stafs must take and present the quantitative data necessary to determine thresholds at face value. As of late 2010, it was rare to see a conidence rating attached to a quantitative source in an assessment report. Perception polls almost always carry statistical conidence bounds provided by the contracted polling organization, while other sources, such as attack reports, some types of economic data, and civil-servant 4 his supposition is based on eforts to conduct such analysis both from within a military intelligence staf in a combat theater and as a researcher. Why Does Centralized Assessment Fail in Counterinsurgency? 175 employment rates, rarely do. his likely leads senior commanders and consumers to believe that assessment data possess a degree of accuracy that would be unachievable even in a peacetime environment.5 Referring to Afghanistan, Anthony Cordesman from the Center for Strategic and International Studies identiies this problem: When numbers are applied to reality, they are always uncertain adjectives—not scientiic quantiication. his is especially true in conlicts like Afghanistan where accurate data collection is often diicult to impossible. Metrics often imply a false precision, particularly since they are rarely shown as full ranges or rounded to relect uncertainty. he problem is made worse by the fact that many users do not really understand probabilities, statistics, or the overall process of quantiication half as well as they understand words. In general, numbers that are unsupported by narrative are a remarkably poor method of communication.6 Not only are quantitative sources ofered at face value, but the assessment reports themselves rarely contain suicient citation; this is particularly true for unclassiied reports released to the public. It is often impossible to ind the original source of reporting used to build theater-level assessments. Sometimes, sources are provided, but these sources are often second- or third-hand aggregations (e.g., for reasons of classiication, a theater-level assessment might cite national-level economic data or simply “CIDNE”). he absence of citation further degrades the transparency and credibility of the assessment process. Weighting Metrics Reduces Transparency and Credibility In Afghanistan, at least one of the analytic techniques that stafs apply to COIN data may do more harm than good. Centralized reports sometimes incorporate weighting of MOEs, MOPs, and indicators. FM 5-0, Appendix H, recommends a mathematical assessment process that includes both a weighting and a rating scheme.7 his research did not reveal any verbatim adoption of the doctrinal model (perhaps a compelling critique of the model’s relevance to COIN), but a number of other weighting schemes have emerged to support EBA at various levels of command in both Iraq and Afghanistan. Figure 8.2 shows another proposed EBA model that incorporates a weighting formula.8 While this model was not designed for COIN and may never have been imple5 And, when policymakers or their stafs discover errors or irregularities in the data, their conidence in DoD assessment reporting is further undermined. 6 Cordesman, 2008, p. 3. 7 See HQDA, 2010. 8 he authors describe the thinking behind the formula: he mathematical mechanics involve an iterative process that repeats the similar steps for each level in the model hierarchy. At the lowest levels, each efect has a number (x) of MOEs associated with it, and each task has a number (y) of MOPs associated with it. In addition, we assign each MOE and MOP a weight relecting relative 176 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 8.2 Effects-Based Mathematical Model with Weighting Scheme MP = w (1)OP(1) + w (2)OP(2) + ... + w iOP(i ) , 1 Mission ∑ wn = 1 where n=1 Equation 4: Mission-level performance scores OP(i ) = w (i ,1)EP(i ,1) + w (1,2)EP(i ,2) + ... + w (i , j)E(i , j) , f Objective where ∑ w(i ,n) = 1 n=1 Equation 5: Objective-level performance scores EP(i , j) = w (i , j,1)TP(i , j,1) + w (i , j,2)TP(i , j,2) + ... + w (i , j,k )TP(i , j,k ) , k where Effect ∑ w(i , j,n) = 1 n=1 Equation 6: Effect-level performance scores TP(i , j,k ) = w (i , j,k ) MOP(i , j,k ) + w (i , j,k )2 MOP(i , j,k )2 + ... + w (i , j,k )9 MOP(i , j,k )9 , 1 1 y Task where ∑ w(i , j,k ) n =1 n=1 Equation 7: Individual-task performance scores SOURCE: Clark and Cook, 2008, Figure 5. RAND MG1086-8.2 mented, it is emblematic of how diicult it can be to add weights to EBA or modiied EBA, and of how weighting can undermine transparency and credibility. his process does not pass the tests of transparency, credibility, or even understandability. Models that require speciic skills and inside knowledge to understand and interpret degrade the value of assessments. In this formula, there is no clear, logical relationship between indicators and weights. he weighted combinations of metrics have no physical or logical meaning. here is no way to incorporate uncoded qualitative data into the formula, so the tens of thousands of narrative intelligence information reports, key leader engagement reports, civil afairs narratives, and other qualitative reports produced in theater would have no value in the model’s output.9 his model may or may not be helpful to a staf that is implementing the strictest interpretation of importance. For each assessment period, we observe values associated with each MoE and MoP and input them into their respective models. [he irst igure, not shown here] outlines the efects-scoring model; [the igure shown here as Figure 8.2] outlines the performance-scoring model. he MOE and MOP scores, between 0 and 1, indicate the level of a particular efect or task, respectively. A score of 1 indicates complete success—at least temporarily. his holds true for all scores at each level. (Clark and Cook, 2008) 9 Bousquet (2009, p. 183) states, “Precise quantitative analysis must yield to a more qualitative understanding via the identiication of more or less hidden patterns in the [complex adaptive] system’s behavior.” Why Does Centralized Assessment Fail in Counterinsurgency? 177 EBA as presented in Chapter Four, but neither this model nor the model in FM 5-0, Appendix H, is relevant to contemporary COIN operations. Even if this quantitative model could be adequately explained and justiied, there is no way to avoid a high degree of subjectivity in this or any other weighting scheme (e.g., the one described in FM 5-0, Appendix H). At some point, someone has to make a decision about what is more important and what is less important to the course of the campaign.10 Typically, a senior policymaker or military commander would make this decision, but subordinate stafs are also free to use their own internal weighting schemes; weighting could be added and hidden by anyone at any stage of the assessment process, at least as it is currently conceived and implemented. Because weighting requires subjectivity that is diicult to explain or justify, it is vulnerable to criticism. And because every consumer of assessment reports brings biases and opinions, consumers also apply their subjective interpretations to the weights (i.e., they might disagree that one thing is more important than another). Policymakers and senior commanders might feel justiied in applying their own weights, thereby changing the entire basis of the assessment. his has happened: Staf oicers in Afghanistan report that centralized weighting schemes were attempted but failed after senior oicers could not agree how each efect should be weighted.11 For a weighting system to be accurate (at least by the standards of that speciic staf) the staf must adjust weighting values to meet changing requirements and conditions. Early in the war, security might be the most important metric, but, later in the war, development might be more important. As importance shifts, weighting schemes would have to relect that shift.12 he mathematical modeling required to account for routine weighting shifts in a number of interrelated metrics over time would be complex to the point that it would be relatively meaningless to policymakers or senior military leaders—interpreting the model would become a subjective exercise. And, because weighting is highly subjective, every shift in weighting would also require subjective judgment. It would quickly become impossible to unravel who decided what was most important at each moment in time and why. If the likely end result of mathematical weighting is a return to subjective debate, then no staf should exert the time and energy required to apply weights to assessments. 10 FM 5-0 (HQDA, 2010, p. B-35) recognizes that even in the course of action development process—a process that takes place before operations, typically in safe and static environments—weighting adds a great deal of subjectivity. 11 12 his statement is based on discussions with coalition staf members in Kabul, Afghanistan, May 2–8, 2010. I observed two major shifts in weighting at the theater level in Afghanistan over a six-month period. hese changes shifted weighting from one line of operation to another. Each shift should have created a large cascading efect throughout the models used to gauge subordinate MOEs and indicators. 178 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Not Relevant to Policy Decisionmaking Relevant reports should efectively and eiciently inform policy consideration of COIN. hey should be suicient to help senior military leaders and policymakers determine resourcing and strategy, and they should satisfy public demand for knowledge up to the point that they do not reveal sensitive information. he other six standards described here shape the degree to which assessment reports are or are not relevant. Chapter One presented a list of general requirements that policymakers might have for relevant COIN campaign assessment: • a close approximation of ground truth upon which decisions can be based • an appreciation for resource requirements (current and predictive) • transparent and credible reporting that can be used to inform Congress and the public • relevance to decisionmaking • predictive analysis (sometimes). Centralized assessment does not provide accuracy that can be substantiated through either an examination of the data or sound analysis, so policymakers can never be sure that it provides a close approximation of ground truth. Optimally, resource requirements are predicated on at least a relatively clear understanding of the threat, progress toward developing the host nation’s government and security forces, the status of the population, and other relevant information. If this information is inadequate for strategy decisionmaking, it is also likely to be inadequate for resource decisionmaking. As established earlier, centralized assessment is neither transparent nor credible, so even if assessment reports appear relevant to decisionmaking, they are not suicient to sustain a long-term COIN campaign that requires U.S. popular support. hus, centralized assessment can provide some predictive analysis, but because it is not predicated on sound data or methodology, this analysis is insuicient for sound decisionmaking. he near-universal dissatisfaction with the COIN assessment process in policy circles emphasizes these indings. Unbalanced Balanced reports relect information from all relevant sources available to military stafs and analysts, including both quantitative and qualitative data. Such reports should contain and relect input from military commanders at all levels of command. hey should also be broad enough in scope to incorporate nonmilitary information and open-source data. Finally, balanced reports include countervailing opinion and analysis, as well as data that both agree with and contradict overall indings. Why Does Centralized Assessment Fail in Counterinsurgency? 179 he mathematical models recommended by the Commander’s Handbook and FM 5-0, and employed in various forms in Afghanistan, rely on core lists of quantitative metrics.13 While FM 5-0 tries to place equal value on qualitative and quantitative data, it also states, “Quantitative indicators prove less biased than qualitative indicators. In general, numbers based on observations are impartial.”14 his belief has been relected at some theater-level stafs in Afghanistan at various times, but it is not correct in the context of loosely structured and nonscientiic COIN assessment. COIN data collection and reporting are not objective eforts. Quantitative data relect subjective bias that is hidden when aggregated. he processes of selecting MOEs and indicators and determining what to collect and report at the tactical level, the collection method, the reporting method, and the way in which data are displayed inherently involve subjectivity. Because stafs tend to collect a great deal of quantitative data but focus on a few key indicators, there is also subjectivity in the decision about what not to count or report. Quantitative and qualitative data are both subjective, but in diferent ways. Because COIN assessment is not controlled scientiic research, the standards associated with scientiic research will not be relected in COIN data.15 heater-level stafs continue to make modest eforts to incorporate qualitative information into assessments, but mathematical models, time-series charts, and core metrics lists require purely quantitative data. In practice, therefore, centralized assessment is driven by quantitative data.16 It is possible to code some qualitative data for quantitative analysis, but coding requires trained analysts, expensive software, and a great deal of time. It is not feasible for even theater-level stafs to code the tens of thousands of qualitative reports fed into centralized databases every year.17 According to an information analyst at a senior staf in Afghanistan, “We can’t use any data that is not in a cohesive or productive format. Everything needs to be similarly coded [i.e., quantiied] or it can’t be inducted into a database.” he analyst also stated that “the data must 13 See HQDA, 2010, Appendix H. A senior information manager and analyst at IJC stated that incidents are run through a mathematical model to “determine causation” (interview with the author, Kabul, Afghanistan, May 2010). 14 HQDA, 2010, p. 6-8. 15 In any event, some scientists might claim that qualitative and quantitative data could be either more or less subjective, depending on how they are produced and obtained. here are ways to collect and analyze qualitative data that are highly structured, and there are various methods to reduce subjectivity in qualitative data. he assumption that one type of data is more or less subjective might apply when comparing a physical experiment to a perception poll, but it does not necessarily apply in the study of a social problem. For this reason, the statement in FM 5-0 is misleading. 16 An examination of the mid-2010 version of the IJC District Assessment Framework Tool, which checks subordinate commanders’ assessments against an objectives metrics list, showed that the majority of the approximately 80 metrics required percentage ratings, binary responses (e.g., present or not present), or some other form of quantitative representation. 17 Even if these reports could be coded, the process of coding would inject an additional layer of subjectivity and a layer of opacity to the inished assessment report. 180 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency be pre-coded before it reaches the our level. Our analysts don’t necessarily have enough context to code raw reporting from the RC-level and below.”18 he RCs do not have the capacity to do this.19 herefore, in practice in Afghanistan, qualitative data are not systematically incorporated into theater-level assessments.20 his also proved true in Iraq.21 Centralized mathematical models, and all centralized assessments for that matter, relect only a subset—perhaps 40–60 percent—of all available information in theater.22 his means that centralized assessment provides a picture that, at best, relects about 60 percent of the events, insights, and relections that are readily available to analysts. Some aspects of COIN that are crucial to success simply cannot be usefully quantiied, even through coding. Unless one places absolute reliance on public opinion polling, it is not possible to accurately quantify insurgent motivations, civilian satisfaction with the government, or government legitimacy. Eforts to quantify the personal capability of key leaders, the potential of a military unit to ight and win, or the degree to which a government institution is prepared to accept responsibility for its portfolio are likely to fall short or mislead. Quantitative metrics are generally incapable of describing the kind of political agreements and development that tend to be the keystone of lasting government stability. An unbalanced approach to assessment cannot hope to provide a policymaker with a reasonably accurate and holistic understanding of the COIN environment or the campaign. Not Suficiently Analyzed here are a lot of things in the calculus of factors afecting progress of the war, there’s no way we know of yet to measure it. —General Creighton W. Abrams23 18 Interview with the author, Kabul, Afghanistan, May 2010. 19 In the case of at least one RC, the entire assessment staf consists of two people. 20 According to FM 3-24 (HQDA, 2006c, p. 4-6), “COIN operations often involve complex societal issues that may not lend themselves to quantiiable measures of efectiveness.” 21 his inding is based on direct observation of and participation in the assessment process in Iraq over the course of three tours from 2003 to 2006. 22 hese percentages are based on a generalized assumption regarding the availability of quantitative versus qualitative data in COIN. hey are based on direct observation of COIN from within a staf headquarters over three tours, research on the Afghanistan campaign, and a study of 89 COIN cases for How Insurgencies End (Connable and Libicki, 2010). his percentage does not, of course, take into account the fact that a great deal of relevant information is not readily available to analysts due to the limitations posed by the COIN environment. 23 Sorley, 2004, p. 363. Why Does Centralized Assessment Fail in Counterinsurgency? 181 Analyzed, inished reports should relect analyses of all available data to produce a uniied holistic assessment. his process should be objective and predicated on at least a general methodological framework that can be modiied to it changing conditions and, if necessary, challenged by consumers. he requirement for holistic analysis is commonly voiced in the assessment literature. But assessment doctrine does not clearly prescribe or deine theater-level analysis of information, only a working group approach to assessment and the application of weights and thresholds. Analysis of raw data and subordinate commanders’ assessments is limited by manpower constraints, the constant demand of external data and brieing requirements, and the inability of assessment staf members to travel outside the theater command’s forward operating bases, even infrequently. heater-level stafs in Afghanistan are particularly prone to resource limitations and demands for data. his section draws on research conducted in Afghanistan in 2010 and 2011. Analysts at top-level stafs in Kabul spend a great deal of time rectifying bad or missing data, retroactively adjusting time-series graphs, responding to directive requirements for raw data from congressional and policy stafs, and trying to convince parallel staf sections to provide relevant data to feed centralized assessment. his was also true in Vietnam and Iraq. he mechanics of centralized assessment take up a great deal of time, and the absence of transparent and credible analysis contributes to a constant demand for raw data by senior policymakers. In the words of an assessment oicer at a theater-level command in Afghanistan, “A lot of what we do is supericial.”24 his oicer’s staf was overwhelmed with requirements for brieing slides and had little time to immerse itself in all-source information.25 Another analyst whose job it was to “redteam,” or provide alternative analysis to RC assessments, was equally overwhelmed. his analyst, responsible for tracking and red-teaming an entire RC, stated, “We’re doers, not thinkers. We’re collecting so much information that we don’t have time to analyze it. We’re just a reporting agency. We put our labels on subordinate products.”26 When analysis does occur, it is typically conducted in an ad hoc and limited fashion. Stafs publish a recurring set of time-series reports intended to show progress or lack of progress along (typically) three lines of operation: governance, security, and development. here is no analytic or statistical method used to compare or combine these reports in a way that would produce a clear quantitative analysis of the entire campaign (but no method could hope to show strong correlation between multiple aggregated variables in COIN).27 Instead, analysts examine individual variables (e.g., 24 Information analyst, interview with the author, Afghanistan, May 6, 2010. 25 Sometimes the staf has only six people, depending on rotations and leave. Since this research was conducted, it is possible that there has been an increase in these levels. 26 27 Information analyst, interview with the author, Afghanistan, May 6, 2010. he exception to this rule was the NC3A efort to produce a combined mathematical report derived from a regression analysis. hat efort was reportedly halted in mid-2010. 182 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency number of attacks reported) and attempt to explain trends through subjective analysis, or they conduct limited comparative studies to show patterns in activity. For example, one staf produced a report that compared time-series attack data with weather and seasonal data to show the efects of weather and seasons on insurgent activity. his report demonstrated a deinable pattern, but this pattern could, at best, be considered little more than an analytic tool for subjective assessment of other time-series trends, or an operational planning tool. he mathematical EBA formulas recommended in the Commander’s Handbook and FM 5-0 have yet to be accepted by theater-level stafs: In 2009, the ISAF commander rejected the most comprehensive EBA analysis model developed by his staf. here is general agreement among military staf oicers interviewed for this study that there is no mathematical solution to COIN. Because it is, in all practicality, a mathematical model, EBA does not ofer a useful analytic process for holistic COIN assessment. Pattern analysis and time-series trend analysis are not holistic analyses. Both pattern analysis and trend analysis are methods that might (with known data sets) be used to develop information to feed holistic analysis, but these eforts do not efectively stand on their own to deliver holistic reports. Any efort to aggregate or conduct regression analysis on quantitative patterns and trends would succumb to the same fate as mathematical EBA solutions. Eforts to analyze centralized aggregated data through less structured approaches (e.g., subjective analysis of aggregated data) are insuiciently transparent and credible to meet policymaker requirements. As of early 2011, there was no proven centralized analytic method for producing a transparent and credible assessment derived from aggregated quantitative data. his monograph recommends the incorporation of all-source fusion methodology from the military intelligence ield to improve assessment analysis. his recommendation is derived primarily from JP 2-0, Joint Intelligence, but it also relies on other military intelligence publications, intelligence community publications, and the professional literature on intelligence analysis.28 It is worthy of noting here that the application of all-source analytic methodology to campaign assessment can improve holistic analysis but it cannot account for all of the inherent laws in centralized assessment. All-source analytic methods would be best matched with layered contextual assessment. Incongruent with Doctrinal Understanding of Counterinsurgency COIN assessment theory should be congruent with U.S. joint and service understanding of warfare, of the COIN environment, and the way in which COIN campaigns should be prosecuted. Appendix E references a 2008 memorandum from the com28 his recommendation is explained in greater detail in Chapter Nine. Why Does Centralized Assessment Fail in Counterinsurgency? 183 manding general of U.S. Joint Forces Command that oicially removes EBO and its associated themes and concepts from the joint lexicon. But the joint doctrine that shaped this assessment process remains centralized, quantitative, and reliant on efectsbased theory and language.29 he Marine Corps continues to cleave to the principles of maneuver warfare but refers to many Army publications that retain efectsbased language. At irst glance, U.S. Army doctrine seems to include some internal contradictions on EBO and EBA. he 2008 version of FM 3-0 clearly states that “Army forces do not use the joint systems analysis of the operational environment, efects-based approach to planning, or efects assessment” and that efects-based theory and language are introduced in doctrine only to familiarize soldiers with joint doctrine. (FM 3-0 was published several months before Mattis issued his letter.)30 However, FM 5-0 prescribes the use of EBO in planning, operations, and assessment.31 Appendix H of FM 5-0 clearly spells out an efects-based process for assessment without ofering an alternative.32 It might or might not be possible to apply a mathematical assessment model like the one prescribed in FM 5-0 or the Commander’s Handbook to a conventional warfare environment. However, it is internally inconsistent to simultaneously accept complexity and chaos (FM 3-24 and FM 5-0), conceptually reject SoSA, ONA, and EBO (FM 3-0), and also dictate the use of EBO and EBA models in COIN (FM 5-0). An examination of the hierarchy of Army doctrine as described in the 2007 Army Modernization Plan clearly gives FM 3-0 (“capstone doctrine”) higher precedence than FM 5-0 (a “reference” document).33 In the Army’s hierarchy of doctrinal publications, all subordinate publications are to mesh with the capstone publications; thus, the rejection of efects-based theory as articulated in FM 3-0 appears to be the Army’s oicial 29 JP 2-0 (U.S. Joint Chiefs of Staf, 2007) clearly spells out methods to examine nodes and links within a SoSA model and describes ways to support an efects-based assessment process. 30 HQDA, 2008, p. D-2. An April 29, 2011, document issued by the U.S. Army Combined Arms Center provides what is termed “unoicial guidance” to training developers on Army terms that have been added, recscinded, retained, or revised. his spreadsheet of terms rescinds “efects based operations,” “efects cell,” and “efects coordinator,” describing each as “non-doctrinal terms.” 31 For example, “By identifying the possible emergence of unintended consequences or threats, commanders consider exploitable opportunities to create efects that reinforce the desired end state” (HQDA, 2010, p. 3-11). 32 FM 5-0 also states, It is often diicult to establish a link or correlation that clearly identiies actions that produce efects beyond the physical domains. he relationship between action taken (cause) and nonphysical efects may be coincidental. hen the occurrence of an efect is either purely accidental or perhaps caused by the correlation of two or more actions executed to achieve the efect. (HQDA, 2010, p. 6-7) 33 HQDA, 2007. 184 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency doctrinal position on EBO and EBA.34 It seems that Army doctrine rejects EBO while retaining EBO language to address some standing joint requirements in the absence of a viable alternative.35 Joint COIN doctrine remains largely efects-based, despite the absence of explicit EBO language in standing publications. JP 3-24 calls for a systems approach to understanding the COIN environment and an efects-based approach to planning, operations, and assessment: Identifying desired and undesired efects within the OE [operating environment] connects military strategic and operational objectives to tactical tasks. Combined with a systems perspective of the COIN environment, the identiication of desired and undesired efects informs a holistic view of the OE. Counterinsurgents plan joint COIN operations by developing strategic and operational objectives supported by measurable strategic and operational efects and assessment indicators.36 JP 3-24 provides more suggestion than direction for assessment, and like some service doctrine, it proposes EBA while recognizing that the complexity of the COIN environment requires modiications to the approach: “[T]he complex nature of COIN operations makes progress diicult to measure.”37 In the same section, it recommends the use of the Interagency Conlict Assessment Framework, but that is a loosely structured guide for preoperational assessment and not a campaign assessment tool or methodology.38 he joint Army/Marine Corps publication FM 3-24 skirts the issue by presenting EBA at face value and then cautioning against overreliance on EBA, without providing an alternative.39 In total, the 2006 version of FM 3-24 devotes three pages to COIN campaign assessment. According to U.S. Army and Marine Corps doctrine, as well as nearly all the literature on COIN examined for this study, COIN is best practiced as a decentralized type of warfare predicated on “mission command.” Decentralization and mission command necessitate a loosely structured, localized approach to prosecuting war. 34 According to the plan (HQDA, 2007): he Army has two capstone manuals regarding doctrine: Field Manual 1, he Army; and FM 3-0, Operations. FM 1 contains the Army’s vision. FM 3-0 provides the principles for conducting full spectrum operations, and describes the operational role of linking tactical operations to strategic aims. It details how Army forces conduct operations in uniied action. hese two doctrinal publications establish the foundation in preparing the Army to dominate land warfare in Joint operations. 35 ISAF is a NATO (not U.S.) command. As of mid-2010, NATO doctrine incorporated the “efects-based approach to operations” (EBAO), a modiied version of EBO that still relies on basic efects-based concepts. 36 Joint Chiefs of Staf, 2009c, p. IX-4. 37 Joint Chiefs of Staf, 2009c, p. X-17. 38 Reconstruction and Stabilization Policy Coordinating Committee, 2009. 39 HQDA, 2006c, p. 5-27. Why Does Centralized Assessment Fail in Counterinsurgency? 185 his approach allows commanders at the tactical level to assess local conditions and tailor the ways in which they address local challenges. his means that there are few if any generalizable or consistent activities across a COIN theater of operations. In COIN, the impact of decentralization of operations on assessment is exacerbated by the complexity and uncertainty inherent in the environment. It is diicult—if not impossible—to develop a practical centralized model for COIN assessment because complex COIN environments cannot be clearly interpreted through a centralized process that removes data from their salient context. he incongruence between decentralized and complex COIN operations and centralized, decontextualized assessment has led military stafs to rely on ad hoc assessment methods that leave policymakers and the public dissatisied with U.S. COIN campaign assessments. FM 3-24 applies to both Army and Marine Corps COIN operations: Mission command is the conduct of military operations through decentralized execution based upon mission orders for efective mission accomplishment. . . . Mission command is ideally suited to the mosaic nature of COIN operations. Local commanders have the best grasp of their situations. . . . hus, efective COIN operations are decentralized, and higher commanders owe it to their subordinates to push as many capabilities as possible down to their level.40 If it is truly the case that “local commanders have the best grasp of their situations,” then local commanders are also best positioned to write assessments; at the very least, their assessments should be prominent in the overall theater process. JP 3-24 states, “Subjective assessment at all levels is essential to understand the diverse and complex nature of COIN problems,”41 and FM 3-0 concurs: “Generally, the echelon at which a speciic operation, task, or action is conducted should be the echelon at which it is assessed.”42 Centralized COIN assessment runs counter to not only these doctrinal statements but also more generally to U.S. joint and capstone doctrine on both warfare and COIN. Further, it is written in a way that seems intended to confound military oicers attempting to create an assessment process. It simultaneously recommends the use of detailed mathematical models and recommends against reliance on mathematical precision and accuracy. As a result, military stafs preparing to deploy to combat tend to ind assessment a confusing and unhelpful distraction rather than a means for 40 HQDA, 2006c, p. 1-26 (emphasis in original). 41 Joint Chiefs of Staf, 2009c, p. X-17. 42 HQDA, 2008, p. 5-17. 186 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency supporting decisionmaking.43 At least in part because of this confusion, they tend to devote only leeting attention to assessment until they are in theater.44 Data Aggregation and the Counterinsurgency Mosaic FM 3-24 and nearly all of the literature on COIN cited in this monograph reinforce the idea that insurgency tends to be a local phenomenon. National, regional, or ethnic issues might provide a common theme to insurgency, but individuals and small groups are the insurgency. With some possible exceptions (e.g., the Chinese communist insurgency), the root causes and motivations that feed insurgent recruitment and support tend to be localized.45 his dynamic has proven to be particularly salient in Afghanistan, a country with a population broken into many distinct ethnic and tribal groups, mostly spread out across thousands of small and often isolated villages. Each village, district, province, and region has unique physical and human terrain, levels of violence, degrees of economic development, and governance. With few exceptions, there are no blanket solutions to any COIN problem: Local concerns tend to require local solutions.46 As demonstrated in places like Vietnam, Afghanistan, Iraq, the Philippines, and Malaya, COIN is a battalion commander’s war.47 he events that transpire in each battalion’s area of operations are also unique. An attack in one district probably will not have the same meaning as an attack in another district. For example, the attack might be tied to an intertribal contracting dispute in one district (one tribe wants the other’s government contract), while it might be clearly tied to a signiicant Taliban presence in another. Shops opening in one market might indicate increasing stability and economic development, while, in another market, this may relect the success of the Taliban shadow government in the absence of Afghan government presence. Following the guidance in FM 3-0, this information—whether quantitative or qualitative—has clear value when assessed at the level at which it is collected. he battalion staf should be able to explain that the intertribal attack does not 43 his last observation was made over the course of two years spent preparing U.S. Army and Marine Corps units from the division to the battalion level for deployment to Afghanistan through structured discussions, briefings, seminars, and workshops. 44 his inding is derived from direct observation of numerous tactical units during their predeployment workups and deployments to Afghanistan and Iraq. 45 Kilcullen, 2009a. However, this is not always the case. he example of the Chinese Revolution stands this assumption on its head. As with nearly all assumptions regarding COIN, this one should be considered generalizable only with caution and with the understanding that each COIN case is unique. 46 47 See Connable and Libicki, 2010, pp. 151–156. In some cases, it is a company or platoon leader’s or platoon commander’s war, depending on force lay-downs and a range of elements in the operating environment (e.g., physical terrain). his monograph focuses on battalion-level operations with the express intent of building staf processes. Writing in the mid-1960s, David Galula (2005, pp. 110–111) describes a “basic unit of counterinsurgency warfare” at the battalion level during earlier phases and down to the squad level during transition. Why Does Centralized Assessment Fail in Counterinsurgency? 187 relect increased insurgent activity and that the market openings can be either positive or negative. As data are aggregated, however, that context disappears. When applied to a core metrics list, more attacks can simply mean that things are getting worse and that all shop openings are beneicial.48 Multilayered data aggregation without embedded analysis leads to an exponential loss of context and concurrent decrease in relevant meaning from level to level. Aggregation without embedded analysis is the primary cause of opacity and loss of credibility in assessment. Figure 8.3 ofers a notional example of the kinds of data aggregation challenges that can develop within a regiment’s (or brigade’s) area of operations. It shows that each of four battalions has reported something diferent about school attendance, but, taken as an aggregate across the regimental area of operations, the judgment is that a large number of children go to school and that attendance relects support for the government. But these data may be painting a misleading portrait. hey are not compared to the total number of children in the area of operations, because no accurate census data are available. herefore, it is not possible to know what percentage of children overall attend school. he Taliban clearly afected attendance in 2nd Battalion’s zone and probably in 3rd Battalion’s, where children are paid to go to school. It appears that no one goes to school in the 4th Battalion’s area, but in reality, attendance is unknown; it could be zero, 50,000, or some number in between. In this notional case, the aggreFigure 8.3 How Aggregation Hides Context How many children attend school? 1st Battalion: • 25,000 attend every day across all districts • Positive view of commander, Afghans • Commander: “School attendance reflects success” 2nd Battalion: • 25,000 attend some days in most districts • Want to attend but Taliban prevents them • Commander: “Insurgent violence reduces attendance” 3rd Battalion: • 50,000 attend 3 days/week across all districts • Do not trust or support the government • Commander: “We pay them to attend school” 1st Regiment: • 100,000 attend school in area of responsibility • School attendance reflects support • Commander: “Attendance is high and positive” 4th Battalion: • 0 attend schools in any district • No data available but they might attend • Commander: “We don’t know how many attend” NOTE: The data in the figure are notional. RAND MG1086-8.3 48 See Kilcullen, 2009a, for a lucid explanation of the impact of aggregating attack incident data. See DownesMartin, 2010b, for another critique of data aggregation in the DAM. While he agreed that context is important, the director of AAG stated that detail at the theater level is not important, and that “more attacks means things are getting worse” (interview with the author, Kabul, Afghanistan, May 2010). 188 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency gated assessment does not relect operationally critical context at the battalion level: he commander is making an assessment that fails to interpret the data in context. Unparsimonious, Costly, Risky, and Contentious What I don’t know I make up, what I do know I don’t bother checking . . . because the metrics system is junk, the scoring mechanism is broken, and we get no feedback. —Senior military oicer on an RC staf, 201049 COIN assessment should be parsimonious. Collection and reporting requirements for assessment should be carefully considered in light of the demands and risk they may leverage on subordinate units. Assessment should rely to the greatest extent possible on information that is generated through intelligence and operational activities without requiring additional collection and reporting. As Chapter Five showed, experts on assessment generally agree that core metrics lists should be parsimonious and that eforts to collect data from subordinate units should be frugal. his expert opinion coincides with U.S. doctrine that states, “Excessive reporting requirements can render an otherwise valid assessment plan onerous and untenable.”50 his simple advice does not go far enough to shape assessment or data collection for assessment. Both the Vietnam and (to a lesser extent) Afghanistan cases show that centralized COIN assessment requires the collection, reporting, and management of enormous amounts of quantitative data. Requirements for data leveraged upon subordinate units rarely relect a thorough and iterative efort to determine the cost and risk that might be associated with the requirement. As a result, subordinate units sometimes ignore requests for data, fabricate data, or endure cost and risk to collect data that are not used to produce relevant assessments. It is possible for assessment stafs to improve the collection requirements process to include considerations of cost and risk, but centralized assessment cannot function without large quantities of aggregated quantitative data. Not all of the data demands for centralized assessment can be met with data produced for intelligence and operations functions, so a considerable amount of recurring collection and reporting will always be necessary to feed these demands. Because centralized assessment is likely to place onerous demand on subordinate units, it is also likely to create some amount of disharmony between subordinate and senior stafs. While this can be alleviated to some extent through clear guidance and iterative discussion, combat tempo and the distributed nature of COIN opera49 Statement by a senior coalition military oicer, Afghanistan, 2010, as quoted in a brieing delivered by Stephen Downes-Martin at the U.S. Naval War College, Newport, Rhode Island, October 19, 2010. Downes-Martin interviewed the oicer in mid-2010. 50 HQDA, 2010, p. H-3. Why Does Centralized Assessment Fail in Counterinsurgency? 189 tions make this type of process diicult and unlikely. In practice, core metrics lists are created through a process that assessment stafs and combat oicers tend to refer to as “BOGSAT,” or “Bunch Of Guys Sitting Around a Table.”51 he subjectivity of core metrics selection is inherent in the process, and tactical-level oicers understand that core metrics selection is subjective. here is an ingrained suspicion of core metrics at the tactical level, and this suspicion is unlikely to be overcome even in the best of circumstances. Centralized assessment is likely to produce staf friction that will lead either to cost and risk of questionable value or to disingenuous reporting. Additional Factors Contributing to the Failure of Centralized Assessment in Counterinsurgency his section addresses a number of additional factors that contribute to the failure of centralized assessment in COIN. hey include overoptimism, the inability to adequately cope with missing data in statistical trend analysis, and the subjectivity inherent in time-series thresholds. It also provides a number of examples of data inaccuracy and inconsistency, and it describes the district assessment process to demonstrate how color-coding and aggregation can undermine transparency and credibility. Overly Optimistic Listen, the president didn’t ask for a “situation” report, he asked for a “progress” report. And that’s what I’ve given him—not a report on the situation, but a report on the progress we’ve made. —Robert W. Komer, as quoted in Vietnam: A History52 he post–World War II American [military reporting] system was receptive only to the recording of sunny hours. All reports were by nature “progress reports.” —Neil Sheehan53 51 I participated in a number of such “BOGSAT” sessions between late 2009 and early 2011. All of these eforts were well-intentioned and professional but ultimately unhelpful. Some of these sessions were unstructured or at low levels of authority, while others were structured and coordinated by high-level authorities. No matter how the sessions were structured, because they were removed from the military planning staf and process, they necessarily boiled down to debates over the partially informed opinions of those present for the discussion. he sessions I attended never relected iteration with ield units, rarely referred to operational planning documents, and never considered the risk of collection through structured and realistic analysis. he best of these eforts relied on the participation of more than 100 top experts and military ofers over a weeklong series of ofsite sessions. his group had access to all relevant documentation, and it included members of the theater and component stafs. Its indings were not adopted. 52 Karnow, 1984, p. 515. 53 Sheehan, 1988, p. 287. 190 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency “Progress reports” are neither unusual in military operations nor necessarily always optimistic. Military operations are conducted to make progress toward objectives; sometimes they succeed and sometimes they fail or struggle to make progress. hese two quotes (above) relect perceptions about the assessment process that are relevant to the way in which assessments are viewed by consumers and the public. Komer’s comment may relect pressure from senior policymakers to provide good news through “objective” analysis, while Sheehan’s comment relects the cynicism that perceived overoptimism in military assessment reports sometimes elicit. Overoptimism in military campaign assessment reporting can undermine transparency and credibility and also contribute to the existing distrust of military assessment reports in some policy circles and among the general public. Sometimes, this overoptimism is self-generated, and sometimes it results from pressure within the policy or military hierarchy. Gelb and Betts describe how assessment stafs during the Vietnam War were subjected to similar pressure to deliver positive assessment reports: Most [assessment analysts] were under varying degrees of pressure, both explicit and implicit, to produce certain kinds of assessments. Whether they biased their reports in a certain direction because someone else told them to do it or because they knew this would be best for them in the long run . . . is not particularly important in this context; the fact is that the pressure was there nonetheless.54 his analysis of Vietnam-era assessment is reinforced by the accounts of those who were involved in the assessment process during this period, as cited throughout this monograph. Testimony in the Westmoreland v. CBS case by retired Marine Corps colonel, COIN expert, and then-Congressman Paul N. McCloskey, Jr., on MACV brieings to visiting policymakers ofers speciic insight into the way in which quantitative data were used to engender positive perceptions: here was heavy stress on numbers, i.e., body count, crew-served weapons captured, strength of VC units, and particularly the favorable trends in those numbers in every category as compared with three months earlier, a year earlier, etc. I do not recall a single unfavorable trend reported to us, and there was a consistent and strong expression that there was “light at the end of the tunnel,” that our “nation building” program was succeeding, that the VC strength was steadily eroding, and that in due course we would be able to return to an advisory status to a strong and stable South Vietnamese government. . . . [I saw a manual] with a title along the lines “Standard Operating Procedure for Handling Visiting CODELs [Congressional Delegations].” he manual explicitly outlined the requirement that CODELs were to be provided only with facts favorable to MACV’s perfor54 Gelb and Betts, 1979, p. 309 (emphasis in original). hey provide additional insight into this dynamic on pages 319–321. Why Does Centralized Assessment Fail in Counterinsurgency? 191 mance and directed withholding facts that would make “MACV’s mission more diicult.”55 It is clear that Vietnam-era analysts were under considerable pressure to deliver positive reports.56 Research into the Iraq and Afghanistan campaigns revealed similar pressures on analysts in both theaters. Downes-Martin also observed overoptimism in Afghanistan assessment reports generated at both the RC and theater levels.57 Observation and interviews for this study showed that the majority of this observed and reported pressure in Iraq and Afghanistan was implicit and probably nowhere near as egregious as during the Vietnam campaign. his kind of pressure is typically nuanced, and it is sometimes unintentional or self-generated. In their discussion of assessments, Gelb and Betts point out that military oicers are trained and educated to be optimistic and forward-leaning and that delivering bad news could represent a “personal failure.”58 his last claim is an overgeneralization of military culture, but it is sometimes accurate. As with Vietnam-era assessment reports, overoptimism in Iraq and Afghanistan assessment reporting is most apparent in the emphasis that is sometimes placed on positive information rather than in an imbalance in the published reports. In other words, reports from both eras often show positive and negative information, but the ways in which the reports are presented may emphasize an optimistic outlook that the report does not necessarily warrant.59 As McCloskey’s testimony shows, oversell was taken to an extreme during the Vietnam War. Karnow also provides a number of examples in which Robert S. McNamara presented overly optimistic public assessments that he then countered with pessimistic but private reports to President Johnson. 60 his oversell backired when the Pentagon Papers were leaked to the public. To a great degree, overoptimism is an issue of perception. In the absence of precise and accurate determinative assessments, only subjective judgment can say whether reports are overly optimistic rather than objectively optimistic. Sheehan judged that Vietnam-era progress reports were overly optimistic based on his perceptions and research of the available data. Although overoptimism may be inherent to some degree, 55 Westmoreland’s memorandum in opposition to CBS’s motion to dismiss, October 19, 1982 (emphasis in original), in Westmoreland v. CBS, 1984. McCloskey ran the COIN school for reservists preparing to deploy to Vietnam and served in Vietnam from 1964 to 1965, prior to being elected to Congress. 56 McNamara testiied in the same case that “General Westmoreland was not under any pressure to deliver good news on the war.” 57 Downes-Martin, 2010b, pp. 2–7. 58 Gelb and Betts, 1979, p. 309. 59 his is a subjective judgment based on nine years of experience reading COIN assessments from Afghanistan and eight years of experience reading COIN assessments from Iraq, as well as a more intensive examination of assessment reporting between November 2009 and March 2011. 60 Karnow, 1984, pp. 341, 357, 440. 192 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency the best way to counter the perception of overoptimism from the public or policymakers is to create depth and transparency in reporting. Inability to Adequately Address Missing Data in Time-Series Trend Analysis his section addresses a signiicant concern in the application of statistical methods to data sets with unknown inconsistencies and gaps. Some amount of accurate (or valid) data is necessary for any kind of statistical analysis. his is certainly true for the kind of time-series analysis that is common in COIN assessment. But nearly every scientiic study is based on the assumption that a percentage of data will necessarily be missing. A study of 300 scientiic articles that had been published within a three-year period showed that a full 90 percent of the studies described in the articles were missing data and that the average amount of missing data across these 270 studies “well exceeded” 30 percent.61 Nonscientiic military assessment of COIN campaigns must also assume that there will be sizable gaps in sets of information and that the unstructured nature of COIN assessment may mean that gaps will be larger than in a controlled scientiic study because it is generally not possible to collect data from all relevant areas due to violence or threat. It will also be very diicult, if not impossible, to provide a sound estimate of the amount of data that is missing because the assessment analysts do not control and typically cannot inspect collection measures in the ield. here is an entire ield of research on the phenomenon of missing data, a statistical term referring to gaps in data sets due to improper or incomplete collection or data corruption. Statisticians have mathematical techniques (e.g., data creation algorithms) designed to ill in these gaps for scientiic analysis. he theory and terminology from missing data analysis is informative. In very simple terms, ignorable, or “missing-atrandom,” data are the kind of data that can be made up for logically and systematically through statistical inference.62 “Nonignorable data” is also somewhat self-explanatory, but it is very hard to make up for nonignorable data using missing data analysis. his chapter argues that each of the individual tiles in the COIN mosaic may be important to the campaign in their own way. In practice, it is typically the people or events that are ignored that turn out to be the most dangerous to the government. It is not enough to state that “nearly all of the tribe supports the government” if 5 percent of the tribe is able to destabilize an entire district. It is not enough to state that an opinion poll covered “most of the country” if the poll-takers were prohibited from entering the 61 McKnight et al., 2007, p. 3. Since this analysis describes missing data problems in scientiic studies, it makes sense that missing data problems in loosely structured COIN assessment would be even more signiicant. However, there is no available empirical evidence to support that conclusion. 62 Inference is used in the broadest sense here. Missing data analysis can be based on imputation, regression analyses, or other approaches. Why Does Centralized Assessment Fail in Counterinsurgency? 193 most dangerous—and potentially most important—areas due to security concerns.63 Eforts to bridge data gaps probably could not account for the damage that might be caused by inaccurate or missing reports.64 he inability to safely jump data gaps poses a serious challenge to researchers trying to apply statistical methods to COIN analysis. Social scientist Paul D. Allison at the University of Pennsylvania (the author of Missing Data) states that missing data methods depend on certain easily violated assumptions for their validity. Not only that, there is no way to test whether or not the most crucial assumptions are satisied. he upshot is that although some missing data methods are clearly better than others, none of them really can be described as good. he only really good solution to the missing data problem is not to have any.65 he inability to control data collection and data management, the lack of reliable data collection methods, and the luid nature of requirements means that it is impossible to accurately track, describe, or account for missing data in a way that would satisfy basic requirements for validity. his means that the most common technique in centralized assessment—time-series analysis—is of questionable utility when integrated into a COIN campaign assessment. Time-Series Thresholds Are Subjective and Unhelpfully Shape Behavior As with all other aspects of centralized assessment, the lack of context confuses the meaning and undermines the value of time-series thresholds. In some cases, thresholds are selected through behind-the-scenes discussion or by the commander’s preference. In a few cases, thresholds are transparently (and commendably) tied to national objectives, but selection remains a subjective judgment call. For example, in Iraq, the coalition set a threshold for electricity that matched prewar levels: If it could achieve this level of service, then it could show that it had put Iraq back on its feet. But this threshold proved elusive as demand rose above prewar levels and was never met; frustration continued unabated as expectations changed. Coalition planners did not factor into the equation the possibility that the removal of sanctions coupled with more plentiful electricity might lead Iraqis to purchase more air conditioners, refrigerators, and tele- 63 his is the most common criticism of poll results in both Iraq and Afghanistan, based on my discussions with consumers of polls in DoD and other government agencies. he second most common complaint seems to stem from a lack of transparency. 64 For example, what if a host-nation military unit exaggerates an incident report to show that it was attacked by 100 insurgents when, in fact, it was attacked by ten? his single report could drastically alter both quantitative and qualitative analysis. What if the same host-nation unit was actually attacked by 100 insurgents but failed to report the attack or the report was lost before it could be oicially recorded? 65 Allison, 2002, p. 2. 194 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency vision sets than they had prior to the war. So, while the original threshold was met, it proved to be relatively meaningless in terms of efect. Assessment of Iraq’s mobile-phone network provides another example of the failure of subjective thresholds to capture actual progress. As violence ebbed in Iraq, Iraqi cell-phone purchases skyrocketed, but the overall number of phones hid the fact that many Iraqis had to purchase three or more phones because there were no available roaming services from region to region.66 he subjective threshold proved to be misleading. Aggregated threshold analyses are susceptible to Simpson’s paradox: hey can show what appears to be forward progress while, according to the same indicator, many of the key areas in the country are sliding backward. 67 For example, a security threshold that sets a maximum acceptable number of violent attacks across the country (e.g., 50 attacks per day) might reveal a lack of progress in a speciic district that is sufering a relatively high ive attacks per day, positive progress at the province level because that same province might be sufering only ive attacks per day (all in the aforementioned district), and then a lack of progress at the national level because the total number of attacks may be well over 50. Without detailed analysis, the national threshold can be misleading. he use of thresholds shapes behavior in ways that are often unforeseen. his was proven in the Vietnam case: Military oicers changed their behavior to increase body counts or show positive HES reporting data because policymakers and senior military leaders had set the performance thresholds for both programs. Strict quantitative thresholds encourage dishonest reporting. And while thresholds might help focus military units on key tasks, so do traditional military mission statements and objectives; the addition of thresholds for operations is superluous and akin to micromanagement. hresholds are poor assessment devices because the data needed to show accurate quantitative progress do not exist in COIN environments. In the absence of a mathematical solution to COIN, it is diicult to envision a systematic, transparent, and credible means of establishing mathematical solutions to speciic aspects of COIN operations. here seems to be no way for a staf to adequately defend the selection of a time-series threshold as anything but a subjective guess as to what level of progress or suppression of activity might be necessary to efect a meaningful change.68 Time-series thresholds may seem to be the most direct and commonsense method for showing 66 his information is based on my experience working with development metrics in Iraq and is derived, in part, from discussions with experts on Iraq assessment between 2004 and 2011. For further information on electricity and cell-phone usage in Iraq, see Brookings Institution, 2010. 67 For an explanation of Simpson’s paradox and its efect on aggregated assessment, see Connable and Libicki, 2010, p. 139. 68 his might not be the case in a conventional campaign designed to degrade a known conventional military force, and it might not apply to some MOPs that are focused on internal U.S. performance. For example, it may be possible to set a defensible threshold for reducing the number of days it takes to move troops into theater in Why Does Centralized Assessment Fail in Counterinsurgency? 195 progress in COIN campaigns, but in practice, they are misleading and often counterproductive. To the credit of commanders and stafs in Afghanistan and Iraq, the use of thresholds appeared to be waning in 2011. Incomplete and Inaccurate Data: U.S. Department of Defense Reports on Iraq Most data on contemporary COIN campaigns (Iraq and Afghanistan) remain classiied or otherwise restricted.69 Even if most of these data were available for unclassiied research, an empirical study of data from a complex insurgency like Iraq or Afghanistan would take months or years and would require extensive ieldwork and interviews with members of tactical units. It is possible, however, to gain insight into data quality by other means.70 A commonly referenced source for oicial, publicly released data are the 9010- and (now) the 1230-series reports produced by DoD for Congress. hese reports are very lightly sourced because they cannot show restricted data, a serious law not in the way the report is written but in the structure of the assessment process. hese reports typically refer to the command that provided the data (e.g., “MNF-I” for Multi-National Force–Iraq or “USFOR-A” for U.S. Forces–Afghanistan) rather than the original source of the information. his makes it nearly impossible for someone outside the U.S. government to check the validity of the data. It is also possible to identify major inconsistencies in historical 1230-series reports that bring into question the overall quality of the data. For example, the November 2006 version of the 9010 report on Iraq shows that there were approximately 1,000 sectarian killings there in October 2006, but the December 2007 report shows that there were approximately 1,700 such killings in October 2006.71 here is no explanation ofered to account for the additional 700 killings; it could be a result of lagging reports or simply a correction of erroneous reporting. here is no way to tell whether either of these reports is accurate, due to the opacity of the sourcing.72 order to improve eiciency. his kind of initiative has little to do with holistic COIN campaign assessment, however. 69 Some aggregated data reports are unclassiied, and, in some cases, speciic data sets are unclassiied. However, to clearly determine quality, it would be necessary to study all types of data from multiple sources that had been captured over an extended period. Even with sampling techniques, a thorough study of this type would require access to restricted data. 70 For example, the September 2009 report lists hundreds if not thousands of discrete elements of information but contains only 35 footnotes referencing a total of three sources, while the March 2010 report on Iraq also jams a staggering amount of data into the narrative text (at least 35 individual pieces of information or conclusions on page 53 alone) but ofers only 24 footnotes referencing three sources throughout the entire report. DoD, 2009b, p. 38; DoD, 2010b, p. 77. 71 72 DoD, 2006, p. 24; DoD, 2007, p. 17. he classiied versions of these reports are more adequately sourced but still not adequate to sustain objective credibility. he most commonly read reports—and the only ones available to the public—are the unclassiied versions provided on the DoD website. 196 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Incomplete and Inaccurate Data: Civilian Casualties he rate of civilian casualties is a commonly used metric in COIN. he thinking behind the use of this metric is that the more civilians who are killed, the less likely the remaining civilians are to trust or support the government because of resulting instability, fear, and a perception that the government is either ruthless or incompetent. Civilian casualty data are often disputed, however. Jonathan Schroden states that, as of 2009, there was no oicial total civilian casualty igure for Iraq and that even if we could consistently separate civilians from insurgents, the actual number of such casualties still may not be easily determined with any degree of accuracy. . . . One efort, called the “Iraq Body Count,” has compiled a running tally via open-source reporting that currently stands around 90,000 [as of 2009]. A second efort by he Brookings Institution relied on a combination of data sources, and gave a total closer to 104,000. A third efort, by researchers at the Johns Hopkins Bloomberg Institute of Public Health, generated a substantially diferent count of over a half-million.73 While the Iraq Body Count (IBC) and Brookings numbers are fairly close to each other, they are still estimates and not actual counts. Schroden points out that oicial Iraqi statistics are corrupt and inaccurate. He also compared these three databases with three others, but none were in agreement.74 Some of the media sources that fed the IBC database were controlled by insurgents or relected inaccurate circular reporting. For example, in 2004, I witnessed an incident that resulted in a false report of 40 civilian casualties, later tallied as accurate by IBC.75 his false report was re-reported in various media outlets and is still part of the IBC total; IBC staf had no consistent way of checking their sources in the ield. A 2011 DoD report identiied 21,000 civilian deaths by IEDs in Iraq between 2005 and 2010, but the DoD spokesman discussing the report stated that the military was “unable to quantify the claim” and that it relected U.S. reporting only.76 he Congressional Research Service lists six diferent sources that estimate total civilian deaths in Iraq; at the time of that review, these totals ranged from 34,832 to 793,663.77 Despite these critical laws in civilian casualty data, they are still highlighted as accurate centralized quantitative metrics rather than rough estimates in the 1230-series reports.78 73 Schroden, 2009, p. 720. 74 Schroden, 2009, p. 721. 75 Connable, 2010. 76 Vanden Brook, 2011. 77 Fischer, 2008, p. 3. 78 See, e.g., DoD, 2010b, p. 29. Why Does Centralized Assessment Fail in Counterinsurgency? 197 Figure 8.4 shows a graph from a Multi-National Force–Iraq brieing to Congress in 2008. It shows casualty statistics, presented in time-series trend format, from January 2006 through March 2008. Two separate data sets are depicted in the graph. he irst, in purple, is the combined total relected in Iraqi and coalition data, while the second, in blue, relects coalition data alone. he graph shows the degree of potential disparity between data sets that address the same category of data. In December 2006, for example, the coalition recorded approximately 1,250 civilian casualties, while the combined igure was approximately 3,700 casualties. his means that Iraqi tabulations of total civilian casualties were about two times higher than the coalition’s overall total—just for this category of data. his disparity throws both estimates into question: Were the Iraqis inlating their igures, or did they have better data than the coalition? Were the coalition forces underreporting casualties, or were they more discerning than the Iraqis? here is no way to tell from the graph or the accompanying report. One or the other igure might be accurate, or neither might be accurate. he graph is instructive not only for an examination of civilian casualty data but also for a broader understanding of how host-nation or third-party reporting can skew centralized data sets. Iraqi and Afghan incident reporting and other types of reporting are often rolled up into coalition data sets without diferentiation in time-series presentations for assessment. Figure 8.4 Civilian Casualty Data, Iraq, 2007 and Early 2008 4,000 Iraqi and coalition data Coalition data only Number of casualties 3,500 3,000 2,500 2,000 1,500 1,000 500 2006 SOURCE: Multi-National Force–Iraq, 2008, p. 2. RAND MG1086-8.4 2007 ct ov D ec Ja n Fe b M ar O N Se p Ju l A ug D ec Ja n Fe b M ar A pr M ay Ju n Ju l A ug Se p O ct N ov Ja n Fe b M ar A pr M ay Ju n 0 2008 198 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Misleading Metrics: Insurgent Recidivism U.S. and coalition forces hold and often release insurgent prisoners in Iraq and Afghanistan as part of routine COIN operations. Some widely publicized problems with the treatment of prisoners early in the wars revealed inadequate prisoner handling practices, poor evidentiary procedures, and a lack of forethought and supervision in the way in which prisoners were processed prior to release. hese problems led to a series of reforms, and a number of programs were implemented in Iraq to improve processes and to reconcile and educate prisoners in an efort to help reintegrate them into society upon release. Military journal articles and oicial reports have trumpeted the improvements wrought by these programs, and it is widely recognized that many of these improvements were genuine and efective in remedying concerns about prisoner treatment. But their contribution to the COIN campaign is less clear. Overall improvement of the prisoner handling and reintegration process was, and is, typically measured through the recidivism metric. his metric showed that very few prisoners who had been released have been reclassiied as recidivists. In early June 2009, the oicial recidivism rate for released Iraqi prisoners was 0.1 percent, compared to 1.2 percent for all of 2007 and 7.7 percent for the three prior years combined.79 According to the DoD semiannual report on the COIN campaign in Afghanistan, the oicial 2010 recidivism rate for Afghanistan was 1.2 percent.80 he U.S. national average for criminal recidivism tends to hover above 60 percent.81 At irst glance, it appears that the prisoner programs in both Iraq and Afghanistan have seen remarkable, if not miraculous, success. hese metrics are intended to show that signiicant progress has been made in reducing participation in the insurgency, thereby indicating some degree of overall success in the campaign. But there are some complexities in recidivism statistics that belie the simplicity of these types of reports. Strictly deined, recidivism means a relapse into a former activity or behavior. By this deinition, to qualify as a recidivist, a former insurgent would only have to commit a new insurgent act; rearrest, reconviction, or reincarceration would not be necessary. he U.S. National Institute of Justice states, “Recidivism cannot accurately be measured just by using arrest data because not all crime is discovered.”82 79 Brown, Goepner, and Clark, 2009, p. 45. 80 DoD, 2010b, p. 61. 81 Langan and Levine, 2002, p. 1. Oicial recidivism studies tend to rely on three years of data collection and additional time for processing and analysis. his report shows that the recidivism rate for prisoners released in 1994 was 67.5 percent over the subsequent three years, while the 1983–1986 study revealed a recidivism rate of 62.5 percent. 82 U.S. National Institute of Justice, 2008. To account for this gap in data, the U.S. government recommends that criminal behavior studies rely not only on arrest data but also on interviews with ofenders, analysis of crime events, and various statistical methods. For practical purposes, though, most studies tend to equate recidivism with rearrest, reconviction, or reincarceration. he rearrest rate for U.S. criminals in a 1994–1997 study was 67.5 percent, and the reincarceration rate for U.S. criminals in a 1994–1997 study by the U.S. National Institute Why Does Centralized Assessment Fail in Counterinsurgency? 199 By this standard, it would be impossible to determine the actual recidivism rate in either Iraq or Afghanistan because only a few of the hundreds of insurgent attacks that take place each week could be clearly tied to individual insurgents.83 In fact, the commander of Combined Joint Interagency Task Force–435 (CJIATF-435), which runs the prison system in Afghanistan, made it clear that he tracked only rearrest (or “recapture”) rates, drawing a clear distinction between rearrest and recidivism: And we refer to it as recapture, not recidivism. Of this year, 550 [prisoners] have been released. here have been four individuals recaptured on the battleield, so less than half of—about half of one percent, or less than one percent, of a recapture rate. Now, I can’t give you recidivism, because maybe someone went back to the ight and was killed and we can’t recapture that. So he wasn’t recaptured.84 However, the DoD report on progress in Afghanistan does not account for any of these nuances or caveats. In contrast to the comments by the CJIATF-435 commander, it describes the CJIATF-435’s success in achieving a 1.2-percent “recidivism rate.”85 It explains the success of its programs by tying this very low rate to its reintegration programs: “CJIATF-435 assesses that reintegration programs are working to prevent previously detained individuals from rejoining the insurgency.” he report does not say how this conclusion was reached, so there is no way to determine correlation between this improperly labeled statistic and the success or failure of any program. Recidivism statistics are often presented in a way that would allow the reader to assume that the military had the means to know the actual rate at which released prisoners returned to committing insurgent acts (an unknowable igure), or perhaps the theater-wide rearrest rate. Even the rearrest or recapture rate would be a diicult igure to obtain from remote tactical units or Afghan units that might not report all arrests or obtain accurate biometric data before releasing prisoners at the local level. of Justice was 25.4 percent. Even by the strictest standard for recidivism (reincarceration), a 0.1- or 1.2-percent recidivism rate is remarkable. 83 It would be possible to argue that criminal behavior and insurgent behavior cannot be equated, and therefore these rates relect success against the insurgency rather than an aberration in normal human behavior. his would be a diicult argument to make, considering the degree to which criminal activity and insurgent activity blend together, but there does not appear to be any empirical evidence equating the two types of behavior for statistical analysis. It might also be possible to argue that Iraqis or Afghans are dramatically less prone to criminal activity than Americans. However, there does not appear to be empirical evidence that would support this argument. 84 DoD, 2010d. He goes on to state that this success is indicative of the success of the task force’s reintegration program, a correlation that relects his informed opinion but is neither clear nor proven empirically. He does not state that insurgents could recommit insurgent acts and escape both capture and death, as happens on a daily basis in both theaters. He does not state whether these statistics relect only theater-level prison operations, or also tactical arrests by infantry battalions in rural areas that might not result in a transfer to central prisons due to lack of evidence or adequate transportation. 85 DoD, 2010b, p. 61. his percentage accounts for a greater number of released prisoners than the 550 cited in the CJIATF-435 commander’s quote. 200 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency In reality, these statistics might only indicate the rate at which released prisoners were reincarcerated at province- or theater-level prisons. he rate of reincarceration might relect the success of reintegration programs, but it also might show that formerly incarcerated insurgents became better at evading capture, that they were improperly identiied upon detention, that they were killed while conducting attacks, that patterns of coalition operation had changed, or any other number of factors. A low rearrest or reincarceration rate might actually mean that security forces are incapable of keeping up with insurgent activity. When carefully analyzed, the oicial Afghanistan recidivism statistic takes on much less (or diferent) importance for campaign assessment than at irst glance. Iraq statistics are similarly opaque. A 0.1-percent recidivist rate for any type of behavior seems to be farfetched at best. It it more likely that this number relects a rearrest or (even more likely) the theater-level reincarceration rate. Because it is presented as a recidivism rate without explanation or caveat, this statistic has the potential to be misleading. Aggregation and Color-Coding: The District Assessment Model his section explores the DAM as it was developed in late 2009 and early 2010 and deployed through late 2010. It refers to this best-known Afghanistan assessment tool to address broader concerns with color-coding and context. Figure 8.5 shows an example of a DAM map. he map shows the color-coded ratings for “key terrain districts” and the extensive amount of space unaddressed by comprehensive assessment. he color codes run from green (good) to red (bad), with white areas showing up as “not assessed.” In this example, the gray areas are key terrain districts that were not assessed in this particular reporting period. he rest of the map area should also be considered white, or not assessed. he data in the igure are notional, but they relect general coverage of Afghanistan assessment during early 2010. Former Congressman John V. Tunney, a critic of paciication assessment systems in Vietnam, said in 1968, “Since paciication is based on the dubious premise that Government control results in political loyalty, reports of progress in paciication [in Vietnam] continue to be misleading for both the American people and policymakers in Washington.”86 he DAM is an improvement over the HES in that it tries to assess loyalty by examining security, development, and governance. As of mid-2010, the district assessment was delivered in several formats: a consolidated color-coded map, a brieing containing a series of color-coded regional maps, and an operations plan assessment report (a narrative version of the district assessment maps). DownesMartin’s research has produced indings on the mapped district assessments that are very similar to those presented in later RAND research. It is therefore appropriate to refer to his insights to summarize concerns with district assessment color-coding: 86 Tunney, 1968, p. 2. Why Does Centralized Assessment Fail in Counterinsurgency? Figure 8.5 The ISAF Joint Command District Assessment Mazar I Sharif Intl. Hairatan Torah Ghundey Shayr Khan Bandar Islam Qal’eh Tor Kham Kabul International Airport Herat Intl. Ghulum Khan Kandahar Intl. Zaranj Wesh (Chaman) Bahram Chah Bypass Corridor Population Population actively supporting sympathizing with government and government and security forces security forces Population neutral/on the fence Population sympathizing with insurgents Population actively supporting the insurgency Not assessed SOURCE: Sok, 2010, p. 36. NOTE: The data in the figure are notional. RAND MG1086-8.5 A color-coded map hides information. he single color-coding is an average (not a summary) of a large number of underlying factors. It is not only possible, but also likely, that an average (i.e. color on the color-coded map) stay the same as some factors improve and others degrade. he color code tells us nothing useful about this efect, and so one must give narrative explanations about the improving and degrading factors. Since smart staf often does this anyway, the color-coded map becomes pointless at best and misleading at worst.87 87 Downes-Martin, 2010b, p. 6 (emphasis in original). He also points out, Finally there is the observed tendency among some RCs to “average colors”. hey present separate colors for their Region for the three LOOs [lines of operation] (Security, Governance and Development), then provide an overall assessment color which happens to be the color of the average point on the color bar chart of the three LOO colors. his is not coincidence, since RCs have been observed during brieings to have diiculty explaining in operational terms why they have given the color to the overall assessment. (Downes-Martin, 2010b, p. 5) 201 202 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Color-coding is the most aggressive form of data aggregation, since the color scheme forces the assessment into only a few very narrow categories (e.g., red, yellow, blue, green). As with thresholds, color-coded assessments are susceptible to Simpson’s paradox. Color codes are, in fact, codes: hey represent a narrow set of numbers (usually three, or ive at most) and demand simpliication and averaging of complex problems. Figure 8.6 illustrates this situation. While the igure uses notional data, it represents a very real problem that RC-level commanders and their subordinates face when trying to assign color codes to key terrain districts, a process that I observed irsthand while at IJC during the inal part of a six-week district assessment cycle. It seems clear that this notional district does not contain a population that would merit the oicial rating deinition level of “neutral or on the fence,” but color-coding demands averaging.88 Taken together, Figures 8.7 and 8.8 reveal additional concerns about data aggregation in the color-coding process. Each smaller square represents a village within a district (also using notional information). In Figure 8.7, color-coding rules essentially suppress information, hiding contradictory data. Figure 8.8 shows how color-coding can suppress critical information, e.g., the most important tribe is on the fence, or the most populous district in the province favors the insurgents. As of early 2011, ISAF was deeply immersed in building a transition model to set the stage for the 2011 strategic timeline. he laws in the district assessment process Figure 8.6 The Effect of Averaging Color Codes in the District Assessment Model Coding by numbers: Is this district “yellow” or hiding “red?” = “Population neutral or on the fence” Population Population actively supporting sympathizing with government and government and security forces security forces Population neutral/on the fence NOTE: The data in the figure are notional. RAND MG1086-8.6 88 See DoD, 2010b, p. 35. Population sympathizing with insurgents Population actively supporting the insurgency Why Does Centralized Assessment Fail in Counterinsurgency? 203 Figure 8.7 The Effect of Aggregating Averaged Color Codes 4 of 4 green = green Theater 3 of 4 green = green Province 11 of 12 green = green District level NOTE: The data in the figure are notional. RAND MG1086-8.7 Figure 8.8 Analysis of the Effect of Aggregating Averaged Color Codes What if 90% of the population live here? What if the smallest but most important tribe lives here? NOTE: The data in the figure are notional. RAND MG1086-8.8 might carry over to the transition process. Downes-Martin’s analysis of the incorporation of the district assessment maps into transition planning seems accurate: he end state [from transition] is, loosely speaking, a Region (or District, or the entire Country) that is suitable for transition to full GIRoA [government of the Islamic Republic of Afghanistan] control, where “suitable” means there is some 204 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency good chance that GIRoA will be able to keep the Region (or District, or the entire Country) stable and secure. But unless we have a credible theory that links the level of active support for an Insurgency to the likelihood of GIRoA successfully running an area, then we have no connection between the rolled-up color-coded assessment and the desired end state. In this case the assessment does not provide senior leaders tasked with judging the suitability of a Region or District for transition with a credible assessment of suitability.89 he senior staf at IJC attempted to make allowances for problems with data quality and aggregation, but even thoughtful iteration between the IJC and RC stafs could not make up for these pervasive laws. By the time the district assessment was delivered to higher headquarters, it was aggregated, averaged, and opaque (to at least some degree). While FM 5-0 exhorts assessment stafs to record justiications for assessments, these justiications were not systematically recorded and were not clearly available to policymakers or the public. It was therefore not possible for senior leaders or journalists to either disaggregate the quantitative models behind the assessments or determine how and why judgments were levied from the battalion to the theater level. It was also not possible to see or understand changes in core metrics requirements, changes in metrics deinitions, changes in theory or models, or adjustments to baseline data (most of which remained unavailable to the general public). Transparency and credibility simply were not achievable with the EBA district assessment process as it existed in 2010. An April 29, 2010, New York Times article shows how a lack of transparency, or at least the absence of embedded contextual narrative, can lead to erroneous interpretations of theater assessments. he article refers to the IJC district assessment and states, “he number of districts sympathetic to the insurgency or supportive of it increased to 48 in March 2010 from 33 in December 2009.”90 his statement does not accurately relect the assessment. According to the head of the IJC assessment section, the IJC added approximately 20 districts to the review between the December and March cycles.91 No clear indication of this change was relected in the brieing slide that contained the color-coded map. To the reporter, who was reporting from Kabul, it looked like things had suddenly taken a turn for the worse when, in fact, the reporting parameters had changed. By the time this monograph is published, the district assessment may no longer be in use in Afghanistan. It is already being supplanted by products like the District Deep Dive, an efort to provide in-depth contextual narrative for transition assess- 89 Downes-Martin, 2010b, p. 7. 90 Rubin, 2010a. 91 Senior oicer, IJC Information Dominance Center Assessments Section, interview with the author, Kabul, Afghanistan, May 2010. Why Does Centralized Assessment Fail in Counterinsurgency? 205 ment.92 Because the District Deep Dive is a relatively new process, an examination of its eicacy or applicability to the overall assessment process was outside the bounds of the current research efort. But despite this probable shift in practice, this examination of the DAM will remain relevant: he dangers of aggregation and color-coding will persist, and the importance of context in COIN is immutable. Color-coding seems to be a default technique for assessment stafs and commanders, so future generations of counterinsurgents should be aware of inherent pitfalls. Policymakers faced with this kind of confusion express skepticism and frustration with operational COIN assessments.93 Chapter Summary By the deinitions proposed in this monograph, the U.S. military has yet to produce a relevant, transparent, or credible COIN assessment methodology. his is not a failure of efort or intellect; in most cases, energetic and intelligent people draft and review the operational assessments delivered to policymakers. Rather, it is a failure of efectsbased theory or centralized pattern and trend analysis to replace or improve upon the fundamental and broadly accepted theories of warfare. While some of the indings in this chapter may not be technically relevant after early 2011, the underlying issues will persist as long as the COIN environment remains uncertain and U.S. doctrine calls for distributed, mission-type operations. It is possible to create a relevant, transparent, and credible assessment process. his monograph proposes options designed to meet the seven standards for assessment articulated in Chapter One. However, these standards are ofered to help frame any solution to the challenge of COIN assessment. Based on the indings presented in this chapter, prospective solutions that lie outside the closed framework of centralized assessment are the most likely to prove successful in COIN. 92 Perceptions of the District Deep Dive expressed by a group of assessment oicials during a December 2010 workshop in he Hague, Netherlands, were generally dismissive. Several oicials saw the efort as rushed, inaccurate, and incomplete. 93 For example, see Abramowitz, 2007, and Walker, 2007. CHAPTER NINE Conclusions, Recommendations, and Options Sound assessment blends qualitative and quantitative analysis with the judgment and intuition of all leaders. Great care must be applied here, as COIN operations often involve complex societal issues that may not lend themselves to quantiiable measures of efectiveness. Moreover, bad assumptions and false data can undermine the validity of both assessments and the conclusions drawn from them. Data and metrics can inform a commander’s assessment. However they must not be allowed to dominate it in uncertain situations. Subjective and intuitive assessment must not be replaced by an exclusive focus on data or metrics. Commanders must exercise their professional judgment in determining the proper balance. —FM 3-24, Counterinsurgency1 hroughout the operations process, commanders integrate their own assessments with those of the staf, subordinate commanders, and other partners in the area of operations. —FM 5-0, he Operations Process2 Military campaign assessment is an arcane process that receives little attention until it fails. he assessment process failed in Vietnam, thereby contributing to the poor policy decisionmaking that led to America’s irst major COIN defeat. Assessments of the Afghanistan campaign sufer from many of the same drawbacks as the Vietnam-era assessments, but the Afghanistan campaign is ongoing. As of this writing, the United States does not intend to fully transition the leading role for the security of Afghanistan to the Afghan government until 2014. In any event, assessment is likely to continue after this transition. his leaves suicient time for the military to improve COIN assessment to support the policy decisions that will shape the end of U.S. involvement in the war. Comprehensive improvements to COIN assessment theory and practice in Afghanistan might also help reshape U.S. assessment doctrine, training, and education over the long term. 1 HQDA, 2006c, p. 4-6. 2 HQDA, 2010, p. 6-1. 207 208 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency his chapter presents conclusions and a summation of the critical analysis presented in this monograph, as well as recommendations and options intended to help guide eforts to improve COIN assessment in Afghanistan and the U.S. military staf process in general. Conclusions Making efective and lasting improvements to the assessment process will require revisiting basic assumptions. his monograph addressed a number of assumptions—some oicial, some implied in the military literature—that underlie the doctrinal and de facto U.S. military COIN assessment process. hey include, but are not limited to, the following: • Assessment is a centralized and quantitative process. • he COIN operating environment is a system that can be understood through systems analysis. • For EBA: It is possible to deliver precise and accurate campaign assessments. • For pattern and trend analysis: Generalized pattern and trend analyses are suficient to support policy decisionmaking in COIN. • Quantitative data and methods are inherently objective and therefore preferable to mixed or qualitative data and methods. hese assumptions have not provided a foundation for efective assessment. Policymakers attempting to penetrate the complexity of COIN are poorly served by centralized and highly quantitative military campaign assessments. Such assessments either attempt to portray accuracy that is beyond the grasp of even a theoretical military collection capability or unintentionally convey accuracy because they are presented through quantitative charts without suicient caveat. Policymakers can be enticed into making overly determinative decisions based on information they believe to be complete and accurate when, in fact, centralized assessment reports are typically neither complete nor accurate. Centralized EBA and pattern and trend assessment do not account for as much as 40–60 percent of available information, because these processes do not systematically incorporate qualitative data. his problem is inherent in a centralized quantitative assessment process and, in the absence of a massive and quite expensive coding efort at multiple layers of command, it will remain an inherent problem. At best, centralized quantitative assessments are informed by qualitative reports, but because assessment does not relect dedicated all-source analysis, this tends not to be the case. he absence of qualitative input into the centralized process corrodes both the transparency and credibility of centralized assessment. Conclusions, Recommendations, and Options 209 he COIN environment can be loosely analogized to a complex adaptive system. But COIN campaigns cannot be efectively understood by applying systems analysis methods (e.g. EBA) to assessment because COIN is not, in fact, an interconnected system that generates obtainable and understandable feedback. Instead, the COIN environment is an unbounded cloud of individuals and groups, perceptions and actions, emergence, and happenstance, all of which are unpredictably whipsawed by chaotic violence and uncertainty. At best, assessment stafs gain an imperfect sense of general patterns and trends from aggregated COIN data, but these patterns and trends do not provide policymakers with a holistic understanding of the environment or the campaign. Chapters Five through Eight explored the subjectivity inherent in all aspects of assessment, as well as the degree to which aggregated quantitative data used to build patterns and trends are incomplete and erroneous. he levels of completeness and quality for nearly all categories of aggregated COIN data are unknown and cannot reasonably be known. his inability to gauge completeness and accuracy unanchors aggregated data sets from ground truth to the point that COIN data do not provide a credible basis for centralized analysis. Compounding this problem is the lack of a credible methodology that could correlate all available quantitative data to produce a holistic centralized assessment; EBA has not proven credible in COIN. As a result, centralized assessment brieings tend to look more like an assembly of dissociated quantitative data reports than holistic assessment. No matter how these reports are packaged, they provide senior military leaders and policymakers with the kind of vague impressionistic understanding of the campaign described by Donald Rumsfeld in Chapter Two. Figure 9.1 provides a generalized schematic of the impressionistic decisionmaking process derived from centralized quantitative assessment. In the igure, the theater-level assessment staf produces a report containing primarily quantitative data and summaries of subordinate reports, but these data and reports are not systematically analyzed to form a holistic picture. Data quality is unknown, and patterns and trends are open to subjective interpretation without clear correlation between data sets. he commander provides insight, which is holistic and informed by the theater assessment report. he decisionmaker sees an array of dissociated, loosely associated, or falsely associated and decontextualized data, patterns, trends, and inished reports (e.g., intelligence reports), as well as the commander’s report. he decisionmaker gains an impression from this mass of information, compares that impression to that of the commander, and makes an impressionistic decision. his process places undue burden on the assessment staf, the military commander, and the policymaker. he staf must scramble to piece together what will necessarily be an incomplete picture from inaccurate and incomplete data, so it never delivers an optimized product. Assessment stafs have almost no opportunity to succeed. he commander provides a uniied picture to the policymaker in the form of a professional assessment, but the value of this assessment rests almost entirely on the 210 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure 9.1 Centralized Impressionistic Decisionmaking Data Report Data Data Data Data Data Data Data Data Report Data Report Data Data Theater commander’s assessment Centralized Impressionistic Decision Data Data Report NOTE: “Data” are raw figures without context or analysis; “reports” are finished but narrowly focused products from subordinate commands, intelligence centers, PRTs, or other organizations. RAND MG1086-9.1 commander’s position (billet) and reputation because there is no structured contextual analytic depth to the assessment. his, in turn, leads to a situation in which an experienced and trusted military commander’s analysis and coup d’oeil are regularly challenged by policymakers, lawmakers, and the public. he policymaker is forced to make decisions that tend to be more impressionistic than necessary and is placed in the position of defending an opaque, less-than-credible assessment to a suspicious public. No matter how capable the military commander or how expert the advice of assessment stafs or intelligence agencies, decisions based on this lat analytic process will appear to be highly impressionistic to the public. All policy decisions are impressionistic to some extent. Even peacetime decisions based on highly accurate quantitative information require some impressionistic analysis. Policymakers try to take in all available input and make what is ultimately a subjective decision: hey have the evidence in front of them, but the inal decision rests with their judgment and intuition. he purpose of assessment and analysis is to reduce the window of uncertainty to help policymakers narrow the range of possible conclusions and options. COIN assessment should support decisionmaking that is relatively less impressionistic and subjective than an uninformed or poorly informed decision: he more transparent and credible the report, the less impressionistic the decision. Few if any strategic COIN decisions should be deterministic.3 3 Some operational and tactical decisions difer from strategic decisions because better data are available. Decisions that might be deterministic include logistics decisions in which data are obtained from known sources and Conclusions, Recommendations, and Options 211 Any military assessment process for COIN should relect the military’s understanding of the COIN operating environment. For the ground combat services most commonly associated with population-centric COIN (the Army and Marine Corps), this understanding is relected in capstone military doctrine: MCDP-1, Warighting, and FM 3-0, Operations.4 Both of these doctrinal publications describe the environment as chaotic and unpredictable, and both services acknowledge that the fog of war can be reduced but never eliminated. Assessment of COIN campaigns should also relect COIN doctrine: FM 3-24/MCWP 3-33.5, Counterinsurgency, and JP 3-24, Counterinsurgency Operations.5 COIN doctrine is less clear and is sometimes contradictory. In places, it coincides with capstone doctrine; for example, it acknowledges the chaotic nature of the COIN environment, it recommends the use of mission-type orders and distributed operations, and it recommends that assessment be conducted at the lowest possible level. FM 3-24 states, “Subjective assessment at all levels is essential to understand the diverse and complex nature of COIN problems.”6 However, it also recommends the use of SoSA and EBA.7 he weight of capstone doctrine and the nearuniversal focus on mission-type orders and distributed operations in other publications on irregular warfare suggest that COIN assessment should relect the more traditional (maneuver warfare) understanding of COIN as opposed to an efects-based approach.8 COIN assessment should relect and embody the enduring realities of chaos, friction, uncertainty, and unpredictability, and it should also relect operations that are (at least in theory) predicated on mission-type orders and distributed command and control. his monograph argues that the centralized and quantitative EBA approach recommended in doctrine is not applicable to COIN, at least not in a way that would efectively support policy decisionmaking. EBA demands “mathematical rigor,” but COIN is more likely to produce immeasurable uncertainty.9 he EBA process was developed to assess conventional warfare, and it is framed in a way that is antithetical to distributed COIN operations. In EBA, battalion-level assessment is little more than outcomes are predictable, operational force disposition decisions, communication decisions, and any other decision in which accurate data are known to be available and the problem is suiciently bounded. 4 Headquarters, U.S. Marine Corps, 1997b; HQDA, 2008. 5 HQDA, 2006c; U.S. Joint Chiefs of Staf, 2009c. 6 HQDA, 2006c, p. 5-26. 7 Comparing JP 3-24 with JP 2-0 shows that joint COIN doctrine depends on determining and measuring efects through the use of joint intelligence preparation of the environment (JIPOE), as described in JP 3-24. JIPOE is an efects-based process predicated on systems analysis and precise node and link identiication and targeting. See U.S. Joint Chiefs of Staf, 2009c. 8 Some EBO advocates state that there is no diference between efects-based and traditional, or maneuver, warfare operations, and some doctrine makes a similar point. his inding relects the indings of this research efort, which recognizes that there is an unsettled debate between EBO advocates and critics. he inding addresses COIN assessment only. 9 HQDA, 2010, p. H-3. 212 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency movement and combat reporting (e.g., “We have crossed phase line x,” or “We inlicted 50 casualties”).10 his system does not allow tactical commanders to inject meaningful insight into theater-level campaign assessment. While EBA is recommended in FM 3-24, EBA also contradicts (or is contradicted by) the same manual, which states, “Local commanders have the best grasp of their situations.”11 EBA also contradicts (or is contradicted by) FM 5-0, which states, “Generally, the echelon at which a speciic operation, task, or action is conducted should be the echelon at which it is assessed.”12 Table 9.1 compares EBA to the nature of COIN as described both in U.S. doctrine and in the literature. his monograph also argues that pattern and trend analysis, while more realistic for COIN than EBA, cannot efectively support policy. When they are not based on exceptionally poor data or generally misleading, patterns can inform operational planning and intelligence analysis. hey can also inform assessment, but they cannot be central to campaign assessment because they cannot be predictive in the absence of other analyses; they do not necessarily show change over time, nor do they necessarily show a shift in conditions. None of these centralized analyses can show defensible causation to support strategic decisionmaking. Trend analysis, and particularly time-series analysis, is of greater interest to campaign assessment because it purports to show measurable change over time. But unlike pattern analysis, trend analysis tied to thresholds does require a degree of accuracy that the COIN environment cannot provide. Trend analysis that purports to deliver even “reasonable accuracy” without the use of thresholds is dependent on data control that typically does not exist in COIN. Both pattern Table 9.1 Effects-Based Assessment Compared to the Nature of Counterinsurgency Effects-Based Assessment Precise Systematic Certain Centralized Imprecise Chaotic Uncertain Distributed Conventional Irregular Effects-based Mission-based 10 U.S. Joint Chiefs of Staf, 2007, p. IV-20. 11 HQDA, 2006c, p. 1-26. 12 Nature of COIN HQDA, 2010, p. 6-1. his is particularly confusing because FM 5-0 also recommends a highly centralized EBA process that relies on a mathematical formula to produce a single assessment number for a campaign. Conclusions, Recommendations, and Options 213 and trend analysis can deliver indings that are misleading, or precise but inaccurate, but neither captures context. Pattern and trend analysis contradicts the same key statements in FM 3-24 and FM 5-0 as EBA. Neither pattern analysis nor trend analysis is suicient to provide or even to support an efective holistic campaign assessment. Centralized assessment places undue burden on subordinate units in the form of core metrics collection and reporting requirements. Centralized lists of “the right things to assess” can become more important than military objectives because quantitative indicators appear to be tangible. It is tantalizing to believe that a commander or assessment staf might be able to show measurable progress in a military campaign that is inherently chaotic and essentially immeasurable. Demand for quantitative data, or even qualitative data that have no bearing on local conditions or objectives, can inadvertently steer tactical units (e.g., an infantry battalion) toward metrics-driven operations and away from the distributed, commander-driven operations called for in U.S. military doctrine. It might be possible to overstate the efect of core metrics lists on operational behavior, but in a worst-case scenario, the military could get caught up in a metrics-driven campaign that produces copious data but not victory (as in Vietnam). While metrics-driven operations could conceivably be productive— indicators and MOEs in doctrine are supposed to be tied to objectives—it is not possible to equate core metrics indicators with military objectives in COIN. Centralized indicators cannot account for variations in tactical conditions or tactical and operational mission-type objectives. Even when commanders and stafs do not conduct metrics-driven operations, they are still faced with collecting and reporting reams of quantitative data for centralized assessment that, in their own words, provide little in the way of useful feedback or support. Collection of some types of assessment data entails physical risk, and most collection eforts require the expenditure of resources (e.g., patrol time, vehicle fuel, maintenance time, administrative time). Core metrics lists levy a “tax” on tactical units that must be paid in time, resources, and, in some cases, life and limb—or be ignored or falsiied. hese costs and risks afect tactical military units (and also PRTs) that are embroiled in complex and dangerous combat operations. If the centralized assessment process were efective, this tax might be justiiable rather than an undue burden, but it is not. A colloquial U.S. military aphorism describes the tactical-level reaction to centralized assessment: “he juice ain’t worth the squeeze.” Table 9.2 compares EBA and pattern and trend analysis to the standards for assessment introduced in Chapter One. he indings in the table relect the indings from a broad examination of both methodologies and a review of their implementation in Vietnam, Afghanistan, and (to a lesser extent) Iraq. he standards, as discussed earlier, are derived from a detailed literature review, direct observation, interviews, a study of 89 insurgency cases, and participation in the assessment process in two separate COIN campaigns. hey are intended to provide a framework for the critical analysis of existing practice and for determining recommendations and options for improvement. 214 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Table 9.2 Comparison of Centralized Assessment Methods to Assessment Standards Effects-Based Assessment Pattern and Trend Analysis Transparent No No Credible No No Relevant No No Balanced No No Analyzed No No Congruent No No Parsimonious No No Standard he table shows that neither EBA nor pattern and trend analysis compares well with the seven standards for efective COIN campaign assessment. An efective assessment process would address each of these seven standards and possibly others not considered in this study. It would provide policymakers with a relevant, transparent, and credible assessment report derived from all relevant and available data; it would provide contextual understanding of the campaign and the operating environment; it would relect trusted commanders’ insights but also internalized checks and balances; it would provide useful feedback and continuity to operational and tactical-level units; it would not place undue collection and reporting burdens on these units; it would contain dedicated and objective analysis; and it would be the product of a whole-of-government approach. If these standards were to replace the assumptions listed at the beginning of this chapter, military campaign assessment would be greatly improved. No assessment process ofers a panacea. Assessing COIN will always be a challenging and laborious process that will deliver imperfect results. Tactical units will always be responsible in some way for helping senior military leaders and policymakers understand the environment and the contextual aspects of the campaign. hese senior leaders and policymakers will never receive a complete picture from a campaign assessment; it will be incumbent upon them to build a more complete picture of the strategic situation by incorporating top-level intelligence analyses, input from trusted advisors, and other reporting into their decisionmaking processes. However, an efective campaign assessment can serve as a platform for sound decisionmaking. A transparent and credible assessment can help sustain popular support for COIN campaigns that historically have lasted about ten years and can also help policymakers identify a failing strategy. Efective assessment can help win COIN campaigns, while inefective assessment can undermine strategy. Assessment can only be efective if the assessment Conclusions, Recommendations, and Options 215 methodology accounts for the realities and complexities of war. he current system does not, but there is suicient time to put an efective process in place to help shape a successful conclusion to the war in Afghanistan. he following sections present recommendations and options for improving the assessment in Afghanistan and in the U.S. military staf planning process in general. he section titled “Options” proposes an alternative to centralized assessment derived from the seven standards for assessment. Chapter Ten presents this process in detail, while Appendixes A and B provide a template and example, respectively. Recommendations his section ofers recommendations for improving COIN assessment doctrine and for improving the military’s approach to incorporating COIN assessment into education and planning. It also recommends a review of the interagency approaches to COIN assessment. Conduct a Joint and Service Review of Military Assessment Doctrine he evidence presented here strongly suggests that the military would beneit from a review of COIN assessment doctrine, which would provide impetus for change. Although doctrine is intended to guide rather than explicitly instruct military planning and execution, a more detailed explanation of the role of assessment and the various approaches to COIN assessment would be helpful for the operational, training, and education communities in each of the services. A review of COIN assessment doctrine would also support improvements in assessment across the full spectrum of military operations (including, e.g., conventional war, noncombatant evacuation operations). Reviews of military assessment doctrine—joint, Army, and Marine Corps, at a minimum—should focus on developing a theory of assessment that better aligns with doctrinal understanding of warfare and of COIN, speciically. Once this theory is established, the review should focus on inding ways to improve assessment processes, methods, and integration. his monograph ofers a framework of standards that could be used to guide the development of a theory of assessment as well as a process and methods. However, this framework merely constitutes a starting point for a comprehensive doctrinal review. In addition, the review should address ways to meet policymaker requirements, incorporate qualitative data, capture context, minimize data collection requirements, provide holistic analysis, integrate and support commanders’ analyses, capture interagency information, and produce adequately sourced and contextual reports. 216 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Incorporate Counterinsurgency Assessment into Training and Education Some military training and education programs incorporate aspects of COIN assessment into their curriculum. For example, the Marine Air-Ground Task Force Staf Training Program instructs Marine Corps stafs on how to assess COIN as part of predeployment training packages for Iraq and Afghanistan. his monograph did not examine the military assessment training and education programs, but a review of the way in which COIN assessment is taught might be beneicial. To maximize efectiveness, training and education on assessment would be seamlessly integrated into existing training and education at all levels of staf functionality. Both a captain at a battalion staf and a colonel at a corps staf should understand the practical need for useful assessment and the underlying theories of assessment doctrine. Conduct an Interagency Review of Counterinsurgency Assessment Research conducted on Afghanistan COIN assessment between 2009 and 2011 revealed little formal and no comprehensive collaboration on assessment between the military and civilian agencies. Any efort to develop an interagency assessment process will necessarily be fraught with bureaucratic complexities, equity issues, and differences in mindset. Currently, policymakers receive separate assessments from each agency (primarily DoD and DoS). However, according to U.S. government and military doctrine, COIN should be prosecuted by a joint military-civilian team working in harmony toward a uniied strategic objective. JP 3-24 and the U.S. Government Counterinsurgency Guide are the best eforts to pull together interagency doctrine on COIN, but neither suiciently addresses assessment. Campaign assessment is mentioned only briely in one paragraph of the U.S. government guide.13 A review of the interagency assessment policy would beneit all agencies as well as policymakers. Incorporate All-Source Analytic Methodology into the Campaign Assessment Process Holistic analysis is generally absent from campaign assessment, at least in part because there is no clearly deined assessment analysis process. Campaign assessment is not intelligence analysis, but all-source intelligence methodology provides an existing and well-established framework for assessment that could be used to both improve and structure campaign assessment reports. All-source analysis for intelligence is best deined as a function of intelligence fusion: Fusion is the process of collecting and examining information from all available sources and intelligence disciplines to derive as complete an assessment as possible of detected activity. It draws on the complementary strengths of all intel13 U.S. Government Interagency Counterinsurgency Initiative, 2009, pp. 47–48. he term “assessment” is used throughout the guide in other contexts. For example, it is used to describe assessment of a prospective COIN campaign. See also U.S. Joint Chiefs of Staf, 2009c. Conclusions, Recommendations, and Options 217 ligence disciplines, and relies on an all-source approach to intelligence collection and analysis.14 he process of intelligence analysis is semistructured. In practice, an intelligence analyst is given an analytic task or a general intelligence requirement. Sometimes, the analyst can generate collection requirements (or the analyst’s command does so) to create tailored information to inform this speciic analytic task. he analyst then acquires and reads and gauges the quality and veracity of as many relevant reports as possible. his process allows the analyst to build knowledge of the subject and to frame analysis to perform the task or meet the requirement. he analyst then writes the report in a narrative format, incorporating evidence from all sources of information available, as well as iterative discussions with fellow analysts and operators. All-source analysis is particularly suited to complex and chaotic environments like COIN. It is structured enough to provide a consistent framework across an entire theater but lexible enough to respond to changes in the type and availability of data that occur on a daily basis in COIN. An experienced all-source analyst is capable of seeking out, analyzing, and incorporating any type of information that might be available, including attack data, human intelligence reports, economic data, open-source press reports, geospatial data, NGO reporting, and host-nation reporting. his kind of analysis is particularly suited to the proposed alternative to centralized assessment presented in the following chapter. However, it could also beneit centralized assessment eforts. Analytic methodology is shaped by guidance in joint doctrine, service doctrine, and intelligence community directives. All-source analysis is semistructured, but these guidelines help bound analysis and improve quality and credibility. Many of these guidelines could be easily applied to assessment. For example, Intelligence Community Directive 203 establishes “core principles of the analytic craft.” hese standards require that analysis be objective, remain independent of political considerations, be timely, be based on all available sources of intelligence, and exhibit proper standards of analytic tradecraft. hese tradecraft standards are exacting, and each of these standards could also be applied to, or modiied for, campaign assessment.15 14 U.S. Joint Chiefs of Staf, 2007, p. II-11. All-source analysis is speciically described as analysis that incorporates “all sources of information, most frequently including human resources intelligence, imagery intelligence, measurement and signature intelligence, signals intelligence, and open-source data in the production of inished intelligence.” (U.S. Joint Chiefs of Staf, 2007, p. GL-5). 15 Oice of the Director of National Intelligence, 2007, pp. 2–4. Tradecraft standards in the directive are as follows: 1. Properly describes quality and reliability of underlying sources 2. Properly caveats and expresses uncertainties or conidence in analytic judgments 3. Properly distinguishes between underlying intelligence and analysts’ assumptions and judgments 4. Incorporates alternative analysis where appropriate 218 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency he greatest beneit of applying all-source fusion methodology to assessment would be in providing a semistructured means for military stafs to create a holistic assessment rather than a disconnected set of patterns, trends, and raw data. All-source fusion analysis will not provide a formula for assessment that produces a clear quantitative inding, but this research found that the state of warfare and technology in early 2011 precludes reliance on any such formula. It also will not substitute for a layered, contextual assessment; all-source analysis, by itself, will simply provide an improved version of the lat impressionistic decision reports described earlier. here are ample resources available to help modify all-source methodology for assessment, including the Sherman Kent School at the CIA, the service intelligence schools, and other organizations that teach or apply all-source fusion methodology. he adoption of all-source fusion methodology should not preclude the application of scientiic methods or scientiic rigor to the assessment process, with the understanding that assessment is not and cannot be conlated with scientiic research. he semistructured nature of all-source analysis leaves considerable room to incorporate other methods and standards. Options his section irst provides an option for improving the existing processes and methods of military COIN assessment that, while available to policymakers and the military, is not supported by this research. It is presented here in the interest of exploring all relevant options to improve COIN assessment. he second option presented here, contextual assessment, is described in the following section and then explained in detail in Chapter Ten. Option One: Improve the Existing Centralized Assessment Process his option assumes that centralized assessment as it currently exists in doctrine and practice will be retained. No matter how assessment is changed and improved, some policy consumers will continue to require the production of time-series graphs and pattern and trend analyses. However, any improvement to centralized assessment would be marginal because it would not address the inherent inconsistency between centralized assessment and COIN. he following recommendations could improve to some extent the transparency and credibility of a centralized assessment process—either 5. Demonstrates relevance to U.S. national security 6. Uses logical argumentation 7. Exhibits consistency of analysis over time, or highlights changes and explains rationale 8. Makes accurate judgments and assessments Conclusions, Recommendations, and Options 219 EBA or pattern and trend analysis—should that process be retained. his is a suggested and not necessarily exhaustive list: • Create a single narrative theater assessment process and deliver a single report from a single theater-level staf to the relevant combatant command. • Publish this report without classiication or caveat, or at least publish a well-cited version of the report without classiication or caveat. • Provide checks and balances on the theater report by – conducting a thorough red-team, b-team, devil’s advocate, or other competing analysis16 – developing a structured alternative assessment process within the intelligence community and at a selected level in DoD above theater. • Restrict the practice of raw data dissemination outside of theater to the greatest extent possible. • Design and implement a single and broadly promulgated set of assessment standards that could be translated and applied from the policy to the tactical level. • Select or clearly describe how to select MOEs, MOPs, and indicators. hen, – clearly explain how MOEs and MOPs are tied to campaign objectives and, in turn, how they will show strategic success or failure – deine the MOEs, MOPs, or indicators in both simple terms and in great detail to ensure that all possible diferences in ield conditions are accounted for – set a method of data reporting from the tactical to the theater level for each MOE and MOP – describe how to set a baseline of data for time-series analysis (or deine the baseline) – set an efect threshold or explain in clear terms how to set an efect threshold. • Provide a simple means for recording disagreements between various levels of command to meet doctrinal standards. Incorporate these reports into the inal theater report. • Eliminate all weighting schemes from the assessment process. • Eliminate from the assessment report all color-coded maps and graphics (i.e., those in which a color equals a number, like in the ISAF DAM and MACV strategic maps). If these reports are required, attach appropriate caveats. • Provide clear and detailed citations for all data, where possible. Where speciic citations are restricted, provide a full list of citations to appropriate government consumers. • Provide a reliability rating for all data, including data sets (such as polls) that already incorporate a reliability rating from a contracted polling agency. 16 See A Tradecraft Primer: Structured Analytic Techniques for Improving Intelligence Analysis (U.S. Government, 2009). 220 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency • Ensure that data collection, reporting, transfer, and management practices deliver and sustain databases that are generally complete and accurate. Find a method to determine the completeness and accuracy of the database, if possible. Explain to consumers the degree of completeness and accuracy of data sets to the greatest extent possible. • Clearly deine and describe the analytic process that the theater staf will use to compare and contrast available data to produce the assessment report. his process should meet clear methodological standards as described by either the intelligence literature or literature from a selected set of scientiic ields. he continued existence of a quantitative, centralized process is likely to make the inclusion of other methods (e.g., all-source analysis) in theater assessment irrelevant. A combined approach will probably do little to improve transparency and credibility, as decisionmakers and the public tend to rely on what appears to be the most concise and accurate reporting. Option Two: Build a Decentralized and Bottom-Up Assessment Process If centralized assessment is untenable, then a decentralized, or bottom-up, process should be considered for COIN. he following list of assumptions should be considered as a potential underpinning for a decentralized approach to COIN assessment. hey could replace the assumptions identiied in the introduction to the section “Conclusions,” earlier in this chapter. hese assumptions are drawn from U.S. military doctrine on warfare, doctrine on COIN, the literature on COIN, and observations and interviews conducted for this study. his option is explored in greater detail in Chapter Ten. • Wars are fought in a complex adaptive environment marked by chaos, friction, disorder, and the fog of war. Largely because COIN campaigns are conducted on both physical and human terrain, the COIN environment lies on the far end of a spectrum of complexity and chaos in warfare. • A complex and chaotic environment undermines data collection and analysis and reduces the ability of a military staf to perceive the operating environment. his efect is exacerbated when staf are far removed from the battleield. • To make up for these limitations, efective COIN campaigns involve distributed, or mission-type, command and control, allowing maximum leeway in operations at the battalion level.17 • Distributed operations succeed because they are fought in ways appropriate for the local context: Tactical commanders have maximum lexibility to ind and 17 In this case, distributed operations is meant to broadly describe the kind of dispersed operations common to COIN. It is not intended to directly relect the concept of distributed operations drawn from defense transformation theory. Conclusions, Recommendations, and Options • • • • • 221 describe solutions that will work in their area of operations but perhaps not in others. Local operations and local solutions require local assessments that defy aggregation into simpliied theater assessment reports. Because COIN is complex and local assessments are not easily or efectively aggregated, theater-level assessments cannot be easily or efectively simpliied. Because COIN deies simpliication through quantitative analysis and similar analytic methods, centralized assessment is not usefully applicable to COIN. If centralized assessment is not usefully applicable to COIN, some form of decentralized assessment might be. Decentralized COIN assessment should be transparent, credible, relevant, balanced, analyzed, congruent, and parsimonious, as deined earlier in this monograph. Some members of the community of experts on Afghanistan COIN assessment believe that a bottom-driven, contextual process is called for. Freek-Jan Toevank, a Dutch operations research analyst tasked with studying assessment in Afghanistan, developed recommendations to improve COIN assessment. Toevank determined that assessments are best conducted at the lowest possible level. In addition, “Assessment should provide authoritative claims on robustness of assessments, including processes, anonymized sources, and methods.”18 According to one former military commander and Afghanistan policy analyst, here is nobody in the province that has a better understanding of what’s going on in the province than the USAID, State, USDA [U.S. Department of Agriculture], representatives, and maneuver battalion commanders. We need to keep metrics simple and as few as possible. Once you get into the nitty-gritty of what is taking place in the provinces, you need to let battalion and PRT commanders develop their own metrics that will help them develop their own assessments.19 he only way to build a bottom-up, transparent, and credible alternative to centralized assessment is to develop a process that is, in its ultimate manifestation, unclassiied and without caveat. his is an inherent risk: Transparency of process must be absolute. his does not mean that intelligence reports, sensitive operations reports, or other sensitive communications need be revealed. It is not necessary to open all the books to the public, so to speak. here are three ways to maximize transparency without compromising security. First, assessments are produced on a time cycle that allows for a thorough security review; the kinds of “running estimates” delivered on a daily or 18 Toevank, 2010. Toevank conducted empirical research over two separate periods between 2004 and 2008 in Kabul and Tarin Kowt, Afghanistan. 19 Statement by a former military commander during a conference brieing, 2010. 222 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency weekly basis through operations channels (distinct from campaign assessments) would remain at the necessary level of classiication. Second, data can be aggregated at the battalion level and still be kept essentially in context; it is not necessary to describe individual events or people in detail. hird, for data that are retained in the report, the time between data collection and assessment reporting (within the assessment cycle— at least a matter of weeks) reduces the tactical value of the information to the insurgents while retaining the operational and strategic value of the assessment. Transparency and the development of a comprehensive, written framework for assessment will contribute to the credibility of the bottom-up option. However, at least the classiied versions of bottom-up contextual assessment will have to contain precise and accurate citation that will allow senior policy stafs to identify sources of information. Reports that are both transparent and credible will be relevant if they also deliver a close approximation of ground truth. In theory, a bottom-up approach might fulill this need more consistently than a centralized approach. Bottom-up assessment should relect all types of data in context, it should be congruent with the fundamental understanding of warfare and COIN as expressed in literature and doctrine, and it should be parsimonious to the greatest extent possible. he following section examines the practicality of bottom-up assessment as it relates to the development of accurate reporting. Practicality of a Bottom-Up Approach Justiication for the development of a bottom-up contextual assessment process rests on the notion that some level of assessment can and should be produced by battalion stafs. In current practice, battalion commanders periodically write a narrative assessment of their battlespace (sometimes, the intelligence or operations summary suices), while their stafs routinely send up brief reports and often reams of raw quantitative and qualitative data to higher headquarters.20 Interviews revealed an impression at some higher military and policy staf levels that battalion commanders write their narratives based on their personal opinions and that data play little or no role at this level.21 In reality, battalion commanders are immersed in data on a daily basis. hey receive a steady stream of reports on violent incidents, civil afairs actions, key leader engagements, human intelligence collection, biographies, friendly operations, ANSF partnering, local economics, and local NGO operations. Proicient commanders also 20 For example, infantry atmospherics reports from local villages, key leader engagement reports, civil afairs reports, speciic and aggregated attack reports, operations reports on friendly movement and actions, progress reports on the development of local security forces, intelligence information reports, and so on. 21 his was not a universal assumption at IJC or ISAF, and, in fact, the IJC commander clearly wanted to rely on commanders’ narrative assessments. However, the IJC assessment process rests on the concept that subordinate commanders’ assessments must be checked against an objectives metrics list. IJC uses the District Assessment Framework Tool to conduct a quantitative check on subordinate assessments. Conclusions, Recommendations, and Options 223 conduct frequent battleield circulation: hey go out and see irsthand what is happening in their battlespace. Good commanders are therefore able to put data in immediate and relevant context. A battalion staf is made up of ield- and company-grade oicers and senior NCOs who are similarly immersed in data.22 hese personnel typically have the same type of training and experience as those of the same rank assigned to the theater staf. Efective battalion commanders analyze data in context, incorporate the advice of their stafs, take into account personal observations gathered from battleield circulation, and then write their assessments. hus, battalion assessments are also holistic assessments. Because commanders are both interpreting data and reporting on their own performance, some level of subjectivity is always present. But because the battalion-level assessment incorporates staf input and data analysis, it relects a blend of all types of information. Because battalion stafs should be more capable of putting data in context than theater-level stafs (“Local commanders have the best grasp of their situations,” according to FM 3-24), battalion stafs are better positioned than the theater staf to assess COIN in their area of operations. In combat, battalion stafs tend to be shorthanded, overworked, and likely to focus only on those tasks that are going to prove useful to their primary mission. hey are unlikely to adopt a complicated assessment reporting format that requires the collection and reporting of data that are not relevant to their operations or area of operations. Nor are they likely to put efort into iling narrative and contextual reports that do not efectively inform their immediate superiors about ongoing situations and near-term requirements. However, they tend to be both capable of and willing to produce reports that make sense to them and that show an impact on their near-term and long-term situations. Battalion stafs are accustomed to illing out narrative operations reports and narrative intelligence reports that explain what has happened in their area and why, and what is likely to happen in the near term. hey are also accustomed to producing requests for resources in narrative format. herefore, battalion stafs are well positioned to write narrative assessment reports that make sense to them, that are relevant to their mission, and that inform their immediate superiors (brigade or regimental staf and the commander) about their requirements. A bottom-up, contextual assessment can capitalize on this existing predilection and capability. 22 his synopsis of battalion-level staf functions is based on my experience with an infantry battalion while serving in the Marine Corps, service on an infantry regiment staf as the assistant operations oicer, debrieings of infantry battalion commanders, observations of battalion stafs in combat over three tours in Iraq, and interactions with battalion stafs during peacetime training during more than 21 years of active military service. 224 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Deflation, Inflation, and Inherent Subjectivity he advantage of context at the battalion level does not mitigate the very real problem of delation and inlation of assessments. Although I am not aware of any published empirical information to support this assumption, there is a strong sense among assessment personnel that commanders tend to assess events negatively when they irst arrive in theater and in a better light when they leave. his dynamic is not necessarily disingenuous; it could simply relect increasing levels of knowledge and comfort with local leaders and events over time. In many cases, the improvement might relect reality.23 However, a senior military oicer with experience in Afghanistan stated, When you come in to command, you’re ired up and ready to go. I think that during the irst two months . . . when you’re trying to igure out what’s going on you tend to temper your expectations. I’ve never heard of a gaining commander looking back and saying, “his guy has it all igured out, I don’t need to make any changes.”24 Figure 9.2 is a simpliied and somewhat ampliied depiction of this efect: he assessment tends to be negative when a staf irst arrives, then rises as the deployment continues. When units perform a relief-in-place transfer-of-authority, the new commander in this notional scenario arrives with a negative bias and then follows the same assessment trajectory as the previous commander. While the widespread belief in a Figure 9.2 Subjectivity in Commanders’ Assessments Better Better Worse Relief in place/ transfer of authority Better Worse Relief in place/ transfer of authority RAND MG1086-9.2 23 his efect was recognized by the IJC and AAG assessment stafs, by members of a 70-person working group on assessments in March 2010, and in many other gatherings of experts. 24 Statement by senior military oicer, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. Conclusions, Recommendations, and Options 225 Manichean division between subjective and objective assessment described throughout this monograph is false, it is fair to say that a single assessment written by a commander and staf with a vested interest in the outcome cannot be considered strictly transparent or credible. It is not enough to trust battalion, regimental, or even regional assessments without additional analysis. he proposed alternative to centralized assessment explored in Chapter Ten is designed to address subjectivity in assessment reporting, but no assessment process can ever eliminate subjectivity. While command outlook and perspective may change considerably during a reliefin-place/transfer-of-authority period, the functional aspects of assessment are likely to relect continuity. Typically, new units pick up and continue the practices of departing units for a period—say, one or two months—before implementing major changes in process and operations.25 his is not always the case, but the saw-tooth changes in subjective opinion relected in Figure 9.2 should not necessarily be conlated with a similarly abrupt change in assessment methods and practice. 25 his analysis is based on direct observation of numerous such transitions in Iraq between 2003 and 2006, as well as discussions with U.S. military oicers on the subject between 2003 and 2011. CHAPTER TEN A Proposed Alternative to Centralized Assessment his chapter explores in greater detail the alternative to centralized COIN campaign assessment proposed earlier in this monograph. Tentatively called contextual assessment, or CA, this concept was developed over the course of this research efort to meet the speciic needs of COIN campaign assessment; it is not necessarily appropriate for other types of military assessment. he process presented here is derived from a number of sources reviewed for this study and is framed on existing models of military reporting. Speciically, the model is based in the description of warfare found in MCDP 1, Warighting, and FM 3-0, Operations, as well as in the literature on the nature of warfare, such as Carl Von Clausewitz’s On War.1 It is also based on the description of COIN in JP 3-24, Counterinsurgency Operations, FM 3-24, Counterinsurgency, and the literature on COIN.2 he CA model is intended to address the seven standards for assessment identiied in this monograph (see Chapter One), and it is speciically designed to capture relevant context in a layered format that also relects critical internal analysis. he speciic framework is based on a survey of both the operational and intelligence reporting formats used by the U.S. military in Vietnam, Iraq, and Afghanistan. hese formats include daily operational reports iled by operations stafs in infantry units, intelligence summaries iled by intelligence stafs, advisor reports (see Appendix C), operational narratives produced at operational-level military stafs, and intelligence analysis fusion products developed by operational and theater-level intelligence stafs. he process described here is intended to it within the observed “battle rhythm” of a contemporary U.S. military force deployed to execute a COIN campaign in a coalition environment. he template and example provided in Appendixes A and B, respectively, are framed to generally align with the format and style of writing identiied in the survey of operational and intelligence reports. he remainder of this chapter describes the step-by-step process for COIN CA in Afghanistan or for a prospective COIN campaign. CA is a framework intended to 1 See Headquarters, U.S. Marine Corps, 1997b; HQDA, 2008; and Clausewitz, 1874. 2 See U.S. Joint Chiefs of Staf, 2009c, and HQDA, 2006c. 227 228 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency draw from the traditionalist approach to COIN—what this approach would imply for assessment to better align assessment with capstone military doctrine and COIN operational doctrine. his chapter also further reines the concepts of both CA and mission-type metrics (MtM). MtM are a subset of CA or a tool intended for use in such an assessment. It is predicated on the idea that there are no “good” or “bad” metrics, only contextualized metrics. In COIN, it is not possible to have a centralized list of the “right” things to count, as the Commander’s Handbook suggests.3 Proposed deinitions are as follows: • Contextual assessment (CA): A transparent, bottom-up method of assessing progress in a COIN campaign that captures embedded contextual data—both qualitative and quantitative—at each layer of assessment, from the tactical to the strategic. • Mission-type metrics (MtM): Relevant qualitative and quantitative indicators that exist,4 can be collected, and can be contextualized at the level at which they are collected. MtM are indicators that show progress immediately against mission or campaign objectives. MtM are identiied by stafs at each level of assessment for each assessment period; they may or may not carry over from one assessment period to the next. MtM have value only in one place and only as long as they provide contextual insight into the success or failure of the COIN campaign. hey are not intended to feed a centralized core metrics list in COIN or other irregular warfare missions that depend on distributed operations. Mission-type is drawn from mission-type tactics or mission-type orders (the same as “mission orders”), the process by which commanders issue lexible guidance to subordinates, particularly in distributed operations.5 In this process, commanders and stafs at each level of command would view MtM as an intelligence analyst would view intelligence sources: hey are either available or unavailable, reasonably or unreasonably collectible, and relevant or not relevant. his lexibility allows the commander and the assessment staf to look at all available information without prejudice and without the need for unnecessary or onerous data collection and reporting at the company level. 3 “In today’s increasingly dynamic operating environment, [commanders] can only gain suicient situational awareness and adapt their current operations, future operations, and future plans if the staf is assessing ‘the right things’ in the operational assessment” (Joint Warighting Center, 2006, p. IV-18). 4 he stipulation to collect only those indicators that exist is intended to ward of what Downes-Martin refers to as “promiscuous metrics collection.” 5 “In the context of command and control, also called mission command and control. Mission tactics involves the use of mission-type orders. Mission-type order: ‘Order to a unit to perform a mission without specifying how it is to be accomplished.’ (Joint Pub 1-02)” (Headquarters, U.S. Marine Corps, 1997b, p. 105; emphasis in original) A Proposed Alternative to Centralized Assessment 229 While any and all information should be considered for assessment and analysis, commanders are free to dictate the consistent collection of like information over time at their level of responsibility. Local commanders are best positioned to direct the collection of information over time for several reasons: (1) they understand the immediate cost and risk of that collection; (2) they and their stafs can analyze that information in context; and (3) they can adjust collection and reporting to meet current local conditions and context. If a battalion commander determines that it would be valuable to know how many attacks were reported over time in the battalion’s area of operations, that commander could assess the cost and risk of collection with input from immediate subordinates. If conditions changed, or it became obvious that the information was no longer relevant, the commander could quickly adjust the collection eforts. here is already a similar process in place for the collection and reporting of mission-relevant information. Both joint and service doctrine calls for CCIR collection and reporting to feed operational planning and decisionmaking. Such information is necessary for commanders to efectively prosecute their operations and accomplish their missions. he requirements are lexible and change as often as necessary to meet changing conditions in the ield. According to JP 5-0, he Operations Process: CCIRs comprise information requirements identiied by the commander as being critical to timely information management and the decision-making process that afect successful mission accomplishment. CCIRs result from an analysis of information requirements in the context of the mission and the commander’s intent. he two key subcomponents are critical friendly force information and priority intelligence requirements. he information needed to verify or refute a planning assumption is an example of a CCIR. CCIRs are not static. Commanders reine and update them throughout an operation based on actionable information they need for decisionmaking. hey are situation-dependent, focused on predictable events or activities, time-sensitive, and always established by an order or plan.6 Joint doctrine already closely links CCIRs with assessment. JP 5-0 states, “he CCIR process is linked to the assessment process by the commander’s need for timely information and recommendations to make decisions.”7 But while joint doctrine recommends a commander-centric, localized, mission-focused, and lexible data collection process for assessment, it goes on to recommend the application of a highly centralized, rigid, and process-centric assessment analysis framework: EBA. he irst part of the joint doctrinal recommendation for information collection and reporting—the description of CCIRs—is congruous with COIN doctrine, is broadly accepted by 6 U.S. Joint Chiefs of Staf, 2006, p. III-27 (emphasis in original). See also HQDA, 2010, p. B-9. CA would expand the concept of CCIRs to encompass all aspects of the campaign, including civilian and economic reporting. 7 U.S. Joint Chiefs of Staf, 2006, p. III-57. 230 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency the military operations and intelligence community, and is also generally accepted and used by tactical units. It would seem logical either to create an assessment collection and reporting process that is analogous to CCIRs or to use the CCIR process to feed assessment. he MtM process could be used in addition to CCIRs, it could be used to modify CCIRs, or it could be deemed suicient to meet assessment requirements without the addition of a new process. Since, by deinition, CCIRs are designed to collect and report data that “afect successful mission accomplishment,” either the second or third option would conceivably be suicient to meet assessment information requirements. CCIRs provide a means for local commanders to shape their own collection requirements and activities, but, in practice, they also tend to relect the CCIR requirements of superior commanders. (For example, a battalion CCIR list might contain CCIRs from a brigade or regimental commander.) his could conceivably provide a pathway for the reintroduction of core metrics for assessment. It might be necessary to separate MtM from CCIRs to ensure that the MtM primarily relect local context. Because CA is an untested, hypothetical process, it would be necessary to resolve these kinds of detailed procedural issues during iterative development. Preferably, no standing metrics requirements would be issued even at the tactical level: Assessments would be fed with available information and with limited, short-duration collection requirements that addressed concerns most relevant to the situation at that time. By dictating the use of bottom-up CA and MtM, the theater command would be exercising prototypical maneuver warfare command and control. In the framework presented in this monograph, CA would begin at the battalion level of command, the irst level at which a cohesive military staf exists (although input from the platoon and company levels could easily be incorporated in formal reports). he assessment would build up from the battalion to the theater level in a step-by-step process. Every step of the process builds a layered assessment designed to ultimately address theater-level strategic objectives. For convenience, this chapter adopts the terminology used to describe current conditions in segments of a COIN theater similar to Afghanistan: districts, provinces, and regions. In other campaigns, these levels may not be directly analogous, but the idea of layered assessment should still hold true. he CA process assumes that mission analysis, planning, and deployment are complete and that clear strategic objectives exist. Each commander will have determined both speciied and implied objectives and tasks and will have written a clear mission statement, and commander’s intent. In other words, everyone involved will understand his or her role in the assessment process. he theater command will have issued its overall instructions for the CA process (including reporting periods), and thus each battalion, while retaining contextual lexibility, will ile reports that are comparable in format. A Proposed Alternative to Centralized Assessment 231 he CA process is founded on the inding that COIN assessment should be bottom-up, and should relect local context and layered analysis. Figure 10.1 presents the outline of a layered contextual assessment process for COIN.8 his building-block approach is relected in the following step-by-step process. Appendix A ofers a series of templates for CA that can be modiied for use at each level of assessment and reporting. Figure 10.1 Contextual Impressionistic Decisionmaking Data Contextual report Data Data Data Data Data Report Data Second-level Contextual report Data Data Contextual report Data All-source analysis Data Data Contextual report Data Data Data Report Data Theater commander’s assessment Contextual Impressionistic Decision Report Data Report Data RAND MG1086-10.1 The Contextual Assessment Process Step 1: Form an Assessment Working Group he battalion staf (and staf through the RC level) establishes an assessment working group (AWG).9 At the battalion level, this is a principals’ working group made up of (at least) the intelligence oicer, operations oicer, logistics oicer, plans oicer, communication oicer, civil afairs oicer (if available), and host-nation security force oicer (or equivalent, if available). Because the working group comprises battalion staf, these oicers should already interact regularly during the course of normal business. It should be easier to provide incentive to key staf to participate in assessments if they control the assessment process at their level of efort and have a clear role in shaping 8 To see the striking contrast with traditional centralized assessment, see Figure 9.1 in Chapter Nine. 9 HQDA, 2010, pp. 6-9–6-10. 232 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency the theater-level assessment. he battalion commander should choose a leader for the AWG to ensure consistent recording and information management. If necessary, the AWG could consist of lower-level oicers, but at some point, a principal’s group should meet to review and discuss the assessment. he battalion commander should carefully choose a staf member to write the report. his person should have good writing ability and, if possible, should have a holistic perspective of the battlespace. In some cases, the commander will want to write the report. At the regiment/brigade and RC levels, the commander might choose a civilian advisor (e.g., the DoS representative) to write the report. hese advisors tend to have a more objective view than some of their military counterparts and are typically trained to write narrative reports.10 It is necessary to keep in mind that because the AWG consists of staf members who work full time for the commander, they are, in essence, grading their boss’s performance. his dynamic exists both in CA and in EBA. Superior commanders and the theater analytic team will have to keep this relationship in mind as they analyze subordinate assessments. Step 2: Shape the Mission-Type Metric Requirements and Acquire Information he commander and staf determine collection requirements for assessment in a structured, step-by-step process designed to keep the collection process parsimonious and targeted to the battalion mission. he irst step would be to conduct a mission analysis and an operational analysis to determine which types of information might inform mission success. his list would relect both mission requirements and the context of the operational environment (i.e., not only what is important but also what matters here and now). he next step would be to determine the availability of existing information through standard reporting (e.g., logistics reports, intelligence reports, civil afairs reports) to eliminate redundancy in collection. he third step would be to develop a draft list of collection requirements tied to the mission and the situation. he fourth step would be to iterate this draft with staf sections at the battalion, as well as with company commanders, to determine feasibility, cost, and risk of collection for each item on the list. he ifth step would be a commander’s review of the list and implementation. Once the list is in place, it would be up to staf sections and subordinate commanders to ind appropriate ways to produce the information. his information would be collected at the battalion level, along with all other relevant and available information. his process would be repeated at least once per reporting cycle (e.g., monthly) and also as often as deemed necessary by the commander. he commander retains the prerogative to change the MtM list (like the CCIR list) at any time without going 10 he recommendation to assign the writing task to the civilian advisor (when one is available) was ofered by Colin F. Jackson of the U.S. Naval War College at a working group of assessment analysts at the U.S. Naval War College in Newport, Rhode Island, on October 19, 2010. A Proposed Alternative to Centralized Assessment 233 through this process, but establishing a pattern of staf analysis over time would help reine these changes once they were made. However, the commander and staf should keep in mind that it is not necessary to create a structured MtM list. Instead, the staf could use available information from operations and intelligence reports based on existing CCIR requirements, as well as information from miscellaneous reporting (e.g., a one-time key leader engagement report), to build the assessment. his would allow for greater lexibility at the company and platoon levels, and would preclude situations in which information requirements might unintentionally drive operations in a direction that did not support the commander’s intent. Step 3: Write the Battalion-Level Assessment he AWG meets and discusses the available data and the current situation and then briefs the commander on its initial assessment. If possible, company commanders and leaders of attached organizations (e.g., a USAID representative or a special operations team) should participate in the AWG meeting to help the AWG ascertain ground truth. he commander can then write an assessment or have the AWG or a designated staf oicer write it for the commander’s signature; each commander must have leeway in this internal process. here are no limits to the length of the battalion-level assessment because the document is intended to capture contextual detail, but it must include, at a minimum, the following elements:11 • a contextual narrative that incorporates both quantitative and qualitative information, explaining how and why the information shows progress or lack of progress toward the battalion’s mission • a justiication for the assessment that includes information about the battalion’s context and explains the thought process behind the commander’s assessment (in a form written or approved by the commander). his second element also provides an opportunity for the commander to ofer predictive analysis similar to that ofered in operations summaries.12 Step 4: Assess Battalion Assessments at the Brigade or Regiment Level he process of forming an AWG is replicated at each level of command. Each battalion sends its assessment—containing all required contextualized information and justiications—up to the AWG at the brigade (Army) or regiment (Marine Corps) 11 At this stage, it might be appropriate to place distribution restrictions on the document to allow for higherheadquarters security review. Eventually, the full report should be released without caveat. 12 he battalion-level assessment could also include written input from companies, if available and applicable, and whole informational reports if necessary and appropriate to support the contextual narrative. 234 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency level.13 he AWG reviews each battalion assessment and then writes a mandatory assessment of the assessment. his gives the higher commanders and stafs an opportunity to agree or disagree with the assessment, expound on the assessment using additional information, and provide any other pertinent comments. Step 5: Shape the Brigade or Regimental Mission-Type Metrics Brigade and regimental AWGs follow the same process as the battalion AWG in selecting MtM. hey incorporate only information that (1) exists,14 (2) can be collected at reasonable cost and risk in relation to the urgency of the requirement, (3) is relevant to the brigade or regiment, and (4) can be contextualized. Contextualizing information becomes increasingly challenging as MtM lists are established at successive layers of command. Stafs and commanders at these higher headquarters have to absorb and understand at least twice, and often three or four times, as much information as at the battalion level, and battleield circulation becomes more problematic and less contextual. MtM information at this level might include NGO reporting, PRT reporting, or other information that might not exist at the battalion level; this remains in the hands of the AWG. Brigade and regimental collection assets might be used to obtain some of this information. Commanders at this level should also keep in mind that it is not necessary to create a formal MtM list. Because brigades and regiments have more robust stafs than battalions, this is an appropriate level at which to implement an all-source assessment analysis process. If implemented in accordance with intelligence literature and doctrine, the process should function with information available from CCIRs and miscellaneous reporting, supplemented by some directed MtM collection. his process should preclude the need for structured assessment information requirements that might unintentionally replicate core metrics. Step 6: Write the Brigade or Regimental Assessment his process also parallels the battalion process. All requirements that apply at that level apply at each successive level of command. However, an additional requirement applies: All battalion assessments, in their entirety, must be included in the single assessment document that is sent to higher headquarters. he document must also include the written assessment of each battalion’s assessment report, or the overall brigade or regimental assessment should address diferences. Some intelligence stafs in Afghanistan (e.g., RC Southwest) already embed subordinate assessments. hus, incorporating subordinate reports at this stage relects existing practice. 13 Or NATO equivalent. 14 Or that can be collected with assets at this level, depending on the commander’s judgment. A Proposed Alternative to Centralized Assessment 235 Step 7: Repeat the Process at the Regional Command Level Each step of this process is repeated at the RC level with the same stipulations. he RC-level assessment that is sent to the theater level will contain each battalion and brigade/regimental assessment from the RC area of operations, all justiications, all assessments of assessments, and any whole information reports that might prove critical in supporting contextual analysis. At no point is any information separated from the single holistic assessment unless it is deemed a security concern. Removal of information from subordinate reports should be documented in the inal serialized theater report. At the RC level and above, the assessment should routinely incorporate input from nonorganic military units and civilian organizations in the region. hese inputs might come from PRTs, special operations units, nongovernmental aid organizations, host-nation elements, and coalition partners. It would be preferable to provide the entire text of these inputs in the inal report because it may be diicult to retrieve nonorganic data once the report is submitted (e.g., PRT reports might never be serialized or retained in military databases). he AWG should include a brief description of each report, its valuation of the report, and its analysis of how the report its into the holistic assessment. Step 8: Conduct the Theater Review he assessment staf reads and reviews all battalion, brigade/regimental, and RC assessments, as well as all additional information contained in the single document passed up from each RC.15 Because, in this construct, the assessment staf is not overly burdened with producing brieings or slides and has the opportunity not only to read operations and intelligence reporting but also to travel in theater, staf members are well positioned to analyze the subordinate assessments.16 Analysis is the assessment staf’s primary purpose. It must be able to leverage its knowledge and general familiarity with each region of the theater to be able to see gaps or weaknesses in subordinate assessments. 15 Approximately 40–60 oicers and NCOs from all staf sections and varied specialties would make up the theater-level assessment staf (they should fall immediately under the director of assessments or equivalent). Each staf section would have a permanent representative on the assessment staf, and principal staf oicers would make up the theater-level AWG (separate from the assessment staf). Members of the assessment staf would have carte blanche to travel within theater, and no fewer than 50 percent of them should be on the road at any one time. he purpose of the assessment staf visits would not be to inspect subordinate commands but to engage with subordinate stafs and gain irsthand context and understanding of battalion, brigade/regiment, and RC-level assessment processes. 16 It may be not necessary to produce additional reporting to feed the theater assessment provided that subordinate assessments contain adequate information and the staf reads the daily operations and intelligence summaries that are produced across the theater. 236 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Step 9: Write the Theater Analysis After conducting the theater review, the theater-level assessment staf writes a theaterlevel analysis of the entire assessment. his report does not need to address each battalion report individually, but it should address the overall situation and any signiicant diferences with subordinate assessments. he theater analysis report does not supplant any subordinate reports; it is added as the next layer in the single assessment document. he report should be framed on all-source analysis methodology. he theater analysts would have a thorough, layered, and contextual report upon which to focus their analysis. hey would conduct all-source analysis to provide a separate and perhaps competing view. his analysis would rely on additional reporting not covered in the assessment (e.g., host-nation institutional reporting, theater-level special operations reporting, coalition reporting) and on the reports cited in the assessment. he analysts would attempt to ascertain bias in the selection of sources and the interpretation of data, as well as in the commanders’ assessments. his need not be a contentious process, and it would be most successful if the analysts were able to travel to subordinate commands to engage with commanders or at least able to ask questions of subordinate stafs. Step 10: Write the Theater Commander Assessment Once the theater-level analytic team has produced its report, the theater commander writes an assessment of that assessment explaining the theater-level “theory of assessment” points of agreement or disagreement with subordinates and the theater analysis. he commander then ofers an overall assessment of the campaign in narrative format. here should be no limit to the length of this narrative; each campaign ofers varying levels of complexity, and each commander must be given the leeway to write as much or as little as necessary. Step 11: Deliver the Report After writing the assessment of the theater-level assessment, the commander signs of on the inal version of the overall assessment. his report includes each battalion, brigade and regimental, and RC assessment; the theater analysis; and the theater commander’s own assessment. Quantitative data are embedded in narrative text at each level of assessment, and all graphics are clearly contextualized at the level at which the information was collected. he report is then delivered to higher headquarters. It will eventually form the backbone of the national-level assessment that is delivered to policymakers and the public. A Proposed Alternative to Centralized Assessment 237 Aggregation of Contextual Reports and Layered Summaries Figure 10.2 shows the layering of reports from battalion to theater level. It depicts the theater analytic assessment as covering all reports up to the commander’s. At the irst level, the battalion report is written and sent to the brigade or regiment. At that level, four battalion reports are consolidated and then sent up to the RC with the brigade or regiment report attached. his process is repeated from the regional to the theater level, where the theater report, theater analysis reports, and commander’s assessment are written. he igure assumes a notional number of battalions per brigade or regiment, brigades or regiments per region, and regions within the theater. Figure 10.2 The Contextual Assessment Reporting Process Battalion report Brigade/ regiment report Regional report This assessment addresses the period This assessment addresses the period between 01 April and 01 May May. Over the past month a great deal of progress has been made towards achieving the mission. This is specifically shown in the reduction of violence that has been directly tied to a successful amnesty program and recent combat operations through intelligence reporting. The economic situation in the area has improved significantly over the past month, as shown in the growth of local markets that has been directly tied to increase in cross-border trade, as identified in civil affairs reporting. This information also shows increased growth Theater report This assessment addresses the period This assessment addresses the period This assessment addresses the period 01 April and 01 May. Over the between between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the been made towards achieving themade This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reductionpast of violence has been month athat great deal of progress has reduction of violence that has been directly tied to a successful amnesty been made towards achieving the the period This assessment addresses directly tied to a successful amnesty program mission. and recent combat operations This is specifically shown the Over the between 01 April and 01in May. through intelligence The has program and recent combat operations reductionpast ofreporting. violence been month athat great deal of progress has economicdirectly situation in the area has through intelligence reporting. The tied tomade a successful amnesty been towards achieving the improvedprogram significantly over the past operations and recent combat economic situation in the area has mission. This is specifically shown in the month, asthrough shown in the growth of local The intelligence reduction ofreporting. violence that has been over the past improved markets that has been directly tied to significantly economic situation in to the has amnesty directly tied a area successful increase in cross-border trade, asover month, as in the growth of local improved significantly theshown past operations program and recent identifiedmonth, in civil as affairs reporting. Thiscombat shown in the growth of has localbeen markets that through intelligence reporting. The directly tied to information also shows increased growth markets that has been directly tied to economic situation in in the area has increase cross-border trade, as increase in cross-border trade, asover the past improved significantly identified in civil reporting. This identifiedmonth, in civil as affairs reporting. This affairs shown in the growth of local information also shows increased growth information also shows increased growth This assessment addresses markets that has been directly tied to This assessment addresses the period 01 April and 01 May. Over the between between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the been made towards achieving themade This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reductionpast of violence that has been month a great deal of progress has reduction of violence that has been directly tied tomade a successful amnesty been towards achieving the the period This assessment addresses directly tied to a successful amnesty program mission. and recent combat operations This is specifically shown the Over the between 01 April and 01in May May. through intelligence The has program and recent combat operations reductionpast ofreporting. violence that been month ahas great deal of progress has the period This assessment addresses economicdirectly situation in to the through intelligence reporting. The tied a area successful amnesty been made towards achieving the between 01 April and 01 May. Over the improvedprogram significantly over the past and recent operations economic situation the hashas mission. Thiscombat isof specifically shown in the past month a greatin deal of area progress month, asthrough shown in the growth local intelligence reporting. The reduction of tied violence that towards has beenachieving been the past improved significantly over the markets that has been directly to made economic situation in to the has amnesty directly tied a area successful mission. This is specifically shown in the increase in cross-border trade, asover month, as in the growth of local improved significantly theshown past operations program and recent reduction of violence that has been identifiedmonth, in civil as af affairs reporting. Thiscombat shown in the growth of has localbeen markets that directly tied to through intelligence reporting. The directly tied to a successful amnesty information also shows increased growth markets that has been directly to has economic situation in tied the area increase in cross-border trade, as program and recent combat operations increase in cross-border trade, as improved significantly the pastreporting. through intelligence The This identified inover civil reporting. identifiedmonth, in civil as affairs reporting. This affairs shown in the growth ofin local economic situation the area has information also shows increased growth information also shows increased growth markets that has been directly tied toover the past improved significantly increase in cross-border month, astrade, shownasin the growth of local identified in civil affairsthat reporting. This markets has been directly tied to information also showsin increased growth increase cross-border trade, as identified in civil affairs reporting. This information also shows increased growth the period increase in cross-border trade, as between 01 April and 01 May. Over the This assessment addresses identified in civil affairs reporting. This the period This assessment addresses the period past month great information also shows increased growth between 01 April and 01 a May. Overdeal the of progress has between 01 April and 01 May. Over the made towards the month a great deal of progressachieving has past month a great deal ofpast progress hasbeen been towards achieving the been made towards achieving themade mission. This is specifically shown in the This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reduction of violence that has been reductionpast of violence that has been month a great deal of progress has reduction of violence that has been directly tied tomade a successful amnesty been towards achieving the the period directly tied to a successful amnesty This assessment addresses directly tied to a successful amnesty program mission. and recent combat operations This is specifically shown the Over the between 01 April and 01in May May. through intelligence The has program and recent combat operations program and recent combat operations reductionpast ofreporting. violence been month athat great deal of progress has economicdirectly situation in to the has through intelligence reporting. The tied a area successful amnesty through been made towards achieving the intelligence reporting. The improvedprogram significantly over the past operations and recent economic the area has mission. Thiscombat isofspecifically shown in in the month, asthrough shown in the growth localsituation economic intelligence The has been situation in the area has reduction ofreporting. violence that improved over the past markets that has been directly tied to significantly economic situation in to the has amnesty significantly over the past improved directly tied a area successful increase in cross-border trade, as month, as shown in the growth of local improvedprogram significantly over the past operations and recent identifiedmonth, in civil as af affairs reporting. Thiscombat shown in the growth month, of has localbeen as directly showntied in the markets that to growth of local through intelligence reporting. The information also shows increased growth markets that has been directly to has economic situation in tied the area increase in cross-border trade, that hasasbeen directly tied to increase in cross-border trade, markets as improved significantly the past identified inover civil reporting. This trade, as identifiedmonth, in civil as affairs reporting. This affairs increase in cross-border shown in the growth of local information also shows increased growth information also shows increased growth markets that has been directly tied to in civil affairs reporting. This identified increase in cross-border trade, as information shows increased growth This assessment addresses the period identified in civil affairs reporting. This also This assessment addresses the period information also shows growth between 01 increased April and 01 May. Over the between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the been made towards achieving themade This assessment addresses the period mission. This is specifically shown the This isthe specifically shown in the between 01 Aprilmission. and 01in May. Over reductionpast of violence has been month athat great deal of progress has reduction of violence that has been directly tied tomade a successful amnesty been towards achieving the the period This assessment addresses directly tied to a successful amnesty program mission. and recent combat operations This is specifically shown the Over the between 01 April and 01in May. through intelligence The has program and recent combat operations reductionpast ofreporting. violence been month athat great deal of progress has the period This assessment addresses economicdirectly situation in to the has through intelligence reporting. The tied a area successful amnesty been made towards achieving the between 01 April and 01 May. Over the improvedprogram significantly over the past and recent combat operations economic situation in the hashas mission. This is specifically shown in the past month a great deal of area progress month, asthrough shown in the growth of local The intelligence reduction ofreporting. violence that towards has beenachieving been the past improved significantly over the markets that has been directly tied to made economic situation in to the has amnesty directly tied a area successful mission. This is specifically shown in the increase in cross-border trade, asover month, as shown in the growth of local improved significantly the past operations program and recent combat reduction of violence that has been identifiedmonth, in civil as affairs reporting. This shown markets in the growth of has localbeen that tied to through intelligence reporting. The directly directly tied to a successful amnesty information also shows increased growth markets that has been directly tied to economic situation in the area has combat increase in cross-border trade, as program and recent operations increase in cross-border trade, asover the past improved significantly through reporting. The This identified inintelligence civil reporting. identifiedmonth, in civil as affairs reporting. This affairs shown in the growth ofin local economic situation the area has information also shows increased growth information also shows increased growth markets that has been directly tied toover the past improved significantly increase in cross-border month, astrade, shownasin the growth of local identified in civil affairsthat reporting. This markets has been directly tied to information also showsin increased growth increase cross-border trade, as identified in civil affairs reporting. This information also shows increased growth Theater-level analysis This assessment addresses the period 01 April and 01 May. Over the between between 01 April and 01 May. Over the month past month a great deal ofpast progress has a great deal of progress has been towards achieving the the period been made towards achieving themade This assessment addresses This assessment addresses the period mission. This is assessment specifically shown the This addresses the period This isthe specifically shown in theOver the between 01 Aprilmission. and 01in May. Over between 01 April and 01 May. reductionpast of violence has been between April and 01 Overhas the month01athat great deal ofMay. progress reduction of month violence that has been directly tied tomade a successful amnesty past month a great deal ofpast progress has a great deal of progress has been towards achieving the directly tied tomade aThis successful amnesty program mission. and recent combat operations been towards achieving the the period been made towards achieving the assessment addresses This is specifically shown in the through intelligence reporting. The has mission. This is assessment specifically shown in the addresses the period program and combat operations mission. This is specifically shown in theOver the reduction ofThis violence that beenrecent between 01 April and 01 May. economicdirectly situation inof the has reduction violence that has been between 01 April and 01 May. Over the through intelligence reporting. Thehas tied to a area successful amnesty reduction of month violence been past a that great deal of progress has improvedprogram significantly over the past directly tied to a successful amnesty past month a great deal of progress has and recent combat operations economic situation intomade the area has amnesty directly tied a successful month, asthrough shown in the growth of local program and recent combat operations been towards achieving the been made towards achieving the intelligence reporting. The This assessment addresses the period improved significantly over the past markets that has been directly tied to through intelligence reporting. The shown mission.in This isarea specifically in the program and combat operations This isthe specifically shown in the economic situation the has between 01 Aprilmission. and 01recent May. Over increase in cross-border trade, as economic situation in the area has reduction of violence has been month, as in the growth of local improved significantly over theshown past through intelligence reporting. Thehas been past month athat great deal of progress has reduction of violence that identifiedmonth, in civil as affairs reporting. improved significantly over theof past directly toThis a successful shown in tied the growth localamnesty been made towards achieving the markets that has been directly tied to has amnesty economic situation intothe area This assessment addresses the period directly tied a successful information also shows increased growth month, as program shown in and the growth of local recent combat operations markets that has been directly tied mission. This is to specifically shown the Over the in cross-border trade, as between 01significantly April and 01in May. improved over the past operations marketsinthat hasincrease been directly tied toprogram through intelligence The has and recent combat increase cross-border trade, as reduction ofreporting. violence been past month athat greatreporting. deal of progress has increase in cross-border trade, asthe situation in has in civil affairs This month, as shown in the growth of local The identified ineconomic civil identified affairs reporting. This through intelligence reporting. directly tied to a area successful amnesty been made towards achieving the identified improved inalso civilshows affairs reporting. Thisshows significantly over the past information increased growth information also increased growth program andThis recent combat operations markets that has beenshown directly tiedarea tothe assessment addresses economic in the hasperiod mission. This isof specifically in the information also shows increased month, asthrough shown in thegrowth growth localsituation intelligence The has inofreporting. cross-border trade, reduction violence that been improved significantly over past markets that hasincrease been directly tied to01 between April andas 01 the May. Over the economic situation in to the has directly tied a area successful amnesty increase in cross-border trade, identified inasover civil affairs reporting. This of local month, as shown in the growth significantly the past This assessment addresses the improved period past month a great deal of progress has program and recent combat operations identified in civilinformation affairs reporting. Thisshows increased growth as shown in thealso growth of has local markets that been directly tiedthe to period This assessment between 01 April andinformation 01 May.month, Over the through intelligence reporting. The addresses also shows increased growth been made towards markets that has been directly to has achieving the past month a great deal of progress has economic situation in tied the area increase in cross-border as May. Over the between 01 April trade, and 01 increase in cross-border trade, asover been made towards achieving the improved significantly the past mission. This specifically shown This assessment addresses the period identified in is civil affairs reporting. Thisin the identified in civil as affairs reporting. This a mission. This is assessment specifically shown the This the period past great month, shown inmonth the growth of local deal of progress has between 01 April addresses andinformation 01in May. Over the information also shows increased growth also shows increased growth reduction of violence been the period This assessment addresses reductionpast of violence has been between April and 01 Over the markets that has been directly tied to that has month01athat great deal ofMay. progress has been made towards achieving the directly tied to a successful amnesty past month a great deal of progress has increase in cross-border trade, as been made towards achieving thedirectly tied to a successful amnesty between 01 April and 01 May. Over the program mission. and recent operations been made towards achieving the This assessment addresses the period identified in civil affairs reporting. mission. This is This specifically shown in the Thiscombat is specifically shown in the through intelligence reporting. The has mission. This is assessment specifically shown in the addresses the period program and recent combat operations past month great deal of progress has information also shows increased growth reduction ofThis violence that been between 01 April 01 a May May. Over the been of and violence that has economicdirectly situation inof the has reduction violence that has been May.reduction Over the between 01 April and 01 May tied to a area successful amnesty through intelligence reporting. The made towards the month a great deal of progressachieving has improvedprogram significantly over the past directly tied to month a successful amnesty past a great deal ofpast progress hasbeen and recent combat operations directly tied to a successful amnesty month, asthrough shown in the growth of localoperations been towards achieving the has program and recent combat been made towards achieving themade economic in the mission. This is area specifically shown in the intelligence reporting. The This assessment addresses thesituation period markets that has been directly tied to through intelligence reporting. The mission. mission. This isarea specifically shown the program recentshown combat operations This isand specifically in past economic situation in the has May. Over the between 01 April and 01in May improved significantly over the reduction of violencethe that has been increase in cross-border trade,in as economic situation the area has reduction of violence has been improved significantly over the past past month athat great deal of progress has reduction of violence that has been through intelligence reporting. The affairs reporting. This identifiedmonth, in civil as af improved significantly over theof past directly to a successful amnesty shown in tied the growth local month, as shown in the growth of local been made towards achieving the the period directly tied toamnesty a successful amnesty This assessment addresses directly tied tosituation a successful information also shows increased growth month, as program shown in and the growth of local recent combat operations economic markets that has been directly tied mission. This is to specifically shown in the Over the in the area has between 01 April and 01 May. markets that has been directly tied to markets that has been directly tied to through intelligence The hasand program and recent combat operations program recent combat operations increase in cross-border trade, as reduction ofreporting. violence been improved over the past past month athat great deal of significantly progress has increase in cross-border trade, asthe situation in has identified ineconomic civil affairs reporting. This through intelligence reporting. Theasreporting. The directly tied to a area successful amnesty increase in cross-border trade, through been made towards achieving the intelligence identified improved inalso civilshows affairs reporting. Thisthe significantly over past operations information increased month, asshown shown inarea thehas growth of local program andgrowth recent economic situation in the mission. Thiscombat isofspecifically in the information also shows increased growth month, asthrough shown in the growth local identified in civil affairs reporting. This economic situation in the area has intelligence reporting. The reduction of violence that has been improved thedirectly past markets that hasover been tied to markets that has been directly tied to significantly economic situation in to the has information also shows increased growth improved the past directly tied a area successful amnesty increase in cross-border trade, asover month, as in thesignificantly growth of localover significantly theshown past increase in operations cross-border trade, as This assessment addresses the improved period program and recent identified in civil as affairs reporting. Thiscombat month, shown in the growth of local month, as shown in the markets that has been tied to growth of local between 01 April andinformation 01 May. Over the through intelligence reporting. The directly also shows increased growth identified in has civil affairs reporting. This markets that has been directly tied to past month a great deal of progress has economic situation inmarkets the area increase in cross-border trade, asbeen directly tied to that has increase cross-border trade, asover thealso been made towards achieving the thein information increased growth improved significantly past shows This assessment addresses period identified in civil affairs reporting. This identified in civil affairs reporting. This increase in cross-border trade, as mission. This is assessment specifically shown the This the period month, as shown in the growth of local between 01 April addresses andinformation 01in May. Over the information also shows increased growth also shows increased growth reductionpast of violence has been between April and 01 Over the markets that has beenidentified directly tied to in civil affairs reporting. This month01athat great deal ofMay. progress has directly tied tomade a successful amnesty past month a great deal of progress has increase in cross-border trade, as been towards achieving the program mission. and recent operations been made towards achieving the information shows increased growth This assessment addresses the period identified in civil affairs reporting. This also Thiscombat is specifically shown in the through intelligence reporting. The has mission. This is assessment specifically shown addresses the period information also shows increased growth reduction ofThis violence that beenin the between 01 April and 01 May. Over the This assessment addresses the period economicdirectly situation inof the has reduction violence that has been between 01 April and 01 May. Over the tied to a area successful amnesty month great between 01 April and 01has May.aOver the deal of progress has improvedprogram significantly over the past directly tied to month a successful amnesty past a great deal ofpast progress and recent combat operations past month a great deal of progress has achieving the month, asthrough shown in the growth of local program and recent combat operations been made towards been made towards achieving the intelligence reporting. The This assessment addresses the period been made towards achieving the markets that has been directly tied to through intelligence reporting. The mission. This isarea specifically shown the mission. This isthe specifically shown in the economic situation in the has between 01 April and 01in May. Over mission. This is has specifically shown in the This assessment addresses the period increase in cross-border trade,in as economic situation the area has reduction of violence been improved significantly over the past past month athat great deal of progress has reduction of violence that has been reduction of violence has been between 01that April and 01 May. Over the identifiedmonth, in civil as affairs reporting. improved significantly over theof past directly toThis a successful shown in tied the growth localamnesty been made towards achieving thedeal directly tied to a operations successful amnesty past month a great progress hasamnesty This assessment addresses period directly tied tothe aofsuccessful information also shows increased growth month, as shown in the growth of local program and recent combat markets that has been directlyThis tiedis to mission. specifically shown in the Overthe program and recent combat operations been made towards achieving between 01 April and 01 May. the marketsinthat has been directly tied to through intelligence reporting. The program and recent combat operations increase cross-border trade, as reduction of violence that has been through intelligence reporting. The mission. is assessment specifically shownhas in the past month aThis great deal of progress This addresses the period increase in cross-border trade, as situation in the has identified ineconomic civil affairs reporting. This through intelligence reporting. The directly tied to a area successful amnesty economic situation in the area has reduction of violence that has been been made towards achieving the between 01 April and 01 May. Over the identified improved inalso civilshows affairs reporting. Thisthe significantly over past information increased growth program andsignificantly recent economic situation in the hashas improved over the past directly tied to aoperations successful amnesty mission. Thiscombat isof specifically shown in the past month a great deal of area progress information also shows increased growth month, asthrough shown in the growth local intelligence reporting. The month, as program shown the growth of local and recent combat operations reduction ofin violence that has been been made towards achieving the past improved significantly over the markets that has been directly tied to economic situation in the area hastied markets that has been through intelligence reporting. The shown in the directly tied to a directly successful amnesty mission. This is to specifically increase in cross-border trade, as month, as shown in the growth of local improved significantly over the past ineconomic cross-border trade, as situation in area has program and recent combat operations reduction ofthe violence that has been identifiedmonth, inincrease civil as affairs reporting. This shown in the growth of has local markets that directly tied to identified inincreased civil affairs reporting. improved significantly over the past through intelligence reporting. The directly tied toThis abeen successful amnesty information also shows growth markets that has been tied togrowth information also shows increased month, asdirectly shown the growth of local economic situation in in the area has combat program and recent operations increase in cross-border trade, as increase in cross-border trade, as markets that has been directly to improved significantly theaffairs pasttied through intelligence reporting. The This identified inover civil reporting. identifiedmonth, in civil affairs reporting. This increase ineconomic cross-border trade, as as shown in the growth ofin local situation the area has information also shows increased information also shows increased growth identified in civil reporting. Thisthe past growth markets that has beenaffairs directly tied toover improved significantly information also shows increased increase in cross-border asin thegrowth month, astrade, shown growth of local identified in civil affairsthat reporting. This markets has been directly tied to information also showsin increased growth increase cross-border trade, as identified in civil affairs reporting. This information also shows increased growth Commander’s report RAND MG1086-10.2 Contextual Assessment in Practice [C]onceptual frameworks must stay open to change on the basis of new information from the external world and avoid at all cost closing on themselves, interpreting all new information through the prism of rigid and untouchable schemata. —Antoine Bousquet17 17 Bousquet, 2009, p. 193. 238 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency he formatted but loosely directed step-by-step process presented here would be the entire assessment process for the theater. It is intended not only to meet the seven standards for assessment developed for this study but also to give structure to commanders’ assessments. Instead of continuing to treat commanders’ assessments as purely subjective opinion and using quantitative data as a means to “check” commanders, this process should help put commanders’ analyses in relevant context. Commanders at each level would be able to anchor their analyses in the layered reports, and the senior-subordinate assessments at each layer would place these analyses in further context. Instead of a distant assessment staf attempting to check a battalion commander’s assessment with inaccurate, incomplete, and aggregated data, the inherent subjectivity in subordinate commanders’ assessments would be addressed transparently through contextual data presentation and superior staf assessment. he theater commander’s assessment would rest upon a solid foundation of layered, contextual, and carefully analyzed assessment reporting rather than a set of dissociated and often inaccurate time-series charts and tables. If the theater commander is able to deliver a report with this kind of granular contextual information and analysis, the demands for raw, aggregated, and decontextualized data might diminish. With the existence of a thorough CA report, there should be no need to build a core metrics list or leverage time-consuming (or risk-inducing) data requirements on subordinate commands.18 Requirements for the theater staf to create executive summaries and brieings—documents that tend to have little context, that sufer from data aggregation, and that are usually poorly sourced—might also diminish because each commander’s assessment at each level is, in efect, a well-sourced executive summary. he inal report would contain detailed narrative and data, as well as several layers of summary assessment. Commanders would continue to produce and receive operations and intelligence reports in accordance with standing orders. CA might produce a report that is hundreds of pages long. No staf will expect a senior oicer or policymaker to read the entire report; that is not its purpose. he purpose of CA would be to ensure that policymakers, senior oicers, or stafs are able to read the entire report or desired sections, that they are able to ind sources of information, and that they understand that summaries are methodically built on contextual data and narrative. Nothing will preclude stafs from excerpting summaries from speciic levels of reporting and providing briefs with scalable detail. An efective staf should have no need for a formal MtM list. he relevant recommendation in this chapter incorporates such a list primarily to help ease the transition from a more centralized process to a truly decentralized but structured process. Envisioning a best-case scenario, stafs would simply collect and analyze data as needed or 18 As discussed earlier, collection requirements can induce risk because it may be necessary for military personnel to physically expose themselves to danger to collect data. For example, if a collection requirement calls for a unit to count the number of water wells in a district, patrols will have to be sent out to count the wells. Members of the patrol could be killed or wounded while conducting this seemingly mundane mission. A Proposed Alternative to Centralized Assessment 239 as the data presented. In this case, MtM would represent a concept rather than a tangible list of collection requirements. his system as described is hypothetical and therefore unproven. here are potential beneits, costs, and risks associated with implementing CA, as discussed next. Potential Beneits and Constraints his section describes the potential beneits of CA, along with some potential downsides associated with each beneit. he following section examines some of the risks and costs inherent in the CA process. he beneits of the process proposed here include the following, with constraints on these beneits in italics: • CA and MtM relect the kind of lexibility envisioned by the developers of maneuver warfare theory as interpreted in joint and service doctrine.19 Increased lexibility in collection and assessment should enhance the ability of tactical and operational units to conduct COIN. Conversely, centralized, decontextualized quantitative analysis will be more challenging once CA and MtM are implemented because centralized quantitative data will be less readily available. • Transparency is nearly absolute for policymakers, and public reports can be highly transparent. his should dramatically improve the credibility of the theater and national assessments and, possibly, the credibility of U.S. regional policy. Transparency carries risks, not all of which can be completely mitigated through a security review. • Because stafs at each level produce their own detailed and contextual assessments, tactical and operational units should beneit from being freed from an assessment process that currently “fails to deliver useful direction to subordinate stafs.”20 he beneits of CA to tactical units will still derive in great part from higherlevel analysis. If no analytic framework is applied to assessment then these beneits may still not be realized. • MtM reduce the impact that core metrics lists have on operational behavior. • he process places data in context, and a detailed, contextual assessment better supports executive decisionmaking and resource allocation. Detailed assessment requires time to develop and, to some extent, to read. Even a short executive summary of a narrative report is longer than a single brieing slide. • Data collection and management requirements are both reduced and redirected to address information that is more relevant and therefore more valuable to assessment. It may not be possible to reduce demands for centralized data in current operations. 19 See, for example, Headquarters, U.S. Marine Corps, 2005. 20 Soeters, 2009, p. 11. 240 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency • A reduction in data management requirements might produce some cost savings, although, as stated, all operations and intelligence information channels must still be maintained. here is a risk that some useful data might not be collected or reported. • Contextualized data assessment could reduce the kind of assessment shopping and raw data analysis currently practiced by consumers who are dissatisied with EBA products. However, this may not be the case if CA or a similar process is not properly and clearly implemented. • CA could improve accountability for campaign progress while ensuring that tactical commanders have a venue to address issues that might be of strategic concern. he availability of a comprehensive document might invite interpretation and exploitation of data by those with an agenda (e.g., those opposing U.S. policy in the region). If not carefully reviewed for sensitive data, CA could compromise operations. • CA and MtM could be quickly implemented because neither requires special training, new data networks, or overly complex formatting: Units from the battalion to the RC level already write assessments of some kind. Improved methods could be introduced over time. Quick implementation will inevitably result in inconsistencies in product and push-back from some units. A slower, more comprehensive rollout of the process would be preferable. Potential Risks or Shortcomings and Mitigating Factors he CA process proposed here is not perfect, nor is it applicable to all theater campaigns. Prior to implementing a new process, decisionmakers will need to consider the potential risks and shortcomings of CA, including the following (mitigating factors are presented in the nested bullets): • Producing an unclassiied assessment of an ongoing military campaign risks exposing sensitive information to hostile actors. – ISAF is already creating unclassiied assessments. he ultimate DoD assessments (the 1230- and 1231-series reports), which should relect all subordinate assessments, are unclassiied and oicially posted on the public Internet. – Assessments should not reveal ongoing or future operations. his information stays within operations channels. – Data can be stripped of sensitive detail, such as names or speciic locations. – Commanders will have to exercise judgment at each level to determine risk. – Classifying an assessment makes it appear that the military has something to hide. • Eliminating core metrics lists and centralized data collection places considerable trust in the hands of subordinate commanders. his could undermine the credibility of the assessment. A Proposed Alternative to Centralized Assessment 241 – his is a “trust-but-verify” system: Each level has built-in checks, and the red team, expanded in size and given increased authority to travel and communicate, adds a top layer of alternative assessment. – he current system, based on core metrics lists and centralized data collection, already sufers from a lack of credibility. Some of the beneit in this system is relative. – Transparency of CA and MtM methods and data should ofset the loss of centralized data analysis. • MtM information cannot be baselined and cannot be adequately tracked in timeseries analysis, the principal method of showing campaign progress. What will happen when units turn over and conduct a relief-in-place? – he core metrics lists at ISAF and IJC have been luctuating, in some cases quite dramatically, over the course of the Afghanistan campaign. When a core metrics list luctuates, it sends ripples across the entire theater and generates its own wave of cascading efects in ISAF and subordinate commands. MtM afect only the staf processes and data collection that are internal to one subordinate command. – Doctrine calls for routine adjustments to the assessment process, measures, and metrics to match operations (FM 3-0, FM 5-0, FM 3-24). – Neither CA nor MtM will prevent subordinate units from baselining and using time-series assessments at the level at which the data are acquired. – However, it is not clear that any assessment staf has accurately baselined a data set in Afghanistan or provided methodologically sound justiication as to how the baseline was deined and the thresholds were established. – It is impossible to eliminate all luctuations in assessment that result from turnovers. Some of these luctuations will be alleviated by the consistent production of a formatted battalion-level assessment; lack of consistency in the current process undermines continuity. Some of these luctuations are also beneicial, however. Barring the existence of indefatigable commanders, it is important to get a new perspective on the problem periodically. he (inevitable) resultant shifts in the MtM list are easier to process in CA than in a centralized, efectsbased process. • Consumers have come to expect raw data, executive summaries, aggregated timeseries brieings, and “objective” centralized data. hey will question why the reports they have been receiving for the past nine years are inadequate. – Providing an efective, transparent, and credible product should help mitigate these concerns. – here may be a false expectation at the theater level and below that all senior policymakers and staf oicers are demanding simpliied assessments. Interactions with senior stafs at DoD and DoS and in Congress indicate that many U.S. decisionmakers want more, not less, detail from the ield. Currently, these 242 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency requirements manifest in requests for raw data in the absence of a trusted holistic assessment. • Even if this system is implemented, senior commanders and stafs will continue to demand raw data collection and reporting, extensive graphic brieings, and the development of stand-alone executive summaries. CA might simply add to the workload of subordinate stafs. – Sustained demands for raw data for centralized assessment, graphics-laden brieings, and stand-alone summaries would indicate that the system is not being successfully implemented. – Ideally, these requirements would be actively restricted at the theater level. Inevitably, CA will reveal many of the laws identiied here. Battalions will report unevenly, and friction might occur as reports are assessed at each level. Some commanders and policymakers will not like the idea of a distributed, lexible indicator process and will continue to demand raw data collection and reporting. hese problems and others should be anticipated and planned for: hey already exist in most places. RCs already report in accordance with their own capabilities and circumstances, and there is no single, clear assessment method for any command below the theater level in Afghanistan. CA does not require additional or specialized training; the process could be implemented quickly, while all-source analytic capability could be added more gradually. he description of CA ofered in this chapter can and should be modiied to meet speciic conditions, but not to the point that transparency and credibility are lost. Based on observation of mid-level and senior stafs in both Iraq and Afghanistan, CA may be counterintuitive to some oicers accustomed to centralizing raw data and producing stand-alone brieings. A measure of command inluence probably will be necessary to ensure that EBA does not creep back into the assessment process once CA has been implemented. Chapter Summary CA ofers the military and policymakers a method for producing a credible report that should inform good decisionmaking and also help sustain public support for policy through transparency. However, applying CA in half-measures will not remedy an overall process that remains mired in efects-based theory and aggregation. If policymakers continue to demand EBA, centralized pattern and trend analysis, and aggregated raw data, they should not expect satisfactory results from military operational assessments. APPENDIX A Contextual Assessment Template his appendix contains a template for CA and MtM. As with the recommendations in Chapter Nine, the template is intended to serve as a springboard for discussion rather than a ixed recommendation. Each command and each campaign will bring speciic requirements that will reshape reporting formats. Report templates are ofered here from the battalion through the theater level. he irst template is also the most simple: It gives overworked and lightly stafed battalions an opportunity to produce a simple and straightforward report with context and citation. Each report builds in complexity, with commanders at each level required to review reports one level down. he theater AWG writes the theater assessment prior to the production of the theater-level analysis. he theater analytic team then reviews all assessments, including the theater assessment. At this point, the theater commander writes an overall assessment. he template headings are in bold text, while discussion of each section is provided in bracketed and italicized text. he assessment documents are narrative text reports, but they may contain graphics. he report should not be produced in a graphic presentation format, such as PowerPoint. Graphic visualization programs constrict narrative, encourage cut-and-paste replication of subordinate reporting, and are diicult to search in contemporary data networks. 243 244 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Battalion Assessment Assessment code: [Each assessment report will be coded for tracking purposes only, e.g., “Report #2-10.”] Area assessed: [should be deined using both a description and a map] Period of assessment: [from-to, information cutof date, date of report] Previous assessment code with comments: [code from the last assessment for this area; should note whether the area of operations or unit has changed] Unit reporting: [unit identiication] [What are the unit’s objectives and mission during the period covered by the report? his is included to ensure that the report is tied directly to objectives and so the consumer can understand what the unit was trying to achieve. What is the commander’s intent, drawn from the operations order? How did the commander intend to achieve these objectives?] Objectives and commander’s intent: Commander’s comments and outlook: [he commander provides insight and then looks forward through the next reporting period, identifying probable trends and possible major events or tipping points. here is no limit to the length of this section.] [explains which approach was taken to assess the AO and why; explains why certain data or types of data were examined; can be written by the commander or staf ] Assessment rationale, including description of MtM: Contextual assessment incorporating MtM: [he staf places any information in context within the overall assessment. he format of this section will depend heavily on theater command direction. For example, it could be broken down by lines of operation, or it could be broken down by district. Either way, it should consist of contextual narrative, and all data should be placed in context. Any graphics should also be contextualized, and relevant caveats should be included in the graphics themselves. Endnote or footnote references identify sources of data. here is no limit to the length of this section.] [All citation is identiied in full. Data gaps are clearly explained. For example, if there has been no friendly presence and no reporting from a particular village, the situation in that village should be identiied as “unknown.” It would also be appropriate to insert requests for intelligence collection in conjunction with the identiication of data gaps. hese requirements would not be published in an unclassiied report.] References and data gaps: Contextual Assessment Template 245 Brigade/Regimental Assessment Assessment code: [Each assessment report will be coded for tracking purposes only, e.g., “Report #2-10.”] Area assessed: [should be deined using both a description and a map] Period of assessment: [from-to, information cutof date, date of report] Previous assessment code with comments: [code from the last assessment for this area; should note whether the area of operations or unit has changed] Unit reporting and deployment period: [unit identiication and from-to dates of presence in the speciied area of operations] Assessment codes for all included subordinate assessments: [self-explanatory] [What are the unit’s objectives and mission during the period covered by the report? his is included to ensure that the report is tied directly to objectives and so the consumer can understand what the unit was trying to achieve. What is the commander’s intent, drawn from the operations order? How did the commander intend to achieve these objectives?] Objectives and commander’s intent: Commander’s comments and outlook: [he commander provides insight and then looks forward through the next reporting period, identifying probable trends and possible major events or tipping points. here is no limit to the length of this section.] [explains which approach was taken to assess the AO and why; explains why certain data or types of data were examined; can be written by the commander or staf ] Assessment rationale, including description of MtM: Contextual assessment incorporating MtM: [he staf places any information in context within the overall assessment. he format of this section will depend heavily on theater command direction. For example, it could be broken down by lines of operation, or it could be broken down by district. Either way, it should consist of contextual narrative, and all data should be placed in context. Any graphics should also be contextualized, and relevant caveats should be included in the graphics themselves. Endnote or footnote references identify sources of data. here is no limit to the length of this section.] [he commander comments on each subordinate assessment one level down; in this case, to the battalion level. Comments relect concurrence or nonconcurrence and explain any diferences of command analysis. Any disagreements should be clearly spelled out. If necessary, this could be handled by a red-team element.] Subordinate assessment comments and justiication: References: [All citations are identiied in full.] 246 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Regional Command Assessment Assessment code: [Each assessment report will be coded for tracking purposes only, e.g., “Report #2-10.”] Area assessed: [should be deined using both a description and a map] Time period of assessment: [from-to, information cutof date, date of report] [code from the last assessment for this area; should note whether the area of operations or unit has changed] Previous assessment code with comments: Unit reporting and deployment period: [unit identiication and from-to dates of presence in the speciied area of operations] Assessment codes for all included subordinate assessments: [from battalion up] Objectives and commander’s intent: [What are the unit’s objectives and mission during the period covered by the report? his is included to ensure that the report is tied directly to objectives and so the consumer can understand what the unit was trying to achieve. What is the commander’s intent, drawn from the operations order? How did the commander intend to achieve these objectives?] [he commander provides insight and then looks forward through the next reporting period, identifying probable trends and possible major events or tipping points. here is no limit to the length of this section.] Commander’s comments and outlook: [explains which approach was taken to assess the AO and why; explains why certain data or types of data were examined; can be written by the commander or staf ] Assessment rationale, including description of MtM: [he staf places in any context within the overall assessment. he format of this section will depend heavily on theater command direction. For example, it could be broken down by lines of operation, or it could be broken down by district. Either way, it should consist of contextual narrative and all data should be placed in context. Any graphics should also be contextualized, and relevant caveats should be included in the graphics themselves. Endnote or footnote references identify sources of data. here is no limit to the length of this section.] Contextual assessment incorporating MtM: Subordinate assessment comments and justiication: [he commander comments on each subordinate assessment one level down; in this case, to the brigade or regimental level. Comments relect concurrence or nonconcurrence and explain any diferences of command analysis. Any disagreements should be clearly spelled out. At this level the command can also comment on battalion-level assessments if it chooses to do so. If necessary, this could be handled by a red-team element.] References: [All citations are identiied in full.] Contextual Assessment Template 247 Theater Analysis Assessment code: [Each assessment report will be coded for tracking purposes only, e.g., “Report #2-10.”] Period of assessment: [from-to, information cutof date, date of report] [code from the last assessment for this area; should note whether the area of operations or unit has changed] Previous assessment code with comments: Theater analysis director: [Identify the leader of the analytic team at the time the report was written.] Assessment codes for all assessments addressed in this report: [from battalion up] Theater analysis comments and outlook: [he analytic team ofers general comments on the overall theater assessment, summarizes points of agreement and disagreement, and ofers either a concurring or competing outlook for the coming reporting period.] [his section analyzes the theater-level assessment in detail, addressing all points of concurrence and nonconcurrence and including reference to data that might support a contrasting viewpoint. his section is written after the theater assessment is written but before the theater commander’s assessment is written. here is no limit to the length of this section.] Analysis of theater assessment: [Once the analytic team has reviewed all subordinate assessments, it can choose to comment on selected assessments down to the battalion level, comment on points of disagreement between subordinate units, or ofer an alternative analysis for any (and, if necessary, all) assessments. here is no limit to the length of this section.] Analysis of selected subordinate assessments: References: [All citations are identiied in full.] 248 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Theater Assessment Assessment code: [Each assessment report will be coded for tracking purposes only, e.g., “Report #2-10.”] Period of assessment: [from-to, information cutof date, date of report] [code from the last assessment for this area; should note whether the area of operations or unit has changed] Previous assessment code with comments: Commanding oficer: [identiies the theater commander] Assessment codes for all included subordinate assessments: [from battalion up, including the theater analysis report] Objectives and commander’s intent: [What are the unit’s objectives and mission during the period covered by the report? his is included to ensure that the report is tied directly to objectives and so the consumer can understand what the unit was trying to achieve. What is the commander’s intent, drawn from the operations order? How did the commander intend to achieve these objectives?] Commander’s comments and outlook: [he commander provides insight and then looks forward through the next reporting period, identifying probable trends and possible major events or tipping points. he theater commander should also comment on the analysis of the theater assessment. here is no limit to the length of this section. his section is included in the report only after the theater analysis team has written its report.] Assessment rationale including description of MtM: [explains which approach was taken to assess the AO and why; explains why certain data or types of data were examined; can be written by the commander or staf ] [he staf places any information from the MtM list or any other information in context within the overall assessment. he format of this section will depend heavily on combatant command and national direction. Any graphics should also be contextualized, and relevant caveats should be included in the graphics themselves. Endnote or footnote references identify sources of data. here is no limit to the length of this section.] Contextual assessment incorporating MtM: [he commander comments on each subordinate assessment one level down; in this case, to the regional command level. Comments relect concurrence or nonconcurrence and explain any diferences of command analysis. Any disagreements should be clearly spelled out. At this level, the command can also comment on battalion-level assessments if it chooses to do so. If necessary, this could be handled by a red-team element. his is an opportunity to produce classiied annexes to the Subordinate assessment comments and justiication: Contextual Assessment Template 249 report, if needed, but the separation of subordinate material for security reasons should be justiied in writing.] References: [All citations are identiied in full.] APPENDIX B Notional Example of a Brigade/Regimental-Level Contextual Assessment his appendix contains a basic example of a brigade/regimental level contextual assessment report. his speciic level of reporting was chosen because it provides suicient opportunity to show detail while also showing analysis of a subordinate assessment. While the report is modeled on the Afghanistan case, data in this example are strictly notional and have no basis in U.S. or NATO military operations, assessment, or oicial documentation of any kind. his example is intended to inform rather than guide: It is not a “perfect” report, since each report in each area at any speciic time will necessarily be unique. It is also incomplete in that it shows examples from certain sections rather than complete narratives. his notional report is written by a brigade-level AWG and the brigade commander. It is a monthly report written at the end of August 2011, assessing progress in a notional “Sector B” that contains 15 districts and “Coastal City,” a notional metropolitan area. his brigade has 16 subordinate elements reporting on each district and Coastal City; some are infantry battalions, some are PRTs that have taken control of districts, and others are special operations units iling reports from districts with no conventional military presence. his would not be an unusual circumstance in Afghanistan or in other theaters. Note that while the summary is relatively easy to read, the detailed reports are riddled with jargon, obscure source titles, and information that is relevant only in a speciic context (e.g., a single attack in a small village). Stafs and commanders must be wary of the excessive use of jargon, but the requirement to accurately source the assessment from ield reports will require some technical language. As long as sources are provided, it will be possible to clear up any confusion caused by the use of military jargon. It should be easy to see where detail could be removed to create an unclassiied or at least a widely releasable classiied report. his may seem like a great deal of narrative text to the nonmilitary reader. However, brigades sometimes produce ten-page intelligence summaries and equally long operational summaries on a daily basis. A daily province-level or regional intelligence summary might consist of 40 or 50 pages of text and graphics. herefore, a 20- to 30-page monthly report assessing 15 districts and a major metropolitan area is not unusual. To mitigate the density of the text, some paragraphs might state, “no change” 251 252 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency from the previous reporting period. However, the use of “no-change” reports is potentially dangerous because it does not force stafs to routinely reconsider their assumptions and analyses. Notional Example of a Brigade/Regimental-Level Contextual Assessment 253 Third Brigade, X Division Operational Assessment for August 2010 Assessment code: B-8-11 [in this example, Sector B, Report 8 of the current iscal year, year 2011] Area assessed: Sector B (Province X) containing Districts 1 through 15 Coastal City metropolitan area Period of assessment: 08/01/11 to 09/01/11 Information cutof date: 09/15/11 Date of report 09/15/11 Previous assessment code with comments: B-7-11 [sector B, month 7, year 2011] Added District 14 and 15 since report B-7-11 Unit reporting and deployment period: 3rd Brigade, X Division 01/01/11 to 01/01/12 Assessment codes for all included subordinate assessments: 1-B-8-11; 2-B-8-11; 3-B-8-11, [and so on, with each initial number denoting a subordinate unit report] Objectives and commander’s intent: Objectives: 3rd Brigade’s objective is to bring its area of responsibility to the point that it is ready for transition to local/national security lead and local governance lead. his point should coincide with the end of the brigade’s deployment period. Concurrently, the 3rd Brigade will develop the capabilities of the host-nation army and police forces to the point that they are capable of taking the lead for border security, combat operations, and local policing at the time of transition. Commander’s Intent: I intend to achieve our transition objective by leveraging a combination of local security forces and local governance to supplement host-nation capabilities. We will focus our main efort on the development of local capability in both security and governance. Our organic forces will both partner with and advise local security forces while providing security bubbles for their progressive development. Battalion commanders will weight their efort toward this mission. he Provincial Reconstruction Team (PRT) will have the lead for the development of host-nation governance, while we will take the lead for the development of local governance. We will make every efort to connect local security and governance to oicial host-nation programs. 254 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Commander’s comments and outlook: Comments: his period of assessment relects mixed results. I am satisied that we are creating lasting security bubbles around key population centers (to include nearly all district centers) and that we have been able to develop local security forces within those bubbles. With exceptions noted below, our local security forces are operating with minimal support and are directly connected with local governance boards. his vital connection between local security and governance will help ensure that the security forces are supported in the absence of central direction and support and that local leaders can exercise some control over their own areas of responsibility. We have been less successful in protecting lines of communication to Coastal City, as the rise in violence there indicates. I believe, however, that this rise in violence is due in great part to the increase in criminal activity as organized gangs ilter into the city and not due speciically to centralized, insurgent-directed attacks. he host-nation military units are progressing quickly, but host-nation police remain a source of concern (see our separate report on security force capabilities). he economy is developing slowly and remains largely reliant on black-market activity. Development eforts have produced limited and usually temporary efect. However, I continue to ind aid and development a useful tool for stability in that it helps us support the development of local security forces. I defer to our other agency partners and the PRT to describe prospects for long-term economic development in Sector B. Some insurgent recruitment success certainly stems from the lack of economic opportunity in our sector, but we believe the majority of insurgents operating in Sector B are motivated by religion, criminal corruption, and ethnic pride (source: Intelligence Reports B-14-1, B-15-7, B-3-2). [All source information, like the rest of the information presented in this example, is notional.] Based on a review of the key-leader engagement reports and sensitive intelligence reporting (source: Intelligence Reports B-KLE-12/15/19/22, B-SI-1, B-SI-29), the political will of local tribal leaders and locally elected oicials seems to indicate increased willingness to support the host-nation government and government security forces. his is exempliied by the recent security agreement between Elder N and Governor M in District 6, which led to both a formal peace accord and the induction of 400 tribal members into the local security force unit there. My personal engagement with leaders across the province reinforces this assessment. here seems to have been a shift in perception among key leaders over the past three months in particular. I attribute this to our persistent operations but also to the aggressive and often ruthless behavior of the insurgents and improvements in government service delivery across the province. Outlook: I have a fair degree of conidence that we will be able to transition nine of 15 districts to the local and host-nation security and governance lead by the end of our deployment period. In the absence of an additional battalion of host-nation army forces, the remaining six districts probably will require another four to six months of development before they are ready for transition of security lead. Coastal City can Notional Example of a Brigade/Regimental-Level Contextual Assessment 255 be made ready for on-time security transition with the addition of six police training teams. A written and serialized justiication for these additional forces is attached to this report and to the oicial request for forces. With the exception of the 10th Army Brigade, Army units will be ready for transition across the board. Coastal City police will not be ready for transition in the absence of the additional police training teams. Local government oicials will be ready to assume control in 12 of 15 districts, while the three remaining districts should be ready for transition under central government control. Continued coalition security and governance support will be required in each district, to varying degrees, beyond the transition point (see below). I anticipate that all 15 districts could be under nominal central government control within two years of transition. Coastal City, on the other hand, will probably not be ready for governance transition for an additional 12–18 months due to extreme government corruption and incompetence. his lag in progress in Coastal City cannot be addressed simply by replacing corrupt oicials. It will just take time to change the culture of corruption in governance and economics. While Governor Z is performing adequately, he can neither govern nor survive without involving himself in corruption to some degree. He and his replacements will remain reliant on the system of patron-client networks that form the backbone of government in Sector B. Patron-client networks are inherently corrupt and opaque and are very diicult to transform (source: 3rd Brigade Governance Advisor Report 6-11). Long-term solutions for the governance problem in Coastal City may lie above the brigade level. With the focus on developing local security and governance, as well as security bubbles around key population centers, metrics were derived from advisors, partnering teams, and security reports. Speciically, this reporting period will focus on the capabilities of local security units and their connection to local governance boards, capabilities of local governance boards (to include performance in the wake of recent local elections in eight of 15 districts), and levels of violent activity—insurgent, criminal, other—in the context of local security force development. Concerns with Coastal City governance and policing will be relected in an increased focus on governance, corruption, and policing metrics in the Coastal City region speciically. [he assessment could then list speciic MtM or simply report them in context in the following section.] Assessment rationale including description of MtM: Contextual assessment incorporating MtM: Overall assessment: Progress in Sector B was mixed in August, but we anticipate that nine of 15 districts (1–7, 10, and 14) will be ready for both security and governance transition by the end of the calendar year. hree additional districts (8, 9, and 15) probably will be ready for governance but not security transition by the end of the year. Districts 11, 12, and 13 continue to be heavily contested by insurgents, and all three districts sufer from endemic corruption, poor security force development, and 256 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency lack of host-nation presence. However, there are mitigating circumstances in each of these three cases that will be addressed below. Governor Z continues to support some criminal activity, but we are aware of this activity and do not believe it counterbalances his overall performance. While violence rose signiicantly in Sector B during the reporting period, the vast majority of these new incidents (280 of 300) were due to increased host-nation security force operations in districts 8, 9, and 15. Most of these incidents resulted in successful outcomes for host-nation forces, although Host-Nation 10th Army Brigade broke under contact and has been dismantled. Coastal City is increasingly violent. While this increased violence undermines the objective of establishing stability in Sector B, it is mostly criminal and not insurgent-related. A stronger police force should be able to reduce violence in Coastal City without signiicant coalition support. Assessment by Lines of Operation—Security: Overall progress on this line of operation is mixed, with improvements in 12 of 15 districts, setbacks in three districts, and setbacks in Coastal City. Violent incident reports, which we anticipate to be 75-percent accurate at our level of analysis, rose from 220 in Assessment Period B-7-11 to 530 incidents in the current period (source: Brigade Incident Report Database 1, entry 8-11). his dramatic increase in violence stems primarily from host-nation army and police operations against insurgents and criminal gangs in districts 8, 9, and 10. Host-Nation Army Brigades 7–12 conducted sweep-and-clear operations followed by stability operations and presence patrols throughout the district centers and in the peripheries (source: subordinate reports 8 through 10-B-8-11). Although the increase in violence was expected, the level of insurgent presence and general instability in these three districts was not anticipated in intelligence reporting (source: see Brigade Intelligence Summary 15-7-10 and District Analysis Report 7-11). Local security forces in Districts 1–7, 10, and 14 are intact and performing adequately with only a few exceptions (source: Special Operations Report B-8-11). Local forces in Districts 3 and 14 are performing exceptionally well due to strong leadership and efective use of aid programs to connect the security forces to the local governance boards and the boards to the host nation. All local security forces continue to receive threats from insurgents to a greater or lesser degree, but these threats have only proven successful in District 8. here is relatively little friction between local forces and the Army, but several incidents in District 2 between the police and local security forces have raised concerns about the future of the program there (source: District Analysis Report 2-11). Assessment by Lines of Operation—Governance: Progress toward governance transition is excellent at district level and below but mixed to poor at the sector (provincial) level. Districts 1–7, 10, and 14 have functioning district governments. Each of these district governments is providing services to the population, including justice, health, and basic civic administration. With the exception of District 5, no district government Notional Example of a Brigade/Regimental-Level Contextual Assessment 257 is collecting taxes. his means that they remain reliant on host-nation government inancing or PRT support. his dependence is a transition hurdle that can be overcome only with improved host-nation support. 3rd Brigade district governance assessment therefore continues to rely on the promised delivery of taxation oicials and budget planners for each district from the host-nation government (source: Regional Command FRAGO [fragmentary order] 72-10). Districts 11–13 continue to sufer from the absence of district-level governance; insurgents provide justice and administrative leadership in each of these districts as of the end of the reporting period (source: District Analysis reports for 11–13 for periods 6–8, and PRT report B-8). District 15 is a unique case: It has a functioning district government, but the government is isolated in the district center due to insurgent activity on the periphery. Governance at the village level is progressing faster than governance at the district, provincial, and metropolitan levels, particularly in Districts 1–7, 10, and 14. In each of these districts, coalition forces have facilitated the simultaneous development of local security forces with local governance boards. Local governance boards establish and control local security, a process that is fostered by the provision of development aid. Civilians in these districts report that they are generally satisied with their local governance boards (source: District Analysis for 1–7, 10, and 14, as well as individual Atmospherics Reports from each district for periods 5–8, Infantry Patrol Reports for periods 5–8, and the Province Poll of 14 August). Local governance in districts 11–13 and 15 is either nonexistent or on the fence and hesitant to engage with the coalition or host nation. Governance in Coastal City—both provincial and metropolitan—continues to be hindered by endemic corruption. [he report would continue to provide depth at this point, including sourcing. Metrics might cover popular perception of justice, perception of governance, and satisfaction with oicial performance, among other points.] Assessment by Lines of Operation—Development: [his section would mirror the other narratives on lines of operation, providing sector-level assessment of development as it relates to the brigade objectives and transition. It will be similarly sourced. Sources might include PRT reports, civil afairs reports, NGO reports, and so.] Assessment by District—District 1: District 1 is generally stable and is steadily progressing toward on-time transition. Violent incidents remain steady at approximately ive per day, but the vast majority of these are half-hearted attacks by small insurgent or criminal groups, or they are—increasingly—failed improvised explosive device attacks that do not result in casualties (source: 1st Battalion Incident Report Roll-Up 1-B-8). he major insurgent attack in Village 23 was dramatic and caused some casualties, but the insurgent group that conducted the attack was wiped out (source: C Company, 1st Battalion Incident Report 1-B-8-82). his turned into a signiicant psychological victory for local security forces and for the local populations in Villages 21–27 (source: 1st Platoon, A Company, 1st Psychological Operation Battalion attachment, Incident 258 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Report 1-B-8-7). Press reports that played up the impact of the attack were generally inaccurate (source: e.g., he Source, “Insurgent Attack in Sector B a Major Setback for Coalition,” 15 August 2011). he district continues to beneit from the presence of Host-Nation 6th Army Brigade Headquarters; the brigade helps provide local security, and its presence is a major deterrent to insurgents and criminals alike. It is possible, however, that this strong army presence has slowed the progress of police development (source: Host-Nation Security Force Report for District 1). Police units are functional on paper but will only just be ready for transition due to lingering issues of corruption (source: 1st Battalion Host-Nation Security Force Report 1-B-8). If the Host-Nation 6th Brigade Headquarters is moved from District 1, the district will sufer a stability setback. Local security forces in villages 1–56 are functional and can help make up for the absence of efective policing in the short term, but the long-term absence of efective police presence in the villages will probably lead to setbacks (source: Company A through D and Weapons Company Reports 1-B-5 through 1-B-8). Governance in the district is adequate, and Governor K continues to work closely with local governance boards and with Governor Z (Intelligence Biography Report of Governor K, 1-B-8). Corruption remains endemic but, according to civil afairs reports, is manageable (source: CA reports 1-B-5 through 1-B-8). [he report may address a few more issues on governance and development but would probably not go on beyond a page and a half of written text. Much of this text would probably be derived from existing operations and intelligence reports.] Assessment by District—District 2, etc.: [Assessments of the other districts would look somewhat similar to the assessment of District 1 but would probably have some signiicant differences in tone, format, sourcing, and information. his should be expected and accepted as long as the reports address progress toward the 3rd Brigade’s objectives. he reports for districts 11–13, where the brigade is sufering from setbacks, will probably be more detailed and will provide more information on security reporting.] Subordinate assessment comments and justiication: District 1: Concur, with additional information. While the attack in Village 23 was a tactical victory and a psychological victory for the population in District 1, at the brigade level, it was a psychological defeat and an operational setback. Our information operations failed to get out in front of the story, and the general consensus at the country level is that this was a major defeat for the coalition. It will not be possible to reverse this consensus without signiicant, visible progress in Villages 21–27. Impact: his attack may force us to shift forces into District 1 at the expense of other districts. We should be able to assess the impact of this incident in the next reporting period. Also, it should be noted that Intelligence Biography Report of Governor K, 1-B-8, has been supplanted by Intelligence Biography Report of Governor K, 1-B-8a, which Notional Example of a Brigade/Regimental-Level Contextual Assessment 259 shows that he has stopped communicating with Governor Z. Impact: possible loss of funds for District 1 development projects. his should also play out over the next reporting period. District 2, etc.: [Remaining assessments might simply state that the brigade commander concurs with the assessment and has nothing additional to ofer, or they might highlight a signiicant diference in analysis. In some cases, the brigade staf might have to write the district-level report due to the absence of forces in a certain district. hat should be very clearly noted in both the previous section and in this section of the report.] [Every report cited in the body text, as well as any additional reports that the consumer might ind useful, should be listed here. Further, because it is often hard to ind nonserialized reports (e.g., a civil afairs village report), wherever possible, these original source reports should be attached to the document or embedded in the document. At the very least, each source should contain URL to help the consumer locate the report on unclassiied or classiied Internet systems.] References: APPENDIX C Phong Dinh Province Report his appendix presents verbatim excerpts from the February 1968 CORDS report on Phong Dinh Province, Vietnam. Written by the province team, this report is a narrative that relects input from U.S., Vietnamese, and subordinate military and civilian sources. It was written at the tail end of the VC’s Tet Ofensive, so it covers a period that is unusually chaotic. he report addresses all aspects of civil-military activity in the province and includes copies of subordinate district reports in their original text. he purpose of presenting this report here is to make available an example of the kind of contextual assessment that has been and can be produced by counterinsurgents. In some ways, the report parallels the CA framework outlined in Chapter Ten, and the CA framework borrows, in part, from this and other province- and district-level reports from the Vietnam, Iraq, and Afghanistan cases. he Phong Dinh report illuminates or reinforces several aspects of the Vietnam case study, as well as key indings presented in this monograph. It shows how quantitative and qualitative data can be captured and explained through narrative context, and it reveals the complexity of the challenges faced by just a segment of the counterinsurgent force. (he Revolutionary Development program reports did not relect other parts of the CORDS program or any tactical or operational unit activities or assessments.) Speciically, the report shows how diicult it can be to pin down quantitative data, like refugee statistics or civilian casualty statistics. his report seems to be a status report and is not necessarily intended to address assessment requirements. herefore, some sections might not be relevant to a holistic theater assessment. It contains recommendations listed by the province senior advisor as “required actions,” so it serves both assessment and operational needs.1 It is not clear whether a civilian or military advisor wrote this report, or whether there was collaboration between a civilian and military oicer to produce it. he report describes various areas within and around Phong Dinh Province. Figure C.1 shows a contem1 hese “required actions” include a request that more ARVN forces be committed to protect a critical road, a suspension of the 1968 Revolutionary Development campaign plan to allow lagging villages to catch up, a revision of the Vietnamese command relationship, reassignment of ranks within the GVN regional force (RF) units, and a supply of seed for 6,000 hectares (MACV, 1968a, p. 18). 261 262 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Figure C.1 Map of Phong Dinh Province, Vietnam Ap Thoi My Binh Thuy Can Tho Thai Lai Ba Se Thoi Thuan Ap Truong Thanh Ap Truong Thun Ap Phu Nhan Thanh Hoa Ap Phu Lo Ap Truong Hiep Rach Go Ap Thonh Loi Thuan Nhon Ap Phu Thanh Xam Dong Phung Hiep Cho Kinh Cung Ap My Ho SOURCE: James M. Kraft Collection, Texas Tech University Vietnam Center and Archives. NOTE: Overlay added to clarify names of districts and key areas. RAND MG1086-C.1 porary annotated map of Phong Dinh Province to help orient the report to the local geography. he report refers to speciic abbreviations and terms: RD revolutionary development; a term used to describe villages or units covered by the CORDS revolutionary development program, with 1967 RD villages having been identiied as part of the program or established in 1967, for example. RD is also a term applied generically to projects associated with the Revolutionary Development program. RF regional forces; local civil defense units oriented at the province level PF popular forces; local civil defense units oriented at the district level New Life Hamlet: Originally “strategic hamlets,” these were hamlets that were either cleared of VC and fortiied or built for relocation of indigenous civilians to isolated and secure areas. Phong Dinh Province Report New Life Hamlet 263 Originally “strategic hamlets,” these were hamlets that were either cleared of VC and fortiied or built for relocation of indigenous civilians to isolated and secure areas. he map and report were retrieved from the Texas Tech University Vietnam Center and Archive’s online resource catalog on February 15, 2011. he declassiied province report was originally archived by the University of California Indochina Archive’s History of the Vietnam War on Microilm. he report is incomplete in that it is missing several pages. Although the portions reprinted here are verbatim, some text deemed overly administrative was removed. Only three of six district reports in the document are republished here. 264 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency CORDS Report—Phong Dinh Province 02/68 Province Monthly Report Phong Dinh Province Period Ending February 29, 1968 1. Status of the Revolutionary Development Plan: a. General he RD program in Phong Dinh Province was dealt a severe setback by the Viet Cong Tet attack on Can ho. In order to defend Can ho City, the Province irst withdrew part and then all of the 2/33 ARVN Battalion protecting the 1967 RD villages of Tan hoi, Giai Xuan, and Cau Nhiem in the early days of the ighting. All but two RD cadre teams were pulled in to assist in the defense of Can ho. A major portion of the RF/PF units committed to protect these 1967 RD task force areas were withdrawn leaving only the 295 RF Company in Tan hoi Village and the 535 RF Company in huan Duc Village. he 826 RF Company moved from Tan Nhon New Life Hamlet to ill the position vacated by the 2/33 ARVN. here units were largely limited to protecting themselves against VC attack and so could provide little protection for the people of the area during this period. In addition, the 293 RF Company was withdrawn from Dinh Mon Village on 19 February since it could not be supported. his removed the last element shielding the 1967 RD area on its west lank. hus, in the face of the emergency in Can ho, the Province denuded the RD areas of much of their external assistance in resisting the return of VC political and military units. Giai Xuan Village, a Phase II 1966 RD village, and parts of Long Tuyen, a 1966 RD village, were the scene of heavy ground ighting and subject to intense aerial bombardment since major VC units used this area as a base of operations in the period following the assault on Can ho. he ighting in this area continued through February and numerous families were forced to lee their homes in both these villages. his situation left much of the 1966 and 1967 RD area either a battleground or a zone exposed to re-iniltration by VC terrorist groups and political cadre. Finally, the conidence of the people living in the RD areas has undoubtedly been shaken by the deep incursions made by major VC units into areas such as Can ho. Unless early action can be taken to return RD forces to these areas in strength, many of the past accomplishments of RD in Phong Dinh may be lost. b. Planning he status of the Phase I 1967 RD village of Giai Xuan and the Phase II 1967 RD villages of Tan hoi, Cau Nhiem, and huan Duc is unknown at this time. No member of this team has been able to visit any of these villages to assess the situation since the opening of the Tet ofensive. here is no RD presence in any of these villages except huan Duc which has two teams in the relatively secure hamlet of An Hung, directly Phong Dinh Province Report 265 across the Can ho river from Can ho city. One battalion or more of Viet Cong occupied Nhon Loc I and Tan Nhon Hamlets (both 1967 New Life Hamlets) for at least several days before launching a surprise attack on Phong Dien District on 24 February. No one in these hamlets reported this threat to the District. A 75mm recoilless rile was set up by the VC in the Tan Nhon RD school and the school was heavily damaged in the subsequent battle. he RD school in Tan Long was ransacked. Other damage in RD areas is unknown. Except for An Hung Hamlet in huan Duc Village, the VC now move through the RD areas with impunity. he RD areas are not yet secure enough for the teams to redeploy into them at this time. Combat operations by the ARVN units are continuing within the Ba Se perimeter and it is hoped that during the month of March adequate security will be re-established permitting the RD teams to move back into the Phase I and II hamlets to resurrect the accomplishments of the 1967 RD program. d. Village Governments Due to the prevailing insecurity and the press of duties connected with refugee relief in Can ho, the Public Afairs Advisor was only able to visit two village governments during February. Only one elected village government, Giai Xuan, was not functioning at the end of the month. he presence of large VC units in Giai Xuan forced the village chairman to lee to the district headquarters at Cai Rang. Chau hanh District reported that one hamlet chief disappeared during the Tet crisis and he is presumed dead or captured. he elected hamlet chief in hoi Trinh B Hamlet in Phuoc hoi Village (considered secure by the GVN) was assassinated on 20 February. Except for the missing hamlet chief, all the local oicials in Chau hanh District (scene of most of the heaviest ighting) are accounted for. he large majority of hamlet oicials in Long Tuyen and An Binh Village are unable to return to their hamlets and so remain near the village oices which are in secure areas. he village governments in An Binh and Long Tuyen Villages appear to be working hard to register the refugees who led the rural sections of these villages. hey have received little assistance from the province and no relief commodities. An Binh Village had to use village funds to print refugee registration forms. 3. Political/Psychological a. General he VC attack on Can ho and the continuing harassment and mortaring of the city has brought home to the people, many for the irst time, that there is a war going on and that, despite their desire to be left alone by both sides, they are being denied this luxury. During the entire month the residents of Can ho and the thousands of refugees who poured into the city have been interested primarily in survival. he fear and terror that originally drove them here have now subsided and have been replaced by more 266 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency generalized feelings of insecurity. Taking advantage of these feelings are those children and old men digging up sand and dirt to put in sandbags sold for 20$ VN apiece. Many people are obviously preparing for the worst. Several are visibly upset over what they consider the inability of the ARVN to protect them [emphasis in original]. hese same people are also highly critical of the ARVN’s engagement with the VC at the University [emphasis in original]. When this was followed by large-scale looting by the ARVN troops, many found it diicult to consider the soldiers as their protectors. Others, however, apparently accept the destruction of the University and the surrounding area by VNAF dive bombers as a necessary corollary of war although they wish something less drastic could have been done. While the refugees are grateful for the emergency assistance provided to them, they want to know when and how the government will help them to rebuild their homes. his is a potential trouble spot which the VC might well try to exploit and the GVN should be aware that the best thing it can do to build an even more solid base of support among the refugees is to fulill its promises of assistance to the refugees now. As to whom the people feel was responsible for violating the Tet Truce and causing all the damage, it all depends on whom you question. Most merely reply that “you know as well as I who was responsible,” which is merely a way of avoiding expressing their true feelings. It appears, however, that most of them accept the fact that it was ultimately the presence of the VC that brought down the bombs and rockets, but then these same people also consider that the ARVN used a “sledgehammer to kill a mosquito.” 4. Security a. General he security situation has deteriorated during the past month as a direct result of the VC Tet ofensive. he paciication efort has come to a complete halt. Route 4 is virtually closed to commercial traic, although some traic does go through on days when military road clearing operations are conducted. Inter-provincial Route 31 is closed as a result of craters and one destroyed bridge. Route 27 to the west remains open and relatively secure. he Xa No Canal remains open but constant surveillance must be maintained to forestall VC tax collection activities. he bleak security situation is improving. he ARVN, RF, and PF have apparently blunted the Tet ofensive, inlicted heavy casualties on the VC, and are now regaining the initiative. Because there was never an apparent triumph, large numbers of VC infrastructure did not surface in this province as in others. Phong Dinh Province Report 267 b. Enemy Situation During the early part of the month, the VC continued their Tet ofensive. In their efort, the enemy made substantial military gains. he VC forced the withdrawal of friendly forces from large portions of the Ba Se perimeter and gained the ability to launch indirect ire attacks on the IV Corps vital installations. Binh huy Airbase has been shelled 10 times, Can ho Airield 11 times, and the IV Corps ammo dump was shelled twice. Two ground attacks were directed against Can ho airield, and one probing action against Binh huy Airforce Base. . . . here has been a degrading trend toward defensive thinking on the part of some Vietnamese leaders. he intensity of the VC attacks has caused several district chiefs to lose all ofensive spirit. hey seem content to wait for the VC to go away. To date RF have sufered relatively light casualties; 43 have been killed. RF and PF have accounted for 141 VC KIA and 47 Viet Cong Cadre; and they have recovered 97 weapons. Company commanders of 4 RF companies have been wounded and, because of the shallow depth of RF leadership, these companies have been hurt badly. . . . 5. New Life Development a. Economics he economic situation in Phong Dinh has not been seriously upset by the VC Tet ofensive even though land communications routes were seriously disrupted. . . . Prices had risen considerably before Tet. Following the VC attack, prices stayed at the Tet level except for 2d grade rice which rose 2 or 3 piastres in price. Prices have now begun to decline and rice in particular is now below the pre-Tet price since the harvest is in full swing. hrough the crisis adequate rice was available in Cai Rang warehouse. . . . e. Public Health A number of diicult public health problems were created by the VC Tet ofensive. he large number of civilian war casualties created by the ighting (approximately 600 war victims were admitted during February) taxed the resources of Can ho hospital, especially since many Vietnamese hospital staf members were unable to return from their Tet vacations for 5–6 days after the attack on Can ho. his great inlux of civilian war casualties meant that approximately 700 surgical cases were handled in February by the USAF Surgical Team and Vietnamese surgeons, a new monthly high for the hospital. . . . f. Public Education he use of 11 public and private schools in Can ho as refugee centers has delayed the resumption of classes following the Tet vacation. he presence of large numbers of refugees in these schools has resulted in some damage to school facilities and general uncleanliness in and around these schools. his Team is now pressing the province to establish a realistic plan to move the refugees out of these schools and reopen them for 268 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency classes as soon as possible. To date no refugees have been relocated. In the rural areas outside Can ho such as My Khanh, An Binh, Nhon Nghia, and Nhon Ai, schools cannot be reopened at this time due to insecurity and failure of teachers to return after Tet. Many schools in rural areas will have to be repaired and several rebuilt before they can be reopened. g. Refugees At the height of the inlux there were 12,250 refugees in 11 GVN supported relief centers in Can ho. his igure has now decreased to 6,400. Figures are not available on refugees outside these camps but there are probably 4,000 more outside their homes. he response of the Province to the inlux of refugees created by the battle for Can ho has been erratic. At least as early as the afternoon of 1 February, GVN oicials including the Deputy Province Chief for Administration, the Social Welfare Chief and several others had assembled at the Province Headquarters to decide on emergency relief measures. . . . By way of contrast, the GVN allowed military units to coniscate the major portion of several thousand loaves of bread purchased that day for distribution to the refugees. As a result, the refugees had to do without for another day. . . . Phong Dinh Province Report 269 District Senior Advisor’s Monthly Summary huan Trung District Period Ending February 29, 1968 District Chief: Captain Ngo Cam he military situation in the District has improved somewhat. Viet Cong incidents have been conined to harassing ire on some of the outposts and propaganda teams entering villages and hamlets. he use of RF/PF units to provide security for the rice harvest was a great success. No Viet Cong incidents were reported in connection with the rice harvest. Programs in the District have slowed down because of the absence of key staf oicers and civilian oicials who were unable to return after Tet. Many refugees in large, house-type sampans have moved into the hoi Lai-hoi Dong Canal. New houses have been constructed in hoi Lai by refugees. hese refugees are from the hi Doi Canal area which has been a target of constant bombing. he only major problem at the present time is a short supply of consumer goods. During one operation on 8 February, the 177 RF Company made contact with one VC company in hoi Khuong (A) Hamlet only 200 meters from the center of hoi Lai Village. As a result, 9 VC were killed, 2 VC were captured, and 11 weapons were captured. hoi Lai Village has received several rounds of a locally manufactured rocket. Most of the rounds were duds. Raymond G. Fischer First Lieutenant, Artillery Acting District Senior Advisor 270 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency District Senior Advisor’s Monthly Summary huan Nhon District Period Ending February 29, 1968 District Chief: Captain Nguyen Ngoc Luu he District town was mortared eight times and there were ifteen attacks on PF outposts during the reporting period. Over ifty roadblocks were erected on Route 31 and ten incidents of cratering and mining occurred. he 75% destruction of the Rach Goi Bridge has stopped heavy traic from making a continuous trip through the District since 3 February. Commercial traic on the Xa No Canal was interdicted by the VC on numerous occasions between 31 January and 18 February and merchants were searched, taxed, and turned back. Since 18 February normal traic is slowly being restored during daylight hours. he District has been in a completely defensive status since 30 January with no participation in division or sector operations. Requests for permission to conduct sub-sector operations have been denied. In this defensive role, however, no outposts were lost or abandoned nor were any weapons lost. here were ten conirmed VC structures and approximately 25 sampans destroyed during the reporting period. Development programs have not progressed at all and planning ceased to exist this month in this District. he one election scheduled for February has been cancelled due to the diiculties of transporting election supplies from Can ho and travel restrictions on the voters. Rice is being harvested now with the help of RF and PF security forces. Only about 35% of potential will be realized this period. Transportation of POL supplies has been a problem this month and has curtailed the activities of the Vietnamese Navy boat platoon and generator-operated communications. ARVN’s inability to provide medical evacuation of wounded soldiers and civilians has caused some morale problems and unnecessary deaths. he understandable lack of air cover and observation aircraft in this District for the outlying areas has allowed the enemy freedom of movement and increased control over the rice harvest in these areas. his situation has eased in the last few days. Richard C. Morris Major, Infantry District Senior Advisor Phong Dinh Province Report 271 District Senior Advisor’s Monthly Summary Phong Dien District Period Ending February 29, 1968 District Chief: Vo Van Dam I. Overall Status of Paciication 1. Military situation—With the advent of the VC Tet ofensive the military situation has deteriorated to its lowest point since the birth of this district in June 1966. he VC have been, and still are able to do as they desire throughout 90% of the district. his includes the placing of 2 to 3 battalions within the district, using its LOCs [lines of communication] freely in and out of the area of the Can ho vital installations, and attacking the [outposts], and the mortaring of the district town at will. 2. he paciication program in both Tan Nhon and Tan Long hamlets has not only been halted, but has been dealt a serious setback. Prior to TET [sic], it was hoped that the RD program in these hamlets would be completed by the end of February, but now an undetermined amount of time will be necessary to repair the damages, regain the conidence of the people and complete uninished projects. II. Problem areas efecting [sic] the present military situation are many. For TET, approximately 25% of the PF and 70% of the sub-sector staf were allowed to take leave; the surprise ofensive staged by the VC; the withdrawal of the 2/33 ARVN battalion and the 294 RF company to Can ho; the superiority of irepower through the better weapons and a more ample ammunition supply enjoyed by the VC; the almost defensive attitude of the Vietnamese leaders within this district and the disapproval by Sector of the majority of sub-sector’s operations requested have all had an inluence on the current situation. III. Eforts to revive the paciication programs shall not and can not begin in earnest until the military situation stabilizes. As to the future of the RD program (Phase II 67) as it applies to this district, all will depend upon decisions made at higher headquarters. IV. Major problems that possibly can be remedied in part or total at this level are: 1. instilling an ofensive attitude in the VN leaders; 2. the reopening of the BA SE road and Cantho [sic] River and other LOCs and 3. in a small way, revive the paciication program. Problems which must be dealt with at higher headquarters are 1. remedying the serious gap which does exist between VC and RF/PF irepower and 2. a study of the feasibility and the establishing of directives to consolidate PF [outposts] and employ an ofensive means of protecting the LOCs, etc. that are now inade- 272 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency quately secured by the [outposts]. I have recommended such a plan to the district chief, but it will take an order from above before this will become reality. William D. Corliss Major, Armor District Senior Advisor APPENDIX D Military Assistance Command, Vietnam, Military Assessment, September 1967 his appendix contains a memorandum bearing a telegram cable to President Lyndon Johnson from GEN William C. Westmoreland, then commanding general of MACV. he report was passed to the President by Walter W. Rostow, then Assistant for National Security Afairs (a post that is now National Security Advisor). Rostow titles his White House memorandum “General Westmoreland’s Activities Report for September.” his title makes the document seem like little more than an administrative report; the original cable from Westmoreland was titled, “Military Assessment for the Month of September.” his is one of General Westmoreland’s campaign assessment reports to the President.1 he excerpts reprinted here are the introduction to the cable and the section on the IV Corps area of operations. Phong Dinh Province was in IV Corps, so this appendix provides some comparison between the province-level reporting in Appendix C and theater-level reporting for the same area of operations. he time frames of the two reports are diferent: Westmoreland’s report covers the month of September, 1967, while the Phong Dinh report in Appendix C covers the post-Tet period in 1968. his 1967 report was chosen because it is more of a routine report than the post-Tet 1968 reports; the MACV reports in the immediate post-Tet period tended to focus almost entirely on military operations and less so on other relevant campaign assessment issues. he purpose of providing this report is to give the reader an opportunity to view a declassiied military assessment. It is also intended to provide additional detail in support of the Vietnam case study in Chapter Six. he text is published verbatim in its original format, beginning with the lead for General Westmoreland’s cable, skipping the I, II, and III Corps summaries, and ending with the IV Corps summary. It also does not include additional functional-area information published at the end of the report.2 he report is marked by an optimistic tone (with some exception), and 1 here was more than one type of assessment report, including the joint report written with U.S. Paciic Command referenced in Chapter Six. However, this cable was clearly titled “assessment.” 2 Included in this information is the report on 7th Air Force activities. hese are primarily input indicators, such as 1,669 reconnaissance missions were lown, 65,896 tons of cargo were airlifted, 406,000 gallons of defoliant were dispersed. his information would also include what could be considered output indicators in the form 273 274 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency it is derived primarily from quantitative input and output indicators. While it does relect Westmoreland’s attrition strategy, there is also some focus on paciication and the development of South Vietnamese governance capacity; Westmoreland emphasizes both of these as areas for improvement in the introduction. He includes a list of military objectives in each corps reporting section, a helpful reminder that was replicated in the CA template (see Appendix A). One of the most notable aspects of the report is that it contains a great deal of tactical detail without presenting any serious strategic analysis. Rostow’s title seems more accurate than Westmoreland’s. No senior policymaker should be inundated with information like, “Each airield has established a joint command post where all internal defense, ambushes, patrols, harassment and interdiction ires and reaction operations are coordinated and controlled.” he report does not tell the President what the quantitative data mean or why they are important. None of the numbers are accompanied by caveats, so the President would have to assume Westmoreland is presenting them as absolutely accurate (e.g., “Small unit operations numbered 56,864”). Assessment language is generally vague not adequately explanatory, nor is it clearly connected to objectives in a way that would help the President understand the direction of the campaign.3 For example: “Progress was made toward improving and maintaining the security of major highways and waterways in the Delta.” he report does not explain why these highways and waterways are important. It also does not explain the colorcoding system used to describe each route (red, amber, or green), and it fails to explain the strategic signiicance of either securing or losing control of these routes to the VC. If the commanding general of MACV felt it important to tell the President that “interprovincial Route 23 from Sa Dec to Cho Mai was changed from red to amber,” then he should explain why this route was of strategic signiicance.4 he report provides almost no predictive analysis, and it does not provide a justiication for the allocation of resources. Instead, it is a neatly packaged list of igures that give the impression of meaning and progress without explaining much of import. In many ways, the detail in this report is similar to the detail in the Phong Dinh Province report. However, subordinate reports are intended for an internal audience, and they are used as a platform for the development of executive reports. Detail that was appropriate in the Phong Dinh report would not be appropriate in the memorandum to the President. he Phong Dinh report also provided more analysis (albeit at the of battle damage assessment: 47 artillery positions destroyed or damaged, 15 37/57-mm anti-aircraft positions destroyed, 140 sampans destroyed. It does not explain how any of these metrics are strategically relevant, nor does it explain how these numbers are relevant except that they had increased or or decreased from the previous month. he supporting information also described how the weather inhibited or did not inhibit operations. 3 4 here are exceptions to this trend, including the report on the recent elections. It is quite possible that Johnson requested this degree of detail and that Westmoreland was simply providing what he was asked to provide. he intent of this analysis is not to blame Westmoreland for delivering an inefective assessment; instead, it is to show what an inefective assessment looks like. Military Assistance Command, Vietnam, Military Assessment, September 1967 275 province level) than this theater-level report. In the excerpt, text added for clariication is in brackets. 276 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Text of Cable from General Westmoreland Subject: Military Assessment for the Month of September [Tuesday, October 10, 1967] he irst week of last month was marked by enemy terrorist rocket and mortar attacks designed to intimidate the people of South Vietnam in order to disrupt the national elections. A determined people answered that challenge unmistakably. Again the enemy has been defeated in his eforts to gain a major victory. Our combined forces are holding the line in some of the bitterest ighting of this conlict along the DMZ, and are seeking to protect the people and destroy the enemy throughout the country. In order to gain and hold the people we must be able to aford them stability and protection. It is in the area of paciication and in the development of South Vietnam’s civil and military leaders that we must progress. [Skip to IV Corps] Fourth Corps he objectives in the Fourth Corps tactical zone for the month of September continued with emphasis on the destruction of Viet Cong main and provincial units and their principal bases; furtherance of the Revolutionary Development efort; upgrading the security and preventing interdiction of the major lines of communication; and improving the defenses of our major airields. Progress continues in improving the defenses of the airields at Can ho, Binh huy, Soc Trang, and Vinh Long. Day and night aerial observation of the areas and approaches to the airields, the conduct of irely missions nightly over these areas, and the construction of permanent and temporary revetments for aircraft protection continues. Each airield has established a joint command post where all internal defense, ambushes, patrols, harassment and interdiction ires and reaction operations are coordinated and controlled. he enemy situation in the Fourth Corps continues to deteriorate slowly. he Viet Cong propaganda and operational eforts against the national elections were virtually unsuccessful as evidenced by a turnout of more than 85 percent of the registered voters. Continued Government of Vietnam pressure, particularly against Viet Cong main force units and base areas, has afected their ability to signiicantly deter the paciication efort and interdict lines of communication. he Viet Cong have not had a major victory this year and intelligence sources conirm that they are experiencing morale problems. With the present U.S./Government of Vietnam force level in Fourth Corps, gains should continue at a slow to moderate pace. he overall morale, combat efectiveness, and ighting spirit of Fourth Corps units continue to be good. Military Assistance Command, Vietnam, Military Assessment, September 1967 277 Republic of Vietnam Armed Forces units conducted a total of 146 major unit operations of battalion size or larger, a 15 percent decrease from last month. hese operations were, however, of a longer average duration. Small unit operations numbered 56,864, the highest number reported to date. Twenty-nine airmobile operations were conducted with a total of 33,970 Republic of Vietnam Armed Forces troops being airlifted. Cuu Long 63, a bilateral U.S./Republic of Vietnam Armed Forces airmobile operation conducted in Base Area 470 as part of Coronado V, resulted in 70 Viet Cong killed in action, 52 Viet Cong captured and 33 weapons captured. here were 23 additional operations penetrating seven other Viet Cong base areas resulting in 118 Viet Cong killed in action, 129 captured and 55 weapons captured. A 4.2 to 1 friendly killed versus Viet Cong killed ratio and 2.5 to 1 weapons captured versus weapons lost ratio compare favorably with previous months. Progress was made towards improving and maintaining the security of major highways and waterways in the Delta. During the month, National Highway 4 remained open to two-way commercial traic with only brief interruptions caused by the destruction of the An Cu bridge in Dinh Tuong Province and four cratering incidents on Route 4. he following changes in line of communication security have occurred in the past 40 days: Highway 4 from Fourth/hird Corps boundary to My ho was changed from amber to green, inter-provincial Route 40 from Vi hanh to Highway 4 was changed from red to amber, inter-provincial Route 23 from Sa Dec to Cho Mai was changed from red to amber, the Mang hit-Nicholai Waterway was changed from red to amber, the My ho River was changed from amber to green, and inter-provincial Route 30 from Kien Von to Hong Ngu (Kien Phong Province) changed from amber to green. he Cho Gao Canal in Dinh Tuong Province and the Dong Kien Canal in Kien Phong Province were reported for the irst time and are carried as green lines of communications. All airields within Fourth Corps are operational. No airields or major lines of communication were closed by the annual lood in September. Paciication programs remain behind schedule; however, Revolutionary Development activity in each of the provinces began to accelerate after the national election period. Red teams continued to be shifted from irst semester to second semester hamlets and by September 30 approximately 60 percent of the shift was completed. here have been no reported Viet Cong attacks on irst semester hamlets. [“Semesters” appear to have been levels of progress in the Revolutionary Development program for New Life hamlets.] he Viet Cong initiated ive incidents against Revolutionary Development cadre resulting in ive Revolutionary Development cadre, ive Regional Force soldiers, and three civilians killed and 21 Revolutionary Development cadre and two Regional Force soldiers wounded. A joint U.S./Government of Vietnam team, headed by Government of Vietnam Brigadier General Hon, completed an inspection of Revolutionary Development activities in all Fourth Corps Provinces. his team reviewed each Province’s potential for 278 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency completing the current 1967 program and previewed the 1968 plans to insure continuity of efort and purpose. he eforts of this team are expected to give added impetus to the paciication [efort] in Fourth Corps. he overall efectiveness of Regional Forces and Popular Forces units remains satisfactory. he number of Regional Forces and Popular Forces desertions average approximately 990 per month, 50 percent less than last year’s rate. he Popular Forces desertion rate is about double that of the Regional Forces. A study has been initiated to identify the reasons for this. In an efort to increase the efectiveness of Regional Forces and Popular Forces units, a program of instruction for in-place training of these units is being developed. he general enemy situation deteriorated slightly as compared with July and August. Although the Viet Cong increased their activities markedly during the election period, they were unable to fulill their plans for disruption of the elections. he rate of incidents reached a high point on September 3 and dropped appreciably following the elections. Of the 728 Viet Cong–initiated incidents occurring during the month, 320 of these happened through September 4. Coordinated Army of the Republic of Vietnam 7th Division and U.S. 9th Division operations conducted in Western [sic] Dinh Tuong Province relieved enemy pressure directed against Highway 4. Documents captured during the period revealed a four-phase Viet Cong plan of operation in Dinh Tuong Province. One phase of this plan, interdiction of lines of communication, was accomplished only to a limited degree; phases Two, hree, and Four (attacks on rear areas, armored cavalry units and infantry units) were not accomplished. No major Viet Cong ground attacks were initiated but harassment tactics, such as shelling of the district towns of Phong Phu, huan Nhon, and Ke Sach in Phong Dinh Province and Song Ong Doc in An Xuyen Province, continued. Agent reports indicate that the Viet Cong have not been able to overcome their shortage of qualiied cadre and that recruitment problems continue to increase. Losses sustained by the Viet Cong totaled 944 killed, 300 captured, and 377 weapons lost compared to August losses of 1005 killed, 332 captured, and 302 weapons lost. Returnees under the Chieu Hoi program totaled 778. In addition to the Chieu Hois, there were 152 Hoa Hao [a Buddhist sect] soldiers who returned to Government of Vietnam control. APPENDIX E Debate over Effects-Based Operations his section examines the debate over the use of efects-based theories in warfare and in COIN. It provides depth to the arguments referenced throughout this monograph and should help explain some of the underlying concerns behind EBA. Chapter Four described efects-based theory and EBO, so this appendix begins by presenting a critical review of efects-based theory. hat discussion is followed by various defenses of EBO published in journals between 2008 and 2011. he appendix concludes with an examination of the internal contradictions in EBO doctrine and the connection between EBO and systems analysis. his research will not settle the debate over EBO, nor does it address the applicability of efects-based theory or assessment to operations other than COIN. EBO is clearly tied to SoSA and it relies on a SoSA understanding of the operating environment. While EBO nodes and links may or may not be intended as a literal interpretation of actual people and tangible targets, doctrinal EBA is a literal interpretation of the operating environment, and it depends on precision and accuracy to be efective. here are internal contradictions in EBO theory as presented in doctrine and described in the literature. The Debate over Effects-Based Operations in Western Literature he debate over EBO is muddied by inconsistencies between what appear to be two loosely bounded groups of EBO proponents. he irst group, represented by early advocates of RMA and EBO, including ADM Bill Owens, describes RMA (and inclusively EBO) as a means of achieving precision and accuracy through both technological advances and changes in operational methods.1 he second group, represented by more contemporary advocates, including P. Mason Carpenter and Tomislav Ruby, contends that EBO was taken too literally by critics and is simply a means for improv- 1 See Owens, 2000. 279 280 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency ing and reining the time-tested theories of traditional operational art.2 Oicial doctrine relects aspects of arguments from both groups and is often internally contradictory. Some manuals clearly state that EBO accepts the immutably complex and chaotic nature of war while also calling for precision, accuracy, and mathematical rigor that seems out of reach in a complex and chaotic environment. he most comprehensive rebuttal to EBO comes from H. R. McMaster. He builds a case against RMA, NCW, ONA, and the other concepts that EBO is associated with.3 McMaster believes that the eforts to transform defense doctrine and policy in the 1990s were tied to a hubristic belief in the value of technology and a lawed association between capitalist business practices and military art and science.4 He states, “he contradiction between the assumption of information superiority in future war and the ‘dynamic, uncertain, and complex’ security environment undermines the intellectual foundation for defense transformation.”5 He addresses strategic simulation in detail and raises concerns that, not only are military simulations based on NCW and EBO lawed, but they also reinforce what he believes are lawed theories. According to McMaster, he concept of efects-based operations assumed near certainty in future war; it treated the enemy as a “system” that could be fully understood through a process called “operational net assessment (ONA).” Because ONA would produce “a comprehensive system-of-systems understanding of the enemy and the environment,” operations could achieve a high degree of speed as well as precision in operational efects. he enemy would be unable to keep pace with the “high rates of change” imposed on him.6 2 See Carpenter and Andrews, 2009, and Ruby, 2008. 3 For deinitions of NCW and ONA see he Implementation of Network-Centric Warfare (DoD Oice of Force Transformation, 2005). Briely, NCW “broadly describes the combination of strategies, emerging tactics, techniques and procedures, and organizations that a fully or even partially networked force can employ to create a decisive warighting advantage. Hannan (2005, p. 27) describes ONA as an analytical process designed within the Department of Defense to enhance decision-making superiority for the warighting Commander. ONA plans to integrate people, processes, and tools using multiple information sources and collaborative analysis. he goal is a shared knowledge environment, with supporting information tools, for planners and decision-makers to focus capabilities. 4 According to McMaster (2003, p. 41), Hubris permeates the language of defense transformation and is particularly evident in the reductive fallacies of information superiority, dominant battlespace knowledge, and their various companion terms. Warnings were ignored. 5 McMaster, 2003, p. 9. 6 McMaster, 2003, p. 74. Debate over Effects-Based Operations 281 McMaster believes that eforts to transform defense doctrine along the lines of NCW and EBO had a deleterious efect on military planning, operations, and, by extension, assessment. He lays out this argument as follows: Because belief in certainty or uncertainty as the dominant condition in war is relative, so are the consequences of that belief. Unique circumstances in combat will shift experiences, capabilities needed, and methods along a continuum between extremes of certainty and uncertainty. Assuming near-certainty, however, generates a series of derivative assumptions and predilections that are likely to lead to diiculties when that base assumption is proven false.7 In essence, McMaster claims that the proponents of EBO have set aside the commonly agreed-upon terms of warfare as expressed in MCDP 1, Warighting, and FM 3-0, Operations; EBO is, to some extent, a rejection of the traditionalist understanding of war. he arguments of McMaster and likeminded theorists gained traction as the wars in Afghanistan and Iraq progressed, but DoD remained—and remains—rooted in NCW and EBO.8 Greenwood and Hammes state, “Warfare has always been interactively complex. his is not a new discovery, but admitting that DoD strayed from this fundamental truth is.”9 he irst real break with NCW and, speciically, EBO did not come until mid-2008. On August 14, 2008, Marine Corps Gen. James N. Mattis, then commanding general of U.S. Joint Forces Command, not only renounced EBO theory but also rescinded oicial recognition of EBO in U.S. joint doctrine: he underlying principles associated with EBO, ONA, and SoSA are fundamentally lawed and must be removed from our lexicon, training, and operations. EBO thinking, as the Israelis found [in 2006 in Lebanon], is an intellectual “Maginot Line” around which the enemy maneuvered. Efective immediately, U.S. Joint Forces Command will no longer use, sponsor, or export the terms and concepts related to EBO, ONA, and SoSA in our training, doctrine development, and support of JPME [joint professional military education].10 7 McMaster, 2003, p. 91. 8 his conclusion is based on the observation that joint doctrine published between 2001 and 2010 continued to incorporate both NCW and EBO concepts. For an example of the persistence of the “see-all, know-all” EBO mindset, see Nakashima and Whitlock, 2011. In this article on the new unmanned surveillance drone Gorgon Stare, a senior U.S. general oicer states, “Gorgon Stare will be looking at a whole city, so there will be no way for the adversary to know what we’re looking at, and we can see everything.” his comment is in line with Owens’s stated philosophy in Lifting the Fog of War (Owens, 2000). 9 10 Greenwood and Hammes, 2009. Mattis, 2008, p. 6 (emphasis in original). As late as mid-2010, Joint Forces Command had responsibility for studying and developing U.S. joint doctrine. 282 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Mattis believed that current doctrine through 2008 (as relected in this monograph) had retained some useful elements of EBO theory. However, he was in line with Marine Corps capstone doctrine (MCDP 1) in stating that all warfare—including conventional conlict and COIN—deies the neat construct required by EBO and SoSA. He states that “[a]ll operating environments are dynamic with an ininite number of variables; therefore, it is not scientiically possible to accurately predict the outcome of action. To suggest otherwise runs contrary to historical experience and the nature of war.” Mattis refers to the Israel Defense Forces’s (IDF’s) experience with EBO during the brief war with Hezbollah in 2006: “Although there are several factors why the IDF performed poorly during the war, various post-conlict assessments have concluded that over reliance on EBO concepts was one of the primary contributing factors for their defeat.” He cites an Israeli general oicer who stated that EBO doctrine was “in complete contradiction to the most important basic principles of operating an army in general . . . and is not based upon, and even ignores the universal fundamentals of war.”11 Retired Marine Corps Lt. Gen. Paul K. Van Riper, the Bren Chair of Innovation and Transformation at Marine Corps University, reinforced Mattis’s decision to unbind EBO from U.S. joint doctrine. Van Riper identiies three types of EBO theory: he irst (and original) was designed to improve air-to-ground targeting, the second to improve U.S. Army ire support coordination, and the third a joint efort designed to envision and exploit a comprehensive picture of the enemy. He states that, as originally envisioned, the irst type of EBO is “only efective with manmade systems that have an identiiable and tightly coupled structure, such as integrated air defenses, distribution networks, and transportation complexes.”12 he third type he labels as “egregious” and a “vacuous concept.” Van Riper contends that EBO proponents do not appreciate the unpredictable nature of cascading efects:13 he nearly limitless ways that an action might ricochet through an interactively complex or nonlinear system mean that for all practical purposes, the interactions within the system exceed the calculative capacities of any computer to follow, at least in any meaningful way. he numbers are so large that even the most advanced computers would take billions of years to process them. . . . In short, ONA and SoSA argue for a pseudoscientiic approach to operational planning. 11 Mattis, 2008, p. 4 (ellipses in original). Mattis cites Kober, 2008; Commission to Investigate the Lebanon Campaign in 2006, 2008; and Matthews, 2008, pp. 23–28, 61–65. his last comment relects the schism that erupted over EBO in the wake of the Hezbollah campaign—some Israeli oicers continue to advocate or at least defend EBO. 12 13 Van Riper, 2009, p. 83. Van Riper, 2009, p. 83. He provides a chess analogy to support his statement regarding efect calculations. Even these calculations assume that the data needed to feed them are available and accurate. Debate over Effects-Based Operations 283 Van Riper draws a distinction between a structurally complex system and an interactively complex system, with the latter deined by emergent behavior. In the structurally complex system, nodes and links are relatively static, while in an interactively complex system, the relationship between elements is “constantly in lux, and links—as conceived of by EBO advocates—are often not apparent and are frequently transitory.”14 Like McMaster, Mattis, Van Riper, and others, Kelly and Kilcullen see COIN as a complex adaptive system rather than a closed or “equilibrium-based system.”15 hey also believe that it is impossible to see the second- or third-order efects from military actions: “Such foresight would require knowledge of all possible efects, including ‘efects of efects.’ . . . Developing such a sequence of knowledge in military operations is almost impossible.”16 hey believe that EBO will remain “at best a worthy aspiration.”17 Milan N. Vego of the U.S. Naval War College sees almost no redeeming value in EBO. In “Efects-Based Operations: A Critique,” he speciically addresses some of the problems with EBO assessment that are relevant to this examination: [B]oth quantitative and qualitative measurements are equally subject to political manipulation, mirror-imaging, and biases. A more serious deiciency of the assessment concept is its almost total lack of sound intellectual framework. EBO proponents assume that the efects of one’s actions could be precisely measured and almost instantaneously known to decisionmakers. his is highly unlikely. his heavy reliance on various quantifying measurements and fast feedback raises the issue of the utility of the efects-based approach, especially at the operational and strategic levels of war.18 Advocacy and defense of EBO range from eforts to reine theory and practice to outright rebuttal of McMaster, Mattis, and Van Riper’s arguments. ADM (ret.) Bill Owens does not address EBO speciically but his argument in support of RMA and its associated concepts is, in essence, a claim that EBO is practicable. Owens was the oicial advocate for RMA in DoD in the late 1990s. In Lifting the Fog of War, he clearly laid out the objective of RMA and its component concepts (including EBO): By 2010—and earlier if we accelerate the current rate of research and procurement—the U.S. military will be able to “see” virtually everything of military signiicance in and above [the operating environment] all the time, in all weather conditions, and regardless of the terrain. We will be able to identify and 14 Van Riper, 2009, p. 84. 15 Kelly and Kilcullen, 2006, p. 66. 16 Kelly and Kilcullen, 2006, p. 71. 17 Kelly and Kilcullen, 2006, p. 72. 18 Vego, 2006, pp. 56–57. 284 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency track—in near real time—all major items of military equipment, from trucks and other vehicles on the ground to ships and aircraft. More important, the U.S. military commander will understand what he sees.19 Similarly, while Alberts, Garstka, and Stein never mention EBO speciically, their detailed defense of NCW is also a defense of its associated terms and concepts. hey ofer a more pragmatic explanation of NCW as it relates to the traditional understanding of warfare: he fact that warfare will always be characterized by fog, friction, complexity, and irrationality circumscribes but does not negate the beneits that networkcentric operations can provide to the forces in terms of improved battlespace awareness and access to distributed assets. While predicting human and organizational behavior will remain well beyond the state of the art, having a better near real-time picture of what is happening (in situations where this is possible from observing things that move, emit, etc.) certainly reduces uncertainty in a meaningful way.20 Paul K. Davis of RAND recognizes the inconsistencies between the more literal interpretations of EBO theory and the realities of the battleield but believes that there are redeeming qualities in EBO. He recommends a family of models and mathematical games to help penetrate the fog of war.21 Davis focuses almost exclusively on conventional warfare. In 2002, then–Air Force Lt Col Christopher W. Bowman, an efectsbased approach proponent, also acknowledged uncertainty in war and believed that assessment was the “Achilles’ heel” of EBO. William J. Gregor, a social scientist at the U.S. Army Command and General Staf College, strongly defends EBO (and therefore EBA) and claims that the critics of EBO theory do not understand its premise.22 Gregor states that EBO models are valid and can be used to see through the complexity of COIN; the military is at fault for not collecting and providing adequate data to feed efects-based operational and assessment models.23 He also believes that senior oicers have gone overboard in rejecting EBO: 19 Owens, 2000, p. 119. 20 Alberts, Garstka, and Stein, 1999, p. 11. 21 Davis, 2001. 22 See Gregor, 2010. 23 he U.S. Army weaves EBO throughout the 2010 version of FM 5-0 (HQDA, 2010). Sections of that ield manual provide a qualiied defense of efects-based theory. While FM 5-0 clearly prescribes the use of EBA, FM 3-0 (published two years earlier and just prior to Mattis’s elimination of EBO from joint doctrine) equivocates: Army forces conduct operations according to Army doctrine. he methods that joint force headquarters use to analyze an operational environment, develop plans, or assess operations do not change this. . . . Army forces do not use the joint systems analysis of the operational environment, efects-based approach to planning, or efects assessment. hese planning and assessment methods are intended for use at the strategic and operational levels by Debate over Effects-Based Operations 285 Because the efect of any action is conditioned by the environment, it is important to develop conidence in the anticipated efect by observing patterns in a large body of data collected over a long period. In both Afghanistan and Iraq, the United States should already have that data. Regrettably, it does not. . . . Currently, there are no consistent measures of efectiveness and only a small body of data with which to judge the likelihood of success. hus, military experience cannot be generalized, and the views of generals are ideographic.24 Notwithstanding this last criticism, EBO and EBA are alive and well. JP 3-0 and FM 5-0 describe a hybrid of efects-based planning processes and assessment process, and efects-based terminology continues to permeate joint doctrine.25 Mattis’s edict elicited a wave of negative reaction from Air Force oicers, including Col P. Mason Carpenter and Col William F. Andrews, who argued that EBO was “combat proven” and “was the basis for success of the Operation Desert Storm air campaign and Operation Allied Force.”26 hey assert that “no one is suggesting certainty or absolute determinism” in EBO, so Mattis’s criticisms were misdirected.27 Air Force oicer Tomislav Z. Ruby contends that Mattis is simply incorrect in his understanding of EBO and that his criticism of efects-based concepts does not, in fact, address EBO at all. He states that “the characteristics that General Mattis refers to in the memorandum [abolishing efects-based language in joint doctrine] were a ‘strawman’ no longer relecting any oicial (read Service) position.”28 He later goes on to posit that the services were not properly applying EBO, an argument echoed by other EBO advocates. Ruby argues that some (including Mattis) confuse EBO with SoSA and that SoSA is only an “analytical tool” to help understand the enemy. he U.S. Joint Staf section responsible for joint doctrine (J7) has argued that joint doctrine does not use EBO language or refer to SoSA explicitly.29 properly resourced joint stafs. However, joint interdependence requires Army leaders and stafs to understand joint doctrine that addresses these methods when participating in joint operation planning or assessment or commanding joint forces. (JPs 3-0 and 5-0 establish this doctrine.) (HQDA, 2008, p. D-2; emphasis added) 24 Gregor, 2010, p. 111. 25 U.S. Joint Chiefs of Staf, 2010; HQDA, 2010. 26 Carpenter and Andrews, 2009, p. 78. In his book Hollow Victory: A Contrary View of the Gulf War, Jefrey Record states that the Gulf War “provided no genuine test of U.S. ighting power” and that it was an anomaly. He adds that the Gulf War “cannot—and should not be permitted to—serve as a model for the future” (Record, 1993, p. 135). he mixed results presented in the Gulf War Air Power Survey also call into question the claim that EBO was “proven” in the Gulf War. It would, however, be fair to say that air power had a signiicant impact on the success of the campaign. 27 Carpenter and Andrews, 2009, p. 81. 28 Ruby, 2008, p. 28. 29 U.S. Joint Chiefs of Staf, 2009a. 286 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency But SoSA is central to EBO: It is the map of nodes and links upon which EBO planning and assessment depends. JPs 2-0 and 2-01.3, published in 2007 and 2009 respectively, contain pictures of the “interconnected operational environment” that is a near-identical copy of the SoSA map in Figure 4.2 in Chapter Four of this monograph.30 he efects-based intelligence process in JPs 2-0 and 2-01.3 is literal. JP 2-0 interprets the SoSA map as a physical map, not as a conceptual prompt. It states, “A systems-oriented [analysis] efort is crucial to the identiication of adversary [centers of gravity], key nodes and links.” A systems-oriented understanding of the battle is derived from SoSA: “he [analyst] must also have a detailed understanding of how each aspect of the operational environment links to the others and how various permutations of such links and nodes may combine to form [centers of gravity].”31 In joint intelligence doctrine, nodes and links are things or people to be identiied and shaped by both kinetic and nonkinetic efects. he language found in unoicial EBO publications and oicial doctrine may not be identical, but the concepts essentially are. he argument that joint doctrine does not contain EBO references is semantic and misleading. Figure E.1 shows the similarities between “system-of-systems analysis” as portrayed in the Commander’s Handbook, the “interconnected operational environment” in JPs 2-0, 3-0, and 5-0, and the “systems perspective of the operational environment” in JP 2-1.3. JP 3-24 incorporates the same diagram into a larger diagram depicting the joint COIN assessment process.32 he diference in titles is rendered meaningless by the obvious similarities in the diagrams. All three describe nodes and links, which are key elements of EBO. Internal disagreement among the various segments of EBO literature makes an accurate assessment of efects-based theory challenging: Some literature presents EBO in a literal fashion, while some journal articles and most responses to Mattis’s memorandum contend that critics take EBO doctrine too literally and therefore do not understand efects-based theory or practice. he original language of defense transformation also tries to hedge on the nature of EBO, stating, “Rather than a new form of 30 U.S. Joint Chiefs of Staf, 2007, p. IV-2; U.S. Joint Chiefs of Staf, 2009b, p. II-45. 31 U.S. Joint Chiefs of Staf, 2007, p. IV-20. 32 Further, JP 2-0 (U.S. Joint Chiefs of Staf, 2007, p. IV-3) clearly describes a concrete and measurable EBO process: Combined with a systems perspective, the identiication of desired and undesired efects can help commanders and their stafs gain a common picture and shared understanding of the operational environment that promotes uniied action. [Commanders] plan joint operations by developing strategic objectives supported by measurable strategic and operational efects and assessment indicators. At the operational level, the JFC [joint force commander] develops operational-level objectives supported by measurable operational efects and assessment indicators. Joint operation planning uses measurable efects to relate higher-level objectives to component missions, tasks, or actions. Debate over Effects-Based Operations 287 Figure E.1 Comparison of Systems Analysis Diagrams in Joint Doctrinal Publications Commander’s Handbook Joint Publication 2-0 Information Social Joint Publication 2-01.3 Information Infrastructure Node Social Military Information Infrastructure Node Infrastructure Social Military Military Node Link Link Operational COG Strategic COG Economic Political Economic Political Joint Publication 3-0 Weakness Link Operational COG Strategic COG Joint Publication 3-24 Economic Joint Publication 5-0 Information Information Social Social Strength Political Infrastructure Infrastructure Military Military Node Link Political Economic Political Economic SOURCES: (top, left to right) Joint Warfighting Center, 2006, p. II-2, Figure II-1 (SoSA); U.S. Joint Chiefs of Staff, 2007, p. IV-2, Figure IV-1; and U.S. Joint Chiefs of Staff, 2009b, p. II-45, Figure II-16; (bottom, left to right) U.S. Joint Chiefs of Staff, 2010, p. II-24, Figure II-6; U.S. Joint Chiefs of Staff, 2009c, p. X-16, Figure X-4 (end-state conditions); and U.S. Joint Chiefs of Staff, 2006, p. III-17, Figure III-2. RAND MG1086-E.1 warfare, EBO is a way of thinking or a methodology for planning.”33 here are two broad positions in the EBO community: 1. EBO as described in the Commander’s Handbook (which is not oicial doctrine) and in some joint doctrine (e.g., JP 2.0, JP 2-01.3) is a guide for concrete action that requires large quantities of accurate and closely analyzed information, and technology can penetrate the fog of war to facilitate this process. 2. EBO is simply a way of thinking about planning and operations; it is not a strict guide for concrete action or assessment, and efects-based theory takes into account the immutably chaotic nature of war. he latter argument is diicult to reconcile with the explicit and directive nature of most oicial efects-based literature, and it does not jive with concrete node-andlink analysis and targeting. here is a clear efort in joint efects-based doctrine to 33 DoD Oice of Force Transformation, 2003, p. 34. 288 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency shape U.S. military operations through the application and interpretation of SoSA. In most cases, the text accompanying these diagrams describes a systematic and practical process whereby stafs attempt to identify actual nodes and links in the environment, shape these nodes and links through efects, and then assess the impact of these efects on the system of systems. If the process is not intended to be taken literally—as the rebuttals to Mattis contend—then this intent is not conveyed in doctrine. he debate over EBO in the U.S. military is ongoing and will not be settled with the conclusions and indings presented, particularly because this research efort focused speciically on COIN assessment. For the purposes of COIN assessment, however, EBO doctrine depends on the ability of military stafs to (1) see and explain all relevant events and actions with both precision and accuracy and (2) produce a clear and accurate picture of the complex adaptive COIN environment. An assessment of an efects-based operation would be able to portray and also explain past events both accurately and in detail. his places EBA on one end of a possible spectrum of assessment approaches: It assumes that suicient accurate data are available to deliver accurate quantitative assessment from the center. Bibliography Historical documents in the Vietnam Center and Archive online database hosted by the Texas Tech University can be accessed via a search at http://www.virtual.vietnam.ttu.edu. Documents in the National Archives online research catalog can be accessed via a search at http://arcweb.archives.gov. Abbaszadeh, Nima, Mark Crow, Marianne El-Khoury, Jonathan Gandomi, David Kuwayama, Christopher MacPherson, Meghan Nutting, Nealin Parker, and Taya Weiss, Provincial Reconstruction Teams: Lessons and Recommendations, Princeton, N.J.: Woodrow Wilson School of Public and International Afairs, January 2008. Abella, Alex, Soldiers of Reason: he RAND Corporation and the Rise of the American Empire, Orlando, Fla.: Harcourt, 2008. Abramowitz, Michael, “Congress, White House Battle over Iraq Assessment,” Washington Post, September 10, 2007. Ackof, Russell L., “Towards a System of Systems Concepts [sic],” Management Science, Vol. 17, No. 11, July 1971, pp. 661–671. As of June 30, 2011: http://ackofcenter.blogs.com/ackof_center_weblog/iles/AckofSystemOfSystems.pdf ———, “From Data to Wisdom,” Journal of Applied Systems Analysis, Vol. 16, 1989, pp. 3–9. ———, “Systems hinking and hinking Systems,” System Dynamics Review, Vol. 10, Nos. 2–3, Summer–Fall 1994, pp. 175–188. Ahern, homas L. Jr., Vietnam Declassiied: he CIA and Counterinsurgency, Lexington, Ky.: University of Kentucky Press, 2010. Alberts, David S., John J. Garstka, and Frederick P. Stein, Network Centric Warfare: Developing and Leveraging Information Superiority, 2nd ed., Command Control Research Program Publications, 1999. As of June 30, 2011: http://www.carlisle.army.mil/DIME/documents/Alberts_NCW.pdf Allison, Paul David, Missing Data, housand Oaks, Calif.: Sage Publications, 2002. Anderson, Edward G., Jr., “A Dynamic Model of Counterinsurgency Policy Including the Efects of Intelligence, Public Security, Popular Support, and Insurgent Experience,” System Dynamics Review, Vol. 27, No. 2, April–June 2011, pp. 111–141. Anderson, Joseph, “Military Operational Measures of Efectiveness for Peacekeeping Operations,” Military Review, Vol. 81, No. 5, September–October 2001. Armstrong, Nicholas J., and Jacqueline Chura-Beaver, “Harnessing Post-Conlict Transitions: A Conceptual Primer,” Peacekeeping and Stability Operations Institute, September 2010. 289 290 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Arquilla, John, and David Ronfeldt, In Athena’s Camp: Preparing for Conlict in the Information Age, Santa Monica, Calif.: RAND Corporation, MR-880-OSD/RC, 1997. As of June 30, 2011: http://www.rand.org/pubs/monograph_reports/MR880.html Arthur, James F., testimony before the House Armed Services Committee on hamlet paciication in Vietnam, March 1970. Available via the Vietnam Center and Archive online database. BACM Research, Vietnam War Document Archive (Disk 1 and Disk 2), set of two CDs containing oicial archival records, 2009. Baranick, Mike, Center for Technology and National Security Policy, National Defense University; David Knudson, Center for Army Analysis; Dave Pendergraft, Headquarters U.S. Air Force; and Paul Evangelista, U.S. Army Training and Doctrine Command Analysis Center, “Improving Analytical Support to the Warighter: Campaign Assessments, Operational Analysis, and Data Management, Working Group 1: Data and Knowledge Management,” annotated brieing, Military Operations Research Society workshop, April 19–22, 2010. As of June 30, 2011: http://www.mors.org/UserFiles/ile/2010%20IW/MORS%20IW%20WG1.pdf Bar-Yam, Yaneer, Complexity of Military Conlict: Multiscale Complex Systems Analysis of Littoral Warfare, New England Complex Systems Institute, April 21, 2003. As of June 30, 2011: http://necsi.edu/projects/yaneer/SSG_NECSI_3_Litt.pdf Bernard, H. Russell, Research Methods in Anthropology: Qualitative and Quantitative Approaches, 4th ed., Lanham, Md.: AltaMira Press, 2006. Bertalanfy, Ludwig von, General System heory: Foundations, Development, Applications, rev. ed., New York: G. Braziller, 1974 Beyerchen, Alan, “Clausewitz, Nonlinearity, and the Unpredictability of War,” International Security, Vol. 17, No. 3, Winter 1992–1993, pp. 59–90. Biddle, Stephen, Military Power: Explaining Victory and Defeat in Modern Battle, Princeton, N.J.: Princeton University Press, 2004. Birtle, Andrew J., U.S. Army Counterinsurgency and Contingency Operations Doctrine, 1860–1941, Washington, D.C.: U.S. Army Center of Military History, 2004. ———, U.S. Army Counterinsurgency and Contingency Operations Doctrine, 1942–1976, Washington, D.C.: U.S. Army Center of Military History, 2006. Blalock, Hubert M. Jr., ed., Measurement in the Social Sciences: heories and Strategies, Chicago, Ill.: Aldine Publishing Company, 1974. Boot, Max, War Made New: Weapons, Warriors, and the Making of the Modern World, New York: Penguin Group, 2006. Borkowski, Piotr, and Jan Mielniczuk, “Postmodel Selection Estimators of Variance Function for Nonlinear Autoregression,” Journal of Time Series Analysis, Vol. 31, No. 1, January 2010, pp. 50–63. Bousquet, Antoine, he Scientiic Way of Warfare: Order and Chaos on the Battleields of Modernity, New York: Columbia University Press, 2009. Bowman, Christopher W., Operational Assessment—he Achilles Heel of Efects-Based Operations? thesis, Newport, R.I.: U.S. Naval War College, May 13, 2002. As of June 30, 2011: http://handle.dtic.mil/100.2/ADA405868 Bracken, Jerome, Moshe Kress, and Richard E. Rosenthal, Warfare Modeling, Danvers, Mass.: John Wiley and Sons, 1995. Bibliography 291 Brigham, Erwin R., “Paciication Measurement in Vietnam: he Hamlet Evaluation System,” presentation at SEATO Internal Security Seminar, Manila, Philippines, June 3–10, 1968. As of June 30, 2011: http://www.cgsc.edu/carl/docrepository/PaciicationVietnam.pdf ———, “Paciication Measurement,” Military Review, May 1970, pp. 47–55. As of June 30, 2011: http://calldp.leavenworth.army.mil/eng_mr/txts/VOL50/00000005/art6.pdf Brookings Institution, Saban Center for Middle East Policy, Iraq Index: Tracking Reconstruction and Security in Post-Saddam Iraq, Washington, D.C., last updated August 24, 2010. See archive of past reports, as of June 30, 2011: http://www.brookings.edu/saban/iraq-index.aspx Brown, James, Erik W. Goepner, and James M. Clark, “Detention Operations, Behavior Modiication, and Counterinsurgency,” Military Review, May–June 2009, pp. 40–47. Bryman, Alan, Saul Becker, and Joe Sempik, “Quality Criteria for Quantitative, Qualitative and Mixed Methods Research: A View from Social Policy,” Social Research Methodology, Vol. 11, No. 4, October 2008, pp. 261–276. Bunker, Ellsworth, “October 2–15, 1968: he Breakthrough in Paris,” oicial telegram from the U.S. Embassy in Saigon, Vietnam, to U.S. Department of State, Washington, D.C., October 2, 1968a, in BACM Research, Disk 2. ———, “Telegram from the Embassy in Vietnam to the Department of State,” oicial telegram from the U.S. Embassy in Saigon, Vietnam, to the U.S. Department of State, Washington, D.C., October 19, 1968b, in BACM Research, Disk 2. Cabayan, Hriar, Rich Contextual Understanding of Pakistan and Afghanistan (PAKAF): A Strategic Multi-Layer Assessment, unpublished workshop report, April 12, 2010a. ———, Subject Matter Analysis PAKAF Rich Contextual Understanding, MG Flynn Final Progress Report, unpublished workshop report, October 25, 2010b. Campbell, Jason, Michael E. O’Hanlon, and Jeremy Shapiro, Assessing Counterinsurgency and Stabilization Missions, Washington, D.C.: Brookings Institution, policy paper 14, May 2009a. As of June 30, 2011: http://www.brookings.edu/~/media/Files/rc/papers/2009/05_counterinsurgency_ohanlon/05_ counterinsurgency_ohanlon.pdf ———, “How to Measure the War,” Policy Review, No. 157, October–November 2009b. As of June 30, 2011: http://www.hoover.org/publications/policy-review/article/5490 Carmines, Edward G., and Richard A. Zeller, Reliability and Validity Assessment, Beverly Hills, Calif.: Sage Publications, 1979. Carpenter, P. Mason, and William F. Andrews, “Efects-Based Operations: Combat Proven,” Joint Force Quarterly, No. 52, First Quarter 2009, pp. 78–81. Cebrowski, Arthur K., and John J. Garstka, “Network-Centric Warfare: Its Origin and Future,” Proceedings, January 1998, pp. 28–36. Center for Army Analysis, “Afghanistan Consolidated Knowledge System,” brieing, August 30, 2010. CIA—see U.S. Central Intelligence Agency. Cioppa, homas, email exchange with the author, October 8, 2010. 292 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Cioppa, homas, Loren Eggen, and Paul Works, “Improving Analytical Support to the Warighter: Campaign Assessments, Operational Analysis, and Data Management, Working Group 4: Current Ops Analysis—Tactical,” brieing, Military Operations Research Society, April 19–22, 2010. As of June 30, 2011: http://www.mors.org/UserFiles/ile/2010%20IW/MORS%20IW%20WG4.pdf Clalin, Bobby, Dave Sanders, and Greg Boylan, “Improving Analytical Support to the Warighter: Campaign Assessments, Operational Analysis, and Data Management, Working Group 2: Campaign Assessments,” working group brieing, Military Operations Research Society conference, April 19–22, 2010. As of September 15, 2010: http://www.mors.org/UserFiles/ile/2010%20IW/MORS%20IW%20WG2.pdf Clancy, James, and Chuck Crossett, “Measuring Efectiveness in Irregular Warfare,” Parameters, Summer 2007, pp. 88–100. As of June 30, 2011: http://www.carlisle.army.mil/usawc/parameters/Articles/07summer/clancy.pdf Clark, Clinton R., and Timothy J. Cook, “A Practical Approach to Efects-Based Operational Assessment,” Air and Space Power Journal, Vol. 22, No. 2, Summer 2008. As of June 30, 2011: http://www.airpower.maxwell.af.mil/airchronicles/apj/apj08/sum08/clark.html Clark, Dorothy K., and Charles R. Wyman, An Exploratory Analysis of the Reporting, Measuring, and Evaluating of Revolutionary Development in South Vietnam, McLean, Va.: Research Analysis Corporation, 1967. Clausewitz, Carl Von, On War, Colonel J. J. Graham, trans., London: N. Trübner, 1874. Colby, William E., and Peter Forbath, Honorable Men: My Life in the CIA, New York: Simon and Schuster, 1978. Commission to Investigate the Lebanon Campaign in 2006, inal report, January 30, 2008. As of June 30, 2011: http://www.mfa.gov.il/MFA/MFAArchive/2000_2009/2008/Winograd%20Committee%20 submits%20inal%20report%2030-Jan-2008 Connable, Ben, “he Massacre hat Wasn’t,” in U.S. Marines in Iraq, 2004–2008: Anthology and Annotated Bibliography, Washington, D.C.: History Division, U.S. Marine Corps, 2010, pp. 75–81. Connable, Ben, and Martin C. Libicki, How Insurgencies End, Santa Monica, Calif.: RAND Corporation, MG-965-MCIA, 2010. As of June 30, 2011: http://www.rand.org/pubs/monographs/MG965.html Cooper, Chester L., Judith E. Corson, Laurence J. Legere, David E. Lockwood, and Donald M. Weller, he American Experience with Paciication in Vietnam, Vol. 1: An Overview of Paciication, Institute for Defense Analysis, March 1972a. ———, he American Experience with Paciication in Vietnam, Vol. 2: Elements of Paciication, Institute for Defense Analysis, March 1972b. Cordesman, Anthony H., he Uncertain “Metrics” of Afghanistan (and Iraq), Washington, D.C.: Center for Strategic and International Studies, May 18, 2007. As of June 30, 2011: http://www.csis.org/media/csis/pubs/070521_uncertainmetrics_afghan.pdf ———, Analyzing the Afghan-Pakistan War, Washington, D.C.: Center for Strategic and International Studies, draft, July 28, 2008. As of September 15, 2010: http://csis.org/iles/media/csis/pubs/080728_afghan_analysis.pdf ———, “he Afghan War: Metrics, Narratives, and Winning the War,” brieing, Center for Strategic and International Studies, June 1, 2010. Corson, William R., he Betrayal, New York: W. W. Norton and Company, 1968. Bibliography 293 Cosmos, Graham A., MACV: he Joint Command in the Years of Withdrawal, 1968–1973, Washington, D.C.: U.S. Army Center of Military History, 2006. As of June 30, 2011: http://www.history.army.mil/html/books/macv2/CMH_91-7.pdf Cushman, John H., Senior Oicer Debrieing Report of Major General John H. Cushman, RCS CSFOR-74, January 14, 1972. As of June 30, 2011: http://www.dtic.mil/cgi-bin/GetTRDoc?Location=U2&doc=GetTRDoc.pdf&AD=AD0523904 Daddis, Gregory A., No Sure Victory: Measuring U.S. Army Efectiveness and Progress in the Vietnam War, New York: Oxford University Press, 2011. Darilek, Richard E., Walter L. Perry, Jerome Bracken, John Gordon IV, and Brian Nichiporuk, Measures of Efectiveness for the Information-Age Army, Santa Monica, Calif.: RAND Corporation, MR-1155-A, 2001. As of June 30, 2011: http://www.rand.org/pubs/monograph_reports/MR1155.html Davis, Paul K., Efects-Based Operations (EBO): A Grand Challenge for the Analytical Community, Santa Monica, Calif.: RAND Corporation, MR-1477-USJFCOM/AF, 2001. As of June 30, 2011: http://www.rand.org/pubs/monograph_reports/MR1477.html DoD—see U.S. Department of Defense. Donovan, David, Once a Warrior King: Memories of an Oicer in Vietnam, New York: McGraw-Hill, 1985. Downes-Martin, Stephen, research professor, U.S. Naval War College, interview with the author, Kabul, Afghanistan, May 5, 2010a. ———, “Assessment Process for RC(SW),” unpublished draft, Newport, R.I.: U.S. Naval War College, May 24, 2010b. Dror, Yehezkel, Some Normative Implications of a Systems View of Policymaking, Santa Monica, Calif.: RAND Corporation, P-3991-1, 1969. As of June 30, 2011: http://www.rand.org/pubs/papers/P3991-1.html Eisenstadt, Michael, and Jefrey White, “Assessing Iraq’s Sunni Arab Insurgency,” Military Review, Vol. 86, No. 3, May–June 2006, pp. 33–51. Eles, Philip, U.S. Central Command, interview with the author, Tampa, Fla., March 24, 2010. Engelhardt, Tom, “Afghanistan by the Numbers: Is the War Worth It? he Cost in Dollars, Years, Public Opinion, and Lives,” Salon.com, September 14, 2009. As of January 27, 2011: http://www.salon.com/news/feature/2009/09/14/afghanistan Enthoven, Alain C., and K. Wayne Smith, How Much Is Enough? Shaping the Defense Program, 1961–1969, New York: Harper and Row/Santa Monica, Calif.: RAND Corporation, 1971/2005. As of June 30, 2011: http://www.rand.org/pubs/commercial_books/CB403.html Fischer, Hannah, Iraqi Civilian Death Estimates, Washington, D.C.: Congressional Research Service, RS22537, August 27, 2008. Flynn, Michael T., Matt Pottinger, and Paul Batchelor, Fixing Intel: A Blueprint for Making Intelligence Relevant in Afghanistan, Washington, D.C.: Center for a New American Security, January 2010. As of June 30, 2011: http://www.cnas.org/iles/documents/publications/AfghanIntel_Flynn_Jan2010_code507_voices.pdf Folker, Robert D. Jr., Intelligence Analysis in heater Joint Intelligence Centers: An Experiment in Applying Structured Analytic Methods, Washington, D.C.: Joint Military Intelligence College, occasional paper no. 7, January 2000. 294 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Galula, David, Counterinsurgency Warfare: heory and Practice, New York: Frederick A. Praeger, (1964) 2005. Gartner, Scott Sigmund, Strategic Assessment in War, New Haven, Conn.: Yale University Press, 1997. Gates, Robert M., Secretary of Defense, statement to the U.S. Senate Appropriations Committee, Subcommittee on Defense, June 16, 2010. As of September 15, 2010: http://appropriations.senate.gov/ht-defense.cfm?method=hearings. download&id=fac53160-5308-4fd-9b61-bb8112599ce2 Gayvert, David, “Teaching New Dogs Old Tricks: Can the Hamlet Evaluation System Inform the Search for Metrics in Afghanistan?” Small Wars Journal Online, September 8, 2010. As of June 30, 2011: http://smallwarsjournal.com/blog/journal/docs-temp/531-gayvert.pdf Gelb, Leslie H., and Richard K. Betts, he Irony of Vietnam: he System Worked, Washington, D.C.: Brookings Institution Press, 1979. Gibson, James William, he Perfect War: Technowar in Vietnam, New York: Atlantic Monthly Press, 2000. Gilbert, Nigel, Agent-Based Models, housand Oaks, Calif.: Sage Publications, 2008. Goldstein, Gordon M., Lessons in Disaster: McGeorge Bundy and the Path to War in Vietnam, New York: Henry Holt and Company, 2008. Graves, Greg, Patricia Murphy, and Frederick Cameron, “Improving Analytical Support to the Warighter: Campaign Assessments, Operations Analysis, and Data Management, Working Group 4: Current Operations Analysis—Strategic and Operational Level,” brieing, Military Operations Research Society, April 19–22, 2010. As of June 30, 2011: http://www.mors.org/UserFiles/ile/2010%20IW/MORS%20IW%20WG5.pdf Greenwood, T. C., and T. X. Hammes, “War Planning for Wicked Problems,” Armed Forces Journal, December 2009. As of June 30, 2011: http://armedforcesjournal.com/2009/12/4252237/ Gregersen, Hal, and Lee Sailer, “Chaos heory and Its Implications for Social Science Research,” Human Relations, Vol. 46, No. 7, July 1993, pp. 777–802. Gregor, William J., “Military Planning Systems and Stability Operations,” PRISM, Vol. 1, No. 3, 2010, pp. 99–114. As of June 30, 2011: http://www.army.mil/-news/2010/07/22/42647-military-planning-systems-and-stability-operations/ index.html Grier, Cindy, Steve Stephens, Renee Carlucci, Stuart Starr, Cy Staniec, LTC(P) Clark Heidelbaugh, Tim Hope, Don Brock, Gene Visco, and Jim Stevens, “Improving Analytical Support to the Warighter: Campaign Assessments, Operational Analysis, and Data Management: Synthesis Group,” brieing, Military Operations Research Society, April 19–22, 2010. As of June 30, 2011: http://www.mors.org/UserFiles/ile/2010%20IW/Synthesis%20Group%20Out%20Brief% 20Final.pdf Haken, Nate, Joelle Burbank, and Pauline H. Baker, “Casting Globally: Using Content Analysis for Conlict Assessment and Forecasting,” Military Operations Research, Vol. 15, No. 2, 2010, pp. 5–19. Halberstam, David, he Best and the Brightest, New York: Random House, 1972. Bibliography 295 Hall, CSM Michael T., and GEN Stanley A. McChrystal, commander, U.S. Forces–Afghanistan/ International Security Assistance Force, Afghanistan, “ISAF Commander’s Counterinsurgency Guidance,” Kabul, Afghanistan: Headquarters, International Security Assistance Force, 2009. As of June 30, 2011: http://www.nato.int/isaf/docu/oicial_texts/counterinsurgency_guidance.pdf Halmos, Paul R., Measure heory, New York: Van Nostrand, 1950. Hamilton, James D., Time Series Analysis, Princeton, N.J.: Princeton University Press, 1994. Hammond, William M., he United States Army in Vietnam: Public Afairs he Military and the Media—1962–1968, Washington, D.C.: U.S. Government Printing Oice, 1988. Hannan, Michael J., “Operational Net Assessment: A Framework for Social Network Analysis,” IOsphere, Fall 2005, pp. 27–32. Harvey, Derek J., director, Afghanistan-Pakistan Center of Excellence, U.S. Central Command, discussion with the author, January 22, 2010. Headquarters, U.S. Department of the Army, Operations Against Guerrilla Forces, Field Manual 31-20, Washington, D.C., 1951. ———, Operations Against Irregular Forces, Field Manual 31-15, Washington, D.C.: Headquarters, Department of the Army, May 1961. ———, Counterguerrilla Operations, Field Manual 31-16, Washington, D.C., February 1963. ———, Civil Afairs Operation, Field Manual, 41-1, Washington, D.C., October, 1969. ———, Intelligence Analysis, Field Manual 34-3, Washington, D.C., March 1990. ———, Mission Command: Command and Control of Army Forces, Field Manual 6-0, Washington, D.C., August 2003. ———, Intelligence, Field Manual 2-0, Washington, D.C., May 2004. ———, he Army, Field Manual 1, Washington, D.C., June 2005. As of September 17, 2010: http://www.army.mil/fm1/ ———, he Operations Process, Field Manual Interim 5-0.1, Washington, D.C., March 2006a. ———, Civil Afairs Operations, Field Manual 3-05.40, Washington, D.C., September 6, 2006b. ———, Counterinsurgency, Field Manual 3-24/Marine Corps Warfare Publication 3-33.5, Washington, D.C., December 2006c. ———, 2007 Army Modernization Plan, Washington, D.C., March 5, 2007. ———, Operations, Field Manual 3-0, Washington, D.C., February 27, 2008. ———, Tactics in Counterinsurgency, Field Manual 3-24.2, Washington, D.C., April 21, 2009. ———, he Operations Process, Field Manual 5-0, Washington, D.C.: March 2010. ———, “Army Doctrinal Term Changes,” unoicial guidelines for Army training developers, spreadsheet, April 29, 2011. Headquarters, U.S. Marine Corps, Small Wars Manual, Washington, D.C., 1940. ———, Intelligence, Marine Corps Doctrine Publication 2, Washington, D.C., June 7, 1997a. ———, Warighting, Marine Corps Doctrine Publication 1, Washington, D.C., June 20, 1997b. 296 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency ———, “A Concept for Distributed Operations,” white paper, Washington, D.C., April 25, 2005. As of July 21, 2011: http://www.marines.mil/unit/tecom/mcu/grc/library/Documents/A%20Concept%20for%20 Distributed%20Operations.pdf Hitch, Charles J., Decision-Making for Defense, Berkeley, Calif.: University of California Press, 1965. Hodermarsky, George T., and Brian Kalamaja, Science Applications International Corporation, “A Systems Approach to Assessments: Dealing with Complexity,” brieing, 2008. As of June 30, 2011: http://www.saic.com/sosa/downloads/Systems_Approach.pdf Howard, Trevor, “Operational Analysis Support to OP HERRICK,” brieing, Defence Science and Technology Laboratory, UK Ministry of Defence, February 13, 2007. HQDA—see Headquarters, U.S. Department of the Army. Hume, Robert, director, International Security Assistance Force Afghan Assessment Group, interview with the author, May 7, 2010. Hunerwadel, J. P., “he Efects-Based Approach to Operations: Questions and Answers,” Air and Space Power Journal, Spring 2006. As of June 30, 2011: http://www.airpower.au.af.mil/airchronicles/apj/apj06/spr06/hunerwadel.html Hylton, David, “Special Inspector General Report on Afghanistan,” NATO Training MissionAfghanistan blog, June 29, 2010. As of June 30, 2011: http://ntm-a.com/wordpress2/?p=1732 Iklé, Fred Charles, Every War Must End, rev. ed., New York: Columbia University Press, 2005. Intelligence, Surveillance, and Reconnaissance Task Force, “Open Sharing Environment: Unity Net,” brieing, July 2010a. ———, “ISR TF Data Sharing Lessons,” brieing, presented at Allied Information Sharing Strategy Support to International Security Assistance Force Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010b. Interagency Language Roundtable, homepage, undated. As of June 30, 2011: http://www.govtilr.org/ International Security Assistance Force, “ISAF Commander Tours U.S. Detention Facility in Parwan,” press release, undated. As of June 30, 2011: http://www.isaf.nato.int/article/isaf-releases/isaf-commander-tours-u.s.-detention-facility-in-parwan. html International Security Assistance Force Afghan Assessment Group, “Transfer of Lead Security Responsibility Efect Scoring Model,” brieing, undated. International Security Assistance Force Headquarters, COMISAF’S Initial Assessment (Unclassiied), August 30, 2009. ———, “Knowledge Management,” brieing, August 2010. International Security Assistance Force Headquarters Strategic Advisory Group, “Unclassiied Metrics,” Kabul, Afghanistan, April 2009. International Security Assistance Force Joint Command, “District Assessment Framework Tool,” spreadsheet, 2010. International Security Assistance Force Joint Command Chief of Staf, Operational Orders Assessment Steering Group, brieing, November 19, 2009. Bibliography 297 International Security Assistance Force Knowledge Management Oice, “Afghan Mission Network,” brieing, September 1, 2010. Introduction to the Paciication Data Bank, U.S. government pamphlet, November 1969. Available via the Vietnam Center and Archive online database. ISAF—see International Security Assistance Force. ISAF AAG—see International Security Assistance Force Afghan Assessment Group. Jacobson, Alvin L., and N. M. Lalu, “An Empirical and Algebraic Analysis of Alternative Techniques for Measuring Unobserved Variables,” in Hubert M. Blalock, Jr., ed., Measurement in the Social Sciences: heories and Strategies, Chicago, Ill.: Aldine Publishing Company, 1974, pp. 215–242. Jaiswal, N. K., Military Operations Research: Quantitative Decision Making, Boston, Mass.: Kluwer Academic Publishers, 1997. Johnson, Neil F., Simple Complexity: A Clear Guide to Complexity heory, Oxford, UK: OneWorld, 2009. Joint U.S. Public Afairs Oice, Long An Province Survey, 1966, Saigon, South Vietnam, 1967. Joint Warighting Center, U.S. Joint Forces Command, J9, Standing Joint Force Headquarters, Commander’s Handbook for an Efects-Based Approach to Joint Operations, Sufolk, Va., February 24, 2006. Jones, Douglas D., Understanding Measures of Efectiveness in Counterinsurgency Operations, Fort Leavenworth, Kan.: School of Advanced Military Studies, U.S. Army Command and General Staf College, May 2006. As of June 30, 2011: http://handle.dtic.mil/100.2/ADA450589 Kagan, Frederick, W., Finding the Target: he Transformation of American Military Policy, New York: Encounter Books, 2006. Kalyvas, Stathis N., and Matthew Adam Kocher, “he Dynamics of Violence in Vietnam: An Analysis of the Hamlet Evaluation System (HES),” Journal of Peace Research, Vol. 46, No. 3, May 2009, pp. 335–355. Kane, Tim, “Global U.S. Troop Deployment, 1950–2005,” Heritage Foundation, Center for Data Analysis, May 24, 2006. As of June 30, 2011: http://www.heritage.org/research/reports/2006/05/global-us-troop-deployment-1950-2005 Kaplan, Lawrence S., Ronald D. Landa, and Edward J. Drea, he McNamara Ascendancy: 1961–1965, Washington, D.C.: Historical Oice of the Oice of the Secretary of Defense, 2006. Karnow, Stanley, Vietnam: A History, New York: Penguin Books, 1984. Keaney, homas A., “Surveying Gulf War Airpower,” Joint Force Quarterly, Autumn 1993, pp. 25–36. Kelly, Justin, and David Kilcullen, “Chaos Versus Predictability: A Critique of Efects-Based Operations,” Security Challenges, Vol. 2, No. 1, April 2006, pp. 63–73. As of June 30, 2011: http://www.securitychallenges.org.au/ArticlePages/vol2no1KellyandKilcullen.html Kiesling, Eugenia C., “On War Without the Fog,” Military Review, Vol. 81, No. 5, September–October 2001, pp. 85–87. Kilcullen, David, he Accidental Guerrilla: Fighting Small Wars in the Midst of a Big One, Oxford, UK: Oxford University Press, 2009a. 298 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency ———, “Measuring Progress in Afghanistan,” Kabul, Afghanistan, December 2009b. As of June 30, 2011: http://hts.army.mil/Documents/Measuring%20Progress%20Afghanistan%20(2).pdf ———, Counterinsurgency, New York: Oxford University Press, 2010. Kinnard, Douglas, he War Managers, Hanover, N.H.: University of New Hampshire Press, 1977. Kober, Avi, “he Israeli Defense Forces in the Second Lebanon War: Why the Poor Performance?” Journal of Strategic Studies, Vol. 31, No. 1, February 2008, pp. 3–40. Komer, R. W., news conference, December 1, 1967. Available via the Vietnam Center and Archive online database. ———, “Text of Ambassador Komer’s News Conference,” transcript, January 24, 1968a. Available via the Vietnam Center and Archive online database. ———, “Memorandum for the Record,” oicial memorandum, November 5, 1968b, in BACM Research, Disk 2. ———, Impact of Paciication on Insurgency in South Vietnam, Santa Monica, Calif.: RAND Corporation, P-4443, August 1970. As of June 30, 2011: http://www.rand.org/pubs/papers/P4443.html Kruglanski, Arie W., and Donna M. Webster, “Motivated Closing of the Mind: ‘Seizing’ and ‘Freezing,’” Psychological Review, Vol. 103, No. 2, 1996, pp. 263–283. Kugler, Cornelius W., Operational Assessment in a Counterinsurgency, Newport, R.I.: U.S. Naval War College, May 10, 2006. Langan, Patrick A., and David J. Levine, Recidivism of Prisoners Released in 1994, Washington, D.C.: Bureau of Justice Statistics, June 2, 2002. As of July 29, 2011: http://bjs.ojp.usdoj.gov/index.cfm?ty=pbdetail&iid=1134 Langer, Gary, brieing on ABC News poll of Afghan perceptions, presented at the Center for Strategic and International Studies, December 2009. Lewy, Guenter, America in Vietnam, New York: Oxford University Press, 1978. Liddell Hart, B. H., Strategy, New York: Praeger, 1967. Loerch, Andrew G., and Larry B. Rainey, eds., Methods for Conducting Military Operational Analysis, Alexandria, Va.: Military Operations Research Society, 1998. Luck, Gary, and Mike Findlay, Joint Operations Insights and Best Practices, 2d ed., Joint Warighting Analysis Center, U.S. Joint Forces Command, July 2008. MACV—see U.S. Military Assistance Command, Vietnam. Mahnken, homas G., Technology and the American Way of War Since 1945, New York: Columbia University Press, 2008. Mann, Edward C. III, Gary Endersby, and homas R. Searle, hinking Efects: Efects-Based Methodology for Joint Operations, Maxwell Air Force Base, Ala.: Air University Press, CADRE paper 15, October 2002. Mao Tse-Tung, On Guerrilla Warfare, Samuel B. Griith, trans., Urbana, Ill.: University of Illinois Press, 2000. Marine Corps Order 1200.17, Military Occupational Specialties Marine Corps Manual, May 23, 2008. As of June 30, 2011: http://www.marines.mil/news/publications/Documents/MCO%201200.17.pdf Bibliography 299 Marlowe, Ann, “Defeating IEDs with Data,” blog post, March 11, 2010. As of June 30, 2011: http://www.worldafairsjournal.org/new/blogs/marlowe/Defeating_IEDs_with_Data Marshall, S. L. A., Vietnam Primer: Lessons Learned, Washington, D.C.: Headquarters, U.S. Department of the Army, 1966. As of June 30, 2011: http://www.lzcenter.com/Documents/12891741-Army-Vietnam-Primer-Pamphlet.pdf Marvin, Brett L., working group presentation at the U.S. Naval War College, Providence, R.I., October 19, 2010. Matthews, Lloyd J., and Dale E. Brown, eds., Assessing the Vietnam War: A Collection from the Journal of the U.S. Army War College, McLean, Va.: Pergamon-Brassey’s International Defense Publishers, 1987. Matthews, Matt M., We Were Caught Unprepared: he 2006 Hezbollah-Israeli War, Fort Leavenworth, Kan.: U.S. Army Combined Arms Center, Combat Studies Institute Press, 2008. As of June 30, 2011: http://purl.access.gpo.gov/GPO/LPS104476 Mattis, Gen. James N., U.S. Marine Corps, “Assessment of Efects Based Operations,” memorandum for U.S. Joint Forces Command, Norfolk, Va., August 14, 2008. As of June 30, 2011: http://smallwarsjournal.com/documents/usjfcomebomemo.pdf McChrystal, Stanley A., COMISAF’s Initial Assessment (Unclassiied), August 30, 2009. McCrabb, Maris, “Efects-Based Operations: An Overview,” brieing for the Air Force Research Laboratory, January 10, 2008. McFadden, Willie J. II, and Daniel J. McCarthy, “Policy Analysis,” in Andrew G. Loerch and Larry B. Rainey, eds., Methods for Conducting Military Operational Analysis, Alexandria, Va.: Military Operations Research Society, 1998, pp. 317–344. McGee, Michael Calvin, “he ‘Ideograph’: A Link Between Rhetoric and Ideology,” Quarterly Journal of Speech, Vol. 66, No. 1, February 1980, pp. 1–16. McKnight, Patrick E., Katherine M. McKnight, Souraya Sidani, and Aurelio Jose Figueredo, Missing Data: A Gentle Introduction, New York: Guilford Press, 2007. McLear, Michael, he Ten housand Day War, Vietnam: 1945–1975, New York: St. Martin’s Press, 1981. McMaster, H. R., Dereliction of Duty: Lyndon Johnson, Robert McNamara, the Joint Chiefs of Staf, and the Lies hat Led to Vietnam, New York: HarperCollins, 1997. ———, “Crack in the Foundation: Defense Transformation and the Underlying Assumption of Dominant Knowledge in Future War,” student issue paper, U.S. Army War College, November 2003. As of June 30, 2011: http://www.au.af.mil/au/awc/awcgate/army-usawc/mcmaster_foundation.pdf McNamara, Robert S., transcript of a news conference, July 11, 1967. Available via the Vietnam Center and Archive online database. ———, “Southeast Asia Operations: Statement by Secretary of Defense McNamara Before Senate Armed Services Committee (excerpts),” testimony on the iscal year 1969–1973 defense program and the 1969 defense budget, February 1, 1968, pp. 263–289. Available via the Vietnam Center and Archive online database. ———, “Aidavit of Robert S. McNamara,” General William C. Westmoreland v. CBS Inc., et al., December 1, 1983. Available via the Vietnam Center and Archive online database. 300 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency McNamara, Robert S., with Brian VanDeMark, In Retrospect: he Tragedy and Lessons of Vietnam, New York: Times Books, 1995. McNamara, Robert S., James G. Blight, and Robert K. Brigham, with homas J. Biersteker and Herbert Y. Schandler, Argument Without End: In Search of Answers to the Vietnam Tragedy, New York: Public Afairs, 1999. Meharg, Sarah Jane, Measuring Efectiveness in Complex Operations: What Is Good Enough, Calgary, Alberta: Canadian Defence and Foreign Afairs Institute, October 2009. As of June 30, 2011: http://www.cdfai.org/PDF/Measuring%20Efectiveness%20in%20Complex%20Operations.pdf Menkhaus, Kenneth J., “State Fragility as a Wicked Problem,” Prism, Vol. 1, No. 2, March 2010, pp. 85–100. Miller, Delbert C., and Neil J. Salkind, Handbook of Research Design and Social Measurement, 6th ed., housand Oaks, Calif.: Sage Publications, 2002. Miller, John H., and Scott E. Page, Complex Adaptive Systems: An Introduction to Computational Models of Social Life, Princeton, N.J.: Princeton University Press, 2007. Ministry of Supply and War Oice: Military Operational Research Unit, successors and related bodies, WO 291, undated. Mission Coordinator, U.S. Embassy, Saigon, South Vietnam, MACCORDS Field Reporting System, July 1, 1969. Available via the Vietnam Center and Archive online database. Mitchell, Robbyn, “Mattis Takes Over as CENTCOM Chief,” St. Petersburg Times, August 12, 2010, p. 1. As of June 30, 2011: http://www.tampabay.com/news/mattis-takes-over-as-centcom-chief/1114800 Mowery, Samuel, Warighting Analysis Division, J8, U.S. Department of Defense, “A System Dynamics Model of the FM 3-24 COIN Manual,” brieing, 2009. Slide shown in Figure 3.1, Chapter hree, as of June 30, 2011: http://www.guardian.co.uk/news/datablog/2010/apr/29/mcchrystal-afghanistan-powerpoint-slide Moyar, Mark, A Question of Command: Counterinsurgency from the Civil War to Iraq, New Haven, Conn.: Yale University Press, 2009. Multi-National Force–Iraq, “Charts to Accompany the Testimony of General David H. Petraeus,” brieing slides, September 10–11, 2007. As of April 1, 2011: http://smallwarsjournal.com/documents/petraeusslides.pdf ———, “Charts to Accompany the Testimony of General David H. Petraeus,” brieing slides, April 8–9, 2008. As of April 1, 2011: http://www.defense.gov/pdf/Testimony_Handout_Packet.pdf Murray, William S., “A Will to Measure,” Parameters, Autumn 2001, pp. 134–147. As of June 30, 2011: http://www.carlisle.army.mil/USAWC/Parameters/Articles/01autumn/Murray.htm Nakashima, Ellen, and Craig Whitlock, “With Air Force’s Gorgon Drone ‘We Can See Everything,’” Washington Post, January 2, 2011. As of January 15, 2011: http://www.washingtonpost.com/wp-dyn/content/article/2011/01/01/AR2011010102690.html National Commission for the Protection of Human Subjects of Biomedical and Behavioral Research, he Belmont Report: Ethical Principles and Guidelines for the Protection of Human Subjects Research, Washington, D.C.: U.S. Government Printing Oice, 1978. As of March 10, 2011: http://www.videocast.nih.gov/pdf/ohrp_belmont_report.pdf Bibliography 301 National Intelligence Council, Estimative Products on Vietnam: 1948–1975, Pittsburgh, Pa.: U.S. Government Printing Oice, April 2005. Nitschke, Stephen G., Vietnam: A Complex Adaptive Perspective, U.S. Marine Corps Command and Staf College, thesis, 1997. Oberdorfer, Don, Tet! he Turning Point in the Vietnam War, Baltimore, Md.: Johns Hopkins University Press, 2001. Oice of the Director of National Intelligence, Analytic Standards, Intelligence Community Directive No. 203, June 21, 2007. As of June 30, 2011: http://www.dni.gov/electronic_reading_room/ICD_203.pdf Oice of the Secretary of Defense, “Metrics Conference Outbrief to IJC/IDC,” brieing, March 18, 2010. Oice of the Special Inspector General for Afghanistan Reconstruction, Actions Needed to Improve the Reliability of Afghan Security Force Assessments, Arlington, Va., June 29, 2010. As of June 30, 2011: http://www.washingtonpost.com/wp-srv/hp/ssi/wpc/sigar.pdf?sid=ST2010062805531 Olson, James S., and Randy Roberts, Where the Domino Fell: America and Vietnam, 1945–1990, New York: St. Martin’s Press, 1991. Owens, Bill, with Ed Oley, Lifting the Fog of War, Baltimore, Md.: Johns Hopkins University Press, 2000. Palmer, Bruce, Jr., he 25-Year War: America’s Military Role in Vietnam, Lexington, Ky.: University Press of Kentucky, 1984. Paul, Christopher, Colin Clarke, and Beth Grill, Victory Has a housand Fathers: Sources of Success in Counterinsurgency, Santa Monica, Calif.: RAND Corporation, MG-964-OSD, 2010. As of June 30, 2011: http://www.rand.org/pubs/monographs/MG964.html Paul, Christopher, William M. Mason, Daniel McCafrey, and Sarah A. Fox, “A Cautionary Case Study of Approaches to the Treatment of Missing Data,” Statistical Methods and Applications, Vol. 17, No. 3, July 2008, pp. 351–372. Perry, Walter L., “Linking Systems Performance and Operational Efectiveness,” in Andrew G. Loerch and Larry B. Rainey, eds., Methods for Conducting Military Operational Analysis, Alexandria, Va.: Military Operations Research Society, 2007. ———, email exchange with the author, February 3, 2011. Perry, Walter L., Robert W. Button, Jerome Bracken, homas Sullivan, and Jonathan Mitchell, Measures of Efectiveness for the Information-Age Navy: he Efects of Network-Centric Operations on Combat Outcomes, Santa Monica, Calif.: RAND Corporation, MR-1449-NAVY, 2002. As of June 30, 2011: http://www.rand.org/pubs/monograph_reports/MR1449.html Perry, Walter L., and John Gordon IV, Analytic Support to Intelligence in Counterinsurgencies, Santa Monica, Calif.: RAND Corporation, MG-682-OSD, 2008. As of June 30, 2011: http://www.rand.org/pubs/monographs/MG682.html Phillips, Rufus, Why Vietnam Matters: An Eyewitness Account of Lessons Not Learned, Annapolis, Md.: Naval Institute Press, 2008. Pool, Ithiel de Sola, Gordon Fowler, Peter McGrath, and Richard Peterson, Hamlet Evaluation System Study, Cambridge, Mass.: Simulmatics Corporation, May 1, 1968. As of June 30, 2011: http://handle.dtic.mil/100.2/AD839821 302 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Prince, W. B., and J. H. Adkins, Analysis of Vietnamization: Hamlet Evaluation System Revisions, Ann Arbor, Mich.: Bendix Aerospace Systems Division, February 1973. As of June 30, 2011: http://www.dtic.mil/cgi-bin/GetTRDoc?AD=AD908686&Location=U2&doc=GetTRDoc.pdf Public Law 110-28, U.S. Troop Readiness, Veterans’ Care, Katrina Recovery, and Iraq Accountability Appropriations Act, May 25, 2007. Quade, E. S., ed., Analysis for Military Decisions, Santa Monica, Calif.: RAND Corporation, R-387-PR, 1964. As of June 30, 2011: http://www.rand.org/pubs/reports/R0387.html Race, Jefrey, War Comes to Long An: Revolutionary Conlict in a Vietnam Province, Berkeley and Los Angeles, Calif.: University of California Press, 1972. Ramjeet, T. J., Operational Analysis in Support of HQ RC(S), Kandahar, Afghanistan, September 2007 to January 2008, in Cornwallis XIII, Farnborough, UK: Defence Science and Technology Laboratory, 2008. Reconstruction and Stabilization Policy Coordinating Committee, Interagency Conlict Assessment Framework, Washington, D.C., 2009. As of July 29, 2011: http://www.state.gov/documents/organization/161781.pdf Record, Jefrey, Hollow Victory: A Contrary View of the Gulf War, New York: Brassey’s, Inc., 1993. Rehm, Allan, ed., Analyzing Guerrilla Warfare, conference proceedings, September 24, 1985. Available via the Vietnam Center and Archive online database. Rennie, Ruth, Sudhindra Sharma, and Pawan Kumar Sen, Afghanistan in 2009: A Survey of the Afghan People, Kabul, Afghanistan: Asia Foundation, December 2009. As of June 30, 2011: http://www.asiafoundation.org/resources/pdfs/Afghanistanin2009.pdf “Rescinding ASCOPEs for PMESII-PT at the Tactical Level; Possibly Good in heory, but What About in Application?” U.S. Army Combined Arms Center Blog, April 20, 2009. No longer available. Ricchiardi, Sherry, “Whatever Happened to Iraq?” American Journalism Review, June–July 2008. As of November 9, 2010: http://www.ajr.org/article.asp?id=4515 Rittel, Horst W. J., and Melvin M. Webber, “Dilemmas in a General heory of Planning,” Policy Sciences, Vol. 4, 1973, pp. 155–169. Rogers, Simon, “he McChrystal Afghanistan PowerPoint Slide: Can You Do Any Better?” Guardian Data Blog, April 29, 2010. As of June 30, 2011: http://www.guardian.co.uk/news/datablog/2010/apr/29/mcchrystal-afghanistan-powerpoint-slide Rosenau, William, and Austin Long, he Phoenix Program and Contemporary Counterinsurgency, Santa Monica, Calif.: RAND Corporation, OP-258-OSD, 2009. As of June 30, 2011: http://www.rand.org/pubs/occasional_papers/OP258.html Roush, Maurice D., “he Hamlet Evaluation System,” Military Review, September 1969. Roy-Bhattacharya, Joydeep, he Storyteller of Marrakesh, New York: W. W. Norton and Company, 2011. Rubin, Alissa J., “U.S. Report on Afghan War Finds Few Gains in 6 Months,” New York Times, April 29, 2010a. As of June 30, 2011: http://www.nytimes.com/2010/04/30/world/asia/30afghan.html ———, “Petraeus Says Taliban Have Reached Out to Karzai,” New York Times, September 27, 2010b. As of June 30, 2011: http://www.nytimes.com/2010/09/28/world/asia/28afghan.html Bibliography 303 Ruby, Tomislav Z., “Efects-Based Operations: More Important han Ever,” Parameters, Autumn 2008, pp. 26–35. Rumsfeld, Donald H., interview with Steve Inskeep, Morning Edition, National Public Radio, March 29, 2005. As of June 30, 2011: http://www.defense.gov/transcripts/transcript.aspx?transcriptid=2551 Schafer, Marvin B., On Rescoring the Hamlet Evaluation System, Santa Monica, Calif.: RAND Corporation, 1968, not available to the general public. Schrecker, Mark, “U.S. Strategy in Afghanistan: Flawed Assumptions Will Lead to Ultimate Failure,” Joint Force Quarterly, No. 59, 4th Quarter, 2010, pp. 75–82. Schroden, Jonathan J., “Measures for Security in a Counterinsurgency,” Journal of Strategic Studies, Vol. 32, No. 5, October 2009, pp. 715–744. ———, email exchange with the author, February 1, 2010a. ———, email exchange with the author, September 29, 2010b. ———, brieing at the U.S. Naval War College, Newport, R.I., October 19, 2010c. ———, email exchange with the author, October 27, 2010d. Shadid, Anthony, “In Iraq, Western Clocks, but Middle Eastern Time,” New York Times, August 14, 2010. As of June 30, 2011: http://www.nytimes.com/2010/08/15/weekinreview/15shadid.html Sharp, U. S. G., and William C. Westmoreland, Report on the War in Vietnam as of 30 June 1968, Washington, D.C.: U.S. Government Printing Oice, 1968. Sheehan, Neil, A Bright Shining Lie: John Paul Vann and America in Vietnam, New York: Vintage Books, 1988. Shimko, Keith L., he Iraq Wars and America’s Military Revolution, New York: Cambridge University Press, 2010. Shrader, Charles R., History of Operations Research in the United States Army, Vol. 1: 1942–62, Washington, D.C.: Oice of the Deputy Under Secretary of the Army for Operations Research, U.S. Army, Center for Military History publication 70-102-1, August 11, 2006. As of June 30, 2011: http://purl.access.gpo.gov/GPO/FDLP521 Smith, Douglas S., commander, 2nd Battalion, 47th Infantry, 9th Infantry Division, interview with Robert L. Keeley, commander, 19th Military History Detachment, Bien Phuoc, Republic of Vietnam, July 1, 1969. As of June 30, 2011: http://www.history.army.mil/documents/vietnam/vnit/vnit457f.htm Smith, Edward Allen, Efects-Based Operations: Applying Network Centric Warfare to Peace, Crisis, and War, Washington, D.C.: U.S. Department of Defense Command and Control Research Program, November 2002. Soeters, Joseph, “Measuring the Immeasurable? he Efects-Based Approach in Comprehensive Peace Operations,” draft revision of a paper presented at the tenth European Research Group on Military and Society conference, Stockholm, Sweden, June 23, 2009. Sok, Sang Am, Center for Army Analysis, “Assessment Doctrine,” brieing presented at the Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. Sorley, Lewis, ed., he Abrams Tapes: 1968–1972, Lubbock, Tex.: Texas Tech University Press, 2004. 304 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency State of Washington, “Recidivism of Adult Felons 2004,” web page, December 2005. As of May 10, 2011: http://www.sgc.wa.gov/PUBS/Recidivism/Adult_Recidivism_Cy04.pdf Stolzenberg, Ross M., ed., Sociological Methodology, Vol. 26, 2006. Sullivan, homas J., Iraq Reconstruction Management Oice, U.S. Embassy Baghdad, “Long Term Plan for IRMO Metrics,” memorandum, August 2004. Sweetland, Anders, Item Analysis of the HES (Hamlet Evaluation System), Santa Monica, Calif.: RAND Corporation, D-17634-ARPA/AGILE, 1968. As of June 30, 2011: http://www.rand.org/pubs/documents/D17634.html Sykes, Charles S., Interim Report of Operations, First Cavalry Division: July 1965 to December 1966, Albuquerque, N.M.: First Cavalry Division Association, undated. Available via the Vietnam Center and Archive online database. Terrif, Terry, Frans Osinga, and heo Farrell, A Transformation Gap? American Innovations and European Military Change, Stanford, Calif.: Stanford University Press, 2010. hayer, homas C., ed., A Systems Analysis View of the Vietnam War, 1965–1972, Vol. 1: he Situation in Southeast Asia, Washington, D.C.: Oice of the Assistant Secretary of Defense for Systems Analysis, Southeast Asia Intelligence Division, 1975a. ———, A Systems Analysis View of the Vietnam War, 1965–1972, Vol. 9: Population Security, Washington, D.C.: Oice of the Assistant Secretary of Defense for Systems Analysis, Southeast Asia Intelligence Division, 1975b. ———, A Systems Analysis View of the Vietnam War, 1965–1972, Vol. 10: Paciication and Civil Afairs, Washington, D.C.: Oice of the Assistant Secretary of Defense for Systems Analysis, Southeast Asia Intelligence Division, 1975c. ———, War Without Fronts: he American Experience in Vietnam, Boulder, Colo.: Westview Press, 1985. ho, Tran Dinh, Paciication, Indochina Monographs, Washington, D.C.: U.S. Army Center of Military History, 1980. As of June 30, 2011: http://cgsc.cdmhost.com/cdm/fullbrowser/collection/p4013coll11/id/1409/rv/singleitem/rec/2 hompson, Robert, No Exit from Vietnam, New York: David McKay Company, Inc., 1970. Toevank, Freek-Jan, “Methods for Collecting and Sharing Data,” plenary presentation, Allied Information Sharing Strategy Support to ISAF Population Metrics and Data Conference, Brunssum, Netherlands, September 1, 2010. Tufte, Edward R., he Visual Display of Quantitative Information, 2nd ed., Cheshire, Conn.: Graphics Press, 2006. Tunney, John V., Measuring Hamlet Security in Vietnam: Report of a Special Study Mission, Washington, D.C.: U.S. Government Printing Oice, December 1968. Ulrich, Mark, U.S. Army and Marine Corps Joint Counterinsurgency Center, “Center of Inluence Analysis: Linking heory to Application,” brieing presented at the National Defense University, 2009. U.S. Agency for International Development, Measuring Fragility: Indicators and Methods for Rating State Performance, Washington, D.C., June 2005. As of June 30, 2011: http://pdf.usaid.gov/pdf_docs/PNADD462.pdf U.S. Army, “FM 5-0 Overview,” Combined Arms Directorate, brieing, April 1, 2010. Bibliography 305 U.S. Army Combined Arms Center, “Understanding the Operational Environment in COIN,” brieing, February 13, 2009. U.S. Army Training and Doctrine Command, Commander’s Appreciation and Campaign Design, version 1.0, pamphlet 525-5-500, January 28, 2008. As of June 29, 2011: http://www.tradoc.army.mil/tpubs/pams/p525-5-500.pdf U.S. Bureau of Labor Statistics, “How the Government Measures Unemployment,” web page, last updated October 16, 2009. As of June 30, 2011: http://www.bls.gov/cps/cps_htgm.htm U.S. Central Command, “CJOA—Afghanistan TF 236 Overview,” unpublished brieing, September 1, 2010. U.S. Central Intelligence Agency, Analysis of the Vietnamese Communists’ Strengths, Capabilities, and Will to Persist in heir Present Strategy in Vietnam, Annex VI: he Morale of the Communist Forces, August 26, 1966. As of March 15, 2011: http://www.vietnam.ttu.edu/star/images/041/04114192001e.pdf ———, he Communist’s Ability to Recoup heir Tet Military Losses, memorandum, March 1, 1968. As of March 15, 2011: http://www.vietnam.ttu.edu/star/images/041/0410586008.pdf ———, A Compendium of Analytic Tradecraft Notes, Vol. 1, Washington, D.C., February 1997. As of June 30, 2011: http://www.au.af.mil/au/awc/awcgate/cia/tradecraft_notes/contents.htm U.S. Defense Logistics Agency, A Study of Strategic Lessons Learned in Vietnam, Vol. VI: Conduct of the War, Book 2: Functional Analysis, May 2, 1980. As of June 30, 2011: http://handle.dtic.mil/100.2/ADA096430 U.S. Department of Defense, Measuring Security and Stability in Iraq, Washington, D.C., November 2006. As of January 13, 2011: http://www.defense.gov/home/features/iraq_reports/index.html ———, Measuring Security and Stability in Iraq, Washington, D.C., December 2007. As of June 30, 2011: http://www.defense.gov/home/features/iraq_reports/index.html ———, Report on Progress Toward Security and Stability in Afghanistan: Report to Congress in Accordance with the 2008 National Defense Authorization Act (Section 1230, Public Law 110-181), Washington, D.C., semiannual report 2, January 2009a. ———, Measuring Stability and Security in Iraq, Washington, D.C., September 2009b. As of June 30, 2011: http://www.defense.gov/home/features/iraq_reports ———, Report on Progress Toward Security and Stability in Afghanistan: Report to Congress in Accordance with the 2008 National Defense Authorization Act (Section 1230, Public Law 110-181), Washington, D.C., April 2010a. ———, Report on Progress Toward Security and Stability in Afghanistan: Report to Congress in Accordance with the 2008 National Defense Authorization Act (Section 1230, Public Law 110-181), Washington, D.C., November 2010b. ———, Measuring Stability and Security in Iraq, Washington, D.C., June 2010c. As of June 30, 2011: http://www.defense.gov/home/features/iraq_reports 306 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency ———, “DoD News Brieing with Vice Adm. Harward from Afghanistan,” November 30, 2010d. As of May 10, 2011: http://www.globalsecurity.org/military/library/news/2010/11/mil-101130-dod02.htm U.S. Department of Defense Oice of Force Transformation, Military Transformation: A Strategic Approach, Washington, D.C., 2003. ———, he Implementation of Network-Centric Warfare, Washington, D.C., January 5, 2005. As of September 16, 2010: http://purl.access.gpo.gov/GPO/LPS57633 U.S. Department of Defense Oice of the Inspector General’s Inspection and Evaluation Directorate, “Combined Forces Command—Afghanistan Management Decision Model,” Arlington, Va., 2005. U.S. Government, A Tradecraft Primer: Structured Analytic Techniques for Improving Intelligence Analysis, March 2009. As of June 30, 2011: https://www.cia.gov/library/publications/publications-rss-updates/tradecraft-primer-may-4-2009. html U.S. Government Interagency Counterinsurgency Initiative, U.S. Government Counterinsurgency Guide, Washington, D.C.: U.S. Department of State, Bureau of Political-Military Afairs, January 2009. As of June 30, 2011: http://www.state.gov/documents/organization/119629.pdf U.S. Joint Chiefs of Staf, Joint Operation Planning, Joint Publication 5-0, Washington, D.C., December 26, 2006. As of June 27, 2011: http://www.dtic.mil/doctrine/new_pubs/jp5_0.pdf ———, Joint Intelligence, Joint Publication 2-0, Washington, D.C., June 22, 2007. As of June 27, 2011: http://www.dtic.mil/doctrine/new_pubs/jp2_0.pdf ———, Joint Doctrine and Education Division Staf, “Efects-Based hinking in Joint Doctrine,” Joint Force Quarterly, No. 53, 2nd Quarter 2009a, p. 60. ———, Joint Intelligence Preparation of the Operational Environment, Joint Publication 2-01.3, Washington, D.C., June 16, 2009b. ———, Counterinsurgency Operations, Joint Publication 3-24, Washington, D.C., October 5, 2009c. As of June 27, 2011: http://www.dtic.mil/doctrine/new_pubs/jp3_24.pdf ———, Joint Operations, Joint Publication 3-0, incorporating change 2, Washington, D.C., March 22, 2010. As of June 27, 2011: http://www.dtic.mil/doctrine/new_pubs/jp3_0.pdf ———, Department of Defense Dictionary of Military and Associated Terms, Joint Publication 1-02, Washington, D.C., November 8, 2010, as amended through May 15, 2011. As of June 27, 2011: http://purl.access.gpo.gov/GPO/LPS14106 U.S. Marine Corps Intelligence Activity, Cultural Generic Information Requirements Handbook (C-GIRH), August 2008. ———, Cultural Intelligence Indicators Guide, October 2009. U.S. Military Assistance Command, Vietnam, CORDS Report—Phong Dinh Province 2/68, province monthly report, February 1968a. Available via the Vietnam Center and Archive online database. ———, MACCORDS-OAD Fact Sheet: RD Cadre Evaluation System, November 15, 1968b. Available via the Vietnam Center and Archive online database. Bibliography 307 ———, Commander’s Summary of the MACV Objectives Plan, 1969a. Available via the Vietnam Center and Archive online database. ———, Directive on MAC-CORDS Field Reporting System, July 1, 1969b. Available via the Vietnam Center and Archive online database. ———, System for Evaluating the Efectiveness of Republic of Vietnam Armed Forces (SEER), November 27, 1969c. Available via the National Archives online catalog. ———, Form Used for Hamlet Evaluation Survey (HES), Report by CORDS Employees, 1969d. Available via the Vietnam Center and Archive online database. ———, Hamlet Evaluation System District Advisors’ Handbook, 1971a. Available via the Vietnam Center and Archive online database. ———, he Hamlet Evaluation System (HES) Reports, oicial memorandum, January 22, 1971b. Available via the Vietnam Center and Archive online database. ———, Hamlet Evaluation System Advisors Handbook, June 1971c. Available via the Vietnam Center and Archive online database. ———, Order of Battle Summary, Vols. 1 and 2, March 1972. Available via the Vietnam Center and Archive online database. U.S. National Institute of Justice, “Measuring Recidivism,” web page, February 20, 2008. As of May 10, 2011: http://www.nij.gov/topics/corrections/recidivism/measuring.htm U.S. Naval Postgraduate School, NPS Academic Catalog, September 10, 2010. As of June 30, 2011: http://www.nps.edu/admissions/Doc/Academic_Catalog_10_SEPT_2010.pdf U.S. Senate Foreign Relations Committee, Vietnam Hearings: Answers to Questions for the Hearing Record on the CORDS Program, transcript, March 1970. Available via the Vietnam Center and Archive online database. USAID—see U.S. Agency for International Development. Vanden Brook, Tom, “IEDs Kill 21,000 Iraqi Civilians 2005–2010,” USA Today, January 11, 2011. As of January 13, 2011: http://www.usatoday.com/news/world/iraq/2011-01-12-1Aied12_ST_N.htm Van Riper, Paul K., “EBO: here Was No Baby in the Bathwater,” Joint Force Quarterly, No. 52, 1st Quarter 2009, pp. 82–85. Vego, Milan N., “Efects-Based Operations: A Critique,” Joint Force Quarterly, No. 41, 2nd Quarter 2006, pp. 51–57. Vincent, Etienne, Philip Eles, and Boris Vasiliev, “Opinion Polling in Support of Counterinsurgency,” in he Cornwallis Group XIV: Analysis of Societal Conlict and CounterInsurgency, Cornwallis Group, 2010, pp. 104–125. As of April 1, 2011: http://www.thecornwallisgroup.org/cornwallis_2009/7-Vincent_etal-CXIV.pdf Wagenhals, Lee W., Alex Levis, and Sajjad Haider, Planning, Execution, and Assessment of EfectsBased Operations (EBO), Rome, N.Y.: Air Force Research Laboratory, AFRL-IF-RS-TR-2006-176, May 2006. As of June 30, 2011: http://handle.dtic.mil/100.2/ADA451493 Walker, David M., DoD Should Provide Congress and the American Public with Monthly Data on Enemy-Initiated Attacks in Iraq in a Timely Manner, Washington, D.C.: U.S. Government Accountability Oice, GAO-07-1048R, September 28, 2007. As of June 30, 2011: http://purl.access.gpo.gov/GPO/LPS87929 308 Embracing the Fog of War: Assessment and Metrics in Counterinsurgency Wass de Czege, Huba, “Systemic Operational Design: Learning and Adapting in Complex Missions,” Military Review, January–February 2009, pp. 2–12. As of June 30, 2011: http://usacac.army.mil/CAC2/MilitaryReview/Archives/English/MilitaryReview_20090228_ art004.pdf Westmoreland v. CBS, 752 F.2d 16, 2d Cir., November 2, 1984. Westmoreland, William C., “General Westmoreland’s Activities Report for September,” memorandum for the President of the United States, October 10, 1967a. ———, “Military Brieing by General William Westmoreland, USA, Commander, Military Assistance Command, Vietnam,” November 22, 1967b. Wheeler, Earle G., Chairman, U.S. Joint Chiefs of Staf, , “Khe Sanh,” memorandum to the President of the United States, February 3, 1968. As of June 27, 2011: http://www.vietnam.ttu.edu/star/images/001/0010113001A.pdf White House, Iraq Benchmark Assessment Report, Washington, D.C.: U.S. Government Printing Oice, September 17, 2007. Wilder, Andrew, “Losing Hearts and Minds in Afghanistan,” Viewpoints—Afghanistan, 1979–2009: in the Grip of Conlict, Washington, D.C.: Middle East Institute, December 2, 2009, pp. 143–146. As of June 30, 2011: http://www.mei.edu/Portals/0/Publications/Afghanistan%20VP.pdf Williams-Bridgers, Jacquelyn, Iraq and Afghanistan: Security, Economic, and Governance Challenges to Rebuilding Eforts Should Be Addressed in U.S. Strategies, testimony before the U.S. House of Representatives Committee on Armed Services, March 25, 2009, Washington, D.C.: U.S. Government Accountability Oice, GAO-09-476T, 2009. As of June 30, 2011: http://purl.access.gpo.gov/GPO/LPS113358 Wirtz, James J., he Tet Ofensive: Intelligence Failure in War, Ithaca, N.Y.: Cornell University Press, 1994. Zehna, Peter W., ed., Selected Methods and Models in Military Operations Research, rev. ed., Honolulu, Hawaii: University Press of the Paciic, 2005. Zhai, Qiang, China and the Vietnam Wars, 1950–1975, Chapel Hill, N.C.: University of North Carolina Press, 2000. Zhu, Hongwei, and Richard Y. Yang, “An Information Quality Framework for Veriiable Intelligence Products,” Massachusetts Institute of Technology, undated. As of June 30, 2011: http://mitiq.mit.edu/Documents/Publications/Papers/2007/An%20IQ%20Framework%20for%20 Veriiable%20Intelligence%20Products.pdf Ziliak, Stephen homas, and Dierdre N. McCloskey, he Cult of Statistical Signiicance: How the Standard Error Costs Us Jobs, Justice, and Lives, Ann Arbor, Mich.: University of Michigan Press, 2008.