Change the Incentives

An Information Theory of Victory

 

Maj. Don Gomez, U.S. Army

 

Download the PDF Download the PDF

 
Students move pieces around the board during a war game based on a Pacific conflict while attending Air War College at Maxwell Air Force Base, Alabama, 21 December 2023

Show me the incentives and I’ll show you the outcome.

—Charlie Munger

 

We’re getting our rear end handed to us in the information space.

—Gen. Glen D. VanHerck, U.S. Air Force

 

Within the information community, it is taken as a near-article of faith that measuring effectiveness is what matters in determining whether an information activity is successful. The following discussion challenges that assertion and offers an alternative framework that could fundamentally alter how information activities are planned, executed, and assessed.

Despite increased discourse concerning the role of information and influence in achieving success in modern conflict, there remains a nagging sense among military leaders, policymakers, and the public that the United States is constantly on the back foot in this arena.1 It is routine for political and military senior leaders to claim that the United States is losing the information war against adversaries who are more nimble, shameless, and aggressive.2 How can it be the case, they openly wonder, that the strongest Nation in the world, which is also home to Hollywood and big-brand marketing talent, cannot compete with the information efforts of hypocritical autocratic regimes, low-budget nonstate actors, or lone wolves leveraging artificial intelligence to pump out cheap propaganda?

The answer to this question is both simple and dull. We are using the wrong bureaucratic incentives. First, humans are messy and do not think, feel, or behave in ways that can be neatly categorized. While this seems intuitive, it does not prevent well-meaning planners, practitioners, and theorists from positing that human activity can be optimally stratified, quantified, and measured. Second, information professionals often conceptualize and demonstrate success in ways incongruent with their senior leadership’s goals and desires; they tend to chase a metric that may be unknown to senior leaders. Changing the status quo and achieving success in the modern information environment requires a shift in thinking away from an outcomes-based model rooted in industrial-age management practices and toward an information-age model that recognizes and accepts the subjective messiness of audiences and incentivizes output in conjunction with a theory of victory.3

Informational Heresy

The notion that measuring qualified output might be preferrable to measuring desired outcomes is likely considered heretical among many information professionals.4 Achieving positive measures of effectiveness is routinely accepted as the gold standard in demonstrating success both internally and to key external stakeholders.5 However, examining how assessments intermingle with incentives within large bureaucracies reveals problems that contribute to the legitimate sense among many leaders that the United States is losing in its information efforts vis-à-vis its adversaries.

To demonstrate that a focus on achieving objective effects limits U.S. information efforts, an exploration of critical definitions is required, particularly nebulous and oft-redefined terms like information. I argue that achieving firm definitions is a distraction that stands in the way of effective operations in this realm. Additionally, a critical examination of prevailing management and assessment practices indicates that incentivizing outcomes tends toward dysfunctional incentive structures that often fail to meet stated objectives. Instead, adapting theory of victory concepts to information activities creates a pathway toward an output-based system that aligns with what many businesses and brands have discovered is critical for sustained long-term success and growth.6 These theory of victory concepts were tested in an experimental information war-game exercise in the fall of 2023 in support of building a deeper understanding of “what winning looks like” in the information space.7 Finally, this discussion concludes with important qualifications and potential recommendations for implementation.

Eternal Term Warfare

There is an increased awareness across the joint force of the importance and relevancy of information in campaigning, competition, and conflict.8 At the same time, there is corresponding confusion over how information concepts fit into greater national strategy, how they are operationalized, and perhaps most importantly, how operations, activities, and investments can be measured to indicate success.9 Much of the confusion begins with the persistent inability to settle on widely accepted and sufficient definitions.

Despite rigorous intellectual effort, definitions regarding information and information activities remain nebulous and dynamic.10 Once a qualifier is affixed to information, it becomes difficult to understand where one effort begins and another ends. How is information warfare different from psychological warfare? What about influence activities or cognitive warfare?11 These terms are constantly deployed and redefined internally and externally with little thought to what they may be subsuming or excluding. This incessant defining and redefining of adjacent words is referred to some as “term warfare.”12 While defining terms is important and can clarify thought, remaining in a constant state of bureaucratic “term warfare” can inject uncertainty, skepticism, and timidity into planning and operations. Meanwhile, adversaries appear to be less interested in what a particular activity is called and more interested in what it can accomplish.

Flying under radar control

For many years, the term “information operations” was used to describe in a general sense the activities taking place and serving as the term for the coordinating function that encompassed other roles of information activity like military information support operations, electronic warfare, or public affairs.13 In recent years, this changed in Army doctrine to information advantage, which differs from joint doctrine’s emphasis on operations in the information environment.14 Further, Army doctrine describes the information efforts of adversaries as information warfare but does not use the same term to describe its own activities.15 Meanwhile, both the Navy and the Air Force have embraced the term information warfare but to different ends.16 To further muddy the waters, academics who study the same activity and journalists who report on it do not make these same distinctions, referring to related activities variably as information warfare, information operations, psychological warfare, or propaganda.17 While the activities taking place have not changed much in recent years with the important exception of the introduction and proliferation of new communication technologies, the terms used to describe them continually change, increasing confusion among both practitioners in charge of their execution and leaders responsible for providing direction and oversight.18

While some might argue that without firm definitions it can be prohibitively challenging to plan and execute effective operations, accepting the nature of these types of activities as inherently murky and arguably undefinable presents an alternative way forward. Reviewing the history of defining information, we see that shifting definitions is one of the only constants. Accepting a sufficient definition for a given time and context is likely to satisfy the needs of the day, fully knowing that as things change, so too will the definitions. Finally, the divergence of opinion on definitions should be welcomed, as this leads to additional research and thought that can advance the discourse. Considering that strong opinions and disagreements abound, to state categorically that an enduring decision has been made regarding a definition would likely hamper future efforts toward growth and innovation. The subsequent discussion accepts the vague and often changing nature of these terms and argues that the constant fight to redefine them only adds to the confusion about what constitutes success. Although unsatisfactory, planners and practitioners can move forward with an understanding that definitions will likely shift with different audiences and contexts. With that, a deeper exploration of the role that metrics and incentives play in information reveals a much larger problem.

Metrics and Incentives: A Broken Cycle of Good Intentions

Leaders expend precious resources to achieve their objectives. Stakeholders up, down, and adjacent to the chain of command want to know if those resources are employed effectively. Demonstrating success is often the key criterion within bureaucracies to gain continued support in executing a plan or course of action. In rigid bureaucracies, demonstrated success serves as a powerful career incentive and often leads to better evaluations, increased promotion potential, desired assignments, and enhanced professional prestige. This confluence of factors—the need to demonstrate success as good stewards of public resources coupled with the incentive structures of large bureaucracies—is a fundamental contributing factor to the inability to compete effectively with adversarial information efforts. This confluence leads to a broken cycle of good intentions, and at the heart of this cycle is a military culture obsessed with metrics.19

While the introduction and proliferation of performance measurement into military activity was well-intentioned, research reveals a chaotic system that often results in misguided efforts that frequently fail to achieve their objectives. At the onset of the Cold War, the U.S. military began adopting emerging business practices that emphasized hyperefficiency and performance measurement to compete with the centralized planning efforts of the Soviet Union.20 As far back as 1956, there were indications that an overreliance on metrics could lead to “dysfunctional consequences” within a system.21 A simple example of this is known as the “ratchet principle,” where workers who meet a certain quota of productivity are rewarded with an increased quota, often without an increase in the means to accomplish the additional work.22 This can lead to workers deliberately ensuring they never meet the initial quota to avoid the imposition of a higher workload.

A well-known military example of dysfunctional consequences in performance measurement is found during the Vietnam War, where the U.S. military measured success by the number of enemies killed and the tonnage of bombs dropped.23 Once measurements are introduced into a system, incentives tend to realign and reward short-term success over long-term progress, regardless of any safeguards implemented by management. In this case, incentives realigned leading to an increase in the number of enemies killed in action and tonnage of bombs dropped as that was what senior leaders valued as indicative of success. This rubric served as a theory of victory—that more enemy dead and more bombs dropped would shift the dynamics of the war toward a U.S. victory.24 Only in this case, the theory of victory was flawed, resulting in tactical efforts that undermined the war effort.25 Additionally, military leaders under both career and operational pressure to produce results may contort themselves and their data to demonstrate success, whether that success is real, exaggerated, or completely false.26

Left Quote

This rubric served as a theory of victory—that more enemy dead and more bombs dropped would shift the dynamics of the war toward a U.S. victory.

Right Quote

In the realm of information, some might argue that this problem simply requires identifying better metrics.27 That is, achieving certainty that the metric measured is precisely correct and its successful manipulation will deliver the objective desired. While this solution is tempting—especially in an era of machine learning, artificial intelligence, and “big data”—even when conducted flawlessly, processes that deal with the thoughts, feelings, and behaviors of humans are rooted in social science approaches that are limited in what they can definitively prove.28 For example, in attempting to determine how to measure the will to fight, researchers at RAND argued that while literature, doctrine, and some of the most prominent military leaders throughout history have stated that the will to fight is the most important factor in war, it remains nearly impossible to prove or measure.29 However, to satisfy the deeply ingrained military desire for objective metrics, the authors offer an impressive model that includes over twenty factors at the individual and unit levels to generate a potential model with quantifiable metrics. Despite this exhaustive and impressive work, the authors conclude that “we can quantify the will to fight in simulations but we can never accurately quantify the will to fight in the real world.”30

Further research into the effects of metrics and performance management systems on bureaucracies reveals systemic and emergent problems. Chief among these is a performance measurement concept known as Goodhart’s law, which states that “when a measure becomes a target, it ceases to be a good measure.”31 The classic example comes from a phenomenon during colonial British rule of India. British officials placed a bounty on the skins of cobras to curb the growing population. Local hunters quickly realized they could exploit the system by breeding cobras at scale and delivering their skins to receive the bounty rather than hunting them—a much less dangerous endeavor.32 While it may be possible to craft the perfect measure to ensure that only the specific desired behavior is enacted, evidence and history show that, in most cases, the measure becomes the target precisely because of the incentive structure built into the system. Considering the wide array of internal and external factors that comingle and interact with a specified metric, it is difficult to predict how all audiences and practitioners will behave to influence it.

In the case of information, identifying assessment criteria, often in the form of crafting measures of effectiveness, is one of the first steps in planning.33 Before deep thought or analysis begins on target audiences, susceptibility, or dissemination methods, planners are already considering what metrics might be used to determine whether the effort will be considered successful. While this may appear logical and forward-thinking, Goodhart’s law states that once the measure becomes the target, it fails to be a good measure. Additionally, practitioners under immense operational pressure to deliver results are incentivized to ensure that the measure moves in the desired direction and will likely dedicate everything they can to make that happen.34 A savvy planner or practitioner may be tempted to craft an information effort that is more likely to generate tangible, observable results rather than a potentially superior effort that is prohibitively difficult to measure.35

Researchers Leo Blanken and Jason Lepore further explore the problem of incentive structures in military operations.36 They argue that in hierarchical systems, the principals who set objectives and policy (i.e., political and military senior leaders) are often far removed from the agents that carry out the tasks (i.e., military planners and practitioners).37 Further, the principals are often unaware that the agents tend to pursue a metric that was generated as a way to demonstrate success as opposed to achieving the actual goal, which may have a distant relation to the established metric. This distance contributes to the confusion between the principal and the agent. While an information practitioner may be able to demonstrate apparent success through the attainment of positive measures of effectiveness (i.e., the desired change is achieved), achieving those metrics was never the goal of the senior leader in the first place, and thus the lingering sense of “losing” in the information environment. This cognitive disconnect between the principal and the agent occurs precisely because of the imposition of a flawed performance measurement system that does not match the task.38

The confluence of performance measurement and metrics, bureaucratic incentive structures, and the unique subjectivity of influencing thoughts, emotions, and behaviors sits at the heart of the problem of achieving success in information efforts. Effective leaders demand results and want to demonstrate that they are good stewards of public resources while also being successful at achieving the goals of their institutions. Good practitioners are diligent in demonstrating that their efforts are effective, typically through communicating the successful attainment of desired outcomes with measures of effectiveness. Unfortunately, research on the subject that stretches back nearly a century indicates that under the best conditions, metrics intermingling with bureaucracies often leads to dysfunction. Add to this the inherently difficult task of measuring abstract concepts like the emotions or thoughts of a target audience in relation to a specific message or information campaign, and it becomes clear that the current system is unlikely to satisfy the needs of all who are involved. This phenomenon is precisely what leads many to rightly conclude that the United States is losing in the information space. To overcome this, an alternative to classic assessments is needed that provides an overarching concept that demonstrates “what winning looks like” in the information environment.

The Need for an Information Theory of Victory

The joint force has recognized that its adversaries are not confined by neat categorizations between war and peace, and that these adversaries routinely engage in forms of warfare below the threshold of armed conflict.39 The shift toward competition provides a useful framework for conceptualizing and planning operations to compete for advantage in the event of war. Inherent to this shift is the important role that information and influence play in this concept.40 While the doctrine, tactics, techniques, and procedures concerning information remain sound, an overarching concept that attempts to conceptualize “what winning looks like” in the realm of information is lacking.41 Borrowing from theory of victory studies and identifying a potential theory of victory for information is the first step in developing an approach that stands a chance of achieving success.

Theory of victory research is a subset of war studies that attempts to fill the gap between crafting effective strategy and achieving the policy goal desired. It attempts to answer the question of how we get from the attainment of the military objective to the achievement of the policy goal. Two war studies scholars, Bradford A. Lee and J. Boone Bartholomees offer complementary approaches to conceptualizing a theory of victory that can be adopted for information. Following a summary of these approaches, they are applied to information to introduce a new model for conceptualizing victory in information.

Soldiers with Task Force Guardian, 41st Infantry Brigade Combat Team, war-game courses of action prior to a combat operation as part of an exercise during the Joint Readiness Training Center rotation

Lee argues that a theory of victory represents “the assumptions that strategists make about how the execution of the military operations that they are planning will translate into the achievement of the political objectives that they are pursuing.”42 While lengthy, that statement captures precisely what a theory of victory is—an assumption about how friendly activity will shift the dynamics of a given system in such a way that the adversary will “give up, go away, or go down swinging.”43 Additionally, Lee recognizes that it is difficult to measure the effects of these activities, with the exception of “first-order military effects” like destroying equipment. Thus, he argues that assumptions are paramount in any theory of victory. While the word assumption often carries a negative connotation due to its ambiguity and introduction of risk, Lee argues that assumptions are essential when conceptualizing victory against a complex adversarial system.44 Furthermore, assumptions should be based on tangible qualities like expertise, experience, data, cultural acumen, etc.; they are not simply gut feelings or the absence of facts. Finally, Lee offers an important caveat for democracies. To sustain continued support toward achieving victory, relevant stakeholders, from military and political officials to the public, must see incremental dividends over time.45 Examples of incremental dividends include the raid that killed Osama bin Laden as part of the larger Global War on Terrorism, and the early disclosure of Russian deception intentions at the outset of the Russia-Ukraine War in 2022.46 These incremental dividends provided a satisfactory and tangible “win” as part of a much longer and more difficult to measure effort against an adversary.

Bartholomees takes a different approach to theory of victory studies, beginning with the claim that “victory in war is at the most basic level an assessment, not a fact or condition.”47 Importantly, he argues that this assessment is subjective, contextual, and hierarchical, and not objective, absolute, or equal among actors. He stratifies the importance of these assessments in the American context, arguing that it is (1) the American public, (2) military and political elites, (3) American partners and allies, and (4) world opinion that determine whether victory was achieved or not, in that order.48 Additional research outside of war studies identifies the concept of intersubjective belief as relevant, where the beliefs of individuals and groups are formed through interaction with one another, and these intersubjective beliefs wax and wane as new information is revealed or norms change over time.49 Altogether, Bartholomees presents the importance of subjectivity, audience, and context in achieving victory.

The combined theory of victory research of Lee and Bartholomees offers a framework toward a potential theory of victory for information. From Lee’s formula for victory, we can propose that a theory of victory for information consists of “the assumptions made about actions/activities taken to influence dynamics within the information environment to achieve a stated objective.”50 And from Bartholomees, we understand that victory is an intersubjective assessment made by various actors in specific times and contexts. This combination of factors—a concept for a theory of victory and an understanding of how victory might be assessed—opens the door for experimentation.

Testing a Theory of Victory in an Information War Game

To test this, I designed an information war game based on theory of victory research and prevailing information concepts. The war game aimed to replicate the complex dynamics inherent in information activities, the incentive structures of bureaucracies, and the subjective assessments of multiple actors. In the war game, two information professionals from opposing states compete for influence over multiple target audiences. Using cards marked with various information activities along with a corresponding value (1, 2, 3), the players attempt to influence generic target audiences to support their side. While this influence effort takes place, additional players take on the roles of political/military elite, partner and ally nations, and world opinion, each with their own ability to register assessments of the two sides. Over the course of multiple rounds players experience how the other players perceive their actions and can adjust their tactics in attempt to “win”—which in this case means maintaining a positive sentiment among a majority of players while also influencing the various target audiences to their side.

An important aspect of the war game concerns the interactions between players. While the influence professionals have full autonomy to play any cards they wish, the political/military elite are responsible for providing additional cards to them at the end of a round. Thus, the political/military elite players have influence over which cards are possible to play in the first place—a dynamic that replicates authorities and permissions in the real world.51 A failed information effort, for example, might result in negative feedback from tangential players, reducing the appetite among the political/military elite for similar activities in future rounds. These dynamics serve as a simulation for what information professionals face when trying to demonstrate success (e.g., effectively influencing a target audience) while their supervisors claim that in the grand scheme, they are “losing” based on the subjective responses of various actors who are often unseen and unaccounted for by the information professional.

While the scope of the war game was small and the results cannot be generalized outside of the context in which it was played, it provided a useful tool for experimentation. Based on previous research, I tested a hypothesis that a high volume of information activity—increased output—would likely contribute to an increased subjective assessment of “winning” among various actors. This was confirmed in the specific context of the war game.52 However, further experimentation is required to generalize and build upon these results. For example, during war-game sessions, it became clear that participants often made subjective assessments based on criteria not controlled for in the game (e.g., their level of familiarity with information concepts). Future research using similar methods could attempt to control and isolate specific criteria to generate deeper insight. Finally, the game itself proved to be a valuable educational tool for demonstrating how information and assessment work in a safe and replicable environment. At the conclusion of a given game, players can state why they thought one side was winning over the other, and with more data, trends are likely to emerge that could inform the development of new concepts.

Not Everything That Counts Can Be Counted

Research informs us that understanding exactly why humans choose to think, feel, and act in certain ways is likely to remain at best an imprecise discipline. Attempts to quantify attributes that are not readily quantifiable introduces the patina of hard science and unqualified certainty. Large bureaucracies—and especially rigid hierarchical bureaucracies like the military—tend to reward short-term success often demonstrated through data manipulation. Incentive structures coalesce around a strong desire at both the individual and organizational level to prove that a course of action is successful. This incentive structure can lead to timidity among information professionals who may choose to pursue a metric that is easy to influence rather than the activity that might truly be effective but difficult to measure.

With that understanding, there is an opportunity to introduce an alternative approach based on two important claims. First, a recognition that as success is currently demonstrated, career and organizational structures incentivize agent behavior that often fails to deliver the stated objective of senior leaders—the agents tend to chase the metric, not the goal. Second, a recognition that despite heroic and continuous efforts, measuring human thought, emotions, and behavior will continue to be an imprecise science. The focus on delivering objective results can prevent organizations from success as they are expected to prove that a course of action will work or did work in order to continue. This dysfunction can be corrected by adopting a theory of victory for information based on qualified assumptions concerning the effects of informational actions/activities and pairing this with an organizational incentive structure that is concerned with qualified output over tangible outcomes. Stated plainly, if organizations measured output in accordance with a theory of victory, planners and practitioners would be properly incentivized to meet the needs of the current information environment. Assessments would then be focused on ensuring the right activities are conducted in accordance with a theory of victory, instead of whether those activities can be definitively linked to tangible outcomes. Finally, information war games can explore and experiment with various criteria that are most likely to lead toward victory.

There are three important qualifications worth considering. First, for a theory of victory for information to be successful, it must be correct. Returning to the Vietnam War example, a theory of victory predicated on body counts and tonnage of bombs dropped was incorrect. That theory did not account for the totality of the dynamics at play, like the role of the American antiwar movement and domestic politics.53 While a theory of victory can be wrong, without one, planners and practitioners are left to do the next best thing in perpetuity. Second, it is possible that as technology and data collection improve, the ability to accurately measure the effectiveness of specific information efforts will also improve. The rapid advances and use cases in artificial intelligence, for example, have generated intense discussion on how new technologies might be used in future information efforts.54 However, the importance of context—especially cultural context—is often absent from this conversation, as new technologies impress stakeholders looking for the next advantage. While technological developments may enhance the ability of data collection and analysis in a data-rich environment, the same may not be the case in a denied area where information activity is taking place.55 Finally, there are cases when achieving strong measures of effectiveness remain the best assessment for determining whether an information effort was successful or not. Instances with a clear behavioral outcome, like surrender or defection, are best accomplished through classic assessment. Adopting theory of victory practices to information does not mean casting aside traditional assessment and the need to determine effectiveness—those assessments are still necessary. However, this provides an alternative framework that offers a way to measure activity that may have an effect absent of clear measures of effectiveness.

The concepts proposed in this research cut against the grain of established practice and are likely to be met with understandable skepticism. While there is a seeming consensus on the importance of information as well as an appetite for increased effectiveness, little is offered that deviates from calls to do more or do better.56 This research offers an alternative. Stakeholders who have oversight on information activity should consider the possibility that the current system as it exists may be flawed and remain open to alternatives. Planners and practitioners in information should consider the concepts described in this research as potential avenues for achieving success. Finally, experimentation—especially in the form of information war games—should be encouraged and incentivized to garner additional insight and criteria toward what winning looks like in the information environment.


Notes

  1. Mark Pomerleau, “Why Is the United States Losing the Information War?,” C4ISRNet, 5 October 2020, https://www.c4isrnet.com/information-warfare/2020/10/05/why-is-the-united-states-losing-the-information-war/.
  2. “Jihad 2.0: Social Media in the Next Evolution of Terrorist Recruitment,” Hearing Before the Senate Committee on Homeland Security and Governmental Affairs, 114 Cong. (7 May 2015), https://www.hsgac.senate.gov/hearings/jihad-20-social-media-in-the-next-evolution-of-terrorist-recruitment/.
  3. This article is based on research conducted at the Naval Postgraduate School as part of a curriculum in the Department of Defense Analysis 2022–2023.
  4. See, for example, Brian Horvath and Jeffrey Sharpe, “PSYOP Needs More Science: The Root Cause of the Branch’s Difficulties with Assessment” (master’s thesis, Naval Postgraduate School, December 2013). Many researchers have studied the problem of assessment in information activity and often come to a similar conclusion: the problem is inadequate assessments that result in poor or nonexistent measures of effectiveness. Within the information community, there is a prevailing narrative that dismisses measures of performance if there is no corresponding measure of effectiveness.
  5. Steven Hendrickson and Riley Post, “A Blue-Collar Approach to Operational Analysis: A Special Operations Case Study,” Joint Force Quarterly 96 (1st Quarter, 2020): 50–57, https://ndupress.ndu.edu/Media/News/News-Article-View/Article/2076070/a-blue-collar-approach-to-operational-analysis-a-special-operations-case-study/. The authors discuss this and offer an alternative.
  6. See Les Binet and Peter Field, The Long and the Short of It: Balancing Short and Long-Term Marketing Strategies (London: Institute of Practitioners in Advertising, 2013). Marketers Binet and Field propose that businesses should commit to a 60:40 split between brand and performance marketing to achieve optimal success. This model emphasizes brand building over performance marketing and roughly correlates to notions of “winning” versus achieving tangible performance objectives like sales.
  7. Alex Deep, “‘What Winning Looks Like:’ Narrative for Integrated Deterrence and Strategic Competition” (Tampa, FL: Joint Special Operations University, 2023).
  8. To Receive Testimony on the Posture of United States Special Operations Command and United States Cyber Command in Review of the Defense Authorization Request for Fiscal Year 2025 and the Future Years Defense Program Before the Committee on Armed Services, 118th Cong. (10 April 2024) (statements of Christopher P. Maier, Assistant Secretary of Defense for Special Operations and Low-Intensity Conflict; and Gen. Bryan P. Fenton, Commander, U.S. Special Operations Command), https://www.armed-services.senate.gov/imo/media/doc/20241.pdf.
  9. Jill Goldenziel, “5 Things to Know About the Pentagon’s Information Strategy,” Forbes (website), 30 November 2023, https://www.forbes.com/sites/jillgoldenziel/2023/11/30/5-things-to-know-about-the-pentagons-information-strategy/.
  10. Sarah P. White, “The Organizational Determinants of Military Doctrine: A History of Army Information Operations,” Texas National Security Review 6, no. 1 (Winter 2022/2023): 51–78, https://tnsr.org/2023/01/the-organizational-determinants-of-military-doctrine-a-history-of-army-information-operations/.
  11. Robin Burda, Cognitive Warfare as Part of Society: Never-Ending Battle for Minds (The Hague: The Hague Centre for Strategic Studies, 6 June 2023), https://hcss.nl/report/cognitive-warfare-as-part-of-society-never-ending-battle-for-minds/.
  12. John Bicknell, host, Cognitive Crucible, podcast, episode 78, “Phoenix Cast Dual Release,” Information Professionals Association, 11 January 2022, https://information-professionals.org/episode/cognitive-crucible-episode-78/.
  13. White, “Organizational Determinants of Military Doctrine.”
  14. Field Manual (FM) 3-13, Information Operations (Washington, DC: U.S. Government Publishing Office [GPO], 2023), https://armypubs.army.mil/epubs/DR_pubs/DR_a/ARN39736-ADP_3-13-000-WEB-1.pdf; Joint Publication (JP) 3-04, Information in Joint Operations (Washington, DC: U.S. GPO, 2022).
  15. FM 3-13, Information Operations, 1-12.
  16. Greg Hadley, “16th Air Force Seeks ‘Unity of Effort’ on Information Warfare,” Air & Space Forces Magazine (website), 13 November 2023, https://www.airandspaceforces.com/information-warfare-16th-air-force-unity-of-effort/; Lauren C. Williams, “The Navy Wants to Make Info-Warfare Training Ubiquitous,” Defense One, 3 April 2024, https://www.defenseone.com/defense-systems/2024/04/navy-wants-make-info-warfare-training-ubiquitous/395434/.
  17. See, for example, the work of the Stanford Internet Observatory: Josh A Goldstein et al., Generative Language Models and Automated Influence Operations: Emerging Threats and Potential Mitigations (Stanford, CA: Stanford Internet Observatory, January 2023), https://cyber.fsi.stanford.edu/io/publication/generative-language-models-and-automated-influence-operations-emerging-threats-and.
  18. Gavin Wilde, “The Problem With Defining ‘Disinformation,’” Carnegie Endowment for International Peace, 10 November 2022, https://carnegieendowment.org/2022/11/10/problem-with-defining-disinformation-pub-88385.
  19. Jody Daniels, “Changing Culture: Moving from Metrics to Readiness” (Fort Liberty, NC: U.S. Army Reserve, August 2022), https://www.usar.army.mil/Portals/98/Documents/CAR/Changing%20Culture%20FINAL%2004052022.pdf.
  20. V. F. Ridgway, “Dysfunctional Consequences of Performance Measurements,” Administrative Science Quarterly 1, no. 2 (1956): 240–47, https://doi.org/10.2307/2390989.
  21. White, “Organizational Determinants of Military Doctrine.”
  22. Ridgway, “Dysfunctional Consequences of Performance Measurements,” 247.
  23. The Fog of War, directed by Errol Morris (New York: Sony Pictures Classics, 2003).
  24. Gregory A. Daddis, Withdrawal: Reassessing America’s Final Years in Vietnam (New York: Oxford University Press, 2017), 7.
  25. Morris, The Fog of War.
  26. Leonard Wong and Stephen J. Gerras, Lying to Ourselves: Dishonesty in the Army Profession (Carlisle, PA: U.S. Army War College Press, 1 February 2015), ix, https://doi.org/10.21236/ADA615274.
  27. Ashley Franz Holzmann and Whitney O’Connell, “Falling Short in Measures of Effectiveness,” Small Wars Journal, 30 August 2016, https://smallwarsjournal.com/jrnl/art/falling-short-in-measures-of-effectiveness.
  28. Stanislav Andreski, Social Sciences as Sorcery (London: Deutsch, 1972), 24.
  29. Ben Connable et al., Will to Fight: Analyzing, Modeling, and Simulating the Will to Fight of Military Units (Santa Monica, CA: RAND Corporation, 2018), https://doi.org/10.7249/RR2341.
  30. Thomas M. Nichols, Winning the World: Lessons for America’s Future from the Cold War, Humanistic Perspectives on International Relations (Westport, CT: Praeger, 2002), 161.
  31. Michael Stumbord et al., “Goodhart’s Law: Recognizing and Mitigating the Manipulation of Measures in Analysis” (Arlington, VA: Center for Naval Analyses, 26 September 2022).
  32. Ibid., 3.
  33. JP 3-04, Information in Joint Operations, VI-4–VI-8.
  34. Wong and Gerras, Lying to Ourselves, 7.
  35. This concept has proven difficult to explain. There are some activities that may be effective but difficult to measure. The common retort is “then how do you know it’s effective,” which is admittedly a difficult question to answer. Yet, that question does not stop politicians, advertising agencies, rebel groups, or online influencers from continuing to push the boundaries in their media campaigns to find success.
  36. Leo J. Blanken and Jason J. Lepore, “Performance Measurement in Military Operations: Information versus Incentives,” Defence and Peace Economics 26, no. 5 (3 September 2015): 516–35, https://doi.org/10.1080/10242694.2014.949548.
  37. Ibid., 4.
  38. Some might argue that the source of the principal-agent problem regarding information is not the lack of an overarching concept like a theory of victory, but it is simply the challenge of communicating complex phenomena to senior leaders in a way that is understood and actionable. While this is certainty important, it does not solve the problem of incentives within bureaucracies.
  39. Andrew Milburn and Shawna Sinnot, hosts, Irregular Warfare Podcast, “Competing for Influence: Operations in the Information Environment,” Modern War Institute, 16 January 2021, https://mwi.usma.edu/competing-for-influence-operations-in-the-information-environment/.
  40. Joint Chiefs of Staff, Joint Concept for Competing (Washington, DC: Joint Chiefs of Staff, 2023 [CAC required]), iii, https://jdeis.js.mil/jdeis/jel/concepts/jcc.pdf.
  41. Deep, “‘What Winning Looks Like.’”
  42. Bradford A. Lee, “Theories of Victory” (presentation, Naval War College, RI, 22 November 2013).
  43. Ibid.
  44. JP 5-0, Joint Planning (Washington, DC: U.S. GPO, 2020), I-6.
  45. Lee, “Theories of Victory.”
  46. Jake Harrington, “Intelligence Disclosures in the Ukraine Crisis and Beyond,” War on the Rocks, 1 March 2022, https://warontherocks.com/2022/03/intelligence-disclosures-in-the-ukraine-crisis-and-beyond/.
  47. J. Boone Bartholomees, “Theory of Victory,” Parameters 38, no. 2 (1 May 2008): 26, https://doi.org/10.55540/0031-1723.2419.
  48. Ibid., 31.
  49. Vivienne Brown, “Intersubjective Belief,” Episteme 16, no. 2 (June 2019): 139–56, https://doi.org/10.1017/epi.2017.29.
  50. Don H. Gomez, “An Information Warfare Theory of Victory” (master’s thesis, Naval Postgraduate School, December 2023), 31.
  51. FM 3-13, Information Operations, 6-7.
  52. Gomez, “An Information Warfare Theory of Victory,” 55.
  53. Daddis, Withdrawal, 13.
  54. Kelley Jhong, “Special Operations Forces Require Greater Proficiency in Artificial Intelligence,” War on the Rocks, 23 February 2023, https://warontherocks.com/2023/02/special-operations-forces-require-greater-proficiency-in-artificial-intelligence/.
  55. Julia M. McClenon, “We’re WEIRD and Our Adversaries Know It: Psychological Biases Leave the United States Vulnerable to Cognitive Domain Operations,” Irregular Warfare Initiative, 7 June 2023, https://irregularwarfare.org/articles/were-weird-and-our-adversaries-know-it-psychological-biases-leave-the-united-states-vulnerable-to-cognitive-domain-operations/.
  56. U.S. Government Accountability Office (GAO), Contested Information Environment: Actions Needed to Strengthen Education and Training for DOD Leaders, GAO-23-105608 (Washington, DC: U.S. GAO, 26 January 2023), https://www.gao.gov/products/gao-23-105608.

 

Maj. Don Gomez, U.S. Army, recently completed an MS in information strategy and political warfare from the Naval Postgraduate School. He also holds a BA in international studies and an MA in Middle Eastern studies. He has operational experience in Iraq and Afghanistan.

 

Legacy Military Review Publication of Enduring Relevance

Book Cover
 

The Counterinsurgency Reader II, published in 2008, was a follow-on edition to a collection of articles published in 2006 that similarly provided a compendium of additional published articles as a resource to help military personnel then engaged in the Global War on Terrorism to develop an understanding of insurgent conflict and irregular warfare.

To read online, visit https://www.armyupress.army.mil/Journals/Military-Review/English-Edition-Archives/MR-Coin-Reader2/.

 

 

Back to Top

November-December 2024