How AI Is Transforming Diplomacy and Conflict Management

20 Feb 2026 13:00h - 14:00h

How AI Is Transforming Diplomacy and Conflict Management

Session at a glanceSummary, keypoints, and speakers overview

Summary

The Belfer Center’s Emerging Tech Program launched the MOVE 37 project to examine how artificial intelligence can be integrated into diplomatic negotiation and policy-making processes [4-5][19-22]. Panelists highlighted that modern negotiations involve dozens of parties, multiple issues, and thousands of documents, creating information overload and time pressure that AI could help manage [41-48][66-72]. Gabriela Ramos illustrated this by describing the UNESCO AI ethics recommendation, which required processing 55,000 public comments and mapping the positions of 193 countries, a task she said would have benefited from more AI support [139-144]. Nandita Balakrishnan noted that AI access varies across academia, government, and industry, and that public-sector analysts still perform labor-intensive manual assessments that could be accelerated with AI tools [170-176][181-189]. Charlie Posniak warned that large language models are opaque, lack verifiable fluency, and cannot alone provide the accountability needed for high-stakes negotiations, emphasizing the need for a broader set of computational methods built over the past 80 years [82-85][87-92]. He identified three technical challenges: representing dynamic, strategic interactions; handling intentional misrepresentation by actors; and defining success criteria for multi-party outcomes [94-100]. To address these, his team proposes a cyclical workflow of research, analysis, strategy formulation, and real-time execution, supported by autonomous research agents, data validation, and live transcription services [102-107][108-115]. All speakers agreed that human authority must remain central, with AI tools kept modular, transparent, and scoped to augment rather than replace negotiators [117-121][284-290]. Robyn Scott presented survey results showing that over 90 % of public servants are optimistic about AI’s potential, yet most pilots lack systematic evaluation and many officials do not understand their own ethical frameworks, underscoring a skills gap [228-236][242-250]. She also cautioned against “sleeping at the wheel”-over-reliance on AI that can lead to false confidence-and advocated keeping users “above the algorithm” to preserve agency [254-259][330-334]. Gabriela stressed that cultural and linguistic diversity must be reflected in training data to avoid bias, citing examples where single-language models mischaracterized negotiators’ perspectives [391-399][423-425]. The panel concluded that while AI can improve data handling, predictive insights, and strategic option generation, its deployment must be carefully governed to maintain accountability, cultural sensitivity, and human judgment [286-290][350-357]. Michael McQuade summarized that MOVE 37 will develop AI-augmented tools, evaluation methodologies, and collaborative networks, positioning the project as an early step toward a new discipline at the intersection of technology and international diplomacy [125-129][204-209][432-438].


Keypoints

Major discussion points


AI as an augment-tool for diplomatic negotiations, not a replacement for humans – the panel repeatedly stressed that negotiations are “fundamentally interpersonal” and that any AI system must keep “human authority … central” and serve as a support rather than a decision-maker [24-30][118-121].


Technical and ethical challenges of deploying AI in diplomacy – concerns were raised about model opacity, accountability, strategic mis-representation, cultural bias, data-poisoning and “sleeping at the wheel” over-reliance [83-86][94-100][322-327][330-334][391-403].


Concrete AI functionalities being explored – the team outlined a task-breakdown (research, analysis, strategizing, execution) and described prototypes such as autonomous research agents, real-time transcription/translation, position-tracking dashboards and predictive geopolitics models [102-108][284-290][286-290].


Capacity-building and institutional adoption gaps – surveys show public-sector optimism about AI but also “pilotitis,” low evaluation rates, limited AI literacy and a large skills gap; the panel highlighted the need for training, clear ethical frameworks, and systematic rollout of pilots [226-242][248-250].


Ensuring cultural and linguistic diversity in AI systems – participants warned that AI must reflect the world’s many languages and cultural perspectives to avoid reinforcing individualistic or biased outcomes; they cited UNESCO’s multilingual work and the Swiss multilingual LLM initiative as examples [391-403][424-425].


Overall purpose / goal


The discussion was convened to launch and shape the MOVE 37 project – an initiative of the Belfer Center’s Emerging Tech Program that aims to design, prototype, and responsibly integrate artificial-intelligence tools into the practice of diplomacy and negotiation. The organizers sought input from scholars, practitioners, and policymakers to map the problem space, identify research and development priorities, and build a collaborative community that will guide the project’s roadmap.


Overall tone and its evolution


Opening segment (0:00-10:00): Formal and upbeat, emphasizing opportunity, collaboration, and the excitement of pioneering a new research frontier [1-7][19-23].


Middle segment (10:00-30:00): Becomes more analytical and cautious, detailing the complexity of negotiations, the technical limits of LLMs, and the ethical risks of opacity, bias, and over-reliance [83-86][94-100][322-327][330-334].


Later segment (30:00-end): Shifts toward pragmatic optimism, focusing on concrete use-cases, capacity-building, and concrete next steps while still acknowledging the need for vigilance and human oversight [226-242][284-290][391-403].


Overall, the tone moves from enthusiastic introduction, through a sober appraisal of challenges, to a constructive, solution-oriented outlook that calls for continued collaboration and responsible development.


Speakers

J. Michael McQuade – Director of the Emerging Tech Program at the Belfer Center, runs the MOVE 37 initiative; expertise in international policy, technology, geopolitics, and AI for diplomacy. [S8]


Charlie Posniak – Full-time fellow and research fellow at the Belfer Center; works on AI-enabled diplomatic negotiation tools and policy guidelines. [S1]


Slavina Ancheva – MPP student and research fellow at the Belfer Center; focuses on framing negotiation complexity and AI augmentation for diplomacy. [S4]


Gabriela Ramos – Former Assistant Director General for Social and Human Sciences at UNESCO; former co-chair of the UN AI Advisory Panel and Spain-India Ambassador for AI; expertise in AI ethics, international negotiations, and UNESCO AI recommendations. [S14]


Nandita Balakrishnan – Director of Intelligence at the Special Competitive Studies Project (SCSP); expertise in intelligence, AI, geopolitics, and public-sector AI adoption.


Robyn Scott – CEO and co-founder of Apolitical; collaborates with Stanford HAI; expertise in government innovation, AI training for public-sector policymakers. [S6]


Audience – Various attendees (e.g., senior advisor Sam Dawes, Indian classical-dance teacher Devika Rao, JPL South Asia staff member Arman); roles not specified.


Additional speakers:


Sam Dawes – Senior Advisor to the Oxford University AI Governance Initiative and Director of Multilateral AI; background in diplomacy (worked for Kofi Annan, UK Foreign Office, Cabinet Office).


Devika Rao – Indian classical dance teacher; involved in cultural-education frameworks linking India and the UK.


Arman – Staff member at JPL South Asia; interested in AI’s impact on balance of power in negotiations.


Full session reportComprehensive analysis and detailed insights

The session opened with J. Michael McQuade, director of the Belfer Centre’s Emerging Tech Programme, which “teaches, trains, and does research on the applications of science and technology for international affairs” and convenes scholars, practitioners and students to explore the intersection of technology, science and geopolitics [1-3]. He announced the launch of the MOVE 37 initiative – a component of the Emerging Tech Programme created “to look at where emerging technologies are creating new policy frontiers… and the implications… for governance, geopolitics, global stability and global conflict” [4-5]. Because artificial intelligence is a “major aspect of our work” in relating technology to modern geopolitical issues, a panel of experts was introduced: Gabriela Ramos (UNESCO), Nandita Balakrishnan (Special Competitive Studies Project), Robyn Scott, CEO and co-founder of Apolitical, and two Belfer researchers, Charlie Posniak and Slavina Ancheva[6-15].


The gathering’s purpose was to mobilise collaborators for a “major new project… looking at the use of artificial intelligence in diplomacy and negotiation” [19-22]. McQuade stressed that the work is not confined to a small Cambridge team but seeks “collaborators, partners, and input from the community… around the world” to shape how AI will be used responsibly in high-stakes diplomatic processes [23-26]. He framed diplomacy as a fundamentally human activity that could be augmented by AI to improve outcomes while preserving human agency [24-30][118-121].


Agenda and framing - Slavina Ancheva set the discussion’s three focal areas: (1) the current complexity of negotiation processes; (2) AI’s potential to alleviate those challenges; and (3) the need to think beyond large-language models (LLMs) toward responsible deployment [40-44]. She asked participants to “close your eyes and imagine” a negotiation with ten agenda items, noting that “it’s not just about those 10 items” because “a lot of other factors… both inside and outside that room” influence the outcome [45-48]. A typical negotiation may involve “seven counterparts from seven political groups, seven different countries… and behind you… 27 other countries you are representing” [48-49], illustrating the multi-layered nature of modern diplomacy. The resulting “information overload” includes “thousands of documents, transcripts, drafts” and is compounded by “finite resources, strategic group-think, and time pressure” [66-72].


Challenges illustrated - Gabriela Ramos described negotiating the UNESCO Recommendation on the Ethics of Artificial Intelligence, a process that involved “193 countries negotiating during COVID” and generated “55 000 comments” [139-144]. She noted that AI could have helped “map the positioning of countries” and provided a “repository of what is the traditional position of certain countries” to streamline briefings and stakeholder outreach [144-145]. Ramos warned that AI tools must avoid “misrepresentation, over-representation of certain cultures, certain languages, assumptions” and that any system should “open a space of human understanding” rather than simply “beat the person in front of me” [354-367]. When asked about cultural inclusivity, she stressed that “culture is expressed by language” and that models need multilingual training to capture philosophies such as Ubuntu, otherwise they risk “maximising individual welfare” at the expense of collective worldviews [391-403][424-425].


Audience questions - Sam Dawes asked (a) how to ensure diverse cultural inputs are embedded in AI models and (b) how to guard against data-poisoning or prompt-injection attacks. Ramos answered that culture is conveyed through language, so training on a wide range of languages and continuously validating source material are essential, and she emphasized the need for continual ground-truth testing to detect poisoning [391-403][424-425][430-432]. Arman raised concerns about the impact on the balance of power when data access is uneven. McQuade responded that the project is examining how AI tools can create competitive leverage but also risk exacerbating asymmetrical information if not widely disseminated [208-212][431-432].


Sectoral perspectives - Nandita Balakrishnan observed that “the public sector has been more in the passenger seat, if not the backseat” while academia and industry enjoy broader toolsets [170-172]. She recounted a mentor pointing out a “ten-year-old data point that completely negates” her analysis, a gap that “AI could have identified and synthesised” [185-189]. From this perspective she argued that AI should be viewed as a “data point” that requires human explanation and accountability, especially in high-stakes policy work [297-304]. Balakrishnan also highlighted AI’s strategic importance for geopolitics, noting that “AI has fundamentally changed the threat landscape” and that “the public sector must leverage these tools… in intelligence, the State Department, commerce, OPM” to stay competitive [194-200]. She cited ongoing projects that use AI to “predict geopolitical events” for both military and diplomatic applications, arguing that demonstrable use-cases are needed to convince policymakers of AI’s value [201-203].


Technical roadmap - Charlie Posniak provided a technical perspective, first dismissing the notion that “you can just ask an LLM” because “their fluency isn’t necessarily verifiable in international and world politics” and the models are “opaque… not always viable” for accountability [82-85]. He reminded the audience of an “80-year-old toolkit” of game theory, decision analysis and machine-learning methods that must be integrated with modern AI rather than replaced by chat-bots [86-92]. Posniak identified three core challenges for diplomatic AI: (1) representing dynamic, strategic interactions that evolve over time [94-98]; (2) handling intentional mis-representation by actors [98-99]; and (3) defining success criteria for multi-party outcomes [100-101]. To address these, his team proposes a cyclical workflow of “research, analysis, strategising, and execution” supported by “autonomous research agents, source validation, real-time transcription and translation services” and modular, transparent tools [102-115][117-121]. He later expanded on concrete functionalities such as “position-tracking dashboards, strategy sandboxes, red-team training and predictive geopolitics models” that can process “vast amounts of unstructured data” [284-290][286-290].


Human-in-the-loop emphasis - During the discussion a brief slip (“woman in the loop”) illustrated the panel’s humor and reinforced the emphasis on keeping humans central to AI-augmented processes [350-352].


Consensus - All panelists agreed that AI must remain an augment-tool rather than a replacement for diplomats. Both Ancheva and Posniak stressed that negotiations are “fundamentally interpersonal” and that AI should “give them the tools to manage these complexities much better” while keeping “human authority… central” [61-62][117-121]. Ramos echoed this, insisting that any AI-driven recommendation must be “questioned” and that “the AI tools… should not be built to beat the person in front of me” [352-367]. Scott reinforced the idea of staying “above the algorithm” – using AI as support rather than surrendering agency – and warned that without careful framing AI could create a zero-sum dynamic that diminishes human agency [254-259].


Capacity gaps - Robyn Scott presented empirical evidence on capacity gaps. A survey of 5 000 public servants showed that “north of 90 % think there is huge possibility in the public sector” yet “only 26 % of them say they understand their own country’s ethical frameworks” and many pilots lack systematic evaluation [226-242][248-250]. She described the phenomenon of “sleeping at the wheel,” where users over-trust AI after it reaches high accuracy, leading to false confidence [330-334]. Her “below-the-algorithm / above-the-algorithm” heuristic urges policymakers to “move people up above the algorithm” to preserve decision-making power [256-259].


Strategic implications - McQuade argued that AI will provide “competitive leverage” and that tools should be “dispersed actively and offensively, not defensively” to reshape power balances in both cooperative and adversarial negotiations [208-212][431-432]. Balakrishnan reinforced this view, noting that AI is now “a foundational way we need to think about geopolitics” and that “you cannot divorce AI… when you’re trying to understand geopolitics and foreign policy” [194-196].


Points of disagreement - Posniak warned that the opacity of LLMs makes them unsuitable for treaty-shaping work because “accountability… is not always viable” [82-85], whereas Scott acknowledged the black-box nature but argued that “the lack of transparency is not insurmountable” and that work-arounds are possible [254-259]. A second tension concerned the level of trust to place in AI outputs: McQuade suggested AI could serve as a trusted augment-tool that aggregates information and offers new levers for negotiators [210-212]; Balakrishnan counter-pointed that AI should remain a “data point” requiring human validation and explanation [297-304]. Finally, Ramos advocated for AI that “opens space for human understanding” rather than being used to “beat” counterparts, while Scott cautioned that without careful framing AI could create a zero-sum dynamic that diminishes human agency [354-367][254-259].


Next steps - MOVE 37 will develop a suite of AI-augmented tools covering the four phases identified by Posniak (research, analysis, strategy, execution) [102-115] and will create “evaluation methodologies” to assess their impact [125-129]. The project will continue “one-on-one interviews” with current and former diplomats to capture “process-level insights” and to inform the design of position-tracking repositories and strategic-option generators [267-278]. It will also pursue multilingual, culturally inclusive datasets, drawing on examples such as the Swiss quasi-governmental LLM trained on over 100 languages for diplomatic use [424-425]. Capacity-building initiatives will be launched to raise AI literacy across intelligence, the State Department and other federal agencies, and pilot programmes will be equipped with systematic evaluation frameworks to close the “pilotitis” gap [242-250][194-200].


Key take-aways


– AI should augment, not replace, human negotiators.


– Transparency, modularity and human-in-the-loop design are non-negotiable.


– Multilingual, culturally representative data are essential to avoid bias.


– Capacity-building and rigorous evaluation are required to move beyond pilots.


– Governance frameworks must balance strategic advantage with ethical safeguards.


Overall, the discussion highlighted strong consensus on AI as an augment-tool, the necessity of transparent, modular, human-centered systems, and the importance of multilingual inclusivity and capacity-building. At the same time, disagreements over model opacity, the appropriate level of trust, and whether AI should be framed as a collaborative aid or a competitive lever indicate that MOVE 37 will need flexible governance structures that balance strategic advantage with ethical safeguards. The panel’s thought-provoking remarks – from Posniak’s challenge to “just ask an LLM” [82-85] to Scott’s “below/above the algorithm” heuristic [256-259] – set the tone for a pragmatic, yet cautious, roadmap toward AI-augmented diplomacy.


Session transcriptComplete transcript of the session
J. Michael McQuade

I’ve been a major figure in international policy for the United States and in education at the Belfer Center, where our objective is to teach, train, and do research on subjects related to the applications of science and technology for international affairs. We have scholars, practitioners, students, all working to address the gaps and the opportunities for technology, science, and geopolitics. I’m very delighted to have everybody here today. The Emerging Tech Program, which I have the honor of running, was launched about a year ago, specifically to look at where emerging technologies are creating new policy frontiers, new opportunities to use technology to engage in policy, and the implications that technologies are creating for governance, geopolitics, global stability, and global conflict.

And the MOVE 37 initiative that we’re here to talk to you about today. is a part of that program. As you can imagine, in a program that’s relating technologies to modern issues around geopolitics, that artificial intelligence is one of the major aspects of our work. We have a terrific panel here today. It’s my pleasure to introduce them by name, and we’ll talk a little bit more about each one in just a moment. The missing chair, which we expect shortly, is Gabriela Ramos, who’s the former Assistant Director General for Social and Human Sciences at UNESCO. Nandita Balakrishnan is the Director of Intelligence at the Special Competitive Studies Project in Washington. And ⁠Robyn Scott is the CEO and co -founder of Apolitical.

And then at the far end, two of my colleagues, researchers on our program at the Belfer Center. Charlie Posniak is a full -time fellow with the program, research fellow with the program. And Slavina Ancheva is a current. student within our program, an MPP student at the Belfer Center. I also want to acknowledge a colleague of ours. who is not here, was not able to get to India in time for the conference, Carme Artigas, who is the former co -chair of the UN AI Advisory Panel, who has been an integral part of starting this work and the ongoing progress we are making. Carme is also the Spain -India Ambassador for AI, High Commissioner in Spain -India for AI, and is here with us in spirit and maybe even on the live stream that we’re doing here today.

So a big shout -out to Karame for all of her help. So why are we here? We are embarking on a major new project, specifically looking at the use of artificial intelligence in diplomacy and negotiation. We are here at a conference about the use of artificial intelligence and the implications of artificial intelligence in so many aspects of society. And our work, is looking at how one will engage non -human intelligences in the process. of diplomacy and negotiation. So we’re here because this is a broad -based project. It is not something that is solely the purview of a small team in Cambridge where we are located, but something that we are looking for collaborators, partners, and input from the community here in the community around the world as we build what will surely be a place where AI is used and surely will be a place where we need to be cognizant and careful about how AI will be used in that exercise.

So the work we do plays on an increasingly bigger role in shaping the relationships between states and within states. We want to have this conversation about the role for AI specifically because of the global nature and the integrated way that AI will play, and more specifically, how will we use artificial intelligence tools to augment humans in what is at its core a fundamentally human process of negotiation. Thank you. diplomacy. Diplomatic negotiations are very high stakes. They are very different, as you will hear from our team, very different than classic one -on -one win -lose negotiations or win -win negotiations. They are very much more complex than that, and that means they require both a unique human touch and a unique application of how artificial intelligence might be used in that process.

But it’s also an area where an enormous amount of potential exists to tackle resource constraints, to find better outcomes, and to use artificial intelligence to enable a more stable and prosperous world, and one for which the follow -up to negotiations can be a subject for the tools and applications of modern technology. How we go about that is crucial, and how we talk about it from the beginning is crucial. There are already a number of tools emerging in this space, but it’s our belief that a more rigorous approach is needed, and you’ll hear some of that in just a moment. So what are we going to do? So what are we going to do? Our team is going to do a brief overview of the way we are thinking about this problem.

And after that, we’re going to have this amazing panel that we will engage in a conversation about views from others who are involved in negotiation and or diplomacy writ large and their views on how technology can be used, be used well or be used not so well, and what the implications of that will be. So with that, let me turn it over to Charlie and Slavina to talk about the project itself. I think Slavina is up first.

Slavina Ancheva

Thank you, Michael. And thank you all for being here this morning. A big welcome. Over the next 10 minutes or so, Charlie and myself would like to present you with a little bit of a framing for the expert discussion that we’ll be getting into right after. We’ll broadly focus on three areas. How do negotiation processes currently look and the complexity that comes with them? The potential for AI to augment many of these challenges and processes. Thinking beyond just LLM. and the need for responsible deployment of these tools. Before we do that, I’d like you to close your eyes and imagine. You walk into a negotiation, you look down at the agenda, and there’s 10 items on it.

But as any good negotiator, you know that it’s not just about those 10 items. It’s a lot of other factors that are happening both inside and outside that room that are affecting how that negotiation process is happening. So for one example, you’re sitting across seven counterparts from seven political groups, seven different countries, and behind you, you have your own team, but also 27 other countries that you’re representing, that you’ve promised a certain outcome or a certain deal. Of course, this is not my story. It’s the story of Carme Artigas that Michael mentioned, who was one of the chief negotiators of the EU AI Act, and later the UN AI Advisory Body and many other negotiations. But it’s not just her story.

It’s the story of many of you. It’s the negotiations that you’ve engaged in at the UN, at COP, bilateral. It’s the negotiations that you’ve engaged in at the UN, interagency negotiations within your own organizations. So you know very well that negotiations are complex and they evolve over time. So what might look like just two states negotiating with each other bilaterally is actually a whole set of issues that are on the table. It could be natural resources, it could be AI, it could be climate, and a whole lot of external and internal stakeholders that are also trying to influence that process. We start to dive into some of this complexity. And more than that, there’s a lot of teams that are sitting behind these principal negotiators, the different departments and agencies that are supporting them with evidence, with documents.

And we’d really like to stress that this is a fundamentally interpersonal process. We’re not looking to replace diplomats or negotiators here, but just to give them the tools to manage these complexities much better. And finally, rarely in this world do we have just two states negotiating nowadays. There’s often a third state at the table. In the case of the EU, maybe 27 member states and, of course, a lot of hundreds of others that could be out there. So with that being said, what are some of the impacts of this complexity? Well, for one, there’s a whole lot of information that needs to be managed. There’s a simple negotiation can generate thousands of documents, transcripts, drafts. On top of that, there’s a certain amount of finite resources that any team has as they grapple with many other challenges throughout the day.

There’s a lot of strategic elements. Sometimes in groups, you might have a group think or herding that leads you in one direction as opposed to exploring your full set of options. And finally, there’s the time pressure. So most negotiations do have some sort of time element and handover element to future teams. So with that being said, how can AI help? And I’d like to turn over to Charlie.

Charlie Posniak

Thanks, Lavinia. So AI systems can now beat some of the best human players at Go, at chess, at video games. At board games, language models, as we’ve heard, have become increasingly competent at delivering a range of sophisticated. legal, academic, technical, software contributions, the pace of change has been staggering. And so what our interdisciplinary team has been looking at is we’re trying to envision a better future for diplomacy where computational methods can transform the practice of diplomatic negotiations and statecraft that Slavina just outlined. So supporting better communications, better resolutions, and processes between states can augment their functions. So we’re trying to chart existing technical tools, develop new ones, and provide a range of policy guidelines to ensure that this happens responsibly, safely, effectively.

So the classic question that we get in response is, why can’t you just ask an LLM? Lots of people are interested in trying to see if language models can simulate diplomacy or if chatbots can guide people through a negotiation all in one step. But ultimately, language models are remarkable, and they used to be carefully scoped for three key reasons. Firstly, their fluency isn’t necessarily verifiable in this international and world politics. Secondly, the opacity where you can’t tell what’s going on inside a model is not always verifiable, is not always viable. Because high -stakes negotiations require… accountability both democratically and internally to understand why recommendations shape treaties in certain ways. And additionally, there’s a toolkit that’s 80 years old here.

We have game theory, decision analysis, machine learning, a great range of theoretical developments that exist precisely to model strategic interactions under uncertainty. And so we see LLMs as playing this role at the heart of a really broad set of learning paradigms that’s tying together. We’re both like supervised and unsupervised, self -supervised learning here. But LLMs provide a really strong way to interact with all of these different learning paradigms and technical architectures that the best advances in AI have been built from. So whether that’s the systems that play chess or Go or board games, these are all pulling together lots of different methods. And if we just rely on chatbots at the heart of things, we miss out on all of the technical developments that the last 80 years have experienced.

But there are three key challenges with trying to expand these techniques into the world of diplomacy and world policy. One is to be able to do it in a way that’s not just a way to do it in a way that’s a way to do it in a way that’s not just a way to do it in a way that’s not just a way to do it in a way that’s not just a way to do it in a way that’s not just a way Firstly, representation. As Sabina was touching on, the game that’s being played here isn’t a board game. These are things that are these interactions are fundamentally changeable over time. The institutions that constrain the actions of states can be made and unmade over the course of a negotiation.

Inference like these are environments where there’s real strategic misrepresentation, where people are lying or deceiving or trying to shape outcomes for their own advantages in ways that the current methods aren’t quite well suited to handle. And finally, there’s as we’re touching on, there’s this sense of specifying success. How can you bring together all of the different counterparts and come up with a relatively set, coherent set of preferences and priorities over the course of a really massive negotiation? So these are three challenges that we’re trying to embark on. And one of the ways that we’re approaching this is by breaking down the the tasks of diplomacy and the tasks of negotiation for AI applications. So just broadly, one of the ways we’ve looked at this, is saying that there are some foundational tasks of research analysis.

analysis, strategizing, and execution that build this evidence base with research, that analysis processes the information that you’ve managed to gather. Strategizing relies on using the analysis and the research to come up with a map from your preferences to your outcomes. And then finally, in the room, executing a negotiation, you’ve got to be able to dynamically adapt and adjust over time. And this isn’t a linear process, but a re -entrant cyclical sense of you have all of these things as they change, feeding up through this knowledge base. And so you need this really strong computational infrastructure to be able to even begin to apply some of the really exciting and fascinating AI and ML methods we’re touching on.

So with this, we see a future where research can be done with autonomous research agents, and you can have source validations and get immediately generated counterpart biographies, analysis of gaps and preferences and evidence bases, strategy sandboxes, red team training, and trying to simulate how both the public and the public can interact with each other. And so you need this really strong data set to be able to do that. And then finally, in the room, executing a negotiation, you’ve got to be able to be able to identify the best way to do that. And so you need this really strong data set to be able to identify the best way to do that. And so you need this really strong data set to be able to identify the best way to do that.

And so and then in real time having transcription and translation services that AI and ML methods are doing a really phenomenal job at. All of these things we think will play a role in this multi -model, multi -method world of computational support for diplomacy negotiations. And so this is just a sense of how we’ve tried to break down this problem and get a grasp on the existing and future technical developments. Finally, we want to end on these three commitments that are central to a lot of the stuff that we’re talking about. One, human authority has to remain central. We can’t have any objection of responsibility over decisions of war and peace. We have to make sure that the tools themselves are modular and transparent so that you can see what’s happening at each stage of the process and which parts of which computational systems are supporting analysis.

And then finally, making sure that augmentation is appropriately scoped for the team, the institution, and the setting that it’s in. So with that, I’d like to hand over to Michael in the panel, our director of the program. And what I hope will be a wonderfull discussion.

J. Michael McQuade

Great. Thanks, Charlie. Thanks, Slavina. So just as everybody’s sort of getting settled in, so we have a plan for a project. We have a vision of how one has a set of signposts and goalposts in what is essentially the ability to augment with intelligence, human intelligence and participation. So there are lots of technical elements of that. We’ll be developing tools. We’re looking at evaluation methodologies, et cetera, et cetera, et cetera. It’s the whole technical side. But one of the benefits of the approach we’re taking is that we have access to a large body of people for whom the day -to -day practice of negotiation diplomacy, not necessarily constrained by the definition of diplomacy, meaning state -to -state to get to an answer, but organization -to -organization, people -to -people, negotiation -to -negotiation.

And I am delighted then to have three people here who can talk a little bit about their views on how artificial intelligence will be used in the process of their work. That allows us to then learn from that experience and how we map that into the Move 37 project. So Gabriela, let me start with you. Welcome. Thank you for negotiating all the traffic to get here. So you’ve been at the center of international policy design and negotiations on issues such as climate change, international taxation, gender equality, artificial intelligence, a whole list of things in a brilliant career. You’ve done this through key roles at UNESCO, but also at the G20 and G7 and at the OECD.

We’re delighted to have you here. And let me ask you just to sort of start the discussion, if you would, to just talk a little bit about what it’s like to sit in the driver’s seat as a mediator trying to bridge sides, and how you would think about AI capability augmenting you in that process.

Gabriela Ramos

Well, thank you. Thank you so much for inviting me. Is it working? For inviting me to this early morning. And I find this topic fascinating. because when you are a diplomat and when you have negotiated many standards or agreements you don’t think about this taxonomy you never think about the taxonomy you just think that you need to get it someday and that you need to find consensus and that you need to find where the problems will be and therefore it’s very interesting that you asked me something to structure better how we do things and I’m going to refer to the negotiation of the recommendation on the ethics of artificial intelligence because that was a very difficult one, 193 countries negotiating during COVID and actually it was very helpful to have a zoom where I could see where all the countries were positioning themselves which actually helped a lot but the interesting thing is that it was about artificial intelligence we were negotiating and we had to map out where countries were and it was very interesting to see that some of the usual suspects that are always blocking the effectiveness of international instruments were aligning with countries that are very supportive of those but that didn’t want to see UNESCO playing a role in this field so I have Russia and I have UK in the same position that helped me because I called the UK and say are you happy to be in the same position and then they just hold one second but the interesting thing is it’s a very heavy document it’s very very because there are so many cultures we have to almost define the step by step and the interesting thing is that when thinking how can AI help us organize better at the moment it did not provide with so many inputs it was 2021 but UNESCO has this idea of being super inclusive we developed the recommendation with all the regions in the world represented and all the disciplines but then we put it out to the world and we receive 55 thousand comments therefore we use AI to integrate them.

That was, no? But then when you think about how do you map the positioning of countries, I think that would have been super useful to have more AI. I used to have full teams providing me with briefings for the people we were going to talk because you need to be conducting a lot of, one thing is the negotiation in the room, another thing is all the legwork that you need to be doing, talking to the different actors, knowing where they stand. And that would have been amazing, just to have a repository of what is the traditional position of certain countries or certain negotiations, which had to do with the substance, but probably has to do also with the positioning of that country in the international context and how much they abide by the rules and how much they support these things.

And then what I find fascinating, but this is always, as my colleagues here said, how you keep the woman, the woman in the loop, I love it. Yes, woman in the loop, not human in the loop, human in the loop. it was a lapsus it’s not lost on me my panelists here it was a lapsus but I like my lapsus the whole point is when you are in front of a person and you’re trying to convince that person that he’s alone nobody’s supporting his position and therefore he should not continue blocking the negotiation how would it be that you can have more information about that person what moves them how can you offer something that will be important for them because this is the kind of things that we do negotiating what would you want to have out of it I know you have your bosses in your shoulder and you need to bring them something to the table but tell them you’re alone that you’re blocking it and imagine you can have the information about that person but that’s also risky because it deals with privacy and all of those things but I feel it would be fantastic because this is strategic thinking and using the right words to get the countries to agree that will bring you on some places and that I think is a very important thing that’s a capacity that can be augmented by AI.

Thank you very much.

J. Michael McQuade

Yeah. Is it on? Yeah. Thank you very much for, actually it’s a terrific transition because Charlie was talking about the complexity of these negotiations, about how they’re not dynamic. I can think of nothing more dynamic than a UNESCO negotiation. Just trying to understand where people’s positions are by itself is a complexity. Trying to integrate 20 or 30 of those positions or 190 of those positions and then trying to find what are the right levers that I might be able to pull. We do this now all the time with people, with you. And the question is how can modern tools help in that process without removing or absolving responsibility for people. So thank you. So Nandita, you have had an amazing career in academia, the public, the private sector.

You’ve been at Stanford. You’ve worked in intelligence and advisory. And you’ve seen all the different things sides of these negotiations, both from the government side, the private sector side, inside and outside. You are currently at the Director of Intelligence at the Special Competitive Studies Project. For those of you who don’t know, SCSP is a major effort funded by Eric Schmidt, sponsored by Eric Schmidt, after the conclusion of the National Security Commission on Artificial Intelligence in the U .S. in the way that technology will be used in competition for economics, national security, et cetera. So it’s a big, broad role. Every day you are negotiating. Every day in your life you have negotiating. So can I just ask you to talk a little bit about your view, both from an SCSP point of view, but also from your career about how you would see this evolving?

Nandita Balakrishnan

Absolutely. And thank you so much for having me. And good morning, everyone. So my career, as you mentioned, has sort of spanned three distinct sectors. And they kind of came at different times, academia, public sector, and private sector. Now, there was a time, I would say, that all… three of these groups would have, the particular ways they would have leveraged technology would have obviously its variations, but the access and adoption of it were much more similar. This is just fundamentally not true of AI. The public sector has been more in the passenger seat, if not the backseat, especially over the last decade. And so what was really interesting to me is I started in academia, then went to the public sector, came out of the private sector, and I saw that dip in my access to AI.

Now, I have been in intelligence, and one important thing about intelligence and maybe a misconception of it is that it is primarily used for military, feeding information for military applications. It’s not true. We are just as vital to diplomatic efforts because every opportunity you’re looking for that something bad could happen, you’re just as much looking at what are the opportunities for something positive to happen. How can you open the negotiating space? So we’re looking at everything from both sides. So I wanted to give that perspective. As an analyst, I can say personally, it was very valuable. valuable to have the rigorous training I had to do things very, very manually. So learning how to write an assessment without the access to AI.

But now that I’m on the back end of it, I can tell you every day I ask myself, like, if I had access to these tools as an analyst, how could I have worked much faster and much smarter? Because at the end of the day, and something that Gabriella was mentioning, there’s a lot of data out there, but a human analyst is never going to be able to manually process most of that by themselves. The story I always like to tell is the very first time I wrote this intelligence piece, I was so proud of it. I thought the argumentation was great. The data I had used was great. I showed it to a mentor, and they said, this is awesome, but you didn’t consider this one piece of data from 10 years ago that completely negates your argument.

And here’s the thing. It’s not that I didn’t do good work. It’s just that there was no way I was going to know that that piece of information exists. Now imagine a tool that can help you not only identify that that data exists, but learn how to synthesize it. Okay. Thank you. Obviously, as Charlie and Slavina mentioned, human in the loop is always going to be important because you want these assessments to have a human level at the end of it. But there is a way to move better and smarter. And this is something that SESP is really advocating for in sort of three distinct ways. So first, at a very meta level, we make the argument that AI has fundamentally changed the threat landscape and the scope for global competition.

It is now kind of the foundational way we need to think about geopolitics, especially as this technology is rapidly evolving. So you really cannot divorce AI and AI adoption when you’re trying to understand geopolitics and foreign policy. Number two, in order to have an ability to make assessments about this emerging technology, to understand geopolitics, you have to have a public sector that is actually leveraging these tools to the best of their ability. Now, there are a lot of ways that AI is being adopted at the public sector. sector, you know, you’re obviously thinking about sort of the, again, military application of drones, but you need to have your day -to -day workflows integrating this technology.

And this is something that we are really focusing on, especially about how to build up AI literacy within the public sector, not just at the military level, but within the intelligence community, within the State Department, and even within, like, commerce, OPM, all the, like, any federal sector employee at some point needs to be moving smarter and faster with AI. And third, we’re looking at the, like, specific use cases. So one of the projects that we were working on last year is looking at how AI can be used for predicting geopolitical events, both for military applications, but also for State Department applications. So, and the reason we do that is because in order to convince the public sector that they should be using AI, you almost need to show them how it could look like 10 years from now as we’re moving to that future.

So by kind of demystifying its use and showing them targeted ways that you can use it, it actually solves your meta -problem of understanding why AI is so important to geopolitics. .

J. Michael McQuade

So I’m hearing a couple of things. I’m hearing this general statement that much of the world is going to be about AI and much of the world is going to be AI creating that world. And that’s the metaphor that comes directly to the project we’re looking at, which is we’re negotiating with AI tools. We have to have a baseline of capability. And yet the landscape in which negotiation and diplomacy are happening is being fundamentally changed by AI itself. And so that whole issue around preparedness and around setting the ground rules. I also heard not just the thought process around artificial intelligence as a trusted agent to accumulate information. Both of you mentioned that. But also as an agent to help understand new pathways for success, new pathways for leverage, whether those are national security or whether those are economic vitality.

The scope of the negotiations doesn’t really change that. So. In the area of. sort of preparation, let me come to you, Robin. Robin is the co -founder and CEO of Apolitical. Apolitical is a global platform for policymakers that specializes in government innovation. She’ll talk about that in just a second. In particular, you have courses in helping governments prepare their workforces for the modern world in which they live. Your AI courses have reached hundreds of thousands of people around the world. And much of what you are trying to do is to prepare the world for the kind of things that we are talking about in Move 37, obviously much broader than just that topic. So let me ask you to talk a little bit about that, if you wouldn’t mind.

What sort of lessons you’ve learned from the field and how you think about policymakers’ willingness or how you change policymakers’ willingness to embark on journeys with new tools and new capabilities.

⁠Robyn Scott

Stanford HAI is one of our collaborators. So we are more context experts than content experts, and we bring the content experts into the middle. So where are we at? Let me give you some data. And this is from a 5 ,000 -person survey that we’ve recently run. Overall, public servants are incredibly optimistic about AI. North of 90 % think there is huge possibility in the public sector. And there are lots of paradoxes here. They’re also wary of it, right? There is a huge value creation opportunity. One figure from BCG estimates that there’s 1 .75 trillion of public sector value to be unlocked if we harness AI in the right way, because AI loves bureaucracy, all these repeatable processes. And about a third of most public officials daily watch And I’m going to do something about it.

And I’m going to do something about it. is research and writing related. AI is great at that. So the prize is very, very big and that’s just the painkiller prize. When you get to the vitamin prize, when you get to what AI could do in terms of predictive policy making and responsive policy and adaptive policy, et cetera, then you get into a space that’s only really bounded by the imagination. So there’s lots of AI talk. There’s less AI action. Increasingly, we’re in a pilotitis zone where almost everyone’s got pilots. 70 % of leaders say they’ve either got AI pilots or plan to launch them this year, but only 45 % of them say they have any plan to evaluate their pilots.

So that’s a pretty big gap to close and we see gaps like this all the time. One of the biggest gaps is leaders not using the technology themselves, which is a real problem because you can’t understand this technology in the abstract. You cannot look over your grandson’s shot. older and see them using it. You’ve got to use it. You’ve got to feel the speed of change. Of the public servants who are implementing AI in the public sector globally, these are people who self -identify of having AI in their jobs. Only 26 % of them say they understand their own country’s ethical frameworks. So approximately three -quarters of all the people rolling out this technology are freestyling. That’s terrifying.

So that’s a skills and knowledge gap not even closed within an institution. It’s not even getting to how do we actually understand the basics of this technology. Just to close, and there’s a whole lot more fascinating data, but one of the things that is increasingly worrying me talking to leaders around the world working on this is that we are now getting quite drunk on the idea of AI agency. Thank you. but we’re not talking about human agency in the process and maintaining it. So I think we risk getting into a zero sum dynamic where, and I think this is relevant to diplomacy, where the agency drains away to AI and that all comes at a cost to humans.

So we need to be building up humans at the same time. And the framing and heuristic I found most helpful for this overall is this idea of this recently merged of being below or above the algorithm. If you’re below the algorithm, you might be an Uber driver being dispatched, an Amazon packing worker being allocated to put stuff into boxes. If you’re above the algorithm, you are using tools to further your goals. We need, when we think about closing that capability gap, and I think in diplomacy, to keep moving people up above the algorithm.

J. Michael McQuade

Great, fantastic comments from all of you, thank you. I’m gonna ask Charlie and Sylvina just a couple of quick comments to make comments on the project. All right. Because then I’m going to come back to the three of you and I’ll prompt you with the question now, which is what would you want to be comfortable with knowing about the tools and the capabilities that you will be asked to use or be offered to use? So, Slavina, can you just follow on Robin’s comments? One of the benefits we have of doing this program at the Belfer Center and the Kennedy School is a large set of people who have done this for a living. This is what diplomats and negotiators have done.

Can you talk a little bit about how we engage that group and what we’re trying to get from that?

Slavina Ancheva

Definitely. So I think, as Michael put it, obviously at the Belfer Center we have quite a variety of current, former diplomats, practitioners, not just from the U .S., but from all over the world. And a large part of the work that we’ve been doing is sitting down for one -on -one interviews with all of them and really getting a sense of how they think. Thank you. just about the content of the major negotiations that they’ve been leading, but more about the process. So very similar to the panel discussion we’re having here today, what are some of the uses that you see one day you could be using AI for? So a lot of what we’ve heard so far, I mean, the position tracking that I think Gabriella referenced.

We’ve heard a lot about historical precedent, the generating of options and strategic options, and really uncovering the deepest interests. And I think where this ties really well to what Robin was saying is a lot of them are also expressing their hesitancy. So they’re being very forthcoming in that, and I think that allows us to take a really sober look at what are the risks of integrating these tools. One of the main questions we get is if you’re using these tools, we ask them, what would you like to know? So exactly what Michael’s saying now. So a lot of these interviews have really been integrated across the different work streams of our project, and we really put diplomats and practitioners at the heart of the rest of the work that we’re doing.

All right. Thank you.

J. Michael McQuade

And Charlie, you talked a little bit before about, you know, there’s an obvious role that LLS has. I think that’s a really good point. And I think that’s a really good point. in helping people accumulate and synthesize a very large amount of information. But there are many more aspects of a negotiation. Can you talk just a little bit about some of the other ways the tools are going to be used in the project that we’re doing?

Charlie Posniak

Yeah, absolutely. Thanks so much, and thanks so much for the comments as well so far. The panelists have touched on a couple of really interesting applications, whether it’s, especially as Robin was talking about, with predictive or adaptive policies, if you’re looking at, and Anita as well, with the predictive geopolitical events. We have a really fascinating array of algorithms that are incredibly competent at these sorts of predictive tasks. And what we now have with the current computational ability and also the ability of language models is we can process vast amounts of unstructured data in ways that makes these algorithms even more accessible to a wider range of people. So I think that that’s one area that I’m particularly excited about.

There’s also a bunch of stuff, as Gabriella was touching on, is how do you take these vast unstructured transcripts and come up with natural… natural language processing additions on top, and can we represent positions that they track? and change over time. I think that is one of the big parts of cognitive load that I think diplomats have spoken about a lot.

J. Michael McQuade

Great, thank you. We’re going to see if questions from the audience are in just a moment, but Nandita, I’m going to start with you. I’m going to, forgive me for doing this, I’m going to characterize your career as an analyst. Every job that I read and see about has been an analyst of some kind. You’re constantly in a place where people are suggesting new tools and new capabilities. So think about AI in the world you’ve inhabited. What’s going to make you most comfortable when somebody shows up and says, here’s the thing, it’s going to make your life better?

Nandita Balakrishnan

That they can explain what the outputs are. So one of the things, like when we were looking at, for example, could we be using AI for predicting geopolitical events, a lot of people, both in industry, in academia, and in the public sector, that are working on these types of projects, they all say the same thing, which is this should be seen as a data point or a shaper of the way you should view the world, not as finished intelligence. Finished intelligence, should always… ultimately be done by a human who is accountable to their policymakers. Now, if you are working in policy, you understand actually how this works. You’ll have your head of state or your head of government come to you and ask you to explain exactly how you got to your assessment.

Right now, we have a human who ultimately has to do that. But what happens when we’re starting to rely more and more on AI tools is you never want to lose that ability to explain the outputs and particularly demonstrate that you’ve looked at all the counterpoints that you could possibly do. Now, this is where I think AI is super helpful because oftentimes when I was trying to figure out, especially in academia, what are all the things I could have done wrong? How could I have measured this differently? You’re always prepared to thinking about all the decisions that you made and how to justify and validate them. But as the scenarios get broader, as they get more complicated, your ability to figure out what the counterarguments are are going to just dwindle over time.

Oftentimes, the argument we make is that humans are biased at the end of the day. We make our decisions differently. We make our decisions differently. We make our decisions differently. We make our decisions differently. We make our decisions differently. We make our decisions based on how we got to where we are today. the experience that you have. So AI can be really, really helpful in helping you sort out the counterarguments, but you still need to understand how those counterarguments work and why ultimately you’ve come to the assessment that you have. So where I would feel super comfortable is this is how I relied on AI. This is how it came to the output that it did.

This is fundamentally why I made the assessment that I did.

J. Michael McQuade

Great. Thank you. And Robyn, I think this is the world you live in every day of helping governments and government officials and civil service workers. So project that onto an AI for diplomacy landscape. What do you think is going to be important to get people to say, I’m going to trust this, I’m going to work it, but I’m at least going to try?

⁠Robyn Scott

Well, at the risk of stating the obvious, I think we should just acknowledge that the people developing these models don’t even have full legibility over how they’re working. So that’s where we’re starting from. So that’s the kind of ceiling. On where we can get to. You can break down the thinking process as it were, but you still have that black box. I don’t think it’s insurmountable. I think some of the things I’m worried about relate to the more psychological aspect of this, and in particular, sleeping at the wheel, this phenomenon where we have this strange relationship with AI where we get false negatives too quickly. So it does a bunch of clever things, except it didn’t do this one thing, and therefore we can’t use it for anything.

And if you check back in in like a month’s time, often it can do the thing. So you have that, the false negative, and then the phenomenon of sleeping at the wheel is where it starts to get very, very good, creeping upwards of like 85%, 90 % accuracy, and then you assume it’s 100 % accurate. And it’s really quite hard to edit your assumptions and say, no, it’s not. And you’ve probably all found this. Because if they’re power users of AI, and I’m one of them, some of them are not. Sometimes it comes across as so smart and so brilliant and comes up with a whole lot of counter -arguments. I use it for sort of kicking the tires from different perspectives all the time on stuff I’m doing.

That it’s almost overwhelmingly smart, and you’re like, it must have covered everything. That’s a default. So I think giving us the human tools and the psychological sort of counter -arguments and weaponry to deal with this is really, really important. I already have a heuristic that whenever I open my phone and I’m dealing with anything with an algorithm, I am in opposition to that algorithm because its interests don’t generally coincide with mine. So I try and get all algorithmic stuff off my phone as a starting point. But the dynamic with AI is a bit different, but I still think you have to have that sort of battle mentality with the technology. So that would be my…

There are many other things to consider, but that’s top of mind.

J. Michael McQuade

I think that’s terrific. And I think this idea of calibrating on the… Like what I want from completeness when I ask for analysis may be very different than what I want from can you just give me some different ideas that I haven’t thought about before. And different stages in a negotiation are going to require different levels of calibration. So, Gabriela, what do you think?

Gabriela Ramos

Well, it’s very difficult to follow these two girls. But the fact is that when you know a little bit more about how these things work, I’m not a technologist. But I have been looking at all of what can go wrong. Misrepresentation, over -representation of certain cultures, certain languages, assumptions. Therefore, if I am negotiating and you’re going to offer me a tool to improve my negotiating skills, I need to be sure that the assumptions that you use to build that tool are not just to beat the person in front of me. or not just to maximize efficiency or not just to do the kind of things that we are teaching the AI to do. And therefore, it’s much more complex.

Because what you want to do is to open a space of human understanding. How do you do that? And therefore, I will be questioning, as Robin said, always questioning, but it’s not what we do. And the other point is that the AI, what is amazing, is that it’s just reproducing cognitive abilities that humans do. So when you go into the using whatever chat box you use to get information, you take for granted what it comes out. What you would never do if you hire somebody in the first week. Even if you have done all the checkpoints for that person to have the capacities that you are looking in the market. So I feel that there is this question of, first, really bringing to the table the AI.

tools that are going to be reliable and trustworthy, and I know that these words are almost a cliche, but the reality is that sometimes they’re not. And the other point is that you can become very lazy. And how do you avoid just to grab the thing and say, that’s perfect? How do you keep that space for ourselves to take the decisions and be not only in the driver’s seat, but actually to think of AI as a supporter cast. And if we get the Oscar, it’s us and not the AI.

J. Michael McQuade

That’s fantastic. I have this mental picture in my mind for those of you who’ve done negotiations in any kind. The first thing you do is you grab a bunch of your team in a room and you say, let’s talk about strategy, and what are we going to do? And Bob in the corner says, here’s an option, and you realize Bob had a bad night last night, so maybe you discount what Bob says. So what happens when the AI says, here’s a thing? I don’t just trust it, it’s a priority. I have to apply a human judgment to what I’m hearing. So terrific points. Okay, we have time for a question or two, and I see one.

Just say who you are, where you’re from, and a quick question.

Audience

Thanks so much, Michael.

J. Michael McQuade

We have a microphone. Thank you.

Audience

Thanks so much. I’m Sam Dawes. I’m a senior advisor to the Oxford University AI Governance Initiative and director of multilateral AI. But my background is in diplomacy, working for Kofi Annan when he was Secretary General, and then for the Foreign Office and Cabinet Office. I wish we had had AI tools back then. So I was really inspired. It’s such a timely, rich panel, so thank you all for that. Something that Gabriella said around culture I think is so important, and I’m thinking about the positives and the risks with applying AI in this space. How can we ensure that the diverse cultural inputs of the world’s most diverse countries, of different societies, are… embedded in the data sets and the models which inform negotiations.

So is that something that UNESCO is working on in the long term and connects to the tools we use? And the second question is around the flip side. If AI is to be a useful neutral mediator in disputes or an assistant to a mediator, a human mediator, then what do we do about data poisoning and prompt injection and those kinds of risks? Thank you.

Gabriela Ramos

Very fast, not on the question, the question of culture. Culture is expressed by language. And therefore the more we can try to represent those languages in the models we use, I think the best we will be prepared to understand it. And I’m fascinated by that. I’m not a linguist, but if I… I would choose another life, I would do that. Because when you hear, for example, there was this Namibian representative during the negotiations of the ethics of AI, and she was saying, I find your draft very individualistic. It’s always about the human. It’s always about the outcomes for people, improving their welfare. And at the end, what I’m thinking about is the Ubuntu philosophy, which is I am because you are, and we are because it’s nature, and we are interlinked.

And therefore, how do you capture this when the models that we are developing are maximizing individual welfare? And so the only answer I have is try to be representative, and I think this is nothing new. We have seen how much these tools can discriminate if you are just built in one language or with the representation of certain characteristics of people. or countries. but really to be sure that you are capturing the richness that comes through language and opening up the sources and that’s the other point the sources this is one thing that I would always ask the answer you’re giving me is based on what sources and that would might help but these are checkpoints that we always need to be testing on the ground

J. Michael McQuade

I think you also if I just raise one other thing I think you raise one other really important point which is you know there’s a whole spectrum of things here there is negotiation because we have a set of interested parties to get to a common good understanding we also have very adversarial negotiations so adversarial negotiations open up this whole possibility of data poisoning of training set differentiation etc. so it’s a very complex world I really appreciate you bringing it up we have time for one more question I think let’s go right here can somebody tell me are we counting down to zero or are we counting down to zero or are we counting down to five Are we okay to keep going to zero here?

Okay, good. We’re going to go to zero no matter what.

Audience

Good morning. Good morning. Namaste. My name is Devika Rao. I meet 300 to 600 people per day, and I work around different languages. So basically I’m an Indian classical dance teacher. Okay. So I have data. I have a human connection. So, and what we want to do, how this cultural education can be supported by AI. So what is the step I can take further? Presently I’m actually working on a framework, cultural framework, which is India and UK POCC 2025, 2030. So I’m also interested in NEP and national health policy because people connected to their health and education. And education, which is the center point. So where I can go. and what kind of co -creation, co -collaboration can happen in this?

J. Michael McQuade

Robyn, is this something you want to jump in on? Maybe give her your email address.

⁠Robyn Scott

I wish I had an immediate response to that. I don’t think there is any default place to go, but I do think this is where the conversation is evolving, and there’s more and more recognition of the cultural oversight and importance. So I would just encourage you to please keep making those points. And I will just make one comment on the first question. The Swiss have built sort of a quasi -Swiss government, quasi -multilateral initiative to build an LLM that is trained from the outset on more than 100 languages, and it is actually run by a friend of mine who’s a former Swiss diplomat, so she’s coming at it very much with a diplomatic context. I’m very happy to make that connection.

Gabriela Ramos

Education. Super. Super complex. Don’t look at the technology. because we always focus on the technology the countries that have introduced so much technology in their educational systems didn’t get better student outcomes because of content we go to the internet and we go to the systems and we try to bring tools to help kids and we never see if they are contextually relevant culturally linked and therefore if you don’t produce the content the tools will not make it.

Nandita Balakrishnan

I’ll just add one last thing I think the way to think about AI is also is it actually solving a problem or are you just trying to introduce it to create a new problem I think this is where you have to think about the point of AI augmentation I think there are a lot of ways we can think about how AI can augment the problem sets that we have but sometimes you don’t actually have the problem that AI is going to solve and you don’t need to force AI to fix it

J. Michael McQuade

Thank you very much Okay we’re going to negotiate If you have a really quick question you can ask it No behind you Thank you it’s got to be quick though

Audience

My name is Arman I’m working for JPL South Asia just a quick question on how do you think this would impact balance of power like given that every country has different access to the kind of data sets that they have and as we saw there can be three states also in the play how it would look like state A knows everything about the rest of the players and the others don’t.

J. Michael McQuade

So we think a lot about I’ll answer this if you guys are okay we think a lot about this in the project which is what’s the evolution of a set of AI tools it’s like everything else that we are here at this conference which is where will tools provide competitive leverage where are the kind of tools in the world we live in ones that should be dispersed actively and offensively not defensively in a world where some of the negotiation is about getting everybody to a positive and some of the negotiation is adversarial so I think it is a huge element of how it will change power structures not just because it’s a thing we think about from a negotiation diplomacy but because the general AI tool.

Okay with that we are just about out of time. I want to thank this amazing panel. Gabriela, Nandita and Robin. I want to thank my colleagues Slavina and Charlie for and I want to thank all of you. We are at the beginning of a long process. When you work at a place like I do at the Belfer you think about projects that have beginnings, middles and ends and you think about projects that can grow into something really really important. So any of you who have interest in what we’re doing please let us know if you feel like you have questions that we ought to be asking or you have answers or you have answers to questions we have asked.

We would love to hear from you as we begin to build what we think is a really important discipline. So thank you and thank you to the sponsors and hosts. I appreciate everybody joining us. Thank you.

Related ResourcesKnowledge base sources related to the discussion topics (30)
Factual NotesClaims verified against the Diplo knowledge base (5)
Confirmedhigh

“J. Michael McQuade introduced the MOVE 37 initiative, a new project by Harvard’s Belfer Centre exploring AI’s role in diplomatic negotiations.”

The knowledge base explicitly describes MOVE 37 as a new Belfer Center project introduced by J. Michael McQuade to explore AI augmentation of diplomatic negotiations [S2].

Confirmedhigh

“Artificial intelligence is a major aspect of the Emerging Tech Programme’s work and can augment human capabilities in diplomacy and negotiation.”

S2 notes that AI is central to the programme’s aim to augment human capabilities in diplomatic negotiations, and S16 lists specific AI applications for negotiation analysis, confirming the claim [S2] and [S16].

Additional Contextmedium

“AI tools can help participants accumulate and synthesize large volumes of information during negotiations.”

S1 highlights the role of large language models in helping people accumulate and synthesize very large amounts of information, providing additional nuance to the claim [S1]; S16 further describes AI-driven data analysis for negotiation scenarios [S16].

Confirmedhigh

“Diplomacy remains a fundamentally human activity; AI should augment but not replace human agency in high‑stakes negotiations.”

Both S17 and S24 stress that AI is a tool that must remain under human control and that the art of negotiation and trust-building are profoundly human, supporting the report’s framing [S17] and [S24].

Additional Contextmedium

“Negotiations often involve many counterparts, multiple countries, and generate massive amounts of documents, creating information overload and strategic group‑think pressures.”

S106 discusses the complexity of international negotiations, including the struggle over process, multiple participants, and extensive documentation, adding context to the described information overload [S106].

External Sources (106)
S1
https://dig.watch/event/india-ai-impact-summit-2026/how-ai-is-transforming-diplomacy-and-conflict-management — And then at the far end, two of my colleagues, researchers on our program at the Belfer Center. Charlie Posniak is a ful…
S2
How AI Is Transforming Diplomacy and Conflict Management — And then at the far end, two of my colleagues, researchers on our program at the Belfer Center. Charlie Posniak is a ful…
S3
How AI Is Transforming Diplomacy and Conflict Management — And the MOVE 37 initiative that we’re here to talk to you about today. is a part of that program. As you can imagine, in…
S4
How AI Is Transforming Diplomacy and Conflict Management — -Slavina Ancheva- Research fellow and MPP student at the Belfer Center, working on the MOVE 37 initiative
S5
https://dig.watch/event/india-ai-impact-summit-2026/how-ai-is-transforming-diplomacy-and-conflict-management — And then at the far end, two of my colleagues, researchers on our program at the Belfer Center. Charlie Posniak is a ful…
S6
How AI Is Transforming Diplomacy and Conflict Management — And the MOVE 37 initiative that we’re here to talk to you about today. is a part of that program. As you can imagine, in…
S7
https://dig.watch/event/india-ai-impact-summit-2026/how-ai-is-transforming-diplomacy-and-conflict-management — And the MOVE 37 initiative that we’re here to talk to you about today. is a part of that program. As you can imagine, in…
S8
How AI Is Transforming Diplomacy and Conflict Management — – Robyn Scott- J. Michael McQuade- Charlie Posniak
S9
tABle of Contents — and Costs, 24 health aff . 1103, 1103 (Sept./Oct. 2005), available at http://content. healthaffairs.org/cgi/reprint/24/5…
S10
MASTERPLAN FLAGSHIP PROGRAMMES — | S/N | Name | State Departments | |——-|———————|-…
S11
WS #280 the DNS Trust Horizon Safeguarding Digital Identity — – **Audience** – Individual from Senegal named Yuv (role/title not specified)
S12
Building the Workforce_ AI for Viksit Bharat 2047 — -Audience- Role/Title: Professor Charu from Indian Institute of Public Administration (one identified audience member), …
S13
Nri Collaborative Session Navigating Global Cyber Threats Via Local Practices — – **Audience** – Dr. Nazar (specific role/title not clearly mentioned)
S14
The digital economy in the age of AI: Implications for developing countries (UNCTAD) — Gabriela Ramos, Assistant Director General for Social and Human Sciences at UNESCO, has highlighted the unique mandate o…
S15
https://dig.watch/event/india-ai-impact-summit-2026/how-ai-is-transforming-diplomacy-and-conflict-management — And the MOVE 37 initiative that we’re here to talk to you about today. is a part of that program. As you can imagine, in…
S16
Negotiations — Negotiation is a complex and dynamic process requiring strategic thinking, psychological insight, and cultural awareness…
S17
Why will AI enhance, not replace, human diplomacy? — AI tools are already here to assist certain aspects of negotiations, from language translation to data analysis. However…
S18
Artificial Intelligence & Emerging Tech — Jörn Erbguth:Thank you very much. So I’m EuroDIG subject matter expert for human rights and privacy and also affiliated …
S19
Cybermediation: What role for blockchain and artificial intelligence? — After explaining in further detail some aspects of NLP, she suggested that these tools can be used to support the work o…
S20
What are diplomatic competencies for the AI era? — In emerging hybrid intelligence models, diplomats must effectively collaborate with AI systems, blending human judgment …
S21
Enhancing rather than replacing humanity with AI — Human judgment stays central, especially for important decisions. AI provides information, analysis, or options, but peo…
S22
Turbocharging Digital Transformation in Emerging Markets: Unleashing the Power of AI in Agritech (ITC) — Moreover, while AI and new technologies have significant potential in agriculture, it is crucial to understand that they…
S23
Seeing, moving, living: AI’s promise for accessible technology — Human oversight and choice must remain central. Users should control when their devices collect data, how that data is u…
S24
AI diplomacy — However, we must remain masters of our tools. The final analysis, the subtle art of negotiation, the building of trust; …
S25
Diplomatic policy analysis — Overreliance on technology:While machine learning and analytics are powerful tools, they are not infallible. Overdepende…
S26
From agentic AI to agreement technologies: LLMs as a new layer in diplomatic negotiation — Another concerns transparency and accountability. For negotiation support, it is not enough that a system produces a pla…
S27
Engineering Accountable AI Agents in a Global Arms Race: A Panel Discussion Report — Moderate disagreement with significant implications. While speakers share common concerns about AI governance, they diff…
S28
WS #288 An AI Policy Research Roadmap for Evidence-Based AI Policy — Alex Moltzau: Yes, so one thing that I didn’t mention that we are working on currently is also these AI regulatory sandb…
S29
Artificial intelligence and diplomacy: A new tool for diplomats? — Artificial intelligence (AI) is transitioning from science fiction into our everyday lives. Over the past few years, the…
S30
WS #97 Interoperability of AI Governance: Scope and Mechanism — Yik Chan Chin: Thank you, Sam, because we know he’s an expert in terms of the UN. So thank you very much for your comm…
S31
How AI Drives Innovation and Economic Growth — The speakers show broad agreement on AI’s transformative potential for development but significant disagreements on impl…
S32
Panel Discussion Summary: AI Governance Implementation and Capacity Building in Government — The discussion revealed a common theme across different contexts: the gap between policy ambition and implementation cap…
S33
AI as critical infrastructure for continuity in public services — Awareness and capacity gaps exist in understanding available standards and building blocks
S34
Artificial intelligence (AI) – UN Security Council — Another significant risk is the potential for bias in AI algorithms, which can reflect existing prejudices and stereotyp…
S35
How to ensure cultural and linguistic diversity in the digital and AI worlds? — Audience:Hamid Hawja is my name, from Morocco, director of Hebdo magazine. I have two questions. First, I’m just wonderi…
S37
Inclusive AI For A Better World, Through Cross-Cultural And Multi-Generational Dialogue — Qian Xiao:OK, well, I’m doing a lot of research on the international governance of AI. And from our perspective, we thin…
S38
AI diplomacy — However, we must remain masters of our tools. The final analysis, the subtle art of negotiation, the building of trust; …
S39
Enhancing rather than replacing humanity with AI — People’s judgment remains crucial, particularly for decisions that involve values, context, or individual circumstances.
S40
Why will AI enhance, not replace, human diplomacy? — In sum, AI is a powerful tool, but it is still just a tool. It is a good master and a bad servant. As we step into AI tr…
S41
Ateliers : rapports restitution et séance de clôture — Joseph Nkalwo Ngoula Merci. C’est toujours difficile de restituer la parole d’experts de haut vol. sans courir le risque…
S42
Open Forum #37 Digital and AI Regulation in La Francophonie an Inspiration and Global Good Practice — Boukar Michel: Thank you, Mr. Henri. Mr. Ambassador in charge of digital, thank you for giving me this opportunity to ta…
S43
How to ensure cultural and linguistic diversity in the digital and AI worlds? — Such inclusivity has far-reaching implications for achieving SDG 4: Quality Education and SDG 10: Reduced Inequalities, …
S44
Artificial intelligence — Content policy Cultural diversity Inclusive finance Multilingualism
S45
Embracing AI in diplomacy: How can Europe prepare for pivotal transformation in global affairs? — Firstly, AI is reshaping the geopolitical environment in which diplomacy operates. It facilitates the redistribution of …
S46
Shaping an inclusive global action to anticipate quantum technologies — Cooperation amongst non-quantum states may shift power dynamics. Moreover, the strategic alignment among the Global Sou…
S47
Revisiting 10 AI and digital forecasts for 2025: Predictions and Reality — Digital networks and AI developments are critical assets for countries worldwide. Thus, they become central to national …
S48
How AI Is Transforming Diplomacy and Conflict Management — And Charlie, you talked a little bit before about, you know, there’s an obvious role that LLS has. I think that’s a real…
S49
Negotiations — Artificial Intelligence (AI)has various applications in diplomacy. It can be used for data analysis to predict the outco…
S50
AI Algorithms and the Future of Global Diplomacy — “But of course, we have specific applications in the foreign office like supporting negotiations.”[14]. “A lot of what d…
S51
World Economic Forum Town Hall on AI Ethics and Trust — Risk Assessment Before Trust Trust requires context and cannot be evaluated without specific use cases. Botsman argues …
S52
AI with Trust — how trust inAIcan be increased through a coordinated interplay between standards, laws and conformity assessment
S53
Searching for Standards: The Global Competition to Govern AI | IGF 2023 — Audience:Thank you. My name is Sonny. I’m from the National Physical Laboratory of the United Kingdom. There’s a few wor…
S54
WS #288 An AI Policy Research Roadmap for Evidence-Based AI Policy — Alex Moltzau: Yes, thank you so much. My name is Alex Maltzau. And I work as a second national expert in the European AI…
S55
WS #283 AI Agents: Ensuring Responsible Deployment — As the session reached its time limit (with Prendergast noting the final 10 minutes), the discussion revealed both the p…
S56
Building the Next Wave of AI_ Responsible Frameworks & Standards — I think there is a significant role the governments, innovation hubs, academia, and startups have to play in developing …
S57
Artificial intelligence (AI) – UN Security Council — Algorithmic transparency is a critical topic discussed in various sessions, notably in the9821st meetingof the AI Securi…
S58
AI as critical infrastructure for continuity in public services — “I believe that there is perhaps awareness challenge as well as the capacity challenge, because I think that this whole …
S59
Open Forum #17 AI Regulation Insights From Parliaments — AI governance requires ongoing education for all stakeholders – politicians, policymakers, and the general public. This …
S60
Gender rights online — AI systems can learnbiasesfrom training data, leading to discriminatory outcomes online, includinggender-based dispariti…
S61
Leveraging AI to Support Gender Inclusivity | IGF 2023 WS #235 — In conclusion, the use of AI language understanding has made significant progress in reducing inappropriate sexual conte…
S62
AI diplomacy — However, we must remain masters of our tools. The final analysis, the subtle art of negotiation, the building of trust; …
S63
Why will AI enhance, not replace, human diplomacy? — AI tools are already here to assist certain aspects of negotiations, from language translation to data analysis. However…
S64
Oman: Nexus between traditional and tech diplomacy — There is much discussion about AI and digital tools changing diplomacy. Yet, it is clear that technology will not replac…
S65
From agentic AI to agreement technologies: LLMs as a new layer in diplomatic negotiation — Another concerns transparency and accountability. For negotiation support, it is not enough that a system produces a pla…
S66
How AI Is Transforming Diplomacy and Conflict Management — “We have seen how much these tools can discriminate if you are just built in one language or with the representation of …
S67
AI Transformation in Practice_ Insights from India’s Consulting Leaders — For consulting firms, the path forward involves embracing AI as an enabler whilst focusing on uniquely human capabilitie…
S68
How AI agents are quietly rebuilding the foundations of the global economy  — AI agents have rapidly moved from niche research concepts to one of the most discussed technology topics of 2025. Search…
S69
Harmonizing High-Tech: The role of AI standards as an implementation tool — Sezio Onoe:Yeah, actually the ITU has published over 100 standards and also the 120, around 120 now ongoing. So actually…
S70
How AI Drives Innovation and Economic Growth — The speakers show broad agreement on AI’s transformative potential for development but significant disagreements on impl…
S71
Panel Discussion Summary: AI Governance Implementation and Capacity Building in Government — The tone was pragmatic and solution-oriented throughout, with speakers acknowledging both challenges and opportunities i…
S72
Workshop 6: Perception of AI Tools in Business Operations: Building Trustworthy and Rights-Respecting Technologies — Ellis emphasized that while organizations are investing in AI technology, there’s a significant skills and capability ga…
S73
AI as critical infrastructure for continuity in public services — Awareness and capacity gaps exist in understanding available standards and building blocks
S74
How to ensure cultural and linguistic diversity in the digital and AI worlds? — Audience:Hamid Hawja is my name, from Morocco, director of Hebdo magazine. I have two questions. First, I’m just wonderi…
S75
Artificial intelligence (AI) – UN Security Council — Another significant risk is the potential for bias in AI algorithms, which can reflect existing prejudices and stereotyp…
S77
WSIS Action Line C8: Multilingualism in the Digital Age: Inclusive Strategies for a People-Centered Information Society — Sofiya Zahova: Thank you, Davide. I’m honored and delighted to join you today on this important panel, but even more ple…
S78
How Multilingual AI Bridges the Gap to Inclusive Access — The tone was consistently collaborative, optimistic, and mission-driven throughout the conversation. Speakers demonstrat…
S79
WAIGF Opening Ceremony & Keynote — The overall tone was formal yet optimistic. Speakers expressed enthusiasm about the potential of digital technologies wh…
S80
Opening — The overall tone was formal yet optimistic. Speakers acknowledged the serious challenges posed by rapid technological ch…
S81
World Economic Forum Annual Meeting Closing Remarks: Summary — The tone is consistently positive, celebratory, and grateful throughout the discussion. It begins with formal appreciati…
S82
Opening address of the co-chairs of the AI Governance Dialogue — The tone is consistently formal, diplomatic, and optimistic throughout. It maintains a ceremonial quality appropriate fo…
S83
Digital trade negotiations- understanding non-participation (National board of Trade – Sweden) — The lack of global representation in digital trade negotiations is problematic as it can lead to a fragmented approach a…
S84
Afternoon session — Israel criticizes the negotiation process, particularly the final stage, as lacking transparency and not reflecting the …
S85
WS #219 Generative AI Llms in Content Moderation Rights Risks — – **Technical Limitations and Trade-offs**: The discussion covered inherent technical challenges including the precision…
S86
Main Session on Cybersecurity, Trust & Safety Online | IGF 2023 — Christopher Painter:Thank you, Olga, and it’s great to be here, and I should say that you’re wondering what I’m wearing …
S87
Transforming Health Systems with AI From Lab to Last Mile — The discussion maintained a cautiously optimistic and collaborative tone throughout. It began with enthusiasm about AI’s…
S88
Strengthen Digital Governance and International Cooperation to Build an Inclusive Digital Future — The recognition of Global South leadership and the importance of environmental sustainability represents a maturing of d…
S89
Agenda item 5: discussions on substantive issues contained in paragraph 1 of General Assembly resolution 75/240 (continued)/4/OEWG 2025 — Israel: Good morning and thank you, Chair. We will present in brief, for the sake of time, some main points of our nat…
S90
Quantum Technologies: Navigating the Path from Promise to Practice — The tone was cautiously optimistic and pragmatic throughout. Panelists demonstrated excitement about quantum’s potential…
S91
High-Level session: Building and Financing Resilient and Sustainable Global Supply chains and the Role of the Private Sector — Conclusively, the discourse maintained an optimistic outlook, acknowledging the transition’s dual aspects of risk and pr…
S92
NATIONAL CYBER SECURITY FRAMEWORK MANUAL — – Hallingstad, Geir, and Luc Dandurand. Cyber Defence Capability Framework – Revision 2. Reference Document RD…
S93
Global Governance of Digital Technologies: A Contemporary Diplomacy Challenge — Technology and international affairs are interrelated. The relationship between technological and international relation…
S94
Creating Eco-friendly Policy System for Emerging Technology — Furthermore, there is an emphasis on inculcating global consciousness, forging new partnerships, and pushing for innovat…
S95
Challenges and Opportunities: Emerging Technologies and Sustainability Impacts  — In summary, the workshop highlighted the need for a comprehensive approach to sustainability in technology, covering inf…
S96
Closing Ceremony — It pushed the discussion towards considering future technological developments and how governance should adapt to them. …
S97
9821st meeting — Ecuador:Mr. President, I thank the United States for convening this important meeting. I also thank the Secretary Genera…
S98
IGF 2023 WS #313 Generative AI systems facing UNESCO AI Ethics Recommendation — Changfeng Chen:Yes, the discussion, the question were very interesting and inspired me to bring up a relative thinking, …
S99
UN General Assembly appoints experts to the Independent International Scientific Panel on AI — The UN General Assembly hasappointed40 experts to serve on a newly created Independent International Scientific Panel on…
S100
GPAI: A Multistakeholder Initiative on Trustworthy AI | IGF 2023 Open Forum #111 — Abhishek Singh:I can take that, no worries. Thank you, Abhishek. The floor is yours. You can give your question. Yeah, t…
S101
Embedding Human Rights in AI Standards: From Principles to Practice — Karen McCabe: But before I go there, first, I want to thank the organizers for this session. It’s really a very importan…
S102
Main Session on Artificial Intelligence | IGF 2023 — James Hairston:I’ll maybe start with two projects that I think begin to get it sort of solving for this, but again, are …
S103
Leveraging the UN system to advance global AI Governance efforts — The current difficulties in achieving consensus in multilateral systems underscore the necessity for inclusive negotiati…
S104
Charting the Course: Discussing the Impact and Future of the Internet Governance Forum — Anriette Esterhuysen:I think the new challenges are endless. Just look at climate change. Look at the impact of climate …
S105
Open Forum #33 Open Consultation Process Meeting for WSIS Forum 2025 — Moving forward, participants suggested several action items:
S106
HUMANITARIAN NEGOTIATION — A great deal of the conflict in any negotiation is often played out in a struggle over process. It may sound ridiculous …
Speakers Analysis
Detailed breakdown of each speaker’s arguments and positions
S
Slavina Ancheva
6 arguments193 words per minute880 words273 seconds
Argument 1
Negotiations involve multiple parties, extensive documentation, strategic dynamics, and time pressure, making them cognitively demanding.
EXPLANATION
Slavina describes diplomatic negotiations as complex processes that involve many counterparties, large volumes of documents, strategic group dynamics, and strict time constraints. This complexity creates a heavy cognitive load for negotiators.
EVIDENCE
She illustrates the complexity by describing a scenario where a negotiator faces seven counterparts from different countries, a supporting team, and the interests of 27 other countries, highlighting the multi-layered stakeholder environment [58-65]. She then enumerates the practical challenges: thousands of documents generated, limited analyst resources, strategic groupthink, and tight time limits that pressure negotiators [66-72].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The complexity of diplomatic negotiations, including many counterparties, large document volumes and tight timelines, is highlighted in the discussion of AI transforming diplomacy [S2].
MAJOR DISCUSSION POINT
Complexity of diplomatic negotiations
Argument 2
The MOVE 37 project will develop tools, evaluation methodologies, and conduct stakeholder interviews to embed AI responsibly in diplomatic workflows.
EXPLANATION
J. Michael outlines the technical agenda of MOVE 37, while Slavina explains how the project gathers insights from practitioners through interviews. Together they emphasize building tools, evaluation frameworks, and stakeholder engagement to integrate AI into diplomacy responsibly.
EVIDENCE
J. Michael notes that the project will create tools, evaluation methodologies, and address technical aspects of AI-augmented negotiation [124-129]. Slavina adds that the team conducts one-on-one interviews with current and former diplomats worldwide to inform the project’s design [267-278].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The description of MOVE 37’s objectives, including tool creation, evaluation and practitioner interviews, appears in the AI-transforming-diplomacy overview [S2].
MAJOR DISCUSSION POINT
MOVE 37 practical implementation
Argument 3
AI should be positioned as an augmenting tool for diplomats, preserving the fundamentally interpersonal nature of negotiations.
EXPLANATION
Slavina stresses that the goal is not to replace negotiators but to give them better tools to manage complexity, keeping the human touch central.
EVIDENCE
She says, “We’re not looking to replace diplomats or negotiators here, but just to give them the tools to manage these complexities much better” [61-62].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The view that AI enhances but does not replace human diplomacy is echoed in discussions about AI’s role as a supportive tool rather than a substitute [S17] and the emphasis on human judgment in AI-augmented decision-making [S21].
MAJOR DISCUSSION POINT
AI as augmentation, not replacement, for diplomatic actors
Argument 4
Responsible deployment of AI in negotiations requires moving beyond large language models to incorporate diverse, domain‑specific methods.
EXPLANATION
Slavina calls for a broader view of AI tools, indicating that reliance solely on LLMs is insufficient for the nuanced demands of diplomatic processes.
EVIDENCE
She outlines the need to “think beyond just LLM” and stresses the need for responsible deployment of these tools [43].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Limitations of relying solely on LLMs and the call for broader, domain-specific AI approaches are noted in the analysis of AI’s role in negotiations [S2] and the broader AI-in-negotiation literature [S16].
MAJOR DISCUSSION POINT
Broadening AI approaches beyond LLMs for negotiation
Argument 5
Negotiations increasingly involve multiple actors beyond bilateral settings, requiring AI tools that can handle multi‑stakeholder dynamics.
EXPLANATION
Slavina notes that modern negotiations often include several states or groups, not just two parties, which adds layers of complexity that AI must be able to manage.
EVIDENCE
She describes scenarios where negotiators face seven counterparts from different countries and mentions that “rarely in this world do we have just two states negotiating nowadays” and that there can be a third state or dozens of members, such as the EU’s 27 members, highlighting the multi-party nature of negotiations [48-65].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The shift from bilateral to multi-party negotiations and the associated information-management challenges are described in the AI-transforming-diplomacy discussion [S2].
MAJOR DISCUSSION POINT
Multi‑party negotiation complexity
Argument 6
Supporting teams behind principal negotiators need AI assistance for evidence gathering and document management.
EXPLANATION
She points out that many departments and agencies provide evidence and documents to lead negotiators, suggesting AI could help these supporting teams handle large information flows.
EVIDENCE
Slavina explains that “there are a lot of teams that are sitting behind these principal negotiators, the different departments and agencies that are supporting them with evidence, with documents” [59-60].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The heavy documentation burden on supporting teams and the potential for AI-assisted evidence gathering are highlighted in the same source on diplomatic complexity [S2].
MAJOR DISCUSSION POINT
AI support for negotiation support staff
C
Charlie Posniak
8 arguments211 words per minute1310 words371 seconds
Argument 1
AI can help manage information overload, synthesize documents, generate strategic options, and support real‑time execution of negotiations.
EXPLANATION
Charlie proposes that AI can reduce the cognitive burden of negotiations by organizing large data sets, producing actionable insights, and assisting negotiators during live discussions.
EVIDENCE
He outlines a three-stage framework-research analysis, strategizing, and execution-supported by autonomous research agents, source validation, real-time transcription and translation, and strategy sandboxes, showing how AI can automate and augment each phase [103-115].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
AI’s capacity to accumulate, synthesize large information sets and support real-time transcription/translation in negotiations is discussed in the AI-diplomacy overview [S2].
MAJOR DISCUSSION POINT
AI‑enabled negotiation support
Argument 2
LLMs suffer from unverifiable fluency, opacity, and lack of accountability; they cannot replace established analytical frameworks such as game theory and decision analysis.
EXPLANATION
Charlie warns that large language models are not sufficiently transparent or accountable for high‑stakes diplomatic work and that traditional tools like game theory remain essential.
EVIDENCE
He points out that LLM fluency is not verifiable, their internal reasoning is opaque, and accountability is lacking for treaty-shaping recommendations [82-85]. He also references the 80-year-old toolkit of game theory, decision analysis, and machine learning that predates LLMs [86-88].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Critiques of LLMs’ opacity and the need to integrate them with traditional frameworks like game theory are presented in the same discussion [S2].
MAJOR DISCUSSION POINT
Limitations of LLMs
Argument 3
Effective diplomatic AI must integrate decades of methodological advances rather than depend only on chat‑bot style interactions.
EXPLANATION
Charlie argues that AI for diplomacy should build on a broad set of learning paradigms and the rich history of computational methods, not just on conversational agents.
EVIDENCE
He explains that LLMs sit at the intersection of supervised, unsupervised, and self-supervised learning, but relying solely on chatbots would ignore the extensive advances made over the past 80 years [89-93].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The argument for building on a long history of computational methods beyond chat-bots aligns with the broader AI-in-negotiation literature [S16].
MAJOR DISCUSSION POINT
Need for broader AI methods
Argument 4
Human authority must remain central; AI tools should be modular, transparent, and augment rather than replace decision‑makers.
EXPLANATION
Charlie stresses that ultimate decision‑making authority must stay with humans, and AI systems should be designed to be understandable and modular so that users can see how each component contributes to analysis.
EVIDENCE
He states that human authority is essential, tools must be modular and transparent, and augmentation should be appropriately scoped to the team and institution [117-120].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The principle that human judgment stays central and AI should be transparent and modular is emphasized in the human-centric AI guidelines [S21].
MAJOR DISCUSSION POINT
Human‑in‑the‑loop governance
Argument 5
Autonomous research agents, real‑time transcription/translation, strategy sandboxes, and red‑team simulations illustrate a roadmap for AI‑supported negotiations.
EXPLANATION
Charlie envisions a future where AI provides autonomous research, instant language services, and simulated environments for testing negotiation strategies, forming a comprehensive support ecosystem.
EVIDENCE
He describes autonomous research agents that generate validated sources, counterpart biographies, gap analyses, strategy sandboxes, red-team training, and real-time transcription/translation capabilities [108-115].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The roadmap featuring autonomous agents, transcription, translation and strategy sandboxes is outlined in the AI-transforming-diplomacy presentation [S2].
MAJOR DISCUSSION POINT
Future vision of AI‑augmented diplomacy
Argument 6
A robust computational infrastructure is essential to process large, unstructured data streams and enable real‑time analysis for AI‑supported negotiations.
EXPLANATION
Charlie argues that without strong data pipelines and processing capabilities, AI cannot effectively support the research, strategizing, and execution phases of diplomacy.
EVIDENCE
He notes that “you need this really strong computational infrastructure to be able to even begin to apply some of the really exciting and fascinating AI and ML methods” [107-108].
MAJOR DISCUSSION POINT
Infrastructure prerequisite for AI‑augmented diplomacy
Argument 7
Established analytical frameworks such as game theory and decision analysis should be integrated with AI tools to model strategic interactions in diplomacy.
EXPLANATION
Charlie argues that the 80‑year‑old toolkit of game theory, decision analysis, and machine learning remains essential and must be combined with modern AI capabilities for robust diplomatic analysis.
EVIDENCE
He references “We have game theory, decision analysis, a great range of theoretical developments that exist precisely to model strategic interactions under uncertainty” and notes that these tools are “80 years old” and should complement LLMs [86-88].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need to combine AI with game theory and decision analysis for strategic modeling is discussed in the same source [S2].
MAJOR DISCUSSION POINT
Integrating traditional strategic models with AI
Argument 8
AI systems must be designed to detect strategic misrepresentation and deception within negotiations.
EXPLANATION
Charlie highlights that negotiations involve environments where parties may lie or deceive, and AI tools need capabilities to identify such behavior to support trustworthy outcomes.
EVIDENCE
He describes “environments where there’s real strategic misrepresentation, where people are lying or deceiving or trying to shape outcomes for their own advantages” and stresses the need for AI to handle these challenges [98-99].
MAJOR DISCUSSION POINT
AI handling of deception in negotiations
G
Gabriela Ramos
7 arguments164 words per minute1455 words531 seconds
Argument 1
The UNESCO AI ethics negotiation illustrated how AI could have streamlined position‑tracking and stakeholder mapping.
EXPLANATION
Gabriela reflects on the UNESCO AI ethics recommendation process, noting that AI tools would have helped map country positions and manage massive public feedback more efficiently.
EVIDENCE
She recounts negotiating the UNESCO recommendation with 193 countries during COVID, using Zoom to see each country’s stance, and receiving 55,000 public comments that were later integrated with AI assistance [141-144].
MAJOR DISCUSSION POINT
UNESCO case study of AI in negotiation
Argument 2
Concerns about misrepresentation, cultural bias, and over‑reliance demand continuous questioning of AI assumptions and safeguards to keep humans in control.
EXPLANATION
Gabriela warns that AI tools can misrepresent cultures or over‑represent certain languages, and stresses the need for ongoing scrutiny to ensure human decision‑making remains primary.
EVIDENCE
She highlights risks of misrepresentation, over-representation of cultures, and the need to question AI assumptions, as well as the danger of becoming lazy and over-trusting AI without verification [352-358][363-367].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The importance of human oversight, bias mitigation and continuous scrutiny of AI outputs is highlighted in the human-centric AI perspective [S21] and the data-governance focus on cultural representation [S23].
MAJOR DISCUSSION POINT
Ethical governance and bias concerns
Argument 3
Multilingual, culturally diverse training data are required to avoid bias; UNESCO’s experience shows language and philosophical diversity must be reflected in models.
EXPLANATION
Gabriela stresses that AI models must incorporate many languages and cultural philosophies to prevent bias, citing UNESCO’s encounter with diverse linguistic perspectives during the AI ethics negotiation.
EVIDENCE
She explains that culture is expressed through language, cites the Namibian representative’s critique of an individualistic draft, references Ubuntu philosophy, and calls for models that represent multiple languages and sources to avoid discrimination [391-403].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need for multilingual, culturally diverse datasets to prevent bias is emphasized in the data-governance and cultural representation discussion [S23].
MAJOR DISCUSSION POINT
Cultural representation in AI models
Argument 4
Overreliance on AI can foster complacency and laziness among negotiators, reducing critical human judgment.
EXPLANATION
Gabriela warns that excessive trust in AI tools may lead negotiators to become passive, undermining the rigorous scrutiny required in high‑stakes diplomacy.
EVIDENCE
She remarks that “you can become very lazy” and stresses the need to keep humans in the driver’s seat, warning against letting AI dominate decision-making [365-367].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Warnings about AI-induced complacency and the necessity of retaining human judgment are echoed in the human-in-the-loop AI guidelines [S21].
MAJOR DISCUSSION POINT
Risk of negotiator complacency due to AI overtrust
Argument 5
AI tools should be designed to foster collaborative understanding rather than to be used merely as weapons to outmaneuver counterparts.
EXPLANATION
Gabriela argues that AI should open a space for human understanding in negotiations, not simply be built to beat the other side or maximise efficiency, and that negotiators must stay in the driver’s seat.
EVIDENCE
She says tools must not be just to “beat the person in front of me” or “maximize efficiency,” but to “open a space of human understanding” and keep humans in control of the process [354-357].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The view that AI should support collaborative understanding rather than be a competitive weapon aligns with the broader argument that AI enhances but does not replace human diplomacy [S17].
MAJOR DISCUSSION POINT
Ethical design of AI for diplomacy
Argument 6
AI could provide negotiators with a searchable repository of historical country positions and past negotiation outcomes to inform strategic preparation.
EXPLANATION
Gabriela notes that having a centralized database of how countries have traditionally positioned themselves would help diplomats anticipate moves and craft more effective arguments during negotiations.
EVIDENCE
She remarks that it would have been “amazing, just to have a repository of what is the traditional position of certain countries or certain negotiations” and that such a tool would aid in understanding both substantive and contextual stances of actors [144-145].
MAJOR DISCUSSION POINT
Historical data repositories for diplomatic strategy
Argument 7
Using AI to gather personal or strategic information about individual negotiators raises privacy and ethical concerns that must be addressed.
EXPLANATION
Gabriela warns that while AI could provide detailed insights into a counterpart’s motivations, such profiling risks violating privacy and requires careful ethical safeguards.
EVIDENCE
She asks “how can you have more information about that person, what moves them… but this is risky because it deals with privacy and all of those things” while discussing the use of AI to influence a negotiator [146-147].
MAJOR DISCUSSION POINT
Privacy risks of AI‑driven profiling
N
Nandita Balakrishnan
9 arguments214 words per minute1355 words379 seconds
Argument 1
AI outputs should be treated as data points that require human explanation and validation; they must not replace human‑crafted intelligence.
EXPLANATION
Nandita argues that AI‑generated assessments are only one input among many and must be accompanied by human justification, especially in high‑stakes policy contexts.
EVIDENCE
She states that AI outputs should be seen as data points, that policymakers need to explain how assessments were derived, and that humans must retain ultimate accountability for intelligence products [297-304].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The principle that AI-generated assessments are only one input and must be validated by humans is reinforced in the human-centric AI guidance [S21].
MAJOR DISCUSSION POINT
Human validation of AI outputs
Argument 2
Expanding AI literacy across intelligence, State Department, and other federal agencies, and demonstrating concrete use cases (e.g., geopolitical event prediction) are key to adoption.
EXPLANATION
Nandita emphasizes the need for AI training throughout the public sector and showcases predictive geopolitical event modeling as a tangible example to encourage uptake.
EVIDENCE
She notes that AI reshapes the threat landscape, calls for AI literacy in intelligence, State Department, and other agencies, and describes a project that used AI to predict geopolitical events for both military and diplomatic purposes [194-202].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Calls for AI literacy across diplomatic and intelligence communities and the importance of concrete use-cases are discussed in the diplomatic competencies for the AI era document [S20].
MAJOR DISCUSSION POINT
AI literacy and use‑case demonstration
Argument 3
Predictive and adaptive policy tools can reshape diplomatic analysis, but must be integrated into everyday workflows with clear human oversight.
EXPLANATION
Nandita points out that AI‑driven predictive tools can improve diplomatic forecasting, yet they must be embedded in routine processes and remain subject to human review.
EVIDENCE
She references the same geopolitical-event-prediction project as an example of predictive capability and reiterates that human-in-the-loop remains essential for accountability and validation [194-202][191-193].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need to embed predictive AI tools into routine diplomatic workflows while retaining human oversight is highlighted in the same competency framework [S20].
MAJOR DISCUSSION POINT
Predictive policy tools and human oversight
Argument 4
AI can surface historically overlooked data, enabling analysts to incorporate long‑term evidence that would otherwise be missed.
EXPLANATION
She illustrates how AI could have identified a ten‑year‑old data point that contradicted an intelligence assessment, highlighting AI’s role in uncovering hidden information.
EVIDENCE
She recounts a story where a mentor pointed out a missing piece of data from ten years ago that “completely negates your argument,” and suggests that an AI tool could have identified that gap [185-189].
MAJOR DISCUSSION POINT
AI as a tool for uncovering hidden historical evidence
Argument 5
AI can increase analyst efficiency by automating manual processes, allowing faster and smarter work.
EXPLANATION
Nandita reflects that without AI she had to perform assessments manually, but with AI tools she could work much faster and smarter, indicating significant productivity gains for analysts.
EVIDENCE
She asks herself, “if I had access to these tools as an analyst, how could I have worked much faster and much smarter?” highlighting the contrast between manual work and AI-assisted efficiency [178-180].
MAJOR DISCUSSION POINT
Productivity gains for intelligence analysts through AI
Argument 6
AI can surface counter‑arguments and alternative perspectives, helping analysts consider a broader evidence base and reduce personal bias.
EXPLANATION
She explains that AI tools can automatically generate possible counter‑points to an analyst’s assessment, prompting the analyst to evaluate and justify their conclusions more rigorously.
EVIDENCE
Nandita states that AI is “really helpful in helping you sort out the counterarguments, but you still need to understand how those counterarguments work” and that this process strengthens the analyst’s justification of their assessment [304-308].
MAJOR DISCUSSION POINT
AI‑assisted critical review of analytical reasoning
Argument 7
AI can facilitate cross‑agency collaboration by embedding tools into the daily workflows of intelligence, State Department, commerce, and other federal entities.
EXPLANATION
She argues that for AI to be effective in geopolitics, it must be integrated across multiple government sectors, enabling shared data and analytical capabilities.
EVIDENCE
Nandita points out the need for “day-to-day workflows integrating this technology” within the intelligence community, the State Department, commerce, OPM, and other federal employees, emphasizing a coordinated adoption across agencies [198-200].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Cross-agency integration of AI tools for coordinated decision-making is emphasized in the diplomatic competencies for the AI era [S20].
MAJOR DISCUSSION POINT
Cross‑sector AI integration in public‑sector decision‑making
Argument 8
Artificial intelligence has fundamentally altered the global threat landscape and must be treated as a foundational element in geopolitics and foreign‑policy analysis.
EXPLANATION
She argues that AI reshapes security, economic, and diplomatic competition, making it impossible to separate AI considerations from geopolitical assessments.
EVIDENCE
Nandita states “AI has fundamentally changed the threat landscape and the scope for global competition. It is now kind of the foundational way we need to think about geopolitics” [194-196].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The assertion that AI is now a core factor in geopolitics and security analysis is reflected in the discussion of AI’s impact on diplomatic competencies [S20].
MAJOR DISCUSSION POINT
AI as a core factor in geopolitics
Argument 9
AI capability constitutes a strategic asset that can provide competitive leverage to states, influencing power dynamics in international negotiations.
EXPLANATION
She emphasizes that nations with advanced AI tools gain advantages, making AI adoption a key factor in diplomatic and security competition.
EVIDENCE
She notes that “AI has fundamentally changed the threat landscape… you really cannot divorce AI and AI adoption when you’re trying to understand geopolitics” indicating its role as a source of competitive leverage [194-196].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The strategic advantage that AI provides to states and its effect on power balances is discussed in the analysis of AI’s impact on global power structures [S22].
MAJOR DISCUSSION POINT
AI as a strategic competitive advantage
R
Robyn Scott
13 arguments0 words per minute0 words1 seconds
Argument 1
Users must stay “above the algorithm,” preserving human agency and avoiding a zero‑sum dynamic where AI dominates the process.
EXPLANATION
Robyn introduces the metaphor of being “below” or “above” the algorithm, urging users to keep AI as a tool rather than allowing it to dictate outcomes, thereby protecting human agency.
EVIDENCE
She describes workers below the algorithm (e.g., Uber drivers) versus users above it who leverage tools for their goals, and warns against a zero-sum dynamic where AI takes over decision-making [256-259][254-255].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The recommendation that human users remain above algorithmic control to safeguard agency aligns with the human-in-the-loop AI principles [S21].
MAJOR DISCUSSION POINT
Maintaining human agency
Argument 2
Public‑sector officials are optimistic about AI but lack systematic evaluation; building AI literacy and robust pilot frameworks is essential.
EXPLANATION
Robyn reports that while public servants see huge AI potential, many lack structured evaluation of pilots, highlighting a gap that must be closed through training and assessment frameworks.
EVIDENCE
She cites a 5,000-person survey showing >90 % optimism, notes the paradox of optimism versus caution, points out that 70 % of leaders have pilots but only 45 % evaluate them, and highlights a skills gap where most implementers do not understand their own ethical frameworks [228-242][243-250].
MAJOR DISCUSSION POINT
AI optimism vs evaluation gap
Argument 3
Data poisoning, prompt injection, and over‑trust in AI outputs pose security and reliability threats; psychological safeguards and “battle‑mindset” against the algorithm are needed.
EXPLANATION
Robyn warns that AI systems can produce false negatives and that users may become over‑confident, advocating for a vigilant, skeptical stance toward algorithmic outputs.
EVIDENCE
She describes false-negative failures, the “sleeping at the wheel” phenomenon where users over-trust high-accuracy models, and proposes a heuristic of treating any algorithmic output as an adversary to maintain critical scrutiny [329-344].
MAJOR DISCUSSION POINT
Security and psychological safeguards
Argument 4
Effective AI policy work requires strong contextual expertise rather than merely content expertise, ensuring that AI tools are tailored to the specific policy environment.
EXPLANATION
Robyn emphasizes that her organization brings contextual knowledge to the table and then integrates content experts, arguing that understanding the policy context is essential for successful AI deployment.
EVIDENCE
She states that “we are more context experts than content experts, and we bring the content experts into the middle” indicating a strategic approach that prioritizes contextual insight over raw content expertise [224].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The emphasis on contextual expertise over pure content expertise for effective AI policy is highlighted in the diplomatic competencies for the AI era [S20].
MAJOR DISCUSSION POINT
Importance of contextual expertise in AI‑enabled policy
Argument 5
AI has the potential to unlock roughly $1.75 trillion of value in the public sector by automating bureaucratic, repeatable processes.
EXPLANATION
Citing a BCG estimate, Robyn argues that the scale of economic benefit from AI in government stems from its ability to handle routine, bureaucratic tasks efficiently.
EVIDENCE
She references a BCG figure that estimates “1.75 trillion of public sector value to be unlocked if we harness AI in the right way, because AI loves bureaucracy, all these repeatable processes” [231-233].
MAJOR DISCUSSION POINT
Economic upside of AI in public‑sector bureaucracy
Argument 6
Leaders must personally use AI tools to develop a concrete understanding, rather than relying on abstract knowledge.
EXPLANATION
Robyn argues that senior officials need hands‑on experience with AI to appreciate its speed and implications, which is essential for informed decision‑making.
EVIDENCE
She notes that “you cannot understand this technology in the abstract” and stresses that “you’ve got to use it” and “feel the speed of change” for leaders to grasp AI’s impact [242-247].
MAJOR DISCUSSION POINT
Need for leader‑level hands‑on AI experience
Argument 7
A large share of public officials lack knowledge of their own national AI ethical frameworks, creating a governance and skills gap.
EXPLANATION
Robyn points out that only a minority understand their country’s ethical guidelines, which risks uncoordinated and potentially unsafe AI deployments.
EVIDENCE
She reports that “Only 26 % of them say they understand their own country’s ethical frameworks, so approximately three-quarters are freestyling,” describing this as a “terrifying” skills gap [248-250].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The gap in understanding national AI ethical guidelines among public servants is noted in the human-centric AI governance discussion [S21].
MAJOR DISCUSSION POINT
Ethical‑framework knowledge gap among public servants
Argument 8
AI excels at research and writing tasks, providing productivity gains for public servants.
EXPLANATION
Robyn points out that AI is particularly strong at handling research and writing, which can help civil servants manage large volumes of documentation more efficiently.
EVIDENCE
She states that AI is “great at” research and writing tasks, highlighting its potential to automate repeatable bureaucratic processes and unlock value for public sector workers [231-236].
MAJOR DISCUSSION POINT
AI as a productivity tool for routine public‑sector work
Argument 9
AI can enable predictive and adaptive policy making, opening a “vitamin” prize of transformative possibilities beyond routine automation.
EXPLANATION
Robyn distinguishes between the immediate “painkiller” of automating bureaucracy and a longer‑term “vitamin” where AI supports predictive, responsive, and adaptive governance, limited only by imagination.
EVIDENCE
She explains that moving to the “vitamin prize” involves AI doing predictive policy making, responsive policy, and adaptive policy, which she describes as a space bounded only by imagination [237-239].
MAJOR DISCUSSION POINT
Future potential of AI for advanced policy functions
Argument 10
Partnering with academic research centres such as Stanford HAI strengthens AI policy work by providing deep contextual expertise and access to cutting‑edge research.
EXPLANATION
Robyn explains that her organization acts as a context expert and brings content experts into the process, and that collaboration with Stanford HAI exemplifies how academic partnerships can enrich the design and deployment of AI tools for policymakers.
EVIDENCE
She states that “Stanford HAI is one of our collaborators” and describes their role as “more context experts than content experts, and we bring the content experts into the middle,” highlighting the value of the partnership [223-226].
MAJOR DISCUSSION POINT
Importance of academic collaboration for AI‑enabled policy making
Argument 11
A multilingual, diplomatically‑oriented LLM initiative led by Swiss actors demonstrates a concrete step toward inclusive AI tools that can serve diverse linguistic and diplomatic contexts.
EXPLANATION
Robyn points to a Swiss‑driven project that is building a large language model trained on more than 100 languages and operated by a former Swiss diplomat, showing how language diversity and diplomatic relevance are being embedded into AI development.
EVIDENCE
She notes that “The Swiss have built sort of a quasi-Swiss government, quasi-multilateral initiative to build an LLM that is trained from the outset on more than 100 languages, and it is actually run by a friend of mine who’s a former Swiss diplomat” [424-425].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The Swiss-driven multilingual LLM project aimed at diplomatic applications is described in the data-governance and multilingual AI discussion [S23].
MAJOR DISCUSSION POINT
Multilingual AI models for diplomatic applications
Argument 12
High optimism about AI in the public sector is not matched by concrete implementation; there is a gap between AI talk and AI action that must be closed.
EXPLANATION
Robyn observes that while public servants express strong enthusiasm for AI’s potential, actual deployment of AI solutions remains limited, indicating a need to move from discussion to tangible projects and measurable outcomes.
EVIDENCE
She cites a 5,000-person survey showing over 90 % of public servants are optimistic about AI, yet she notes that “there’s lots of AI talk. There’s less AI action.” She also highlights the prevalence of pilots (70 % of leaders have them) but the low rate of systematic evaluation (only 45 % evaluate), illustrating the implementation gap [228-236][241-242].
MAJOR DISCUSSION POINT
Implementation gap between AI optimism and action
Argument 13
AI tools in education must be culturally and contextually relevant; otherwise they fail to improve student outcomes.
EXPLANATION
Robyn stresses that deploying AI technology in schools is insufficient on its own; the tools need to be aligned with local cultural norms and content relevance to be effective. Without such alignment, even sophisticated AI systems will not lead to better learning results.
EVIDENCE
She observes that countries which introduced large amounts of technology into their educational systems did not achieve better student outcomes because the content was not culturally or contextually appropriate, and that tools must be paired with locally relevant content to be useful [426-430].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The necessity for culturally and contextually appropriate AI tools in education is emphasized in the cultural relevance and digital divide analysis [S23].
MAJOR DISCUSSION POINT
Cultural and contextual relevance of AI in education
J
J. Michael McQuade
6 arguments182 words per minute2899 words954 seconds
Argument 1
The MOVE 37 project will develop tools, evaluation methodologies, and conduct stakeholder interviews to embed AI responsibly in diplomatic workflows.
EXPLANATION
Michael outlines the project’s technical roadmap, emphasizing tool development, evaluation strategies, and direct engagement with diplomats to ensure responsible AI integration.
EVIDENCE
He states that the project will create tools, evaluation methodologies, and leverage a large network of negotiation practitioners, while Slavina adds that one-on-one interviews with diplomats inform the design [124-129][267-278].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The description of MOVE 37’s objectives, including tool creation, evaluation and practitioner interviews, appears in the AI-transforming-diplomacy overview [S2].
MAJOR DISCUSSION POINT
MOVE 37 implementation plan
Argument 2
Establishing a new discipline for AI‑augmented negotiations, with ongoing collaboration among scholars, practitioners, and policymakers, is the overarching goal.
EXPLANATION
Michael frames the initiative as the foundation of a new field that brings together academia, policy, and practice to shape AI‑supported diplomacy over the long term.
EVIDENCE
He describes the vision of a discipline with clear start-middle-end phases, invites ongoing participation, and thanks collaborators, underscoring the intent to build a lasting community [124-129][436-440].
MAJOR DISCUSSION POINT
Creation of AI‑augmented diplomacy discipline
Argument 3
AI tools can reshape global power dynamics by providing competitive leverage to states that adopt them, potentially widening asymmetries in diplomatic negotiations.
EXPLANATION
Michael notes that the diffusion of AI capabilities will affect how power is distributed, with AI‑enabled states gaining strategic advantages in both cooperative and adversarial negotiations.
EVIDENCE
He states that “we think a lot about… where will tools provide competitive leverage… how it will change power structures” and that AI tools will be “dispersed actively and offensively” affecting negotiations [431-432].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The potential for AI to alter international power balances and give strategic advantage to AI-enabled states is discussed in the analysis of AI’s impact on power structures [S22].
MAJOR DISCUSSION POINT
AI’s impact on international power balance
Argument 4
Human judgment must filter AI‑generated recommendations in negotiations to maintain accountability.
EXPLANATION
Michael stresses that when AI proposes options, negotiators need to apply human judgment rather than accept AI outputs uncritically, preserving democratic and internal accountability.
EVIDENCE
He says, “When the AI says, here’s a thing? I don’t just trust it, it’s a priority. I have to apply a human judgment to what I’m hearing” indicating the need for human oversight of AI suggestions [371-374].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The requirement that human judgment oversee AI recommendations to preserve accountability aligns with the human-in-the-loop AI guidelines [S21].
MAJOR DISCUSSION POINT
Human‑in‑the‑loop governance for AI‑augmented diplomacy
Argument 5
AI can serve as a trusted information aggregator, helping negotiators identify leverage points and new pathways to successful outcomes.
EXPLANATION
Michael suggests that AI should be viewed as a reliable source that consolidates data and insights, enabling negotiators to discover strategic options they might otherwise miss.
EVIDENCE
He says “we are looking for collaborators, partners, and input… and the follow-up to negotiations can be a subject for the tools and applications of modern technology” and later notes “the trusted agent to accumulate information… new pathways for success” [210-212].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The role of AI as a trusted aggregator of information for strategic insight is highlighted in the AI-transforming-diplomacy discussion [S2].
MAJOR DISCUSSION POINT
AI as an information‑gathering aid
Argument 6
Establishing baseline capabilities and clear ground rules for AI use is essential to ensure responsible deployment in diplomatic contexts.
EXPLANATION
Michael stresses the need for predefined standards and capabilities so that AI tools augment negotiations without undermining accountability or creating ambiguity.
EVIDENCE
He remarks that “we have to have a baseline of capability” and that “the question is how can modern tools help in that process without removing or absolving responsibility for people” indicating the importance of ground rules [208-212].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The call for baseline capabilities and governance rules for AI in diplomacy is reflected in the human-centric AI governance recommendations [S21].
MAJOR DISCUSSION POINT
Baseline capabilities and governance for AI in diplomacy
A
Audience
1 argument158 words per minute399 words151 seconds
Argument 1
Audience concerns highlighted the need for transparent source attribution and safeguards against malicious manipulation of training data.
EXPLANATION
An audience member raised questions about ensuring cultural diversity in datasets and protecting AI models from data poisoning and prompt‑injection attacks.
EVIDENCE
The participant asked how to embed diverse cultural inputs in models and how to guard against data poisoning and prompt injection in AI-mediated negotiations [386-390]. Gabriela responded by emphasizing language representation, source transparency, and testing safeguards [391-403].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Calls for transparent data provenance, cultural diversity in datasets, and protection against data poisoning are discussed in the data-governance and cultural representation literature [S23] and reinforced by human-centric AI safeguards [S21].
MAJOR DISCUSSION POINT
Audience demand for data governance and cultural representation
Agreements
Agreement Points
AI should augment, not replace, diplomats; human authority and judgment must remain central.
Speakers: Slavina Ancheva, Charlie Posniak, Gabriela Ramos, J. Michael McQuade, Robyn Scott, Nandita Balakrishnan
AI is not replacing diplomats but giving them better tools to manage complexities [61-62] Human authority must remain central; tools should be modular and transparent [117-120] Need to keep humans in the loop and question AI assumptions to avoid misrepresentation and over-reliance [352-358][363-367] Human judgment must filter AI-generated recommendations to preserve accountability [371-374] Users must stay “above the algorithm” to preserve human agency [256-259] AI outputs are data points that require human explanation and validation [297-304]
All speakers emphasized that AI is a supportive tool while final decision-making must stay with humans, requiring transparency, questioning, and human judgment throughout the negotiation process [61-62][117-120][352-358][371-374][256-259][297-304].
POLICY CONTEXT (KNOWLEDGE BASE)
Both the AI diplomacy discourse and scholarly commentary stress that AI is a preparatory tool while final decisions must rest with human diplomats, preserving the art of negotiation and trust-building [S38][S39][S40].
Building AI literacy and capacity across the public sector is essential for effective adoption.
Speakers: Nandita Balakrishnan, Robyn Scott
Expanding AI literacy across intelligence, State Department and other federal agencies is key, with concrete use-cases like geopolitical event prediction [194-200] Leaders must personally use AI tools; there is a large skills gap and many pilots lack systematic evaluation [242-247][228-242]
Both speakers highlighted the need for widespread AI training and hands-on experience among officials, noting current gaps in skills and evaluation of pilots [194-200][242-247][228-242].
POLICY CONTEXT (KNOWLEDGE BASE)
Capacity-building was highlighted as a key challenge for Francophone states in digital-AI regulation forums [S42] and reiterated in UN-AI infrastructure discussions emphasizing skills and standards development [S58]; parliamentary insights further call for continuous education of all stakeholders [S59].
Multilingual and culturally diverse training data are required to avoid bias and ensure inclusive AI tools.
Speakers: Gabriela Ramos, Robyn Scott
Multilingual, culturally diverse training data are required to avoid bias; language expresses culture and must be represented in models [391-403] Swiss initiative building an LLM trained on >100 languages for diplomatic use demonstrates a concrete step toward inclusive AI [424-425]
Both speakers stressed that AI systems must incorporate many languages and cultural perspectives to prevent discrimination and reflect global diversity [391-403][424-425].
POLICY CONTEXT (KNOWLEDGE BASE)
Ensuring linguistic diversity aligns with SDG-4 and SDG-10 recommendations and broader AI content-policy goals for multilingualism and cultural inclusion [S43][S44]; gender-bias research also underscores the need for diverse datasets to mitigate discriminatory outcomes [S60][S61].
AI can alleviate information overload in negotiations by synthesizing documents, generating strategic options, and supporting real‑time execution.
Speakers: Charlie Posniak, Slavina Ancheva, J. Michael McQuade, Nandita Balakrishnan, Robyn Scott
AI can manage information overload, synthesize documents, generate strategic options, and support real-time execution [103-115] Negotiations involve massive documentation, strategic dynamics and time pressure, creating heavy cognitive load that AI could help mitigate [66-72] AI is a trusted agent to accumulate information and reveal new pathways for success [210-212] AI can surface historically overlooked data, improving analyst efficiency [185-189] AI excels at research and writing tasks, offering productivity gains for public servants [231-236]
All speakers recognized AI’s potential to handle large data volumes, produce analyses, and aid negotiators during live discussions, thereby reducing cognitive burden and improving outcomes [103-115][66-72][210-212][185-189][231-236].
POLICY CONTEXT (KNOWLEDGE BASE)
Case studies of AI-enabled diplomacy demonstrate its role in aggregating large document sets, modeling negotiation scenarios, and offering actionable options, thereby reducing cognitive load for negotiators [S48][S49][S50].
Responsible AI deployment requires transparent, modular tools, baseline capabilities, and systematic evaluation.
Speakers: Charlie Posniak, J. Michael McQuade, Robyn Scott, Gabriela Ramos
Tools must be modular, transparent and augment rather than replace decision-makers [117-120] Establishing baseline capabilities and clear ground rules is essential for responsible deployment [208-212] There is a gap between AI optimism and actual evaluation of pilots; systematic assessment is needed [228-242] Continuous questioning of AI assumptions is necessary to keep humans in control [352-358]
Speakers agreed on the need for clear standards, modular transparency, and rigorous evaluation to ensure AI is used responsibly in diplomatic contexts [117-120][208-212][228-242][352-358].
POLICY CONTEXT (KNOWLEDGE BASE)
The UN Security Council stresses algorithmic transparency and rigorous testing as prerequisites for trustworthy AI systems [S57]; responsible-deployment frameworks and emerging standards discussed at IGF and EU workshops further call for modular, benchmarked tools and systematic assessment [S55][S56][S53][S54].
AI constitutes a strategic asset that can reshape global power dynamics and provide competitive leverage to states.
Speakers: J. Michael McQuade, Nandita Balakrishnan
AI will change power structures and provide competitive leverage; tools will be dispersed actively and offensively [431-432] AI has fundamentally changed the threat landscape and is a strategic competitive advantage in geopolitics [194-196]
Both highlighted that AI adoption will affect international power balances, giving advantage to states that effectively integrate AI into diplomatic and security strategies [431-432][194-196].
POLICY CONTEXT (KNOWLEDGE BASE)
Analyses of AI’s geopolitical impact note its capacity to shift economic and societal power, creating new centres of influence and serving as a strategic lever for national advantage [S45][S47].
Similar Viewpoints
Both stress that without widespread AI competence and hands‑on experience among officials, AI initiatives will remain ineffective and untested [194-200][242-247][228-242].
Speakers: Nandita Balakrishnan, Robyn Scott
Expanding AI literacy across agencies and need for leaders to use AI personally [194-200][242-247] Large skills gap and lack of systematic evaluation of AI pilots [228-242]
Both call for clear, transparent standards and baseline capabilities to guide AI integration in diplomacy [117-120][208-212].
Speakers: Charlie Posniak, J. Michael McQuade
Need for modular, transparent tools and baseline capabilities for responsible AI use [117-120][208-212]
Both agree that linguistic and cultural diversity must be embedded in AI models to ensure fairness and relevance in diplomatic contexts [391-403][424-425].
Speakers: Gabriela Ramos, Robyn Scott
Importance of multilingual, culturally diverse data to avoid bias [391-403] Swiss multilingual LLM initiative as a concrete example [424-425]
Both recognize the heavy information burden of negotiations and see AI as a means to reduce cognitive load and improve decision‑making [66-72][103-115].
Speakers: Slavina Ancheva, Charlie Posniak
Negotiations are complex and generate massive documentation; AI can help manage this complexity [66-72] AI can manage information overload and support real-time execution [103-115]
Unexpected Consensus
Cultural and linguistic inclusivity in AI models.
Speakers: Gabriela Ramos, Robyn Scott
AI must represent many languages and cultural philosophies to avoid bias [391-403] Swiss project building a multilingual LLM for diplomatic use [424-425]
Despite coming from different professional backgrounds (UN diplomacy vs. policy-implementation consultancy), both converged on the necessity of multilingual, culturally aware AI, highlighting a cross-sector recognition of this issue that was not anticipated earlier in the discussion [391-403][424-425].
POLICY CONTEXT (KNOWLEDGE BASE)
Policy documents on AI content and multilingualism call for inclusive model design to serve diverse linguistic communities and reduce inequities [S43][S44].
Treating AI outputs as supplemental data points rather than definitive intelligence.
Speakers: Nandita Balakrishnan, Charlie Posniak
AI outputs should be seen as data points requiring human validation [297-304] LLMs have unverifiable fluency and opacity; they cannot replace established analytical frameworks [82-85]
Both emphasized the limitation of LLMs and the need for human oversight, aligning on the view that AI should augment rather than replace traditional analytical methods-a convergence not explicitly stated before their individual remarks [297-304][82-85].
Overall Assessment

The panel displayed strong consensus on keeping humans central to AI‑augmented diplomacy, the necessity of building AI capacity and literacy, the importance of multilingual and culturally diverse data, and the need for transparent, evaluated, and modular AI tools. There was also agreement that AI will become a strategic asset influencing power dynamics.

High consensus across most themes, indicating broad alignment among scholars, practitioners, and policymakers on the principles governing AI use in diplomatic negotiations. This convergence suggests that future initiatives like MOVE 37 are likely to adopt human‑in‑the‑loop designs, prioritize capacity building, and address cultural inclusivity, while also preparing for the strategic implications of AI on international power structures.

Differences
Different Viewpoints
Acceptability of opacity in AI models for diplomatic use
Speakers: Charlie Posniak, Robyn Scott
LLM fluency is not verifiable, opacity hampers accountability for high-stakes negotiations [82-85] Full legibility over model workings is not possible, but the lack of transparency is not insurmountable [323-328]
Charlie argues that the opacity of large language models makes them unsuitable for treaty‑shaping work because their reasoning cannot be verified or held accountable. Robyn acknowledges the black‑box nature but contends that it does not preclude use, suggesting work‑arounds are feasible. The two therefore disagree on how much opacity can be tolerated in diplomatic AI tools.
POLICY CONTEXT (KNOWLEDGE BASE)
UN discussions flag algorithmic opacity as a risk, emphasizing the need for transparency and verifiable evaluation in diplomatic AI applications [S57]; responsible-deployment dialogues similarly caution against opaque systems [S55].
Degree of trust to place in AI‑generated assessments
Speakers: J. Michael McQuade, Nandita Balakrishnan
AI can be a trusted information aggregator that offers new levers and pathways for negotiators [210-212][371-374] AI outputs should be treated only as data points that require human explanation and validation; they must not replace human-crafted intelligence [297-304]
Michael envisions AI as a reliable source that negotiators can lean on for strategic insight, whereas Nandita stresses that AI outputs must always be justified and overseen by humans, limiting the level of trust placed in them. Both see value in AI but diverge on how much autonomous reliance is acceptable.
POLICY CONTEXT (KNOWLEDGE BASE)
World Economic Forum and AI-trust frameworks argue that trust must be grounded in context-specific risk assessments and supported by standards, laws, and conformity mechanisms [S51][S52].
Framing AI as a collaborative aid versus a competitive lever
Speakers: Gabriela Ramos, Robyn Scott
AI tools should not be built to ‘beat’ counterparts; they must open space for human understanding and keep humans in control [354-357][363-367] There is a risk of a zero-sum dynamic where AI dominates the process; a ‘battle-mindset’ against the algorithm is needed to avoid complacency [254-259][424-425]
Gabriela stresses that AI should foster collaborative understanding and avoid being used as a weapon against other negotiators. Robyn, while also warning against over‑reliance, frames AI more as a competitive technology that could create a zero‑sum environment, advocating for a defensive stance toward algorithms. Their goals of responsible AI overlap, but their conceptual framing diverges.
POLICY CONTEXT (KNOWLEDGE BASE)
Strategic analyses portray AI as a competitive asset reshaping power balances [S45], while governance sessions advocate for collaborative, responsible use of AI agents to support-not replace-human decision-making [S55].
Unexpected Differences
Readiness to codify AI governance standards versus practical evaluation gaps
Speakers: J. Michael McQuade, Robyn Scott
We need to set baseline capabilities and ground rules for AI-augmented negotiations now [208-212] A large share of pilots are unevaluated, and most officials lack understanding of ethical frameworks, indicating that solid standards are premature [241-242][243-250]
It is surprising that Michael, leading the MOVE 37 initiative, pushes for immediate baseline rules, while Robyn, who works closely with public‑sector pilots, highlights the lack of evaluation and ethical‑framework awareness, suggesting the field is not yet ready for firm standards.
POLICY CONTEXT (KNOWLEDGE BASE)
IGF and EU policy roadmaps highlight a tension between the push to formalise AI standards and the reality of limited evaluation data, calling for evidence-based benchmarks and iterative testing before full codification [S53][S54][S55][S56][S58][S59].
Overall Assessment

The panel largely concurs that AI can augment diplomatic work, but key tensions arise around model transparency, the level of trust to place in AI outputs, and whether AI should be framed as a collaborative facilitator or a competitive tool. These disagreements reflect differing risk tolerances and disciplinary backgrounds (technical vs policy).

Moderate – while there is broad consensus on the need for human‑in‑the‑loop and responsible deployment, the speakers diverge on practical governance (opacity, trust, framing). The disagreements suggest that any MOVE 37 implementation will need flexible guidelines that accommodate both high‑accountability requirements and the pragmatic constraints of existing AI technology.

Partial Agreements
Both agree that AI should support negotiators with richer contextual and strategic information, but Gabriela emphasizes a searchable historical database while Charlie stresses formal analytical models; they differ on the primary type of tool to prioritize.
Speakers: Gabriela Ramos, Charlie Posniak
A repository of historical country positions would aid negotiators in preparation [144-145] Integrating game-theoretic and decision-analysis frameworks with AI is essential for strategic modelling [86-88][98-100]
Both want responsible AI deployment, but Michael assumes a baseline can be defined now, whereas Robyn points out that the current pilot landscape is too unevenly evaluated to support firm baselines.
Speakers: J. Michael McQuade, Robyn Scott
Establish baseline capabilities and clear ground rules for AI use in diplomacy [208-212] Many AI pilots lack systematic evaluation, creating a gap that must be closed before standards can be set [241-242][243-250]
Takeaways
Key takeaways
Diplomatic negotiations are highly complex, involving many parties, massive documentation, strategic dynamics, and time pressure, creating a heavy cognitive load. AI can alleviate information overload, synthesize documents, generate strategic options, provide real‑time transcription/translation, and support execution, but must be integrated with existing analytical frameworks (game theory, decision analysis). Large language models alone are insufficient due to unverifiable fluency, opacity, and lack of accountability; a broader suite of AI methods and decades‑old techniques is required. Human authority and human‑in‑the‑loop must remain central; AI tools should be modular, transparent, and used to augment—not replace—decision‑makers. Maintaining human agency (“above the algorithm”) is essential to avoid a zero‑sum dynamic where AI dominates the process. Significant gaps exist in AI literacy, systematic pilot evaluation, and ethical governance within the public sector; training and clear evaluation frameworks are needed. Risks of cultural bias, data poisoning, prompt injection, and over‑reliance on AI outputs were highlighted; multilingual, culturally diverse training data and robust safeguards are required. The MOVE 37 project will develop tools, evaluation methodologies, stakeholder interviews, autonomous research agents, strategy sandboxes, and red‑team simulations to embed AI responsibly in diplomatic workflows. Collaboration among scholars, practitioners, and policymakers is crucial to establish a new discipline of AI‑augmented diplomacy.
Resolutions and action items
Commit to keeping human authority central and ensuring AI tools are modular and transparent (as stated by Charlie Posniak). Develop a suite of AI tools for research, analysis, strategy, and execution phases of negotiations (MOVE 37 project). Create and deploy evaluation methodologies for AI pilots in the public sector (Robyn Scott’s observation). Conduct one‑on‑one interviews with current and former diplomats to inform tool design (Slavina Ancheva). Build AI literacy programs across intelligence, State Department, and other federal agencies (Nandita Balakrishnan). Assemble a multilingual, culturally representative training dataset for negotiation‑support models (Gabriela Ramos). Establish red‑team and sandbox environments for testing AI‑augmented negotiation strategies (Charlie Posniak). Invite interested stakeholders to join the project and provide feedback (Michael’s closing invitation).
Unresolved issues
How to systematically ensure that AI models capture the full cultural and linguistic diversity of all negotiating parties. Concrete mechanisms to protect against data poisoning, prompt injection, and other adversarial attacks on negotiation‑support AI. Standardized procedures for evaluating AI pilots and translating pilot results into policy adoption. Balancing power asymmetries when some states have far greater data access and AI capabilities than others. Defining clear criteria for when AI should be used for strategic option generation versus when human judgment must dominate. Methods for transparent source attribution in AI‑generated analyses to satisfy accountability requirements.
Suggested compromises
Use AI as a decision‑support tool while retaining final authority and accountability with human negotiators. Maintain modular AI components that can be inspected and turned off, allowing negotiators to stay “above the algorithm.” Limit AI deployment to specific, well‑defined tasks (e.g., document synthesis, translation) while keeping higher‑level strategic choices human‑driven. Implement pilot projects with built‑in evaluation and human‑oversight checkpoints before wider rollout. Adopt a “battle‑mindset” toward AI: continuously question outputs and verify against independent sources rather than accepting them unquestioningly.
Thought Provoking Comments
Why can’t you just ask an LLM? … Their fluency isn’t necessarily verifiable in international politics, the opacity makes accountability hard, and relying only on chatbots would miss out on 80 years of game theory, decision analysis, and other technical developments.
Challenges the simplistic notion that large language models alone can handle diplomatic negotiations, highlighting issues of verification, transparency, and the rich existing methodological toolbox.
Shifted the discussion from enthusiasm about AI tools to a more cautious, nuanced view, prompting subsequent speakers (e.g., Robyn and Nandita) to address trust, evaluation, and the need for human oversight.
Speaker: Charlie Posniak
The heuristic of being ‘below the algorithm’ (e.g., an Uber driver dispatched by a system) versus ‘above the algorithm’ (using tools to further your goals). We must move people up above the algorithm in diplomacy.
Introduces a clear, relatable framework for thinking about human agency versus algorithmic control, emphasizing empowerment rather than replacement.
Reoriented the conversation toward preserving and enhancing human decision‑making power, leading to deeper discussion about AI literacy, pilot evaluation, and the risk of “drunk on AI” attitudes.
Speaker: Robyn Scott
When I wrote my first intelligence piece, a mentor pointed out a ten‑year‑old data point that completely negated my argument. An AI tool that could surface such hidden evidence and synthesize it would have made the analysis faster and smarter.
Provides a concrete, personal illustration of AI’s potential to overcome human cognitive limits and data‑access bottlenecks in intelligence work.
Illustrated the practical value of AI for analysts, reinforcing the earlier point about augmenting—not replacing—human judgment and prompting calls for explainability and transparency.
Speaker: Nandita Balakrishnan
We received 55,000 public comments on the UNESCO AI ethics recommendation and used AI to integrate them. I wish we had AI to map the traditional positions of countries and give me a repository of their negotiation histories.
Shows a real‑world application of AI in a massive, multilingual policy process and raises the strategic question of how AI could support position‑tracking while flagging privacy and bias concerns.
Moved the discussion from abstract theory to a tangible use‑case, sparking follow‑up questions about cultural representation, data poisoning, and the ethical limits of profiling negotiators.
Speaker: Gabriela Ramos
Culture is expressed by language. To capture philosophies like Ubuntu, we must ensure models are trained on many languages and sources, otherwise they will reflect individual‑welfare bias and miss collective worldviews.
Links linguistic diversity directly to cultural fairness in AI models, highlighting a subtle but critical source of bias in diplomatic AI tools.
Expanded the conversation to include multilingual fairness and the need for diverse training data, leading to mentions of the Swiss multilingual LLM project and reinforcing the earlier point about representation.
Speaker: Gabriela Ramos (response to audience question)
We think a lot about how AI tools could provide competitive leverage and whether they should be dispersed offensively, not just defensively, because they will reshape power structures in both cooperative and adversarial negotiations.
Frames AI deployment as a geopolitical issue, moving the dialogue from technical feasibility to strategic implications for global balance of power.
Served as a turning point that broadened the scope of the panel, prompting participants to consider equity, access, and the macro‑level consequences of AI‑augmented diplomacy.
Speaker: J. Michael McQuade
Overall Assessment

The discussion was steered by a series of pivotal remarks that moved it from a high‑level enthusiasm about AI to a layered, critical examination of its role in diplomacy. Charlie’s warning against over‑reliance on LLMs introduced the need for verification and accountability, which set the stage for Robyn’s agency heuristic and Michael’s geopolitical framing. Personal anecdotes from Gabriela and Nandita grounded the debate in real‑world practice, highlighting both the promise (handling massive comment volumes, surfacing hidden data) and the perils (privacy, cultural bias, power imbalances). Audience questions about cultural representation and data poisoning, answered by Gabriela, further deepened the conversation around fairness and security. Collectively, these comments redirected the panel toward a balanced view that AI should augment, not replace, human negotiators, that transparency and multilingual inclusivity are essential, and that the strategic distribution of AI capabilities will shape future power dynamics.

Follow-up Questions
How can we ensure that the diverse cultural inputs of the world’s most diverse countries are embedded in the data sets and models that inform negotiations?
Cultural representation is essential for fairness, legitimacy, and effectiveness of AI‑mediated diplomatic processes.
Speaker: Sam Dawes (Audience)
What safeguards are needed to address data poisoning, prompt injection, and other adversarial risks when AI is used as a neutral mediator or negotiation assistant?
Ensuring the integrity and security of AI tools is critical to maintain trust among parties and prevent manipulation of outcomes.
Speaker: Sam Dawes (Audience)
What concrete steps can practitioners like Devika Rao take to advance AI‑supported cultural‑education frameworks and foster co‑creation collaborations?
Guidance is needed for individuals working at the intersection of culture, education, and AI to translate ideas into actionable projects and partnerships.
Speaker: Devika Rao (Audience)
How will AI impact the balance of power given unequal access to data sets and AI capabilities among states, especially in multilateral negotiations?
Understanding power asymmetries is vital to design equitable AI governance mechanisms and avoid exacerbating geopolitical tensions.
Speaker: Arman (Audience)
Develop a comprehensive, searchable repository of historical country positions and negotiation precedents to support AI‑driven diplomatic tools.
Such a knowledge base would enable negotiators to quickly assess counterpart stances, reduce cognitive load, and improve strategic preparation.
Speaker: Gabriela Ramos; Slavina Ancheva
Research effective AI‑literacy and training programmes across all public‑sector agencies (beyond defence and intelligence) to enable informed AI adoption.
Broad AI competence is required for policymakers to evaluate, trust, and responsibly integrate AI into diplomatic workflows.
Speaker: Nandita Balakrishnan
Design robust evaluation frameworks for AI pilots in the public sector to measure impact, identify failures, and guide scaling decisions.
Many pilots lack systematic assessment, limiting learning and the ability to demonstrate value or address shortcomings.
Speaker: Robyn Scott
Create multilingual, culturally inclusive large language models (e.g., the Swiss initiative) tailored for diplomatic contexts.
Multilingual models ensure that non‑English perspectives are accurately captured, reducing bias and enhancing global relevance.
Speaker: Robyn Scott
Investigate methods to keep AI tools modular, transparent, and accountable, preserving human authority throughout the negotiation process.
Transparency and explainability are prerequisites for trust, legal compliance, and ethical deployment in high‑stakes diplomacy.
Speaker: Charlie Posniak
Explore AI techniques for generating strategic options, counter‑arguments, and preference mapping to aid negotiators in complex, multi‑party settings.
These capabilities can help negotiators navigate information overload, avoid groupthink, and identify creative solutions.
Speaker: Charlie Posniak; Gabriela Ramos
Study the psychological effects of over‑reliance on AI (e.g., “sleeping at the wheel,” false‑negative bias) and develop safeguards to maintain critical human oversight.
Preventing complacency ensures that AI augments rather than replaces human judgment, preserving decision quality.
Speaker: Robyn Scott
Assess the feasibility and implications of using AI to predict geopolitical events for proactive policy‑making and diplomatic strategy.
Predictive analytics could give governments early warning and strategic advantage, but requires validation and ethical oversight.
Speaker: Nandita Balakrishnan
Determine appropriate calibration levels of AI assistance at different negotiation stages (pre‑negotiation preparation vs. real‑time execution).
Different phases demand distinct types of AI support; calibrating these helps align tool output with negotiators’ needs and maintains control.
Speaker: J. Michael McQuade (prompting Gabriela)

Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.