AI Safety at the Global Level Insights from Digital Ministers Of

20 Feb 2026 10:00h - 11:00h

AI Safety at the Global Level Insights from Digital Ministers Of

Session at a glanceSummary, keypoints, and speakers overview

Summary

The panel opened with Yoshua Bengio stressing the importance of independent scientific assessments of AI capabilities and risks, acknowledging deep uncertainty and pledging continued support for such reporting [1-5]. Lee Tiedrich introduced a multidisciplinary panel-including Singapore’s Minister Josephine Teo, Professor Alondra Nelson, and AI Security Institute director Adam Beaumont-to examine the safety report and asked about notable changes between 2025 and 2026 [7-19].


Bengio identified the rapid rise of autonomous AI agents that can operate for extended periods, access credentials and the internet, and interact with each other as a key emerging risk that reduces human oversight [20-31]. Minister Teo framed Singapore’s position as a small state that must balance adoption with safety, citing recent legislation that obliges platforms to remove harmful AI-generated images and highlighting AI’s dual role as both a cyber-threat and a target of attacks [36-70]. Nelson explained that the report aims to provide a “ground truth” by aggregating evidence-informed scenarios, deliberately avoiding prescriptive policy while exposing systemic risks such as loss of autonomy and social cohesion, and calling for stronger political will to act on the findings [78-103].


Beaumont emphasized the urgency of cybersecurity concerns, describing the institute’s pre- and post-deployment testing, red-team exercises, and the open-source Inspect framework as tools to raise evaluation standards and encourage a broader ecosystem of independent auditors [107-127][214-215]. Addressing the jagged performance of general-purpose models, Bengio argued for per-capability risk assessment and rigorous scientific communication to prevent false claims that could mislead policymakers [132-150]. Teo suggested that policymakers should create mandatory safety standards akin to product testing (e.g., IKEA’s durability checks) and explore mechanisms such as insurance schemes and pragmatic research roadmaps to bridge the gap between science and deployment [160-186].


Both Bengio and Nelson noted a missing “policy-options” layer between scientific evidence and legislative action, proposing that scientists outline possible courses without dictating choices to aid decision-makers [244-250]. Lee highlighted the need for greater AI literacy and for translating evaluation metrics across cultures and norms, drawing on Singapore’s governance experience [32-35]. He also raised the evidence gap in rapidly evolving risks, to which Beaumont replied that clear, transparent evaluation criteria and the development of third-party auditors are essential to fill that gap [211-215]. Nelson added that adopting standards similar to those used in human-genetics risk assessment and funding common-good research could strengthen the evaluation ecosystem [252-266].


The discussion concluded with consensus that multi-sector collaboration, standard-setting, and international agreements are essential to manage AI’s systemic and catastrophic risks, especially as sovereignty concerns demand cooperative safety frameworks rather than isolationist walls [292-303][304-311].


Keypoints


Major discussion points


Rapid emergence of autonomous AI agents and multi-agent systems creates new safety challenges.


Yoshua highlighted that “advances in agency of the AI systems” mean “more autonomous… less oversight” and that agents can be given credentials and Internet access, leading to “concerning” interactions when they are let out into the world [20-31]. Josephine echoed this, noting the need to “think about how these AI agent systems are being architected” and to put “guardrails around it” [66-70].


Translating scientific findings into concrete, thoughtful policy guardrails.


Josephine stressed that policymakers must turn safety insights into “operable guardrails,” often via standards, regulations, and laws, but “thoughtfully” to avoid stifling innovation [50-55]. She cited Singapore’s new law imposing obligations on services that host harmful AI-generated images [58-64] and discussed the exploration of “insurance schemes” and industry-wide standards as mechanisms to align incentives [168-173]. Yoshua added that scientific rigor and humility are essential so that “policy decisions… are not based on false claims” [138-150].


Building an independent evaluation ecosystem to close the evidence gap.


The discussion began with a call for “independent scientific assessment” of AI risks [1-5]. Alondra explained that the report deliberately stays at “what do we know” and is moving toward “real-time” evidence [88-94]. Adam described concrete steps: the open-source Inspect framework, extensive red-team testing, and the goal of fostering a “wide ecosystem of third-party evaluators” akin to accounting auditors [214-225]. Lee’s follow-up question framed this as a need to define “step one… step two… and by whom” [221-225].


Broadening the risk lens beyond catastrophic scenarios and emphasizing global cooperation.


Alondra argued for attention to “systemic risk,” including loss of autonomy, manipulation, job anxiety, and social cohesion, noting that these compounding harms threaten democracy [191-199]. Adam highlighted the dual-use nature of AI in cybersecurity and bio-security, stressing the urgency of understanding “the confluence of some of these risks” [119-124]. Yoshua and Josephine both stressed that AI risks transcend borders, calling for international agreements, shared standards, and collaborative governance rather than “walls” of sovereignty [292-302][227-233].


Overall purpose / goal of the discussion


The panel convened to unpack the findings of the newly released AI safety report, assess emerging technical risks, and explore how scientific evidence can be turned into practical, policy-ready tools and regulations. Participants aimed to identify research and evaluation gaps, propose mechanisms for trustworthy deployment, and chart a collaborative path forward for governments, industry, and the research community.


Overall tone and its evolution


The conversation maintained a collaborative and constructive tone throughout, marked by mutual respect among academics, policymakers, and industry leaders. It began with a forward-looking, cautious optimism about scientific assessment [1-5], moved into urgent concern over autonomous agents and systemic risks [20-31][191-199], shifted to pragmatic problem-solving regarding policy translation and evaluation tools [50-64][214-225], and concluded with a unifying call for international cooperation and responsible governance [292-302][227-233]. While the tone remained respectful, the urgency and emphasis on concrete action grew as the discussion progressed.


Speakers

Alondra Nelson – Area of Expertise: Science, technology, and social values; AI policy and ethics.


Role/Title: Harold F. Linder Chair and Professor at the Institute for Advanced Study, where she leads the Science, Technology, and Social Values Lab; former Deputy Director of the White House Office of Science and Technology; senior advisor on the AI safety report. [S1][S3]


Josephine Teo – Area of Expertise: Digital development, AI governance, cybersecurity.


Role/Title: Singapore Minister (Minister for Communications and Information) leading Singapore’s digital development, smart-nation strategy, public communications, and cybersecurity initiatives. [S4]


Yoshua Bengio – Area of Expertise: Deep learning, AI research, AI safety.


Role/Title: Professor, University of Montreal; AI pioneer and Turing Award laureate; co-chair of the AI safety report. [S9]


Adam Beaumont – Area of Expertise: AI security, risk assessment, red-team testing.


Role/Title: Director, UK AI Security Institute (AI Security Institute). [S13]


Lee Tiedrich – Area of Expertise: AI policy, governance, evaluation frameworks.


Role/Title: Researcher at the University of Maryland; moderator/facilitator of the panel discussion. [S15][S16]


Participant – Area of Expertise: (not specified)


Role/Title: Audience member who asked questions during the Q&A.


Additional speakers:


Melinda – Area of Expertise: (not specified)


Role/Title: Unidentified participant addressed during the Q&A segment.


Full session reportComprehensive analysis and detailed insights

The session opened with Yoshua Bengio emphasizing that policymakers worldwide need independent scientific assessments of AI capabilities, harms, and existing mitigation measures, and warning that the future will contain “unknown unknowns” such as psychological effects that must be prepared for on the basis of the best scientific knowledge [1-5].


Lee Tiedrich introduced the multidisciplinary panel: Minister Josephine Teo (Singapore, digital development, smart-nation strategy, cybersecurity), Professor Alondra Nelson (Institute for Advanced Study, senior advisor to the report), and Adam Beaumont (director of the UK’s AI Security Institute, the first government-backed body for safe, beneficial AI) [11-16]. He asked Yoshua to highlight the most significant changes between 2025 and 2026 [18-19].


Yoshua identified the rapid emergence of autonomous AI agents as the key shift. These agents can operate for hours or days, hold credentials, and access the Internet, thereby reducing the “human-in-the-loop” oversight that characterises today’s chat-bots [22-29]. He warned that once deployed, agents begin to interact with one another, a nascent but concerning phenomenon [30-31][S29][S62].


Lee noted the need for greater AI literacy so the public understands what agents can and cannot do [32-35].


Minister Teo likened AI governance to aviation safety: Singapore does not build aircraft, yet must ensure safe manufacturing, maintenance, and air-traffic management. She highlighted Singapore’s new law that imposes statutory obligations on platforms to remove harmful AI-generated images once notified [58-64], and stressed that AI is both a cyber-threat and a cyber-target, especially for multi-agent systems. She called for thoughtful guardrails, standards, and insurance-type schemes to manage these risks [65-70].


Alondra Nelson described the report as a ground-truth, evidence-based resource that deliberately does not prescribe policy. It uses OECD-style scenarios to provide foresight and expands its scope to systemic risks-loss of human autonomy, manipulation, job anxiety, and threats to social cohesion and democracy [78-85][90-93][191-199][202-204][88-102]. She warned of a collective-action problem and urged the community to agree on a small set of evaluation standards, recommending public-sector funding (analogy to the Human Genome Project’s “LC program”) to support upstream safety research [221-225][256-259].


Adam Beaumont outlined the AI Security Institute’s work: pre-deployment testing, post-deployment red-team exercises, model-card disclosures, grant-making, and the open-source Inspect framework that enables third-party auditors to evaluate models [107-127][214-215]. He emphasized the need for clear measurement goals, transparent reporting, and the development of a wide ecosystem of independent evaluators, akin to accounting auditors [214-225].


Addressing the jagged performance of general-purpose models, Yoshua warned that AI can be extremely capable in some domains while weak in others, creating dual-use dangers. He called for per-capability risk and intention assessments, and stressed scientific rigor, humility, and group review to avoid false claims that could mislead policymakers [132-150].


When asked how to translate science into practice for organisations without scientific staff, Minister Teo used an IKEA analogy: safety should be certified by standards and insurance-type mechanisms so end-users need not verify safety themselves. She mentioned Singapore’s national AI R&D plan that funds responsible-AI research, the development of testing frameworks and toolkits, and the importance of insurance schemes discussed at Davos [161-166][169-186][180-184].


Yoshua suggested adding an optional “policy-options” layer to the report-scientifically grounded choices and their likely consequences-while explicitly not prescribing a single course of action[244-250]. Alondra agreed that the report should remain evidence-based and highlighted the need for standardised evaluation protocols and public funding to support them [256-259].


Adam argued that evaluation responsibility should be shared among government, industry, academia, civil society, and individuals, recommending regulatory sandboxes, joint funding programmes, and collaborative pilots to build the ecosystem [267-277].


In the audience Q&A on AI sovereignty, Yoshua replied that true sovereignty means partnerships and international agreements, not isolationist “walls”, and called for mechanisms to verify compliance[292-300]. Minister Teo concurred, stating that a self-contained “sovereign AI” is unrealistic and would leave nations behind [304-312].


Points of Consensus (all speakers):


1. Urgent need for guardrails, standards, and independent evaluation of autonomous agents and multi-agent systems [20-31][68-70][214-216][256-259];


2. Scientific rigor and evidence-based, non-prescriptive reporting as the foundation for policy [138-150][76-77][88-102];


3. Dual-use cyber- and bio-security threats amplified by increased autonomy [65-70][119-124][22-31];


4. International cooperation over isolationist sovereignty[292-302][304-312];


5. Practical tooling (e.g., Inspect, insurance schemes) to give end-users confidence[155-159][161-166][214-215].


Remaining Open Issues


– Concrete standards and guardrails for agent credential access;


– Effective labeling or watermarking of harmful AI-generated content;


– Mechanisms for international verification of AI-safety agreements;


Standardised metrics for assessing intent and capability in jagged models;


– Clear allocation of responsibility among governments, industry, and individuals;


– Development of real-time, longitudinal evaluation methods to keep pace with rapid model improvements [58-64][68-70][292-302][132-150][214-225].


Action Items


1. Develop and promote third-party evaluation frameworks (e.g., expand the Inspect ecosystem) [107-127][214-215];


2. Enact thoughtful, targeted standards and regulatory sandboxes that balance innovation with safety [155-159][161-166][169-186];


3. Increase funding for research on multi-agent systems, cybersecurity, and bio-security (including insurance-type incentives) [107-127][180-184];


4. Continue collaborative updates of safety research priorities (Singapore’s responsible-AI programme as a model) [180-184];


5. Produce scientifically grounded policy-option briefs that outline possible actions and their consequences without prescribing a single path [244-250];


6. Foster cross-sector partnerships to co-design evaluation standards and verification protocols [214-225][256-259][292-302].


The discussion underscored that, while there is strong agreement on the challenges posed by autonomous agents, the need for rigorous independent evaluation, dual-use threats, and global cooperation, the path forward requires careful balancing of timely guardrails with ongoing research and the establishment of shared standards. Continued, science-driven dialogue will be essential to navigate the “unknown unknowns” ahead.


Session transcriptComplete transcript of the session
Yoshua Bengio

continue rapidly for policymakers across the globe to rely on an independent scientific assessment of what AI can do and what it can cause and what we can do already to try to mitigate this. I’m committed to continue supporting such reporting. As you know, we’re heading into a future with many unknown unknowns, things that we could not even imagine a year ago, like the psychological effects are happening, and there will be other surprises in the future. And so we must accept the prevailing uncertainty and collectively prepare for all plausible scenarios according to the scientific community. So thanks, and looking forward for the continued discussion. Thank you.

Lee Tiedrich

Oh, it’s working. Oh, okay. Well, thank you, Yashua, for your leadership and for giving us an overview of the safety report. And now we’re going to dig into the safety report in more detail. And to do this, we’ve got an amazing panel. To my left, we have Minister Josephine Teo from Singapore, who leads the Singapore government’s efforts in digital development, public communications and engagement, smart nation strategy, and cybersecurity. We’re also joined by Professor Alondra Nelson, who holds the Harold F. Linder Chair and leads science, technology, and social values lab at the Institute for Advanced Study, where she’s been on the faculty since 2019. And Alondra also contributed significantly to the report as a senior advisor. And then we also have Adam Beaumont, who is the director of the UK’s AI Security Institute.

The first and biggest government -backed organization dedicated to ensuring advanced AI is safe, secure, and beneficial. And I’m Lee Tedrick with the University of Maryland, and I also had the honor of serving as a senior advisor on the report. So to get us started, I’ll send the first question to Yashua. You talked about how the technology has evolved quite rapidly and continues to evolve rapidly, and you highlighted some of the significant changes. But are there any particular changes that really stand out to you as being significant in 2026 as compared to 2025?

Yoshua Bengio

Yes. I think in terms of risk management and potentially policy, the advances in agency of the AI systems is something we should pay a lot more attention. The reason is simple. Having AIs that are more autonomous means less oversight. So right now when you interface with a chatbot, of course the human is in the loop, right? It is a loop. And then usually you take what the AI is proposing, and then you humanize. You even do something. something with it. Agents are a different game where the agents will work on a problem for you for hours, days, and they will be given credentials. They will be given access to the Internet. So we need to have AI technology that will be much more reliable and avoid some of the issues we’re seeing today before this can be deployed in a way that’s safe and accepted because businesses and users at some point will be concerned that they can’t trust this technology with all the credentials that we might give them.

And then we’re also seeing things that are, I think, somewhat unexpected but not yet sufficiently studied, which is once we kind of let out these agents into the world, they start interacting with each other. And I think it’s about early days, but what we’re seeing is a bit concerning.

Lee Tiedrich

Yeah, I know it’s certainly gotten a lot of attention in the press, and I think it highlights the need to increase AI literacy too so people understand what these agents can and cannot do. For Minister Teo, Singapore has been at the forefront of AI governance from the ASEAN AI Governance Guide to the Singapore Consensus on AI Safety. And one of the things that Yashua highlighted that the report talks about is, you know, the need to translate some of the evaluation for different cultures and different norms and also to be able to put it into practice. Based on Singapore’s experience, what does it look like to take the science and actually put that into tools and practice that people around the world can use?

Josephine Teo

Thank you very much. Perhaps I will offer a perspective as a small state in a part of the world that has a lot of interest in the adoption of AI technologies, but perhaps is still only becoming much more aware of the extent of the risk. And so in my interaction with the international community, I think that the role of AI is really important. with my counterparts, I often share with them a perspective. They would have visited Singapore. They would have, you know, traveled in and out of our air hub. And I explained to them that Singapore does not own aircraft technologies. We, Boeing does not belong to us, neither does Airbus. But we have to be concerned about the safety of how these aircraft are manufactured.

We have to be concerned about maintenance, repair, and overhaul. We have to be concerned about air traffic management. If we didn’t have all of these elements in place, it’s very hard to see how you can have a thriving air hub, you know, and be responsible for the lives of millions of people passing through the airport. So that’s the reason why we think we have to be invested in the conversation and the efforts to bring about AI safety. If we want to see wide adoption in our region, then we must equally be aware of how the air traffic is manufactured. So the risk can be mitigated. The second point I’d like to make is that ultimately as policymakers, our objective in understanding the safety aspects must translate into how we can put them into operable guardrails.

And very often this would mean standards that are being imposed. This would mean regulations and laws. But we have to do it in a thoughtful way because we still do want to benefit from this technology. So if we are not targeted in the way we implement these requirements, then what we might achieve is not just the impact to the pace of innovation. What we could end up is a situation where we have given a false promise to our citizens, giving them the impression that we have protected them when in fact we haven’t actually done so. So that’s why I think we need to be thoughtful. interest is also that when there is clarity about what needs to be done, we want to be able to move very quickly.

Joshua talked about the misuse of AI, for example, to use them for generating images that often target women and children. And what we did was that last year we introduced a new law. It imposes statutory obligations on the services that bring these images and make this content available to vast numbers of people. They’ve always said that we are not responsible for the generation of such content. And so that’s something that we take on board. But having been notified of the existence of such harmful content, then there is an obligation for you to remove it. So this new law that we passed imposes such an obligation. And Joshua also talked about the financial… in the reports, our AI and cybersecurity is intersecting in very, very concerning ways.

For example, AI being used to target systems, and so AI is a threat. Now, however, we also see that AI itself can be a target of cyber attacks. And when AI becomes the target of cyber attacks, particularly for multi -agent systems, those kinds of risks can easily go out of hand. So even as the Singapore government is experimenting with the use of AI, we want to be very thoughtful about how these AI agent systems are being architected and what exactly goes into the decision -making process as to the agency that is being granted. Is there a way to put guardrails around it? So I would just say that AI as a threat, AI as a target, and where we really need to cooperate and do much better in that, is using AI as a tool.

to fight these threats. So those are the kinds of things that within the ASEAN community we hope to be able to make progress on.

Lee Tiedrich

That’s great, and thank you. And it’s a great segue to Alondra. You’ve worked not only in academia but have had high -level positions in the United States government, and a lot of your work is focused on the relationship between science, technology, and public accountability. And the report is really intended to inform policymakers and inform the broader community and intentionally does not take the next step of advising policymakers on what to do. And I’d be interested in your thoughts as to both the structure of the report and drawing that line, and importantly, what’s next? What should policymakers be thinking about as they read and digest the report?

Alondra Nelson

Yeah, thank you so much, and thank you all for being here. So let me just start by thanking Yoshua again because I was at Bletchley. We were at Bletchley Park. We were having a… conversation and one of the things I said when I spoke there was that we are going to need new democratic institutions for this moment. One of those are certainly the ACs, but one of those is this report, right? Like our ability to have a ground truth as a global community about the risks are deeply important for any future that we’re going to have with AI that’s beneficial. So, and I know that takes a lot of work and so thank you Yoshua for doing that.

And in the course of doing that and it’s serving as a senior advisor have seen how they’ve created a whole new system. I mean, you know, some of it comes out of CS culture, some of it comes out of research culture that we know, but they literally have created a new institution to help us kind of think through what’s the best information, how do you make evidence -based claims about the state of science and the midst of kind of radical uncertainty and that’s a new task for researchers of across our fields and disciplines. So I just want to tip my hat to you and make sure that people actually know how much work you’ve put into this.

So, you know, I think the report, its mandate, and I think it does a really good job of exactly not crossing that line, Lee, which is to say, what do we know? What’s the best of what we know? What are some, I mean, this report, I think, for the first time uses some OECD scenario, so it’s sort of reaching a little bit to evidence -informed kind of foresight and forecasting. And, you know, it really responds to, I think, the fact that a lot of our information about what’s happening in AI comes from journalism. It’s a very hard time to be a journalist right now, so this is not a knock on journalists, but it’s just to say that we don’t have globally, you know, the kind of sort of horizon of information that we really need in the policy space to make good policy decisions.

That said, you know, states will have lots of different policies and concerns that they want to make, so it’s not the mandate of the report to sort of direct how people should think about the evidence, but it is to say there’s more than anecdotal journalism here, and this is the best of what we know. In this moment, Yoshua mentioned there’s sort of updates that are happening, so the report is the team is also getting better at getting the information and more closer to real time. So I think it establishes that ground truth that’s so important for AI, particularly in the context not only of uncertainty, as I said, but of lots of hype that we’re sort of reading about and hearing about every day.

But I would also say that the report does a good job of, at the end of each section, making some nods to policy. So these are – so what should policymakers make of these scientific insights? And so it does a very good job at sort of steering what the implications of the fact that we have now growing uses of multiagent systems. How might you need to think about that? How might you need to think about the fact that there are growing sort of biosecurity and cybersecurity risks, for example? And then, Lee, to your point about what needs to be done, I think we all know what needs to be done. And I think – I hope that the report, because it is not anecdotal, not whim, allows there to be some – stronger political spines and some more political will.

to make the hard decisions that we need to make in the regulatory and policy space, both in individual nation states and I think also as a global community. So if it can be a resource for helping policymakers make good and strong and evidence -based arguments, and also I think allowing governments to support the funding of the creation of more evidence, I think it will be all to the good and obviously moving into the space of some sort of guardrails and regulatory regime is what needs to happen is the next step.

Lee Tiedrich

Thank you, Alondra. And also it’s a great segue over to Adam because the report also identifies what are some of the key research gaps and what are some of the key gaps in the evaluation ecosystem. And so for you, Adam, as the leader of the AC, what jumped out to you? What did you do in terms of? of risks and what are your priorities in terms of how to start addressing those risks going forward based on the report?

Adam Beaumont

Yeah, thank you very much. And I wanted to reiterate thanks to Joshua and the panel and also to call out the work of AC in supporting the secretariat of that for the past couple of years. And I know there are a couple of lead writers in the audience too. So it’s really great to see just the collaborative effort that’s happened around the world on that. I think it’s so important for enabling policymakers to have an objective, independent data science report. In AC, you asked me about which kind of risks jump out most. It’s quite hard to pick from our research staff. There’s about 100, so it’s like naming which is your favorite child. But there are a few from my – favorite’s a strange word.

There are a few that really jump out to me with my background in national security. And Joshua, you spoke. You’ve spoken a bit about this already. in cybersecurity and in biological capabilities. Both of those are very dual use. But I think in cybersecurity, we’ve seen such rapid development in the capability of the models, even in the last few weeks and months. And I think the report does a great job of explaining how that capability can assist in cyber operations at many different stages in that life cycle or different tasks. We’re not yet seeing that fully autonomous, though. And I think that is the area that concerns me that we’re trying to research and understand right now is, what does the confluence of some of these risks look like when combined with more autonomy, particularly in the genetic AI scenarios?

And some of the things we’re doing in AC about that, I guess we’re quite well known for our pre -deployment testing of frontier AI models. We also do post -deployment. You can see some of the impact of that in the model cards. Some of the companies published. we do a lot of red teaming and with that we’re trying to strengthen the safeguards of the models that are being provided but also raise the bar for the level of security research that’s happening so this week we published research around some of our methods on how we do that where we want to both responsibly disclose that but also grow the number of people that are working to help raise the bar we also use grant making and try to raise the level of investment happening in this space and then we’re trying to develop the way that we do evaluations to adapt to the way that models are improving capabilities, for example you get different results if they use more tokens with inference time scaling so we’re trying to make sure that our valuations account for that or by using cyber ranges rather than just capture the flag type scenarios so I care about all of those different risks that we are researching But the one I’m watching right now is probably on cybersecurity.

Lee Tiedrich

Thank you. And back to you, Yashua. One of the things that you had mentioned in the overview is the jagged performance of the general purpose AI models. And I’d be interested in your thoughts on how that impacts the evaluation science. If you have a general purpose model and it’s good at some tasks but not others, should evaluators be thinking about things differently?

Yoshua Bengio

Yes. Also, I think the general public and the media needs to escape this vision of an AGI moment. Because if AI continues to have these jagged capabilities, it means that we could well be in a world where AI already has dangerous capabilities and dual use for some things. At the same time as it might be really weak on other skills. And so the… The thing that matters at this point is in this world… that continues is very careful scientific evaluation of you know per scale per ability risk and capability right uh by the way that includes capability and intention something i didn’t mention too much in my presentation we’re seeing a lot of concerns with ais having goals that we would not like them to have um and in spite of our instructions acting uh against um their moral alignment training um so yeah this this is we we can’t stay at this very abstract i mean maybe like a few years ago thinking about agi was like a reasonable abstraction reaching human level but now it’s kind of meaningless because you know we’re gonna have things that can be extremely stupid in some ways maybe weak in some ways and already dangerous in the wrong hands in some other ways so we we have to be more technical and more precise you in talking about the risk And also, if you’re a business and you want to deploy, you also want to know, is the AI going to be good for what I’m trying to do?

I want to add one thing about the report spirit, about the report’s rigor. That’s not directly connected to your question, but I think it’s really important. There is a central requirement for science. When we talk about rigor, what does it mean? What it really means for every scientist, when they put something in writing or something official, they should not make a claim that could be false. They should only be claiming things that they’re totally sure about. Especially in the context. Where policymakers are going to use that information. You don’t want decisions to be taken based on false claims. And, of course, opinions abound in our world, especially because they impact people’s interests. And this is why it’s so important that we can ground our policy decisions in scientific evaluation.

And what it really means is this. It means a kind of humility and honesty, even when you may be biased in one way or another. To stick to those facts. And you need a group of people, because each of us can be personally biased, right? I am. Everyone is. It’s human. A group of people who can catch each other’s maybe going across that red line of rigor and not making statements that couldn’t be defended very strongly.

Lee Tiedrich

Thank you. And a very, very important point. I think… I think in addition to the policymakers needing to be able to use this information, I, through my work, end up talking to a lot of organizations, nonprofits, small and medium -sized businesses. And what I hear a lot is, like, it’s great. Like, you have to start with the science, and that is ground zero. But then for some of those other organizations, they need the tooling. They’re not going to have a whole scientific staff on how do we put that into practice. And I’m just wondering from the government’s perspective, Minister Teo, what are your thoughts on how we might be able to advance some of the tooling to take this great learning and make it easier for companies and other organizations to actually deploy?

Josephine Teo

I was at a similar session recently, and this topic came up. And the way I think about it is I use IKEA as an example. You know, when you go to IKEA, you buy furniture, and IKEA promises you that… furniture has been tested. So, you know, if it’s a couch, it has been jumped on, I don’t know, 25 ,000 times, and it didn’t break, you know, and so your kids are not going to be hurt if they jumped on it too, well, up to 25 ,000 times. And if you think about a user on the receiving end of this technology, it is, I think, quite unreasonable to expect them, you know, to have to impose safety conditions on their own.

They are simply not in a position to do so. They don’t have the power to decide, you know, what gets sold to them and what does not get sold to them. So we as policymakers must recognize that there is a huge gap between those that we are encouraging to adopt AI tools, adopt AI technology in various contexts. We must think about… Where are the right points to make… these requirements mandatory when it is perhaps not so much requirements that are mandatory, but it is useful for industries to come together. For example, in Davos, we discussed the possibility of insurance schemes, creating the right incentives for AI model developers. And I think that there is no easy landing point just yet.

But if we fail to engage in these conversations in a rational way, then I think we are even further behind in trying to manage the risks. So I would say that the thoughtfulness has to be applied at many different levels. There needs to be continued research in AI safety. And so I’m very happy that we are continuing to have this conversation. Thank you. in Singapore and we hope to update where are the areas of safety research that should be prioritised. I think this year, I certainly agree with you, multi -agent systems is going to come up quite prominently. But we cannot just stop there. We also have an ongoing programme. We started by setting aside commitments under our own national AI R &D plan and in fundamental research, one of the areas that we are very interested in is responsible AI.

So you need the two to go hand in hand. But can you not have some testing frameworks and toolkits to begin with? We think that that is also not helpful. It is more pragmatic to try and to recognise the shortcomings of those testing tools and then to invest further effort in promoting more thoughtful, thoughtful ways of looking at the research. of these systems and how to mitigate against them. Ultimately, we should try and get to a point where the end user has assurance of safety, that they don’t have to be thinking so hard about whether the proper tests have been applied. We’re not there yet, but I think we need to find a way to work out the roadmap.

Lee Tiedrich

That’s very interesting. You can also think of analogies in the medical context. We don’t always understand how the medicine works, but we have assurance that if it’s prescribed for us, it’s going to work well. Turning back to you, Alondra, there’s been a lot of conversation around catastrophic risks, and the report is intentionally broader than just catastrophic risks. I’d be interested in your thoughts as to whether that was a good place to draw the line and what some of the benefits are of broadening our aperture beyond just the catastrophic risks.

Alondra Nelson

Thank you. Certainly, the reason that I continue to be involved with this is because… under Yoshua’s chairmanship of the report, that it is attentive to a broader set of risks. So there’s a section of the report that’s called systemic risk, and I think what we haven’t quite pieced together is that particularly if we care about democracy, if we care about social cohesion, it is not the individual risk, like we all have our favorites or unfavorites, Adam, to your point. It is the compounding of those risks together. Like we are careening without seatbelts in a car quickly in a society in which all of these risks and harms are happening simultaneously. So that is a very dangerous world for social cohesion.

That is not a society that’s healthy, and that’s not healthy for democracy. And so I think the attention to the broader set of risks, which include things like loss of human autonomy, what does it mean when you’re not in charge of your own decision -making, what does it mean when you’re not in charge of your own decision -making? What does it mean when sycophancy and other sorts of, I think, outputs mean that you are being manipulated in some way through the use of the tools and technologies? How do we think about the fact that there might be job loss or job displacement, the anxiety that it creates? I mean, talk about a lack of social cohesion.

The anxiety it’s already creating in a lot of societies about people’s livelihoods and their abilities to protect them and their families and their well -being. So I think what the report does incredibly well under a kind of large banner of safety is to think at a 30 ,000 -foot level, if you take all of the chapters together, about what are those compounding risks? What does it look like if all of those risks sort of move together simultaneously? And therefore, it is equally important to think about that technology in a healthcare space that’s malfunctioning, giving a misdiagnosis, as important as it is in some ways to think about a bio -risk. And so I think that’s important. And I’m…

I’m really gratified that the report continues to be anchored in that broader aperture of risk.

Lee Tiedrich

I would agree, too, because I think a lot of those risks are, especially with agents, they’re here today and they’re just going to continue to increase, and we do need to keep the focus on them.

Yoshua Bengio

Just a small comment about the systemic risks. Of course, I completely agree, but I want to point out one factor that makes them potentially catastrophic, except maybe at a slower pace, is because so many people are going to be using these systems, and the global dynamics and social dynamics are so difficult to anticipate and could be incredibly impactful, both on the positive and negative side.

Lee Tiedrich

I think Yashua and Alondra’s comments tee up the next question for Adam. These risks are evolving quite rapidly, and one of the things that the report, I think, emphasizes is we have an evidence gap. for researchers to keep up and it’s hard to do longitudinal studies in a very short period of time. I’d be interested in your perspectives from the ACs. How do you address that as you start thinking about real -world evaluation today and how does that impact the approach to evaluation and what might the ACs be able to do to help fill some of this evidence gap?

Adam Beaumont

some of our learnings and some of them are quite simple there are things like if if you’re evaluating something be really clear what is it you are trying to measure and make sure your evaluation is actually getting after the thing that you are focused on as some can be quite misleading in the way that they are organized but in addition to areas where we had good consensus around best practices we also highlighted areas where there’s still uncertainty or we need more research and again we want to communicate that and be very transparent so that more people can join in as we do see this as requiring like many great minds around the world and they just aren’t enough safety and security researchers to do that all in one place but in addition to talking about the practice of evaluation we’re also trying to provide tooling for other organizations to do that and one of the things I’m very proud of the AC developed in the UK was the inspect framework there’s been open sourced and is used really extensively by different companies, organisations in government, outside government.

And the thing I would love to see over this coming year is how we can really grow a wide kind of ecosystem of third -party evaluators that can offer that independence and bring rigour and scientific method to the way that we measure these capabilities and then can communicate about them. And just I’m going to ask one quickfire question for the whole group and then I’m going to open it up for Q &A, so start thinking about your questions. But, you know, I’m interested, Adam, and I think it touches on some of the themes of like how do we take the science and bring it to practice and how do we actually create this evaluation ecosystem.

So step one is developing the science. Step two is then figuring out, well, how do we actually evaluate this? And then there’s the, you know, by whom. And how do you see an evaluation ecosystem? How do you see the ecosystem emerging? Do you see governments being the evaluators? Do you see this going more like we have with accounting, where you have third -party certified auditors doing the evaluations? I’d be interested in each of your thoughts. And maybe start with Minister Tia, and then we can go down the line.

Josephine Teo

Well, certainly in the ASEAN context, I would advocate for an approach that deals with near and present dangers that everyone is dealing with. The risk of not focusing on what’s most prominent in people’s minds today, policymakers’ minds today, is that the conversation may feel too theoretical, and we may lose interest and momentum, and we don’t even build the foundations of cooperating in a meaningful way. And what are some of those areas where AI intersects with? AI being used, misused, for harming people in terms of the content creation. I think that’s one. Almost every single policy. that I come across is very, very upset by the fact that they have to address their constituents’ concerns about all these harmful images that are being created with the help of AI.

It’s very offensive to our societies. And if we are not able to work on these areas in a meaningful way, in a practical way, then I think we risk losing my colleagues’ attention. So what can we do? We have to then seriously ask, is watermarking the correct approach of dealing with it? Is there some other way of labeling AI -generated content? Is that even the right direction that we should be moving on? The other area is that I think it will be very prominent, and that is the use of AI in cybersecurity. I don’t think at this point in time AI as a threat is adequately addressed. AI as a target is even further from that.

It’s in people’s minds. of the conversation in the areas that my colleagues care about, I think stands a better chance of anchoring their attention and creating meaningful opportunities for us to say, here are the ways you can test for it, and here are the tools that can be applied. They won’t be perfect, but they are important stuff.

Yoshua Bengio

So I want to mention maybe a totally different aspect that’s orthogonal to this. As I’ve been thinking about the process of bringing the science to have an impact with policymakers, I feel like there is a step in between what we’ve done and the actual political decision making, and that is using scientifically grounded policy options. So the report doesn’t go into recommendations, and I think that was a great mandate that we started from, but I think there is something in between taking the policy decisions and this. Thank you. grounded in what the scientists see and the people like economists and social scientists, based on this, what are reasonable options for policymakers without saying you have to take this one?

You could do this, you could do nothing. And what are the consequences that are expected based again on the science without making an actual recommendation? Because in the real world, I understand policymaking is hard because you always have a tension between different values and objectives and interests. We shouldn’t make those choices, but we can help make it easy for policymakers.

Alondra Nelson

I think I would offer we’re just getting started with evaluations and assessments. And so I wouldn’t want to put a thumb on a scale and pick one. I mean, I think that we actually have to try a lot of different things. I also think to the extent that we have a body of knowledge around evaluation that is coming from ACs and policymaking, and other researchers. Thank you. that, you know, I worry that we’re going to have a collective action problem and so that everyone’s doing their own different kind of evaluation. And I think what we will need to fundamentally do is make, as a research community, a few choices about, you know, something closer to a standard, like this is the way that we are, the few ways that we’re going to proceed.

So I think there’s that. I do think that it needs to be obviously multi -sector. It’s a fairly obvious point. How do you do that is an open question. I wrote a piece in Science a few months ago where I suggested that we might think about the LC program for human genetics and genomics in which, you know, 3 % of the Human Genome Project research budget in 1990, 1991, was dedicated to upstream research of potential risks and harm of human genetics. So that doesn’t present risks and harms, but it means that you go in upstream to projects thinking about them as a part of the research and design, often before deployment. And it doesn’t mean that you can prevent things like someone doing illegal human genome gene editing, right?

But I think that you do have a global community that has thought about it and is ready to have a conversation and knew in the case of the human genome, the human gene editing, that it was wrong and why it was wrong and we had discussed it. So I think that there are, you know, lots of models that government’s deeply important here and that, you know, I think that there are schemes that would require, I think, the public sector to, you know, place a little money in the space of a sort of common good or a comments for research to understand and advance much more in the evaluation and assessment space.

Adam Beaumont

Yes, you asked who should be involved in evaluation or where should be done and I guess my answer to that is should it be government, should it be industry? It’s kind of all of the above and I really agree with you that we’re very early in the journey and there’s still a lot of uncertainty but I do think there’s a role for governments to play, there’s a role for industry, there’s a role for researchers, civil society but also individuals and we saw that at the start of the year when people are very willing to trade away. They can trade away all their keys, passwords, anything for the… enjoyment of agent autonomy. And that reminds me a lot of the early days of cybersecurity where we need to grow ecosystems.

Individuals have a responsibility as much as governments and I’m sure over time we’ll see more institutions and organisations grow that help do that. But the key to it has got to be collaboration. So on a practical level, things like regulatory sandboxes or like policy lab type things where you can try limited pilot approaches seem to be good. We’re trying a bit of that in the UK. Things like joint funding programmes that bring researchers, policy makers together to kind of iterate options again seems a good idea. But I strongly agree we’re just early in the journey. We should keep options open.

Lee Tiedrich

Thank you. I think we have time for one or two questions. Wow, we have a lot of hands. What I’m going to do is call on two people. We’ll kind of combine the questions and we’ll let the panel. Well, I wish I had more time. So we’ll take one here and one over there. Go ahead. Right here in the second row. Can someone bring a microphone over? Or a speaker. It’s not a very good move, right?

Participant

Can you hear me?

Lee Tiedrich

Yes.

Participant

So I have a question. So, like, now we hear a lot about, like, the rise of business and sovereignty, like, everywhere, and, like, a lot of more countries are trying to claim it in some ways or another. And I would be really curious to hear, like, how, at least in the AI safety field, how are you seeing that impact and which other safety concerns are most pressing, like the grown -up of the window based on that first?

Yoshua Bengio

Yeah, so I think we should be careful about what sovereignty means. It doesn’t mean building walls around your country. It means making sure your country will retain the ability to, you know, take its decisions. And, you know, succeed economically. economically and politically. And often that means the opposite of walls around your country. It means making partnerships with others that increase your chances of, you know, not ending up in a bad place. And that includes agreements on safety, right, because many of the risks we’ve discussed, you know, they’re not limited by borders. We can collaborate on the safety technology with multiple countries. We can have the kinds of agreements that Singapore has been leading where multiple parties, you know, from many different countries agree on principles.

And eventually we will need international agreements and we will need technology for verification of these agreements. We are far from that, but that’s the only kind of world where, you know, I would want my children to live. Where AI is not used to dominate others and we don’t see, like, reckless behavior across. the world.

Josephine Teo

Ms. I’m so glad that Yoshua has offered a view that to me is a very sound approach. You said earlier that what we want is a world where every country can be at the table, not on the menu. And that’s exactly how you can preserve sovereignty, even with AI developments. The idea that you get sovereign AI by confining everything to your own shores, I think it gives a false sense of security. Firstly, it’s not achievable. Secondly, the idea that you can do so, I think, would mean that for many countries where the most sophisticated applications will have to originate from elsewhere, that just cuts you off. It cuts you off from being able also to make progress, and that puts you even further behind.

So how is that sovereign? So it has to be a topic that is dealt with thoughtfully. It’s not a term to be bandied about too easily.

Lee Tiedrich

Melinda, Adam, any thoughts? Okay. Yeah, so we unfortunately are running out of time, but I would love to thank our panelists for being here today and sharing the report. And I hope all of you will read the report and continue to engage with us because, as we said, there’s a lot more work to be done. Thank you very much. Thank you all. Thank you. you you Thank you. Thank you.

Related ResourcesKnowledge base sources related to the discussion topics (13)
Factual NotesClaims verified against the Diplo knowledge base (6)
Confirmedmedium

“Lee Tiedrich introduced the multidisciplinary panel including Minister Josephine Teo and Professor Alondra Nelson (and Adam Beaumont)”

The knowledge base lists Josephine Teo and Alondra Nelson as participants in the AI safety discussion, confirming their presence on the panel [S1].

Confirmedhigh

“Yoshua Bengio identified the rapid emergence of autonomous AI agents as the key shift in AI between 2025 and 2026”

Sources describe a critical technological shift toward proactive, autonomous AI agents capable of independent decision-making, matching the report’s description of autonomous agents as a major change [S81] and [S82].

Additional Contextmedium

“These autonomous agents can operate for hours or days, hold credentials, and access the Internet, reducing human‑in‑the‑loop oversight”

The knowledge base notes that autonomous agents can act independently and perform tasks without continuous human supervision, providing background for the claim about extended operation and reduced oversight [S81] and [S82].

Confirmedmedium

“Once deployed, agents begin to interact with one another, creating a nascent but concerning phenomenon”

Discussion of standards and protocols for agents to work together indicates that multi-agent interaction is an emerging issue, confirming the report’s point [S86].

Confirmedmedium

“Lee noted the need for greater AI literacy so the public understands what agents can and cannot do”

UN and other forums emphasize the importance of data and technological literacy for AI understanding, supporting the call for broader AI literacy [S88] and [S90].

Confirmedhigh

“Minister Josephine Teo likened AI governance to aviation safety, using Singapore’s experience with aircraft safety as an analogy”

The policymaker’s guide explicitly describes Minister Teo’s aviation safety comparison to illustrate AI governance challenges [S6].

External Sources (90)
S1
AI Safety at the Global Level Insights from Digital Ministers Of — -Alondra Nelson: Professor who holds the Harold F. Linder Chair and leads science, technology, and social values lab at …
S2
A Digital Future for All (afternoon sessions) — – Alondra Nelson – Harold F. Linder Professor, Institute for Advanced Study Alondra Nelson: I do. I do. I mean, I thin…
S3
Global Perspectives on Openness and Trust in AI — -Alondra Nelson- Former deputy director of the White House Office of Science and Technology under President Biden
S4
AI Impact Summit 2026: Global Ministerial Discussions on Inclusive AI Development — -Josephine Teo- Role/title not specified (represents Singapore)
S6
S7
Transcript from the hearing — Let me introduce the witnesses and seize this moment to let you have the floor. We’re honored to be joined by Dario Amad…
S8
UN Secretary-General unveils Science and Technology Advisory Board — The United Nations Secretary-General, António Guterres, announced the creation of aScientific Advisory Boardto provide i…
S9
Driving U.S. Innovation in Artificial Intelligence — 17. Yoshua Bengio – Professor, University of Montreal
S10
WS #280 the DNS Trust Horizon Safeguarding Digital Identity — – **Participant** – (Role/title not specified – appears to be Dr. Esther Yarmitsky based on context)
S11
Leaders TalkX: Moral pixels: painting an ethical landscape in the information society — – **Participant**: Role/Title: Not specified, Area of expertise: Not specified
S12
Leaders TalkX: ICT application to unlock the full potential of digital – Part II — – **Participant**: Role/Title not specified, Area of expertise not specified
S13
AI Safety at the Global Level Insights from Digital Ministers Of — – Alondra Nelson- Adam Beaumont – Yoshua Bengio- Alondra Nelson- Adam Beaumont
S14
Published by DiploFoundation (2011) — Malta: 4th Floor, Regional Building Regional Rd. Msida, MSD 2033, Malta Switzerland: Rue de Lausanne 56 CH-1202 Ge…
S15
Welfare for All Ensuring Equitable AI in the Worlds Democracies — – Lee Tiedrich- Amanda Craig Deckard – Lee Tiedrich- Sachin Kakkar
S16
Agents of Change AI for Government Services & Climate Resilience — – Lee Tiedrich- Srinivas Tallapragada Tiedrich advocates for developing comprehensive global standards through internat…
S17
The Dawn of Artificial General Intelligence? / DAVOS 2025 — Nicholas Thompson: Yoshua? Yoshua Bengio: All right, there are several things that Andrew said that I think are wrong…
S18
AI Development Beyond Scaling: Panel Discussion Report — – Yejin Choi- Yoshua Bengio – Yoshua Bengio- Eric Xing – Yoshua Bengio- Eric Xing- Yejin Choi Choi advocates for cont…
S19
Main Session on Artificial Intelligence | IGF 2023 — Seth Center:IAEA is an imperfect analogy for the current technology and the situation we faced for multiple reasons. One…
S20
European Tech Sovereignty: Feasibility, Challenges, and Strategic Pathways Forward — The conversation maintained a constructive tone, with participants balancing criticism of European shortcomings with opt…
S21
Agentic AI in Focus Opportunities Risks and Governance — “If the data can be manipulated, if the lineage of data is not properly understood, if it is not really governed, if the…
S22
Generative AI: Steam Engine of the Fourth Industrial Revolution? — Technology is moving at an incredibly fast pace, and this rapid advancement is seen in various sectors such as AI, semic…
S23
How AI Drives Innovation and Economic Growth — Rodrigues emphasizes that while early AI discussions were dominated by fear about job displacement and technological thr…
S24
https://dig.watch/event/india-ai-impact-summit-2026/ai-safety-at-the-global-level-insights-from-digital-ministers-of — And very often this would mean standards that are being imposed. This would mean regulations and laws. But we have to do…
S25
Emerging Shadows: Unmasking Cyber Threats of Generative AI — Data poisoning and technology evolution have emerged as significant concerns in the field of cybersecurity. Data poisoni…
S26
AI malware emerges as major cybersecurity threat — Cybersecurity experts areraising alarmsas AI transitions from a theoretical concern to an operational threat. The H2 202…
S27
Tech Transformed Cybersecurity: AI’s Role in Securing the Future — Lastly, the analysis highlights the interdependence of cybersecurity and AI for the safety of digital assets. Both are c…
S28
UN OEWG hosts inaugural global roundtable on ICT security capacity building — The UN recently hosted the inauguralGlobal roundtable on ICT security capacity buildingunder the auspices of theOpen-End…
S29
Challenging the status quo of AI security — Multi-agent systems are rapidly being deployed across organizations, creating urgent need for coordination standards and…
S30
Science as a Growth Engine: Navigating the Funding and Translation Challenge — A lot of research had been done before. So to explain this, to really take the society serious, because in the end, it’s…
S31
High-Level Dialogue: The role of parliaments in shaping our digital future — There is insufficient interaction between those making policy decisions and the scientific community that understands th…
S32
Closing the accountability gap: A proposal for an evidence-led accountability framework — 7 Jacqueline Eggenschwiler, Accountability Challenges confronting Cyberspace Governance , Journal on Internet Regulatio…
S33
In brief — Humanitarian actors need to be aware of the different nuances of the term ‘evidence-based’, particularly w…
S34
Diplomatic policy analysis — Global collaboration:Policy analysis helps identify shared interests and opportunities for cooperation, fostering consen…
S35
Plenary: Sustainability at Risk: Drawing Insights from Climate Talks to Elevate Cybersecurity — Emphasis is placed on collaboration and a global perspective when addressing cybersecurity needs in the Global South. Jo…
S36
Opening of the session — Greater international cooperation is necessary in the context of threats.
S37
Interdisciplinary approaches — AI-related issues are being discussed in various international spaces. In addition to the EU, OECD, and UNESCO, organisa…
S38
A Global Human Rights Approach to Responsible AI Governance | IGF 2023 WS #288 — Different governments and countries are adopting varied approaches to AI governance. The transition from policy to pract…
S39
Building Inclusive Societies with AI — Government role as facilitator rather than direct implementer in startup and private sector initiatives Multi-stakehold…
S40
WS #294 AI Sandboxes Responsible Innovation in Developing Countries — Natalie Cohen, Head of Regulatory Policy for Global Challenges at the OECD, positioned sandboxes within broader regulato…
S41
Sandboxes for Data Governance: Global Responsible Innovation | IGF 2023 WS #279 — Ensuring fairness and avoiding regulatory capture are identified as important considerations in sandbox implementation. …
S42
AI Safety at the Global Level Insights from Digital Ministers Of — This comment established the central theme for the entire discussion. It shifted the conversation from abstract AI safet…
S43
Ensuring Safe AI_ Monitoring Agents to Bridge the Global Assurance Gap — It’s a really foundational area for collaboration for all of us. Now, my view is that if we do get assurance, and by rig…
S44
From summer disillusionment to autumn clarity: Ten lessons for AI — The 40-member Scientific Panel will produce annual reports that synthesise research on AI’s risks, opportunities, and im…
S45
Diplomatic reporting — Contextual analysis:Beyond raw information, diplomatic reports offer context by analysing the implications of events for…
S46
Towards 2030 and Beyond: Accelerating the SDGs through Access to Evidence on What Works — Andrea Cook: Thank you, Ambassador Rae, for your insightful opening remarks and throwing down the challenge. We really…
S47
WS #288 An AI Policy Research Roadmap for Evidence-Based AI Policy — Eltjo Poort, Vice President Consulting at CGI in the Netherlands, supported this view: “Regulation does not hamper innov…
S48
Agentic AI in Focus Opportunities Risks and Governance — “If the data can be manipulated, if the lineage of data is not properly understood, if it is not really governed, if the…
S49
UNGA/DAY 1/PART 2 — The advancement of AI is outpacing regulation and responsibility, with its control concentrated in a few hands. (UN Secr…
S50
Advancing Scientific AI with Safety Ethics and Responsibility — And also, very importantly, how we have to also see it from the context of, you know, people doing their own thing, DIY …
S51
AI Meets Cybersecurity Trust Governance & Global Security — “AI governance now faces very similar tensions.”[27]”AI may shape the balance of power, but it is the governance or AI t…
S52
Open Forum #58 Collaborating for Trustworthy AI an Oecd Toolkit and Spotlight on AI in Government — Lucia Rossi: Thank you, Yoichi, and good afternoon to the audience here and online. It’s a pleasure being here at the IG…
S53
Practical Toolkits for AI Risk Mitigation for Businesses — Discovering and highlighting business incentives, particularly trust, consumer adoption, as well as headline risks might…
S54
AUDA-NEPAD White Paper: Regulation and Responsible Adoption of AI in Africa Towards Achievement of AU Agenda 2063 — Additionally, in an AI-driven economy, it will be necessary to take practical steps to implement policy considerations t…
S55
UNSC meeting: Scientific developments, peace and security — Dual-use nature of technologies presents notable risks
S56
Can National Security Keep Up with AI? / Davos 2025 — AI technology has both beneficial and potentially harmful applications. This dual-use nature creates dilemmas and challe…
S57
Agenda item 5: discussions on substantive issues contained inparagraph 1 of General Assembly resolution 75/240 part 1 — Ukraine: Mr. Chair, Ukraine aligns itself with the statement delivered by the European Union. We would like to make so…
S58
Discussion Report: Sovereign AI in Defence and National Security — This comment addresses a key concern about AI sovereignty leading to fragmentation, instead positioning it as a foundati…
S59
Global AI Policy Framework: International Cooperation and Historical Perspectives — Mirlesse outlines practical steps for implementing open sovereignty, emphasizing domestic AI deployment in key sectors w…
S60
Ensuring Safe AI_ Monitoring Agents to Bridge the Global Assurance Gap — Larter emphasised that the emerging agentic economy requires new technical protocols for agents to communicate with each…
S61
Challenging the status quo of AI security — Agent identity management presents fundamental challenges including defining what constitutes agent identity, establishi…
S62
AI Safety at the Global Level Insights from Digital Ministers Of — Both speakers identify multi-agent systems as a priority area of concern, with Bengio noting concerning interactions bet…
S63
Why science metters in global AI governance — Because there’s not enough past evidence to be sure that a particular tipping point is going to happen. So the situation…
S64
WS #453 Leveraging Tech Science Diplomacy for Digital Cooperation — Armando Guio: Thank you. Thank you very much, Sofie. And it’s great, of course, to see you all and to share this panel w…
S65
Keynote-António Guterres — “First, creating an independent international scientific panel on AI.”[10]”We must replace hype and fear with shared evi…
S66
Closing the accountability gap: A proposal for an evidence-led accountability framework — 7 Jacqueline Eggenschwiler, Accountability Challenges confronting Cyberspace Governance , Journal on Internet Regulatio…
S67
Open Forum #42 Global Digital Cooperation: Ambition to Country-Level Action — Margarita Gomez: Thank you. Thank you so much. It’s a pleasure to be here and thank you everybody that is joining on…
S68
https://dig.watch/event/india-ai-impact-summit-2026/transforming-health-systems-with-ai-from-lab-to-last-mile — So there’s a massive gap there. And then we’re also now starting to see in different contexts anecdotal evidence of wher…
S69
Opening of the session — Greater international cooperation is necessary in the context of threats.
S70
WS #64 Designing Digital Future for Cyber Peace & Global Prosperity — Genie Sugene Gan: Thank you. Well, I think one success story, I think from our lived experience would be our No More…
S71
Plenary: Sustainability at Risk: Drawing Insights from Climate Talks to Elevate Cybersecurity — Emphasis is placed on collaboration and a global perspective when addressing cybersecurity needs in the Global South. Jo…
S72
Cybercrime and Law Enforcement: Conceiving Jurisdiction in a Borderless Space — Cooperation at various levels, sectors, and regions is vital in addressing cyber threats. Ghana’s Cyber Security Act of …
S73
Insights from AI experts’ testimonies before US Senate — Leading AI experts testified before the US Senate Judiciary Committee andoffered their opinions on the emerging AI techn…
S74
Opening of the session — Mauritius recognized that while technologies are inherently neutral, the rapid advancement and convergence of emerging t…
S75
AI-driven Cyber Defense: Empowering Developing Nations | IGF 2023 — He introduces a panel of experts from different fields
S76
World Economic Forum Panel on Quantum Information Science and Technology — This World Economic Forum panel discussion brought together leading experts to explore quantum information science and t…
S77
WS #266 Empowering Civil Society: Bridging Gaps in Policy Influence — Stephanie Borg Psaila: Thanks, Kenneth. I’ll reflect on a few comments that our colleagues have made, and I’ll start wit…
S78
Chief Economists’ Briefing: What to Expect in 2025? / DAVOS 2025 — Fernando Honorato Barbosa: Yeah, so again, it’s the second change we’re seeing since the pandemic. The pandemic was t…
S79
Agenda item 5: discussions on substantive issues contained in paragraph 1 of General Assembly resolution 75/240 (continued)/5/OEWG 2025 — The Chair expressed concern about the lack of progress towards consensus, urging delegates to show more flexibility in t…
S80
Open Forum: A Primer on AI — Artificial Intelligence is advancing at a rapid pace
S81
WS #283 AI Agents: Ensuring Responsible Deployment — Prendergast frames agentic AI as a critical technological shift where AI has evolved beyond reactive tools to become pro…
S82
Building Trusted AI at Scale Cities Startups & Digital Sovereignty – Keynote Jeetu Patel President and Chief Product Officer Cisco Inc — Patel describes a rapid progression of AI from chat‑based bots to agents that can perform tasks autonomously, and antici…
S83
The fading of human agency in automated systems — Crucially, a human presence does not guarantee agency if the system is designed around compliance rather than contestati…
S84
https://dig.watch/event/india-ai-impact-summit-2026/how-the-global-south-is-accelerating-ai-adoption_-finance-sector-insights — And I think that’s true in the short term when the ecosystem is getting prepared. But in longer term, frauds and mis -se…
S85
AI agent autonomy rises as users gain trust in Anthropic’s Claude Code — A new study from Anthropicoffersan early picture of how people allow AI agents to work independently in real conditions….
S86
U.S. AI Standards Shaping the Future of Trustworthy Artificial Intelligence — <strong>Sihao Huang:</strong> of these agents work with each other smoothly. And protocols are so important because that…
S87
AI agents face prompt injection and persistence risks, researchers warn — Zenity Labs warned at Black Hat USA that widely used AI agents can behijacked without interaction. Attacks could exfiltr…
S88
Artificial intelligence (AI) – UN Security Council — Furthermore, there was a consensus on the necessity for enhanced data literacy and data management skills. As AI systems…
S89
WS #100 Integrating the Global South in Global AI Governance — Jill: Thank you, for the opportunity and also for the question, by the way. So, IEEE, as you say, is a standards organi…
S90
Mediation and artificial intelligence: Notes on the future of international conflict resolution — The idea of technological literacy is certainly not a new one. 31 As a starting point, it can be defined as ‘having know…
Speakers Analysis
Detailed breakdown of each speaker’s arguments and positions
Y
Yoshua Bengio
6 arguments140 words per minute1253 words534 seconds
Argument 1
Increased risk due to reduced human oversight as AI agents gain more autonomy (Yoshua Bengio)
EXPLANATION
As AI agents become more autonomous they operate with less direct human supervision, which raises the chance of unintended actions. This shift from human‑in‑the‑loop chatbots to self‑directed agents creates new safety challenges.
EVIDENCE
Bengio explains that “Having AIs that are more autonomous means less oversight” and contrasts current chatbot interactions, where a human remains in the loop, with agents that are given credentials and internet access, highlighting the reduced supervision and the need for more reliable technology before deployment [22-29].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
External sources note that autonomous systems can operate beyond immediate human supervision and create new safety challenges, echoing Bengio’s concern about reduced oversight [S1], [S17].
MAJOR DISCUSSION POINT
Risk of reduced oversight
AGREED WITH
Josephine Teo, Adam Beaumont
Argument 2
Multi‑agent systems beginning to interact autonomously, raising safety concerns (Yoshua Bengio)
EXPLANATION
When multiple autonomous agents are released, they start communicating and acting with each other without human control, which could lead to emergent risky behaviours. This phenomenon is still in early stages but already shows concerning signs.
EVIDENCE
Bengio notes that once agents are let out “they start interacting with each other… it’s early days, but what we’re seeing is a bit concerning” [30-31].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The risk of emergent, unpredictable interactions among multiple agents is highlighted as a priority concern in the ministerial insights [S1].
MAJOR DISCUSSION POINT
Emergent interactions among agents
Argument 3
Scientists should offer scientifically grounded policy options that outline consequences without dictating choices (Yoshua Bengio)
EXPLANATION
Bengio argues that scientific reports should provide policymakers with evidence‑based options and likely outcomes, while refraining from prescribing exact actions. This helps bridge the gap between science and policy without overstepping into advocacy.
EVIDENCE
He states that there is “a step in between… using scientifically grounded policy options… you could do this, you could do nothing… consequences… based again on the science without making an actual recommendation” [244-250].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The report’s approach of presenting evidence-based options without prescribing actions is described in the ministerial briefing [S1].
MAJOR DISCUSSION POINT
Science‑informed policy options
DISAGREED WITH
Lee Tiedrich, Alondra Nelson
Argument 4
Jagged capabilities of general‑purpose models demand per‑capability risk and intention assessment (Yoshua Bengio)
EXPLANATION
Bengio points out that large models exhibit uneven performance across tasks, so evaluation must consider each capability and its associated risks and intentions individually. This granular approach is needed to understand both dangerous and weak aspects of the technology.
EVIDENCE
He says “we need very careful scientific evaluation per scale per ability risk and capability… includes capability and intention” while warning that models can be dangerous in some areas yet weak in others [132-135].
MAJOR DISCUSSION POINT
Need for per‑capability risk assessment
Argument 5
Sovereignty should emphasize collaborative agreements and verification mechanisms rather than isolationist “walls” (Yoshua Bengio)
EXPLANATION
Bengio contends that national sovereignty in AI should be about forming international safety agreements and verification standards, not building protective barriers. Cooperation across borders is essential because AI risks transcend national boundaries.
EVIDENCE
He explains that sovereignty “doesn’t mean building walls… it means making partnerships… agreements on safety… international agreements… verification of these agreements” [292-300].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Bengio’s view that sovereignty means partnership rather than walls is reflected in the discussion of international safety agreements [S1] and European tech-sovereignty debates [S20].
MAJOR DISCUSSION POINT
Cooperative approach to AI sovereignty
AGREED WITH
Josephine Teo, Alondra Nelson
Argument 6
Scientific rigor and collaborative peer review are essential to avoid false claims in AI safety reporting
EXPLANATION
Bengio emphasizes that scientists must only assert statements they are fully certain of, practicing humility and honesty. He argues that a group of reviewers is needed to catch individual biases and ensure that policymakers receive trustworthy, evidence‑based information.
EVIDENCE
He outlines a central requirement for science that rigor means not making potentially false claims, stresses the need for humility and honesty, and calls for a group of people to catch each other’s biases and prevent statements that cannot be strongly defended [138-152].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Bengio stresses the need for rigorous, collaborative validation to prevent unsupported claims, as outlined in the ministerial summary [S1].
MAJOR DISCUSSION POINT
Importance of scientific rigor and collaborative validation
J
Josephine Teo
7 arguments150 words per minute1690 words672 seconds
Argument 1
Need for guardrails in AI agent architecture to prevent misuse and unintended behavior (Josephine Teo)
EXPLANATION
Teo stresses that as governments experiment with AI agents, they must design clear safeguards within the agents’ decision‑making processes. Guardrails are required to limit misuse and ensure trustworthy operation.
EVIDENCE
She remarks that Singapore wants to be “very thoughtful about how these AI agent systems are being architected… Is there a way to put guardrails around it?” highlighting the need for protective measures [68-70].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Guardrails for agentic AI are called for in multiple sources, emphasizing their importance for safety [S21], [S22], and the ministerial insights [S1].
MAJOR DISCUSSION POINT
Architectural guardrails for AI agents
AGREED WITH
Yoshua Bengio, Adam Beaumont, Alondra Nelson
DISAGREED WITH
Adam Beaumont
Argument 2
Policymakers require thoughtful, targeted standards to avoid false promises while still reaping AI benefits (Josephine Teo)
EXPLANATION
Teo argues that regulations must be carefully crafted so they protect citizens without stifling innovation. Over‑ambitious or poorly targeted standards could give a false sense of security.
EVIDENCE
She notes that standards and regulations need to be “thoughtful… otherwise we give a false promise to our citizens… we must be thoughtful… when there is clarity we want to move quickly” [50-56].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need for thoughtful, well-targeted regulation is discussed in the policy guidance notes [S24] and the OECD coordination guide [S6].
MAJOR DISCUSSION POINT
Balanced, thoughtful regulation
Argument 3
AI serves both as a threat and a target in cybersecurity, with emerging bio‑security implications requiring coordinated response (Josephine Teo)
EXPLANATION
Teo highlights that AI can be used to launch attacks and can also be attacked itself, especially in multi‑agent contexts, creating compounded security challenges. Coordinated action is needed to address both dimensions.
EVIDENCE
She states “AI as a threat, AI as a target… particularly for multi-agent systems, those kinds of risks can easily go out of hand” and adds that AI is a threat in cybersecurity and bio-security, requiring cooperation [65-68][70-71].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
AI’s dual role as threat and target, especially in cyber- and bio-security contexts, is documented in analyses of AI-driven malware and data-poisoning attacks [S25], [S26], [S27] and reinforced in the ministerial briefing [S1].
MAJOR DISCUSSION POINT
Dual role of AI in cyber‑ and bio‑security
AGREED WITH
Adam Beaumont, Yoshua Bengio
Argument 4
Investment in testing frameworks, insurance schemes, and industry‑wide tooling to support safe deployment (Josephine Teo)
EXPLANATION
Teo proposes mechanisms such as insurance schemes and collaborative testing frameworks to incentivize safe AI development. She emphasizes that tooling and standards must evolve alongside the technology.
EVIDENCE
She references discussions at Davos about insurance schemes, the need for industry collaboration, and calls for continued research and pragmatic tooling, noting the shortcomings of current testing tools and the need for a roadmap [169-184].
MAJOR DISCUSSION POINT
Funding and tooling for safe AI deployment
AGREED WITH
Adam Beaumont
Argument 5
Achieving “sovereign AI” via self‑containment is unrealistic; multilateral principles are needed to stay competitive and safe (Josephine Teo)
EXPLANATION
Teo argues that trying to keep AI entirely within national borders creates a false sense of security and hampers progress. Instead, shared international principles are required for both competitiveness and safety.
EVIDENCE
She says “the idea that you get sovereign AI by confining everything… is not achievable… it cuts you off… we need multilateral principles” and links this to preserving sovereignty through cooperation [304-312].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The impracticality of isolationist AI strategies and the call for multilateral cooperation are highlighted in European sovereignty discussions [S20] and the ministerial insights on collaborative sovereignty [S1].
MAJOR DISCUSSION POINT
Limits of self‑contained AI sovereignty
AGREED WITH
Yoshua Bengio, Alondra Nelson
Argument 6
Statutory obligations should require platforms to remove harmful AI‑generated content
EXPLANATION
Teo describes a new Singapore law that imposes duties on services that host AI‑generated images targeting women and children, making them responsible for removing such content once notified.
EVIDENCE
She explains that the law imposes statutory obligations on services that make harmful content available, shifting responsibility from the generator to the platform and requiring removal upon notification [58-63].
MAJOR DISCUSSION POINT
Platform accountability for harmful AI content
Argument 7
AI can be employed as a defensive tool to counter AI‑driven threats
EXPLANATION
Teo notes that while AI poses threats, it can also be used to fight those threats, suggesting a dual role for AI in security strategies.
EVIDENCE
She states that AI is both a threat and a target, and that there is a need to cooperate to use AI as a tool to fight these threats [70-71].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The potential to use AI defensively against AI-based threats is mentioned in cybersecurity analyses of AI’s role in security operations [S27].
MAJOR DISCUSSION POINT
Using AI defensively against AI‑based threats
A
Alondra Nelson
6 arguments193 words per minute1537 words476 seconds
Argument 1
The report provides evidence‑based grounding without prescribing specific policies, enabling stronger political will (Alondra Nelson)
EXPLANATION
Nelson praises the report for staying within the evidence‑informed domain, avoiding direct policy prescriptions, and thereby giving policymakers a solid factual base to build political resolve.
EVIDENCE
She notes the report “does a really good job of exactly not crossing that line… not prescribing… evidence-based… allows stronger political spines” and emphasizes that it offers more than anecdotal journalism [88-102].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The report’s non-prescriptive, evidence-based stance is highlighted in the ministerial summary and the OECD guide to AI safety coordination [S1], [S6].
MAJOR DISCUSSION POINT
Evidence‑based, non‑prescriptive reporting
AGREED WITH
Yoshua Bengio, Lee Tiedrich
Argument 2
Call for standardized evaluation methods and collective action to prevent fragmented, inconsistent assessments (Alondra Nelson)
EXPLANATION
Nelson warns that without common standards, researchers will produce divergent evaluations, leading to a collective‑action problem. She advocates for agreed‑upon methods to ensure consistency.
EVIDENCE
She expresses concern about “a collective action problem… each doing their own different kind of evaluation” and calls for “a few choices about… a standard” [256-259].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
A collective-action problem around divergent evaluations and the call for common standards are discussed in the ministerial briefing [S1] and the OECD coordination guide [S6].
MAJOR DISCUSSION POINT
Need for standardized evaluation
DISAGREED WITH
Adam Beaumont
Argument 3
Systemic, compounding risks (e.g., loss of autonomy, manipulation, job displacement) threaten social cohesion and democratic health (Alondra Nelson)
EXPLANATION
Nelson frames AI risks as systemic, where multiple harms interact and erode social cohesion, autonomy, and democratic stability. She stresses that these intertwined risks are more dangerous than isolated catastrophic events.
EVIDENCE
She describes systemic risk as “compounding… loss of human autonomy, sycophancy, job loss, anxiety… threatens social cohesion and democracy” [192-199][200-203].
MAJOR DISCUSSION POINT
Compounding systemic AI risks
Argument 4
The report’s inclusion of systemic risk perspectives broadens focus beyond isolated catastrophic scenarios (Alondra Nelson)
EXPLANATION
Nelson highlights that the report expands the risk narrative to include systemic and societal dimensions, not just extreme catastrophic outcomes, providing a more comprehensive view of AI safety.
EVIDENCE
She states that the report “continues to be anchored in that broader aperture of risk” and that it “think at a 30,000-foot level… compounding risks” beyond singular catastrophes [191-197][202-206].
MAJOR DISCUSSION POINT
Broadening risk scope
Argument 5
Global verification standards and shared research funding are essential for coordinated safety efforts (Alondra Nelson)
EXPLANATION
Nelson argues that worldwide verification mechanisms and pooled funding are needed to create a common safety infrastructure and to sustain research that underpins policy decisions.
EVIDENCE
She mentions that the report “allows stronger political spines” and calls for “global community… funding of the creation of more evidence” and later cites the need for “public sector… common good… research funding” [102-103][263-266].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need for international verification mechanisms and pooled research funding is emphasized in the ministerial insights [S1], the IAEA analogy on verification challenges [S19], and the OECD coordination framework [S6].
MAJOR DISCUSSION POINT
International verification and funding
AGREED WITH
Yoshua Bengio, Josephine Teo
Argument 6
New democratic institutions are needed to govern AI safety effectively
EXPLANATION
Nelson argues that the emergence of advanced AI requires the creation of novel democratic bodies and mechanisms to ensure transparent, accountable governance.
EVIDENCE
She recounts her remarks at Bletchley Park that we will need new democratic institutions for this moment, citing the report itself as one such institution and emphasizing the importance of a global ground truth on AI risks [81-84].
MAJOR DISCUSSION POINT
Institutional innovation for AI governance
A
Adam Beaumont
8 arguments176 words per minute1145 words388 seconds
Argument 1
Autonomy amplifies cybersecurity and bio‑security threats, especially when agents are combined with dual‑use capabilities (Adam Beaumont)
EXPLANATION
Beaumont points out that autonomous AI agents, especially those with dual‑use potential in cybersecurity and genetics, can create powerful new threats when combined, accelerating risk profiles.
EVIDENCE
He notes that “both of those are very dual use… we have seen rapid development… confluence of some of these risks when combined with more autonomy, particularly in the genetic AI scenarios” [119-124] and also references AI as a threat and target [65-67].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Dual-use risks of autonomous agents in cyber and genetic domains are documented in reports on AI-driven malware and data-poisoning attacks [S25], [S26], [S27] and echoed in the ministerial discussion of dual-use threats [S1].
MAJOR DISCUSSION POINT
Dual‑use risk of autonomous agents
AGREED WITH
Josephine Teo, Yoshua Bengio
Argument 2
Development of a third‑party evaluation ecosystem (e.g., inspection frameworks, auditors) to bring rigor to regulatory decisions (Adam Beaumont)
EXPLANATION
Beaumont describes efforts to create an ecosystem of independent evaluators, using tools like the open‑source Inspect framework, to provide transparent, rigorous assessments for policymakers.
EVIDENCE
He mentions the open-sourced Inspect framework used widely and expresses a desire to “grow a wide kind of ecosystem of third-party evaluators” that bring independence and scientific rigour [214-216][221-225].
MAJOR DISCUSSION POINT
Third‑party evaluation ecosystem
DISAGREED WITH
Alondra Nelson
Argument 3
AC’s pre‑ and post‑deployment testing, red‑team exercises, and the open‑source Inspect framework address current evaluation gaps (Adam Beaumont)
EXPLANATION
Beaumont outlines the AI Security Institute’s comprehensive testing pipeline, including pre‑deployment checks, post‑deployment monitoring, red‑team challenges, and the publicly available Inspect framework, to fill evaluation shortcomings.
EVIDENCE
He details “pre-deployment testing… post-deployment… model cards… red-team… inspect framework open sourced and used extensively” [124-127][214-215].
MAJOR DISCUSSION POINT
Comprehensive testing pipeline
Argument 4
Clear definition of measurement goals and transparent communication are crucial for reliable evaluations (Adam Beaumont)
EXPLANATION
Beaumont stresses that evaluators must first specify what they intend to measure and then ensure their methods align, while communicating uncertainties openly to avoid misleading results.
EVIDENCE
He advises “be really clear what is it you are trying to measure and make sure your evaluation is actually getting after the thing… transparent communication” [214-215].
MAJOR DISCUSSION POINT
Goal‑oriented, transparent evaluation
AGREED WITH
Lee Tiedrich, Josephine Teo
Argument 5
Dual‑use nature of AI in cybersecurity and genetics heightens the potential for large‑scale harm (Adam Beaumont)
EXPLANATION
Beaumont emphasizes that AI technologies can be repurposed for both offensive cyber operations and biological applications, creating a heightened risk of widespread damage.
EVIDENCE
He links “cybersecurity and biological capabilities… dual use… confluence of risks… particularly in genetic AI scenarios” [119-124].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Large-scale dual-use threats are highlighted in analyses of AI-powered ransomware and bio-security concerns [S26].
MAJOR DISCUSSION POINT
Large‑scale dual‑use threats
Argument 6
Cross‑sector collaboration (government, industry, academia, civil society) and regulatory sandboxes foster cooperative risk mitigation (Adam Beaumont)
EXPLANATION
Beaumont advocates for multi‑stakeholder partnerships, including regulatory sandboxes and joint funding programmes, as practical ways to test and refine AI safety measures.
EVIDENCE
He mentions “collaboration… regulatory sandboxes… policy lab… joint funding programmes… early in the journey” as mechanisms to bring together diverse actors [267-277].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Cross-sector partnerships and regulatory sandboxes are recommended in the OECD guide to AI safety coordination [S6] and the UN OEWG roundtable on ICT security capacity building [S28].
MAJOR DISCUSSION POINT
Collaborative risk‑mitigation mechanisms
DISAGREED WITH
Josephine Teo
Argument 7
Targeted grant‑making can expand security research capacity for AI
EXPLANATION
Beaumont highlights that the AI Security Institute uses grant programmes to increase investment in security research, aiming to raise the overall level of expertise and safeguards.
EVIDENCE
He mentions that the institute uses grant making to raise the level of investment in the space and to strengthen security research capabilities [214-215].
MAJOR DISCUSSION POINT
Funding mechanisms to boost AI security research
Argument 8
Cyber‑range environments provide realistic evaluation of AI security capabilities
EXPLANATION
Beaumont points out that evaluating AI models using cyber‑range scenarios, rather than simple capture‑the‑flag exercises, yields more accurate assessments of how AI can be used in real‑world cyber operations.
EVIDENCE
He explains that the institute is adapting evaluations to use cyber ranges instead of just capture-the-flag type scenarios to better capture model capabilities [124-125].
MAJOR DISCUSSION POINT
Realistic testing environments for AI security evaluation
L
Lee Tiedrich
2 arguments130 words per minute1147 words526 seconds
Argument 1
Practical tooling for businesses is essential; likened to IKEA’s safety‑tested furniture, users should not bear the burden of safety verification (Lee Tiedrich)
EXPLANATION
Lee raises the need for user‑friendly tools that allow companies to adopt AI safely, comparing the desired assurance to IKEA’s tested furniture that consumers can trust without conducting their own safety tests.
EVIDENCE
He asks about tooling for organizations [155-159] and Josephine responds with the IKEA analogy, describing how furniture is “tested… you don’t have to impose safety conditions on your own” [161-166].
MAJOR DISCUSSION POINT
Need for accessible safety tooling
AGREED WITH
Josephine Teo, Adam Beaumont
Argument 2
Bridging scientific findings to actionable policy tools should avoid prescribing specific actions
EXPLANATION
Lee stresses that the report’s role is to inform policymakers with evidence‑based insights without dictating exact policies, thereby supporting informed decision‑making while preserving policy autonomy.
EVIDENCE
He notes that the report is intended to inform policymakers and the broader community and intentionally does not take the next step of advising policymakers on what to do, highlighting the importance of providing evidence without prescribing actions [76-77].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The report’s intent to inform without prescribing is noted in the ministerial briefing and the OECD coordination guide, which stress evidence-based policy support [S1], [S6].
MAJOR DISCUSSION POINT
Science‑to‑policy translation without prescriptive recommendations
Agreements
Agreement Points
Need for clear guardrails, standards and third‑party evaluation for autonomous AI agents
Speakers: Yoshua Bengio, Josephine Teo, Adam Beaumont, Alondra Nelson
Increased risk due to reduced human oversight as AI agents gain more autonomy (Yoshua Bengio) Need for guardrails in AI agent architecture to prevent misuse and unintended behavior (Josephine Teo) Development of a third‑party evaluation ecosystem (Adam Beaumont) Call for standardized evaluation methods and collective action to prevent fragmented assessments (Alondra Nelson)
All four speakers stress that as AI agents become more autonomous they must be bounded by well-defined safeguards, transparent standards and independent evaluation mechanisms to avoid unsafe behaviour. Yoshua notes the loss of oversight and the need for reliable technology before deployment [22-29][30-31]; Josephine explicitly asks for guardrails around agent decision-making [68-70]; Adam calls for clear measurement goals, transparent communication and a growing ecosystem of independent auditors [214-216][221-225]; Alondra warns of a collective-action problem and urges a few agreed-upon standards [256-259].
POLICY CONTEXT (KNOWLEDGE BASE)
This need aligns with the UN call for universal guardrails and common standards for AI safety [S49], reflects Minister Josephine Teo and Adam Beaumont’s endorsement of a third-party evaluation ecosystem and regulatory sandboxes [S42], and is supported by OECD discussions on sandboxes as regulatory experimentation tools [S40][S41]; clear regulatory guardrails are also cited as a catalyst for innovation [S47].
Scientific rigor and evidence‑based, non‑prescriptive reporting to support policy making
Speakers: Yoshua Bengio, Lee Tiedrich, Alondra Nelson
Scientific rigor and collaborative peer review are essential to avoid false claims in AI safety reporting (Yoshua Bengio) Bridging scientific findings to actionable policy tools should avoid prescribing specific actions (Lee Tiedrich) The report provides evidence‑based grounding without prescribing specific policies, enabling stronger political will (Alondra Nelson)
The three speakers agree that the report should remain a rigorous, evidence-based resource that does not dictate policy choices. Yoshua stresses humility, honesty and group review to prevent false claims [138-152]; Lee notes the report’s role is to inform, not advise, policymakers [76-77]; Alondra praises the report for staying within the evidence-informed domain and not crossing the line into prescription [88-102].
POLICY CONTEXT (KNOWLEDGE BASE)
The 40-member Scientific Panel’s mandate to produce policy-relevant but non-prescriptive reports provides a direct precedent for this approach [S44]; a global human-rights AI governance roadmap similarly stresses evidence-based, flexible guidance [S38]; and clear, evidence-based regulatory guidance is highlighted as reducing uncertainty for organisations [S47].
Dual‑use nature of AI creates heightened cybersecurity and bio‑security threats, especially when combined with autonomy
Speakers: Josephine Teo, Adam Beaumont, Yoshua Bengio
AI serves both as a threat and a target in cybersecurity, with emerging bio‑security implications requiring coordinated response (Josephine Teo) Autonomy amplifies cybersecurity and bio‑security threats, especially when agents are combined with dual‑use capabilities (Adam Beaumont) Increased risk due to reduced human oversight as AI agents gain more autonomy (Yoshua Bengio)
All three highlight that autonomous AI systems can be weaponised (threat) and also become vulnerable (target), magnifying cyber and bio risks. Josephine describes AI as both a threat and a target, especially for multi-agent systems [65-68][70-71]; Adam points to rapid capability growth and the confluence of dual-use risks with autonomy [119-124]; Yoshua links autonomy to reduced oversight and emerging safety concerns [22-31].
POLICY CONTEXT (KNOWLEDGE BASE)
UN Security Council discussions note that dual-use technologies pose notable security risks [S55]; AI-cybersecurity analyses emphasize inclusive governance to mitigate such threats [S51]; recent reports document a surge in malicious cyber activity targeting critical infrastructure, underscoring the urgency [S57]; and the dual-use dilemma is highlighted in security-focused forums [S56].
International cooperation rather than isolationist “walls” is essential for AI sovereignty and safety
Speakers: Yoshua Bengio, Josephine Teo, Alondra Nelson
Sovereignty should emphasize collaborative agreements and verification mechanisms rather than isolationist “walls” (Yoshua Bengio) Achieving “sovereign AI” via self‑containment is unrealistic; multilateral principles are needed to stay competitive and safe (Josephine Teo) Global verification standards and shared research funding are essential for coordinated safety efforts (Alondra Nelson)
The speakers converge on the view that AI governance must be multilateral. Yoshua argues sovereignty means partnerships and international agreements, not walls [292-300]; Josephine says self-containment is unattainable and stresses multilateral principles [304-312]; Alondra calls for worldwide verification mechanisms and pooled funding to support safety research [102-103][263-266].
POLICY CONTEXT (KNOWLEDGE BASE)
Debates on sovereign AI stress cooperation over fragmentation and promote international collaboration [S58]; the Global AI Policy Framework advocates open sovereignty and coordinated standards [S59]; multi-stakeholder approaches are repeatedly endorsed across UN, OECD and other bodies [S37][S38][S39]; and low public trust in sole government regulation reinforces the case for cooperative models [S40].
Provision of practical, user‑friendly tooling for businesses and organisations to adopt AI safely
Speakers: Lee Tiedrich, Josephine Teo, Adam Beaumont
Practical tooling for businesses is essential; likened to IKEA’s safety‑tested furniture, users should not bear the burden of safety verification (Lee Tiedrich) Investment in testing frameworks, insurance schemes, and industry‑wide tooling to support safe deployment (Josephine Teo) Clear definition of measurement goals and transparent communication are crucial for reliable evaluations (Adam Beaumont)
All three stress that end-users need ready-made, trustworthy tools rather than having to design safety checks themselves. Lee asks how tooling can be advanced for organisations [155-159]; Josephine uses the IKEA analogy to illustrate the need for pre-tested solutions and mentions insurance and testing frameworks [161-166][169-184]; Adam emphasizes clear measurement objectives and transparent evaluation tools [214-215].
POLICY CONTEXT (KNOWLEDGE BASE)
Toolkits for AI risk mitigation have been developed to give businesses actionable guidance [S53]; clear, user-friendly regulatory guidance is shown to accelerate safe innovation [S47]; the need to support small-scale commercial activities outside formal oversight is highlighted in discussions of DIY AI science [S50]; and sandboxes provide hands-on environments for testing tools [S40].
Funding mechanisms (grant‑making, insurance schemes) are needed to expand AI safety research and deployment
Speakers: Josephine Teo, Adam Beaumont
Investment in testing frameworks, insurance schemes, and industry‑wide tooling to support safe deployment (Josephine Teo) Targeted grant‑making can expand security research capacity for AI (Adam Beaumont)
Both speakers highlight financial instruments to boost safety capacity. Josephine discusses insurance schemes and the need for pragmatic funding for testing tools [169-184]; Adam notes the institute’s grant-making programme to raise investment in security research [214-215].
POLICY CONTEXT (KNOWLEDGE BASE)
Policy papers on AI in Africa call for financing schemes, including credit and insurance mechanisms, to support safe AI deployment [S54]; UN discussions on accelerating the SDGs stress the importance of funding evidence-based interventions, which includes AI safety research [S46].
Similar Viewpoints
Both see autonomy of AI agents as a source of new safety challenges that require explicit safeguards before wide deployment [22-29][68-70].
Speakers: Yoshua Bengio, Josephine Teo
Increased risk due to reduced human oversight as AI agents gain more autonomy (Yoshua Bengio) Need for guardrails in AI agent architecture to prevent misuse and unintended behavior (Josephine Teo)
Both advocate for a coordinated, standards‑based evaluation ecosystem to ensure rigorous, comparable safety assessments [256-259][214-216].
Speakers: Alondra Nelson, Adam Beaumont
Call for standardized evaluation methods and collective action to prevent fragmented, inconsistent assessments (Alondra Nelson) Development of a third‑party evaluation ecosystem (Adam Beaumont)
Both emphasize that the report should inform policymakers without dictating policy choices, preserving political autonomy while supplying evidence [76-77][88-102].
Speakers: Lee Tiedrich, Alondra Nelson
Bridging scientific findings to actionable policy tools should avoid prescribing specific actions (Lee Tiedrich) The report provides evidence‑based grounding without prescribing specific policies, enabling stronger political will (Alondra Nelson)
Unexpected Consensus
Both a government minister (Josephine Teo) and a security‑industry leader (Adam Beaumont) endorse the creation of a broad, third‑party evaluation ecosystem and regulatory sandboxes as a way to manage AI risk
Speakers: Josephine Teo, Adam Beaumont
Investment in testing frameworks, insurance schemes, and industry‑wide tooling to support safe deployment (Josephine Teo) Development of a third‑party evaluation ecosystem (Adam Beaumont)
While Josephine’s focus is on policy and industry collaboration, she nonetheless supports mechanisms (insurance, testing frameworks) that resemble the independent evaluation infrastructure championed by Adam, indicating a cross-sector convergence that was not explicitly anticipated. Both refer to practical, third-party tools and sandbox-type approaches to ensure safety [169-184][267-277].
POLICY CONTEXT (KNOWLEDGE BASE)
Minister Teo explicitly discussed guardrails for multi-agent systems and Beaumont supported a third-party evaluation ecosystem at the AI Safety Global Level dialogue [S42]; sandboxes are positioned as part of broader regulatory experimentation frameworks [S40]; and global assurance initiatives stress inclusive, third-party monitoring to build trust [S43].
Overall Assessment

The panel shows strong convergence on four main themes: (1) the urgent need for guardrails, standards and independent evaluation of autonomous AI agents; (2) the importance of scientific rigor and evidence‑based, non‑prescriptive reporting; (3) recognition of dual‑use cyber‑ and bio‑security threats amplified by autonomy; (4) the necessity of international cooperation and multilateral frameworks for AI sovereignty. Additional consensus appears around practical tooling for businesses and the role of targeted funding mechanisms.

High consensus – the speakers from academia, government, and industry largely agree on the problem definition and on broad strategic directions, though they differ on implementation details. This alignment suggests that forthcoming policy initiatives can draw on a shared understanding of risk, the need for standards, and the value of collaborative, evidence‑driven approaches.

Differences
Different Viewpoints
Extent of policy guidance that the safety report should provide
Speakers: Yoshua Bengio, Lee Tiedrich, Alondra Nelson
Scientists should offer scientifically grounded policy options that outline consequences without dictating choices (Yoshua Bengio) The report intentionally does not take the next step of advising policymakers on what to do (Lee Tiedrich) The report does a really good job of exactly not crossing that line … not prescribing … evidence‑informed (Alondra Nelson)
Yoshua argues that the report should include a middle layer of scientifically-grounded policy options that spell out likely outcomes while stopping short of direct recommendations, whereas Lee and Alondra stress that the report must remain strictly non-prescriptive and avoid any policy advice. This creates a clear split on how far the scientific assessment should go toward guiding policy decisions. [244-250][76-77][88-102]
POLICY CONTEXT (KNOWLEDGE BASE)
The Scientific Panel’s mandate for policy-relevant but non-prescriptive reporting illustrates the tension over how much guidance to embed [S44]; multi-stakeholder AI policy roadmaps emphasize evidence-based yet flexible guidance [S38]; and clear regulatory guidance is argued to reduce uncertainty and spur innovation [S47].
Who should conduct AI evaluation and what mechanisms should be used
Speakers: Adam Beaumont, Alondra Nelson
Development of a third‑party evaluation ecosystem (e.g., inspection frameworks, auditors) to bring rigor to regulatory decisions (Adam Beaumont) Call for standardized evaluation methods and collective action to prevent fragmented, inconsistent assessments (Alondra Nelson)
Adam promotes building an ecosystem of independent third-party evaluators supported by open-source tools such as the Inspect framework and regulatory sandboxes, while Alondra warns that without agreed-upon standards the field will suffer a collective-action problem and calls for a few common evaluation approaches. The disagreement centres on whether to prioritize an open, pluralistic evaluator market or to first establish shared standards. [214-216][221-225][256-259]
POLICY CONTEXT (KNOWLEDGE BASE)
OECD sandboxes involve third-party evaluators and stress mechanisms to avoid regulatory capture [S41]; multi-stakeholder governance models propose shared evaluation responsibilities among government, private sector and civil society [S39][S51]; and OECD discussions outline a broad evaluation ecosystem involving diverse actors [S40].
Timing and urgency of implementing guardrails for AI agents
Speakers: Josephine Teo, Adam Beaumont
Need for guardrails in AI agent architecture to prevent misuse and unintended behavior (Josephine Teo) Cross‑sector collaboration, regulatory sandboxes and early‑stage research are needed before robust guardrails can be set (Adam Beaumont)
Josephine calls for immediate, thoughtful guardrails around AI agents and mentions industry-wide tooling and insurance schemes as near-term solutions, whereas Adam stresses that the field is still in its early stages and that robust safeguards should follow further research, pilot sandboxes, and ecosystem development. This reflects a disagreement on how quickly concrete safeguards should be deployed. [68-70][267-277]
POLICY CONTEXT (KNOWLEDGE BASE)
UNGA statements call for immediate universal guardrails and common standards [S49]; Minister Teo highlighted the pressing need for guardrails for autonomous systems [S42]; and AI safety literature stresses that lack of guardrails can lead to risky outcomes, urging swift action [S48].
Role of government versus multi‑stakeholder approaches in AI evaluation
Speakers: Adam Beaumont, Josephine Teo
Cross‑sector collaboration (government, industry, academia, civil society) and regulatory sandboxes foster cooperative risk mitigation (Adam Beaumont) Policymakers must craft thoughtful, targeted standards and regulations to avoid false promises while still reaping AI benefits (Josephine Teo)
Adam argues that evaluation should be a shared responsibility across all sectors, including governments, industry, and civil society, while Josephine emphasizes a government-led, standards-focused approach to ensure safety. The unexpected tension lies in two senior AI-safety figures advocating different primary actors for evaluation and regulation. [267-269][50-56]
POLICY CONTEXT (KNOWLEDGE BASE)
IGF sessions underline the preference for multi-stakeholder approaches over single-entity solutions [S39][S51]; sandboxes are designed to balance government oversight with private-sector innovation [S40][S41]; and a global human-rights AI governance framework advocates inclusive, multi-stakeholder governance [S38].
Unexpected Differences
Urgency of implementing AI‑agent guardrails versus continuing research
Speakers: Josephine Teo, Adam Beaumont
Need for guardrails in AI agent architecture to prevent misuse and unintended behavior (Josephine Teo) Cross‑sector collaboration, regulatory sandboxes and early‑stage research are needed before robust guardrails can be set (Adam Beaumont)
It is surprising that two senior AI-safety leaders differ on how quickly concrete safeguards should be rolled out: Josephine pushes for immediate, industry-wide guardrails, while Adam argues the field is still too early for firm safeguards and should focus on research and pilot sandboxes first. [68-70][267-277]
POLICY CONTEXT (KNOWLEDGE BASE)
AI safety panels note the need to deploy guardrails while research continues, reflecting a dual-track approach [S44]; evidence that clear guardrails can accelerate innovation while research proceeds is highlighted in regulatory discussions [S47]; and calls for immediate safeguards alongside ongoing research are voiced in AI safety forums [S48].
Primary actor for AI evaluation (government‑led vs. multi‑stakeholder)
Speakers: Adam Beaumont, Josephine Teo
Cross‑sector collaboration, regulatory sandboxes and third‑party evaluation ecosystem (Adam Beaumont) Policymakers must craft thoughtful, targeted standards and regulations (Josephine Teo)
While both aim for safe AI, it is unexpected that they diverge on who should lead the evaluation effort: Adam envisions a shared, multi-stakeholder ecosystem, whereas Josephine places the government at the centre of setting mandatory standards. [267-269][50-56]
POLICY CONTEXT (KNOWLEDGE BASE)
Discussions on sovereign AI stress the importance of international, multi-stakeholder evaluation rather than sole government control [S58][S59]; regulatory sandboxes illustrate shared evaluation responsibilities among diverse actors [S40]; and inclusive governance frameworks propose joint actor models to prevent capture and ensure fairness [S51].
Overall Assessment

The discussion revealed moderate disagreement centred on how far scientific reports should go in guiding policy, the preferred architecture of AI evaluation (standardised versus third‑party ecosystem), and the timing of implementing guardrails for autonomous agents. While participants share the overarching goal of safe AI deployment, they diverge on the mechanisms, actors and urgency required to achieve it.

The level of disagreement is moderate: it does not fracture the dialogue but highlights distinct strategic preferences that could affect coordination, speed of regulation and the design of evaluation infrastructures. If unresolved, these differences may lead to fragmented standards, delayed safeguards, and uneven international cooperation in AI governance.

Partial Agreements
Both agree that AI agents must be made safe before widespread adoption, but Josephine focuses on embedding technical guardrails within the agents themselves, whereas Adam emphasizes external evaluation mechanisms, sandboxes and third‑party audits as the path to safety. [68-70][267-277]
Speakers: Josephine Teo, Adam Beaumont
Need for guardrails in AI agent architecture to prevent misuse and unintended behavior (Josephine Teo) Cross‑sector collaboration, regulatory sandboxes and third‑party evaluation ecosystem to ensure safe deployment (Adam Beaumont)
Both stress that the safety report should remain non‑prescriptive, providing evidence without direct policy prescriptions, thereby supporting policymakers while preserving their autonomy. [76-77][88-102]
Speakers: Lee Tiedrich, Alondra Nelson
The report is intended to inform policymakers and intentionally does not advise what to do (Lee Tiedrich) The report does a really good job of exactly not crossing that line … not prescribing … evidence‑informed (Alondra Nelson)
Takeaways
Key takeaways
Autonomous AI agents are rapidly gaining capabilities, reducing human oversight and creating new safety challenges, especially when multi‑agent systems interact. The report provides an evidence‑based scientific assessment of AI risks without prescribing specific policies, aiming to strengthen political will and inform policymakers. Jagged performance of general‑purpose models requires per‑capability risk and intention assessment rather than a single safety metric. Systemic and compounding risks—loss of autonomy, manipulation, job displacement, and threats to democratic cohesion—must be considered alongside catastrophic scenarios. AI acts both as a threat (e.g., in cyber‑operations, bio‑security) and as a target of attacks, amplifying dual‑use concerns. International cooperation and a re‑interpretation of sovereignty—favoring collaborative agreements and verification mechanisms over isolation—are essential for effective AI safety governance. A robust, third‑party evaluation ecosystem (including standards, audit frameworks, and open‑source tools like the Inspect framework) is needed to translate scientific findings into practice. Practical tooling for businesses (analogous to safety‑tested IKEA furniture) is critical so end‑users are not burdened with verifying AI safety themselves.
Resolutions and action items
Develop and promote third‑party evaluation frameworks and auditors, building on AC’s open‑source Inspect framework. Encourage governments to create thoughtful, targeted standards and regulatory sandboxes that balance innovation with safety. Invest in research on multi‑agent systems, cybersecurity, and bio‑security risks, including funding mechanisms and insurance schemes. Continue collaborative work to update safety research priorities, with Singapore focusing on responsible AI and multi‑agent system testing. Produce scientifically grounded policy option briefs that outline consequences without prescribing specific choices. Facilitate cross‑sector partnerships (government, industry, academia, civil society) to co‑design evaluation standards and verification protocols.
Unresolved issues
Specific standards and guardrails for AI agents’ autonomy and credential access remain undefined. How to effectively label or watermark AI‑generated harmful content and enforce compliance across platforms. The precise mechanism for international verification of AI safety agreements and the role of sovereign regulation. Standardized metrics for measuring intent and capability in jagged AI models are still lacking. Allocation of responsibility among governments, industry, and individuals for AI safety and cybersecurity remains unclear. Longitudinal, real‑world evaluation methods to keep pace with rapid model improvements have not been finalized.
Suggested compromises
Adopt targeted, thoughtful regulations that protect citizens while preserving AI‑driven economic benefits. Use regulatory sandboxes and policy labs to pilot safety measures before wide deployment. Combine mandatory safety standards with industry‑led insurance and incentive schemes to share risk. Balance sovereignty concerns by pursuing multilateral safety agreements rather than isolationist policies. Implement a phased approach: develop scientific assessments → create policy option briefs → allow governments to choose among vetted options.
Thought Provoking Comments
Having AIs that are more autonomous means less oversight. Agents will be given credentials and Internet access, and we are seeing them interact with each other, which is concerning.
Highlights a new class of risk—autonomous multi‑agent systems that operate without a human in the loop—shifting the conversation from current chatbot safety to future systemic threats.
Prompted Lee to raise AI‑literacy concerns, led Josephine to discuss guardrails for agents, and set the agenda for later focus on multi‑agent system risks throughout the panel.
Speaker: Yoshua Bengio
Singapore does not own aircraft technologies, but we must ensure safety of manufacturing, maintenance, and air traffic management. Likewise, we introduced a law imposing obligations on services that host harmful AI‑generated images.
Provides a concrete policy analogy that links traditional safety regulation to AI, showing how targeted legislation can address harms without stifling innovation.
Served as a practical example of turning scientific insights into enforceable standards, influencing later discussion on mandatory guardrails, tooling, and the need for clear regulatory pathways.
Speaker: Josephine Teo
We are going to need new democratic institutions for this moment; the report provides a ground truth for the global community about AI risks.
Frames AI safety as a governance challenge requiring new institutional structures, emphasizing the report’s role as a shared evidence base rather than a policy prescription.
Set the overarching framing for the panel, legitimized the report’s scope, and underpinned later remarks about systemic risk, multi‑sector collaboration, and the need for evidence‑based policy.
Speaker: Alondra Nelson
We do pre‑deployment testing, post‑deployment testing, red‑team exercises, and we released the open‑source Inspect framework for third‑party evaluators.
Moves the discussion from abstract risk to concrete, actionable evaluation infrastructure, demonstrating how the AI security community is already building tools for systematic assessment.
Catalyzed the conversation about building an evaluation ecosystem, inspired suggestions for third‑party auditors, and linked back to Alondra’s call for standardized evidence.
Speaker: Adam Beaumont
If AI has jagged capabilities, we could have dangerous abilities in some areas while being weak in others; we need careful scientific evaluation per scale per ability, including intention.
Challenges the simplistic AGI narrative, introduces the concept of capability‑specific risk assessment, and stresses scientific rigor and humility.
Led Lee to ask about how jagged performance affects evaluation science, prompted Alondra to stress the need for rigorous, non‑anecdotal evidence, and deepened the technical discussion of risk metrics.
Speaker: Yoshua Bengio
We must look at compounding systemic risks—loss of autonomy, manipulation, job anxiety—that together threaten democracy, not just isolated catastrophic events.
Broadens the risk lens from singular catastrophic scenarios to societal‑level systemic threats, linking AI safety to democratic health and social cohesion.
Shifted the tone from technical threats to societal impact, encouraging participants to consider policy measures that address multiple, interacting harms.
Speaker: Alondra Nelson
Think of AI like IKEA furniture that’s been tested; users shouldn’t have to impose safety themselves. We may need mandatory standards and insurance schemes to give assurance.
Uses a relatable analogy to argue for market‑based safety guarantees and proposes insurance as a mechanism to align incentives.
Inspired dialogue on practical tooling, standards, and the role of government versus industry, feeding into later discussion on evaluation frameworks and third‑party certification.
Speaker: Josephine Teo
There should be a step between the scientific report and policy decisions: scientifically grounded policy options that outline possible actions and consequences without prescribing a specific choice.
Identifies a missing bridge between evidence and policy, proposing a neutral, option‑based approach that respects political pluralism while grounding decisions in science.
Guided the conversation toward how to translate evidence into actionable policy, influencing suggestions about standardization, policy labs, and collaborative evaluation efforts.
Speaker: Yoshua Bengio
Sovereignty isn’t about building walls; it’s about partnerships, international agreements on safety, and verification mechanisms across borders.
Reframes AI sovereignty as cooperative rather than isolationist, emphasizing the necessity of global governance structures.
Shifted the final segment toward global collaboration, echoed by Josephine’s remarks, and reinforced the panel’s consensus that AI safety requires international coordination.
Speaker: Yoshua Bengio
Overall Assessment

The discussion was steered by a handful of high‑impact remarks that repeatedly reframed the problem space—from Yoshua’s warning about autonomous agents and jagged capabilities, to Alondra’s call for new democratic institutions and systemic‑risk perspective, and Josephine’s concrete policy analogies and tooling proposals. Each of these comments opened new sub‑threads (AI literacy, regulatory guardrails, evaluation ecosystems, and global governance) and prompted other panelists to expand, challenge, or operationalize the ideas. Collectively, they moved the conversation from a broad safety report overview to a nuanced, multi‑dimensional agenda that links technical risk assessment, rigorous scientific standards, practical regulatory tools, and international cooperation.

Follow-up Questions
How can we develop practical tooling and standards to help companies and organizations deploy AI safely, similar to the IKEA analogy?
Bridging the gap between scientific insights and operational safeguards is needed so SMEs can adopt AI without having to build safety measures themselves.
Speaker: Josephine Teo
What should the evaluation ecosystem look like? Who should conduct AI safety evaluations – governments, industry, or third‑party auditors?
Clarifying governance and responsibility for independent, rigorous AI evaluations is essential for trustworthy deployment and regulatory compliance.
Speaker: Lee Tiedrich, Josephine Teo, Yoshua Bengio, Alondra Nelson, Adam Beaumont
How can we create scientifically grounded policy options that translate scientific findings into actionable choices without prescribing specific policies?
Policymakers need evidence‑based option sets that outline possible actions and consequences, filling the gap between the report’s science and concrete policy decisions.
Speaker: Yoshua Bengio
How to address AI as both a cybersecurity threat and a target, especially with multi‑agent systems?
Dual‑use risks where AI can be used to attack or be attacked require focused research on defenses, safeguards, and mitigation strategies.
Speaker: Josephine Teo, Adam Beaumont
What are the emergent risks from interactions among autonomous AI agents?
Agents that can act autonomously and interact with each other may produce unforeseen harmful dynamics, a research area that is still early and under‑studied.
Speaker: Yoshua Bengio
How to evaluate jagged capabilities of general‑purpose models across diverse tasks?
General‑purpose models exhibit uneven performance; evaluation frameworks must account for task‑specific strengths and weaknesses to assess risk accurately.
Speaker: Yoshua Bengio
What standards or upstream research funding models (e.g., similar to the Human Genome Project) should be established for AI safety?
Dedicated budget and standardized research agendas can anticipate and mitigate risks before deployment, mirroring successful models from genetics.
Speaker: Alondra Nelson
What mechanisms (e.g., international agreements, verification technologies) are needed to ensure AI safety across borders?
AI risks are transnational; collaborative frameworks and verification tools are required to uphold safety while respecting national sovereignty.
Speaker: Yoshua Bengio
Is watermarking or other labeling of AI‑generated content effective for mitigating harmful content?
Evaluating technical solutions like watermarking is crucial to curb the spread of harmful AI‑generated images and misinformation.
Speaker: Josephine Teo
Can insurance schemes incentivize safe AI development?
Financial mechanisms such as insurance could align developer incentives with safety standards, but their design and impact need study.
Speaker: Josephine Teo
How to fill the evidence gap and conduct longitudinal studies given rapid AI evolution?
Rapid capability changes outpace traditional research cycles; methods for continuous, real‑time evidence collection are needed.
Speaker: Adam Beaumont
How to improve AI literacy among the public to understand agent capabilities?
Public understanding of what AI agents can and cannot do is essential for responsible adoption and to prevent misuse.
Speaker: Lee Tiedrich
How do systemic, compounding risks affect democracy and social cohesion?
Combined risks (autonomy loss, manipulation, job displacement) could destabilize societies; research is needed on their aggregate impact.
Speaker: Alondra Nelson
What are the implications of rising business and national sovereignty trends on AI safety, and which safety concerns become most pressing?
Geopolitical shifts may reshape safety priorities; understanding these dynamics is important for global coordination and risk mitigation.
Speaker: Participant (question)

Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.