Setting the Rules_ Global AI Standards for Growth and Governance

20 Feb 2026 17:00h - 18:00h

Setting the Rules_ Global AI Standards for Growth and Governance

Session at a glanceSummary, keypoints, and speakers overview

Summary

The panel convened to discuss why AI standards are essential for aligning global AI development with safety, trust, and inclusive outcomes [4-5]. Participants defined standards variously as benchmarking methodologies that quantify risk uncertainty (ML Commons), safety guidelines that follow product release (Qualcomm), and normative governance frameworks that set global “good” baselines (Singapore government) [13-15][17-24][28-33]. Microsoft described its internal Responsible AI Standard as a tool to align product, engineering, and sales teams around common expectations, while urging external standards to create a shared language across the ecosystem [41-46]. OpenAI’s AI Standards Lead emphasized translating internal risk-management practices into a common language for customers and building interoperability to foster consumer trust [56-59]. The Indian Bureau of Standards highlighted standards as mechanisms for consumer confidence and quality assurance, linking national work to ISO’s SC42 efforts [61-64].


A recurring challenge identified was determining “what is good enough,” requiring consensus that includes industry, regulators, and broader stakeholders rather than a single perspective [96-103][108-114]. Panelists agreed that standards must be open and inclusive so smaller firms can adopt them without building proprietary processes, a point underscored by Qualcomm’s call for open governance models [169-185]. Measuring AI performance was described as developing taxonomies, datasets, and evaluators that estimate uncertainty under defined assumptions, recognizing that different sectors may accept different risk thresholds [251-259]. The group noted that standards should complement, not replace, regulation, providing technical expectations that regulators can reference even when formal rules are absent [77-90][214-223].


Looking ahead, participants expect a rise in certification schemes that signal consensus on “good enough” and modular, interoperable standards that can evolve with advancing models [336-340][388-392]. Future-proofing will rely on process-oriented standards that remain applicable as AI capabilities change, while specific evaluation methods will be updated over time [346-354]. Accelerated development of testing methodologies within ISO processes was cited as a priority to keep pace with rapid AI innovation [378-382]. The panel concluded that despite the nascent state of AI standardisation, collective action across industry, policy, and standards bodies is vital to build trust and enable responsible AI deployment [462-470].


Keypoints


Major discussion points


Standards are seen as essential for building trust and aligning “what good looks like” across the AI ecosystem.


The moderator frames the need to demystify standard-setting and stresses global cooperation and inclusion [4-5]. Panelists echo this: Rebecca describes benchmarking as a way to measure risk, a major adoption barrier [13-15]; Amanda explains Microsoft’s internal responsible-AI standard that aligns product, engineering and sales teams and calls for a common external language [41-46]; Chris notes that standards solve a collective-action problem and give legitimacy that pure industry or government actions lack [108-114]; Esther adds that standards translate risk-management practices into a language of consumer trust and interoperability [57-59].


Defining and measuring standards is technically difficult and requires consensus on “good enough.”


Rebecca points out the recurring question of what constitutes “good enough” and stresses the need for a broad, multi-stakeholder consensus [97-102]. Lee lists concrete focus areas-testing, transparency disclosures, and incident reporting-as early priorities for standardisation [80-90]. Rebecca further explains that benchmarking must provide a methodology, taxonomy and reference implementations, yet the core challenge is estimating uncertainty under defined assumptions [250-259]. Chris expands on this by distinguishing high-level process standards from technical benchmark standards that must evolve with model capabilities [291-298].


Inclusive, global cooperation among industry, policy makers, and standards bodies is crucial.


Bhushan highlights the mix of “standard setters and measurers” from industry and policy [34-36]. Kshitij describes India’s AI governance framework and the inter-connectedness of ISO, ML Commons, IEEE and other bodies, stressing the need to adapt global standards to local use-cases [207-212]. Etienne stresses that open, inclusive governance (e.g., ML Commons, ISO) lets smaller firms participate and rely on standards without building their own risk-management systems [176-185]. Lee notes that regulators can reference technical standards to define expectations, and even without regulation standards help differentiate trustworthy providers [214-218].


Future outlook: faster development, certification, interoperable modular standards, and addressing concrete challenges such as language bias.


Bhushan envisions certification that signals “good enough” and a move toward consensus-based benchmarks within two years [337-340]. Chris argues that process standards are relatively future-proof, while specific evaluations must be updated as models advance [347-354]. Lee reports ongoing work on testing methodologies that she hopes to push through ISO within a year [378-382]. Amanda calls for a modular, interoperable standards ecosystem that avoids reinventing the wheel for each new use-case [388-393]. Audience concerns about language bias are addressed by Esther (multilingual evaluation suites) and Etienne (need for reusable safety tests across languages) [441-447][452-460].


Overall purpose / goal of the discussion


The panel was convened to demystify AI standard-setting, explain why standards matter for safety, trust and market adoption, identify the technical and governance challenges in creating and measuring those standards, and outline a coordinated path forward that brings together industry, regulators, standards organisations and civil-society stakeholders.


Overall tone


The conversation maintains a collaborative and solution-focused tone throughout. Early remarks are introductory and aspirational, quickly moving to constructive exchanges about concrete challenges (testing, transparency, measurement). When discussing obstacles-such as the “good enough” dilemma, speed of standard development, and skill gaps-the tone becomes more urgent but remains collegial. The closing segment retains optimism, emphasizing shared commitment to faster, interoperable standards and collective action. No major shifts to conflict or negativity are observed; the tone stays professional, forward-looking, and inclusive.


Speakers

Speakers (from the provided list)


Bhushan Sethi – AI transformation consultant; moderator of the panel.


Rebecca Weiss – Executive Director of ML Commons, an AI benchmarking organization and engineering consortium. [S1]


Etienne Chaponniere – Vice President of Technical Standards at Qualcomm. [S4]


Lee Wan Sie – Singapore government official working on AI governance and policy; focuses on setting global AI norms. [S8]


Amanda Craig – Leader of Microsoft’s Public Policy team for AI and the Office of Responsible AI. [S2]


Joslyn Barnhart – Works at Google DeepMind on AI standards, governance, and policy. [S10]


Chris Meserole – Executive Director of the Frontier Model Forum, advancing Frontier AI safety and security. [S12]


Esther Tetruashvily – AI Standards Lead at OpenAI. [S6]


Kshitij Bathla – Representative of the Bureau of Indian Standards (BIS), National Standards Body of India; represents ISO ICJTC1 SC42. [S17]


Audience – Various audience members asking questions (e.g., on language bias, auditability, privacy governance). [S19][S20][S21]


Additional speakers (not in the provided list)


Juan C. – Unnamed panel participant referenced by Amanda Craig; contributed a comment on standards aligning around “what good looks like.”


Full session reportComprehensive analysis and detailed insights

Opening & Goal – Bhushan Sethi, an AI-transformation consultant, opened the session by stating that the panel’s aim was to demystify AI standard-setting, explore global cooperation, and define “what good looks like” for AI development [2-3][4-5][8-10].


Speaker definitions (in speaking order)


Rebecca Weiss (ML Commons) – Standards are benchmarking methodologies that define risk-measurement and provide technical artefacts for integration into development pipelines [13-15][250-261].


Etienne Chaponnière (Qualcomm) – Unlike telecom standards, which are mandatory for product shipment, AI safety standards usually follow product releases and focus on safety [16-24].


Lee Wan Sie (Singapore) – Standards set global norms and common technical processes for AI governance, aligning “what good looks like” across jurisdictions [26-33][80-90].


Amanda Craig (Microsoft) – Microsoft’s internal Responsible AI Standard aligns product, engineering and sales functions; external standards are needed to create a shared market language [41-46].


Joslyn Barnhart (Google DeepMind) – Regulation is already referencing standards that have not yet been created, creating an urgent need for industry-driven standardisation [48-51].


Chris Meserole (Frontier Model Forum) – Standards solve a collective-action problem by providing an open, credible process that levels the playing field [52-55][108-110].


Esther Tetruashvily (OpenAI) – Standards translate internal risk-management practices into a common language for customers and enable ecosystem interoperability [56-59].


Kshitij Bathla (Bureau of Indian Standards) – Standards are tools that build consumer trust, assure quality, and must be adaptable to Indian-specific use-cases while aligning with ISO [61-64][207-212][206-210].


Core themes


Trust & “good enough” – The panel repeatedly stressed the need for credible, non-subjective reporting and a consensus on what constitutes “good enough” for different sectors [96-102][112-114][134-136][215-224].


Measurement & benchmarking – Rebecca detailed the benchmark components (taxonomy, dataset, evaluator) and highlighted uncertainty estimation as the main technical challenge [252-261][255-259]; Chris distinguished high-level process standards from the scientific benchmarks needed to operationalise them [291-298]; Amanda emphasized that shared metrics are essential to assess progress beyond the “nascent” stage [274-277].


Inclusivity & open governance – Etienne, Kshitij and Lee emphasized that open governance models (ML Commons, ISO, IEEE) enable smaller firms to adopt standards without building bespoke risk-management systems [176-185][207-212][215-224].


Regulation vs. market – Joslyn noted that regulators cite yet-to-exist standards; Chris explained that regulators often off-load risk-management requirements to the standards process, making standards a de-facto regulatory tool [48-51][194-196]; Lee argued that standards can serve as market differentiators even without legal mandates [215-224].


Regional perspectives


India (Manav mission & BIS) – Kshitij described the “Manav” human-centric vision, India’s AI Governance Guidelines, and BIS’s work to align national standards with ISO/IEC JTC1/SC42 outputs while incorporating India-specific risk considerations [207-212][206-210].


Singapore – Lee reported ongoing work on testing methodologies that she aims to submit to ISO within the next year, underscoring the panel’s consensus on the need for accelerated timelines [376-382].


Audience Q&A


Skill-gap & auditability – An audience member asked how governments can audit industry-driven assurance programmes given technical skill gaps [398-405]; Chris replied that the openness and legitimacy of formal standard-setting bodies mitigate this risk [112-114]; Lee added that certification can provide an independent assurance mechanism [215-224].


Language bias – A participant queried multilingual bias in India; Esther explained OpenAI’s use of multilingual evaluation suites (MMLU and Indian-dialect tests) and called for broader community participation [441-447]; Etienne noted that reusable safety-test frameworks are needed for many languages [452-460].


Minimum-consensus vs. absolute requirements – Joslyn answered that regulators will likely accept standards that provide a concrete “minimum bar” rather than overly abstract criteria [436-438].


Two-year outlook & action items


– Bhushan envisaged a rise in certification schemes that codify consensus on “good enough” within the next two years [336-342].


– Chris advocated for future-proof, process-oriented standards with evaluation methods that evolve as model capabilities advance [347-354].


– Lee aims to accelerate ISO-level testing work within a year [376-382].


– Amanda called for a modular, interoperable standards ecosystem to avoid reinventing the wheel for each new use case [388-393].


– Etienne reiterated the importance of open standards that keep costs manageable for smaller companies [176-185].


Conclusion – The panel concluded that AI standards are essential for translating high-level norms into verifiable practices, building trust across consumers, enterprises and regulators, and addressing the collective-action problem of rapid AI innovation [462-470]. Unresolved challenges include auditability, defining “good enough” for diverse risk tolerances, and developing comprehensive multilingual evaluation frameworks. The discussion underscored that coordinated, multistakeholder effort is vital for standards to become a durable foundation for responsible AI deployment.


Session transcriptComplete transcript of the session
Bhushan Sethi

I’m going to provide a brief introduction and then I’ll have my panelists introduce themselves and we’ll get into the discussion. So I’m a consultant around AI transformation. I help companies implement AI, drive the return on investment in a responsible way with AI. What’s really important about this discussion is we need to demystify what we mean by standard setting. There’s been a whole lot of discussion at this week’s summit around the importance of global cooperation, that the importance of inclusion around AI, driving solutions that meet everybody’s needs. The tech CEOs spoke about it yesterday. World leaders have spoken about it. We’re here in India where it’s about planet and people and prosperity. So that’s what the discussion is going to be about.

And we are going to have time for Q &A at the end. But I’m going to have my panelists introduce themselves first in the order that they’re sitting to introduce themselves and also talk about what standards mean for them? What lens they’re looking at from a standard perspective around AI?

Rebecca Weiss

Hello, my name is Rebecca Weiss I’m the executive director of ML Commons we are an AI benchmarking organization we are an engineering consortium that focuses on that problem and so for us as a technical standards organization around benchmarking what that means for us is two things one, we want to define the methodology for measurement and two, we want to create the technical artifacts that allow for engineers to integrate this methodology into their development life cycle. So for us, when we see what’s happening in the world today, the ability to measure risk is a big barrier to adoption and that ability to understand and estimate the uncertainty around the behavior of an AI system is something where we think benchmarking can help.

So, I will actually we have a large panel so I’m going to let everyone else have a chance to talk and I’m sure more will come out in our dialogue.

Etienne Chaponniere

My name is Etienne Chaponniere I work for Qualcomm. I’m a vice president of technical standards And so what we do within that role is, effectively, we have a team going to technical standards for AI, and we actually try to coordinate where is it that we need to go, how is it that we need to make sure that we understand what it means to be compliant. I come from a world of telecom, as Qualcomm can evoke to some folks. And for us, it’s a very different thing, right? For the telecom world, you cannot ship a product unless you comply to a standard because you need it for interoperability. In the world of AI standards, it’s a bit different.

So we’re talking more about safety standards, and those typically tend to trail the products. The products are out there, and then they’re going to comply to standards at some point when the standards are available. What matters, however, what is common in all of this is that the standards need to be available at scale for everyone and in a way that engineering teams can do it easily, at least from the product side. So I think I’ll leave it at that, and, yeah, that’s it.

Lee Wan Sie

I’m Wan Lee from Singapore government. I work in AI governance and policy. So many things, but specifically for standards, what it means to us is setting norms. That means alignment globally on what good looks like. And specifically in the area of AI governance, then a lot of it has to do at this stage in terms of common methodologies and processes that we have to follow. So, but it’s still technical. It’s not a checkbox, but hopefully that helps us all align to what good looks like. Thanks.

Bhushan Sethi

And maybe before the next introduction, just so you can get a flavor, we have standard setters and measurers. We have people in industry and we have people who play in the policy and the regulatory environment. And that’s the importance around this topic.

Amanda Craig

Thank you. Hi, everyone. I’m Amanda from Microsoft. I lead the public policy team with AI. And the Office of Responsible AI at Microsoft. I think Juan C. said it well when she described standards as really, like, aligning around what good looks like. And I would offer, you know, we actually at Microsoft in our office, we define something called our responsible AI standard that applies to all of our internal kind of product groups, our engineering function, our sales function. And if you think about, like, the role of that internal standard is to align all of the internal stakeholders we have around what good looks like. Like, externally, we need the same sort of mechanism, right? And that’s the role that standards can play in the broader ecosystem.

So we want to partner with our industry colleagues, and we want to partner with governments and others around the world to be able to define what good looks like so we can all have that common language instead of expectations.

Joslyn Barnhart

Hello. Jocelyn, Google DeepMind, where I also work on issues of AI standards, governance, and policy. building on what’s been said. So I think that was an interesting point that often technical standards come first and process and safety standards often come later. In the space of AI at the moment, actually, regulation has gone ahead and jumped to, you know, we’ve regulated and essentially made reference to standards that do not yet exist. So for places like Google DeepMind who have not invested heavily in the standard space in the past, this is now of an utmost priority because we actually need this to assist with implementation and compliance. So that is a primary goal on our side.

Chris Meserole

I’m Chris Meserole,. I’m the executive director of the Frontier Model Forum. Our mission is to advance Frontier AI safety and security, and we work with many of the leading Frontier AI developers and employers, including several colleagues on the stage today, to advance, you know, best practices for risk management. For Frontier AI in particular, there’s a kind of unique and a set of unique and novel risks that over the last couple of years. the community has really started to develop and converge around a set of best practices that now I think need to start to graduate into actual formal standards, and I think that’s kind of why we’re here. That’s why we’re very interested in the standard -setting space.

Esther Tetruashvily

Hi, everyone. My name is Esther Tetruashvily, and I’m the AI Standards Lead at OpenAI. Echoing many of the things that have already been said, I think standards for us, especially as a frontier AI lab, is about translating some of our practices for risk management into the language of risk management for customers across the supply chain, and it’s also about creating a language for consumer trust and assurance. It’s also about, in the age of agents, thinking about interoperability and helping everyone benefit from this ecosystem that we’re developing here. So I’m really excited to be here and to talk about these issues with you all. Thank you.

Kshitij Bathla

Hello, everyone. I’m Kshitij Bathla from Bureau of Indian Standards, the NETS. National Standards Body of India, and here representing ISO ICJTC1 SC42, because BIS, European Standards, is a part of the SC42. and for us I would say standards are the tools which enables consumers’ trust in whatever ecosystem for which they are developed as well as enable us for the industry to get it done to ensure the quality and the consumer trust. That’s the main focus area for us. Thank you.

Bhushan Sethi

So let’s start with why we need standards. Why are we even here? Because there’s a lot of confusion between standards, regulation, legislation. Are we going to get global cooperation around these things? Maybe should it just from a standard setting perspective and then maybe from a regulatory perspective. Why are we here? What’s the problem we’re solving and for whom?

Kshitij Bathla

So I would say the problems, there are multiples in the standards domain. Specifically, it always starts with what we are tackling with. What is AI? That was the primary focus of the JTC1 and SE42 when it started. So it defined what is AI. what is generative AI now they are talking about what is agenting AI as of now talking about so I think the most of the specific points that needs to be taken care is what is coming next and to keep pace with that and apart from once it comes to that when we have kind of mentioned that what it is all about then how do we verify and validate whatever is being said that this is a system which is having AI for example I would say someone says they have an equipment call it washing machine or is equipped with AI but is it actually equipped with AI or it’s just a normal logic system so this is something that we are trying to do the standardization.

Bhushan Sethi

So it’s about trust it’s about verifying the tech firms here represented are moving very fast with the model development so it’s like we need standards there from aregulatory perspective what would you add there?

Lee Wan Sie

I think the most important thing I wouldn’t say from a regulatory perspective. Maybe in terms of why, from an AI policy perspective, we think standards are helpful. Like I said, it’s about defining alignment in what should be in, let’s say, transparency. So I think if you say what would be the top three things today that we want to think about testing, setting for standards would be one, testing. How do you do testing for AI? Whether it’s AI models or AI applications, I think that’s one area. Because then it defines what good testing can look like. Two, perhaps in transparency, what would disclosure look like? Everyone has their own way of sharing the information that they want to share.

One way is to standardize it so it’s easier for the readers, people who are consuming this information to understand. And I’m saying this in very, very broad terms. I mean, it depends on which reader you’re talking about, who’s going to consume. just in broad terms, perhaps one way of standardizing it. Maybe the third way could be in how you’re reporting or monitoring incidents. But it’s still very, very early days. But that’s where standards, again, in terms of alignment, that might be one that would be useful to find alignment in these areas.

Bhushan Sethi

So ,how do we report? How do we disclose? How do we make it credible? And so it’s not a subjective tick -the -box exercise, etc. From a standard setting, Chris and Rebecca, from a standard setting perspective, what would you add to that before we have kind of the industry view?

Rebecca Weiss

I’m happy to add to this. So I think there’s been a theme that has come across in this panel a couple of times, which is what is good enough? And I think in order to define that, a standard represents a consensus about what is good enough. The problem that we have is who contributes to that consensus. It shouldn’t probably be exclusively an industry perspective. You need to have more stakeholders or more constituencies that need to be represented in that definition. And then on top of that, what is good enough, as I think Jocelyn mentioned earlier when we were talking before this panel, there’s a scientific element to that. How do you define the characteristics of a system such that you can actually create?

the kind of uncertainty estimation that lives up to a statistical guarantee, but then there’s also the political element to that, which represents a whole set of issues that I’m actually not qualified to talk about, so I will pass it to Chris.

Chris Meserole

I think it’s worth backing up from this thing. One of the original questions was, what are standards for? Is Chris’s mind working? I was just saying, one of the things we should maybe do is back up a little bit to this question of what are standards for, and I think a big part of what standards are for is to try and solve this collective action problem. There’s a kind of unique set of risks that we are worried about. We want to make sure everyone’s on the same page so that no one kind of actor is disadvantaged or advantaged compared to others. Having standards for how we’re going to manage risks across an ecosystem are extremely useful for that, so there’s a policy dimension to it.

There’s also an adoption dimension to it, right, because people want to know that there’s kind of… of a common way across industry of handling a certain class of risk. And I think being able to set standards and have a formal standard -setting body, to one of the points that was made earlier, by definition a standard -setting body is open, right? So there’s a legitimacy and a credibility to standard -setting bodies that you don’t have if it’s just industry or just government in many cases. And I think, you know, all of those kind of factors coming together are exactly why we’re so keen on kind of pushing forward the standards discussions.

Bhushan Sethi

Yep. So maybe from a hyperscaler perspective, maybe Esther, then Jocelyn, and we can kind of like play it clear, the difference, how is this showing up kind of at your firms and how are you thinking about this?

Esther Tetruashvily

Yeah, no, that’s a great question. I think from sort of a market adoption perspective, a lot of our technology, like general purpose AI models or foundation models, are being integrated into existing ecosystems or on top of. stacks. And there’s a lot of confusion in terms of risk controls and risk management about what that means. We have our own risk management processes. They have their own risk management processes. And one of the barriers to adoption is having a common language to talk about how do you map those controls onto one another. There’s a separate challenge, I think, of who is best positioned to control a particular risk. What are the risks? What are the net new risks?

What are the risks that are already existing where we don’t need to create something net new? And so for us, it’s both an imperative in some ways to kind of translate what we’re doing in terms of managing risks into the language of upstream, downstream customers so that they can understand and map those same practices onto their controls. And then we kind of can create a universal language that can ease trust and assurance in an easy, rockable way across the market. There’s also just space for, I think several people have talked about. Regulations moving ahead. of the standards, where we are still developing methodologies, what is standardizable in what we’re doing, recognizing where the science is not cut up yet, and where we maybe are in a place of more maturity.

Bhushan Sethi

And maybe just to bring it to life for the audience, given the huge amount of subscribers you have in India, around the world, growing every day, what’s changed in the standard vernacular at OpenAI?

Esther Tetruashvily

In terms of our adoption, or in terms of how we’re distributing it?

Bhushan Sethi

Yeah, the prominence of it, how people are thinking about it, the importance of the topic.

Esther Tetruashvily

So I think there’s both an aspect of it that’s like, what does already exist that we can use that can reassure customers that we are following the best practices for the industry, say for privacy or cybersecurity. There’s an existing risk management standard, ISO 42001, that OpenAI just got certified in. And that definitely signals something to the market. And to customers. Then there’s also sort of a transparency. element, right? We have our safety frameworks, we update them, we disclose information about in our model cards performance on a variety of metrics. And then there’s certain things we do to kind of elevate and help stakeholders across the spectrum in terms of how to build evaluations. So we currently published a safety hub that gets updated regularly that kind of tells how we’re performing in a variety of metrics and what are the best methodologies and how to work with this.

Bhushan Sethi

Great. So Joslyn, can you bring to life how Google DeepMinds are thinking about standard setting in that context?

Joslyn Barnhart

Yes. I’ll take it back to what Chris was talking about in terms of collective action problems. So some of the mitigations we’re talking about associated with some of the more extreme risks that Frontier AI poses can be quite costly. And so I do think that there is just a strong industry incentive to work together to resolve this collective action problem. Again, as Chris said, doing this through standards through an open, legitimate process seems to be incredibly impactful. Again, like the… The worst… thing for adoption would be a safety incident. So again, we have a collective incentive as an industry to make sure that we raise the floor to avoid that on all of our behalves.

So I do think that that is seen, you know, I think standards at this point are seen as a very clear and important strategic play for making, you know, essentially clearing the path for rapid adoption.

Bhushan Sethi

Amanda, how do they show up at Microsoft right now? Can you hear the question? How do they, how do these standards show up at Microsoft? Amanda’s going to speak about Microsoft experience.

Amanda Craig

Thank you. Yeah, I was going to start by just thinking about, at Microsoft, at Google, at other places, it’s not a totally new kind of process that we’re going through, right, in terms of thinking about standards and the importance of standards for adoption of this technology, sufficient trust in order to have adoption and in order to really enable compliance. I mean, I think Esther made a really good point. and sort of acknowledging that, you know, especially as we are deploying this technology, we are working with customers that have their own set of standards and regulation, and part of the challenge that we find ourselves, like, facing right now in AI governance is we have a lot of high -level norms and expectations that, again, are not so different from the patterns we’ve seen before.

Basically, we want to know how AI providers are managing risk, but we are in the early days of defining really what that means in practice in a really detailed way, especially, like, across the AI value chain. So what are model developers really responsible for doing for risk management? What are application developers really responsible for doing? How does that dock in to what deployers of those applications that are oftentimes implementing existing standards and meeting existing regulatory requirements? How does all that fit together? And, again, you know, we’ve done this with other digital technologies as well, like software, like cloud services, where we’re ultimately trying to define in practice what are the challenges that we’re facing right now.

is everyone responsible for doing? How do we have a common language to be able to talk to each other among sort of providers or the supply chain of technology and those that are ultimately deploying it? But we actually really do need the standards to support that, right? Because otherwise we are stuck at the sort of like high level conversation about norms around we want to evaluate risk. We want to figure out what the kind of right transparency practices are. Or we can find ourselves in this sort of deep technical weeds but like sort of having a place in between that is really at the level of standards, of technical standards, really helps drive that kind of common set of expectations so that you can have trusted.

Bhushan Sethi

So we need them. They’re important. We’ve got to drive adoption. There’s a collective action agreement here. From a Qualcomm perspective, SCM, bring to life the business model, how you use this in engineering your products.

Etienne Chaponniere

Yeah, so I think there’s one thing that I’d like to note. I think there’s one thing that I’d like to note here. As Qualcomm, we basically provide chipsets, right? We’re not building chipsets. We’re not building chipsets. We’re not building big models. What matters to us still is the fact and the reason why we’re engaged in those standards, whether it’s in ISO, Sentinelic for Europe, ML Commons, when it’s other type of standards, is effectively the fact that it provides scale in the sense of providing scale not only across the globe but also allows any different type of companies to benefit from it. I mean, let’s be clear, right? If you look at the companies who have the type of resources to either set up their own standards and risk management systems internally, they’re typically pretty big companies.

Now, the thing with AI is that there’s a huge amount of companies who are being created every day, and they don’t have the resources to put this together. And so there’s two conditions for making sure that the type of standards that are being put together are, one, inclusive, is that they’re open, as Rebecca, you were alluding to before. And so, whether it’s ML Commons, which has a very open governance model, or ISO, or Sentinelic in Europe, there needs to be an opportunity for everyone to participate. So that’s the first step. However, we know, and that’s the reality, that not everyone has the means to participate. Because they’re like super focused, they need to bring up their own LLM for that particular use case or maybe very general use case, and they just don’t have the resources to do this.

So from that standpoint, having the standard as effectively a mechanism for them to go directly to product and know that they’re going to comply with what the, effectively, world or the community has set up is really important. So from Qualcomm, the reason why we want to participate is to enable this type of accessibility to companies which are not always the biggest one.

Bhushan Sethi

Yep. So agreement that we need them. Before we go into how we set standards, how we measure and benchmark them, and Rebecca will bring that to life, a wildcard question is, there could be a lot of people listening to this to say, the world is not connected and cooperating around this. We don’t have global regulations on AI. But yet we have… industry leaders, standard setters, vehemently agreeing. How should the audience think about that? Is there a disconnect there or would anyone like to comment on that?

Chris Meserole

I would actually put, so part of one of the reasons why I think we’re all so interested in standards is one of the things you have, one of the things you’re seeing is multiple jurisdictions saying some version of we think that there are new risks with frontier AI. We as the government are concerned on behalf of our citizens that we are kind of attending to those risks across industry. Those risks and how to manage those risks are probably best left to be developed or kind of managed through the standard setting process, but they aren’t always setting the standards. So in the United States, there’s a couple of different states, for example, within the United States that have passed requirements for frontier AI developers.

to have a frontier AI framework, but they don’t specify what should actually be in the framework. They kind of offload some of that to the standards process, which is why I think it’s so important to have these standards in place. Like, there’s a clear kind of policy and regulatory interest in there being mechanisms by which some of the risks that may come with frontier AI are managed, but we need to kind of color in the lines a little bit exactly, like, you know, how we’re all going

Bhushan Sethi

And before we go to Rebecca, just from an India perspective, PM Modiji talked about Manav yesterday and the AI vision. Through there, there was a lot of focus on validity and governance, so standards were implied there. Do you want to just bring to life kind of how India thinks about this before we go to Rebecca and talk about measurement?

Kshitij Bathla

So I would say the Manav mission, it’s welfare, human -centric, and all those aspects are there. And from the governance perspective, also what is going on is that the government is not going to be able to do anything about it. we as of now the India AI governance guidelines are there. This is providing you a framework that these are the things that you should look into. Just providing a reference to. So in this direction the Indian government as of now is moving into. Coming into the from the perspective of standardization and at the national level as well as the ISO level I am adding to the question that you asked previously. That standards bodies are interconnected with each other.

The ISO there is a license mechanisms. We have the ML Commons as the license there. The IEEE is there. All bodies are there. So they are all interconnected there and whatever is coming as of out of these bodies is an outcome which is based on the studies. but done by various forums it’s not only the one I would say just the ISO body or not so in this direction the Indian standards that we are working on we are developing are also in the direction because here is something which is global we can’t have cells was specifically for India there could be the risks there could be specific use cases that are India specific for that those we need to have some specific guidance but more or less everything is the global thing that we are trying to look

Bhushan Sethi

into and then adapt those with the specific use cases that we need to right so we need global we need to adapt that to kind of local kind of conditions and use cases so let’s get a bit more technical Rebecca like why is this hard how do we measure it like how does it compare to benchmarking maybe Rebecca and then and then from a regulatory perspective did you want to make

Lee Wan Sie

I just want to respond to Chris comment and your question about you know if there’s no regulations then why do we care about standards right I mean, sure, I think there will be regulators who will say, yes, turn to the technical standards to define the expectations, which I think is a fair point that Chris made. But even when there’s no regulations, I think the standards still are useful. I mean, Esther just mentioned that OpenAI is certified for 42 ,001. You didn’t need to do that, but why did you do it, right? And Entropy has done that as well. And I think the idea is that perhaps there’s also a way to differentiate for organizations, for enterprises. And it doesn’t have to be the frontier model labs only.

It could be app developers and so on. A way to differentiate themselves and say that, look, I’m adhering to a global standard. I’m demonstrating that I have actually implemented something that’s good enough. I’ve addressed a risk in this way. I think that’s one good…

Bhushan Sethi

Do you want to make a quick comment? Yes, do you want to make a response to everything we’re getting to? Sorry, Rebecca. Please.

Lee Wan Sie

I just want to respond to Chris’ comment and your question about, you know, if there’s no regulations, then why do we care about standards, right? I mean, sure, I think there will be regulators who will say, yes, turn to the technical standards to define the expectations, which I think is the fair point that Chris made. But even when there’s no regulations, I think the standards still are useful. I mean, Esther just mentioned that OpenAI is certified for 42 ,001. You didn’t need to do that, but why did you do it, right? And Entropy has done that as well. And I think the idea is that perhaps there’s also a way to differentiate for organizations, for enterprises. And it doesn’t have to be the frontier model labs only.

It could be app developers and so on. A way to differentiate themselves and say that, look, I’m adhering to a global standard. I’m demonstrating that I have actually implemented something that’s good enough. I’ve addressed it. I’ve risen this way. I think that’s one good… reason for standards, even if there’s no regulatory cover. So the certification assurance part is helpful. Yeah, I just wanted to add that as a little bit of colour just to give some benefits to the standards community that is still kind of very…

Bhushan Sethi

Thank you. Bringing the regulatory perspective and kind of the Singapore experience. So let’s get into measure. And the fellow panellists, if you want to respond to anything, just give me the signal. We’re going to make this an interactive conversation. So Rebecca, how do we measure this?

Rebecca Weiss

Well, solve all the problems in one definition. No, I’m kidding. But as I said earlier, benchmarking consists of two things. It consists of a methodology, at least from our perspective, the way that we do benchmarking consists of a measurement methodology, and it consists of reference builds, implementations of that methodology so that engineers can use that. And the definition of a benchmark, as we’ve been trying to operationalize this in places like ISO and others, is a taxonomy, a data set, and an evaluator system. And the point of all of that construct is, as Etienne pointed out, this allows for you to scale this kind of approach towards the type of deployments that we’re expecting to see in these types of AI settings.

The challenge behind all of this is that what you’re really trying to do is estimate uncertainty. Uncertainty. You’re trying to provide a sense of, I’m not going to tell you that your system is, quote -unquote, safe or not. What I’m going to tell you is, under these considerations, under these conditions, under these assumptions, the estimated likelihood of a particular risky behavior is X. And then it is up to you as a risk management professional, a deployer, a developer, it’s up for you to decide, is that enough? Is that good enough for your needs? And I don’t think it’s going to be the same for different sectors. I think sometimes. Sectors will have a much higher bar for the amount of uncertainty.

that needs to be estimated, and then other sectors will probably be like, that’s good enough for me. I don’t necessarily need to get much further than what you are offering right off the date. So we can go into all of the different questions that are made open, but those particular areas related to developing that taxonomy, developing those data sets, and developing those evaluators, the best practices and the standards to make it clear that this is the best in the industry, this is the way that it is, that’s what we need to get better at.

Bhushan Sethi

Yeah, so what I’m hearing is we need clarity. Clarity of the taxonomy, clarity of what we’re measuring, and it needs to be verifiable and credible. From an industry perspective, would anyone like to pick up, like, how’s that going to work? What’s in place now? What some of the challenges might be? How do you get organizational buy -in? Anything to add from an industry? Amanda, do you want to start us off?

Amanda Craig

Sure. I mean, I think there’s work to do across all the elements that Rebecca just laid out, and it’s really a reason why we are really invested in working with M .L. Cummins, because I think we need places that are bringing industry and and and civil society and stakeholders together to actually work through these problems and resolve these hard questions in ways that are really going to be sort of valid and reliable broadly. And so I think that’s really the work still ahead, but I think we are also making good progress, right? And thanks to ML Commons for helping to facilitate that. My thought on this is that we’ve been talking for years now about how nascent this field is and that actually to judge if we are actually making progress, this too could be standardized, right?

Like we don’t have common ways of assessing are we still in a nascent stage? What levels of uncertainty do we have? So to Rebecca’s point, I think this is absolutely essential so we can all align exactly on have we made some progress? We’ve made sufficient progress to start relying on these things. To what degree can we rely on them for important decision -making around deployments?

Esther Tetruashvily

yeah I think I’ll just add if we take this back down to the basics I think whether you’re an enterprise customer or you’re a consumer of our products you just want to know is this thing going to be accurate can I rely on this thing is this going to get me into trouble if I incorporate this in my workflows am I going to carry some sort of liability and at the core of standards is figuring out a way to have a common mechanism to provide an answer of reassurance you can trust us here’s a measurement certified by somebody else that this thing is reliable that this thing is accurate that I can rely on this thing and I can use this thing and I think we’re in this moment where we’re still trying to figure out as an industry and as a community about what that’s going to look like and so whether it’s advancing the measurement science because we currently don’t have enough of that in order to make sure that we can give an estimate of what is accurate what is reliable what is safe for specific risks or on the other side, what are the risks that we care about?

I think some risks might be some countries, some jurisdictions might have one list of risks. Other countries might have a different list of risks. And then there’s going to be a question of, like, how do you control for that, right? And that’s kind of what Rebecca Nemel -Commons and many others are working on, is how do you provide some sort of mechanism of credibility that says we’ve measured this, this thing is safe, that can then be certified, could be, you know, understood in the same way for everyone. So at the end of the day, in order for us to really unlock the value of this new technology that is transformative, I think many of us who are here today for the Indian Impact Summit recognize that potential.

We all also need to kind of answer those questions, and standards are the way you facilitate it.

Bhushan Sethi

Yeah, and so there’s a theme of trust that’s going through this. So maybe, Chris, add to that, and then I’ll add to that into a comment from a quote,

Chris Meserole

Yeah, just briefly, I think I also just want to situate how kind of benchmarking standards and some of the scientific questions we’ve been talking about fit in. Like there’s I think we’ve been talking a lot about different types of standards. I just want to clarify that there’s like a kind of broader, high -level set of process standards where you kind of say, all right, for this class of risk, what we’re going to do is we’re going to identify what the risk is. We’re then going to evaluate what that risk might actually be. And then we’re going to put in place certain kinds of mitigations and controls. Those are kind of, it’s a process for how you’re going to walk through risk management for something.

That absolutely needs to be standardized. But then even within that, once we get to, all right, once we have agreed on what the risk is that we’re trying to evaluate, how do we actually do that? And that’s where the standards come in for the benchmarks that we want to see developed. And that’s where some of these scientific questions, I think, really come into play because we need to have, you know, those kind of credible scientific evaluations and tests for the whole kind of broader risk management effort to hang together. And it’s, you know, again, critical, I think, for this whole process.

Bhushan Sethi

Yes, this has got to live next to the risk. Risk management, identification, mitigation strategy in any company. Go ahead, Jocelyn.

Joslyn Barnhart

I just had briefly. I think the possibility for comparison across models is also something that’s super important here. I think there’s an important safety dimension there. If we actually are all measuring the same thing and can give consumers some relative assessment of safety, of quality, this is actually going to potentially contribute to a race to the top as opposed to the bottom. And so we’re solving it.

Bhushan Sethi

So that’s the question of who we’re solving for. Two of the panelists have mentioned consumers. It’s not just about enterprise. It’s not just about government. It’s all about consumer trust. Essie, what would you add?

Etienne Chaponniere

What I wanted to add is the fact that here when we’re talking in general about trying to create standards to resolve the type of safety risk that we’re going to see, it’s just also to reassure the audience that it’s not that we’re trying to solve every single risk that happens. There is a huge amount of existing standard bodies, whether it’s in ISO and SensenELEC and other places, where they already have identified risk for their particular verticals or their particular… not silos, but the particular industries, those are already at work, right? So how they’re going to use AI, how the AI is going to be effectively, the AI safety is going to be translated to their own processes.

Those things are already happening, right? So it’s not only the people on this panel who are working on this, the entire community of standards, whether it’s in automotive, radio equipment directive, everything is already, everybody’s already looking at that, right? In the end, the difficult part is going to be to make sure that there is a commonality in terms of the type of techniques that we’re using whenever there’s an automated technique that we can use. Because from an industry standpoint, what is really useful, in particular if you’re a smaller company, is to make sure that you can run something efficiently and it addresses as much as the use cases that you run as possible. So that is an important thing that we need to keep in mind when we’re doing this.

So it’s why, I mean, from Qualcomm, obviously, we don’t address every single thing, but we want to make sure that at least in the areas we’re involved, there’s going to be as much as a commonality in terms of the measurement techniques that we’re going to use.

Bhushan Sethi

So consensus around the need to do it, consensus around the fact that it’s hard, but it’s important for consumers and business and investors. But Jocelyn made a point that we’ve been talking about how this is a nascent topic, et cetera. I want to look forward. What over the next two years does this look like? What have we got to get right? The models are changing. There could be regulation that changes. There could be changes around China, U .S. operating in different ways. What does this topic look like? How do we make sure we stay the course on this topic? Anyone want to offer a perspective as we look forward? And then we’ll start wrapping up.

And thinking about questions so we can get questions from the audience. I’ll take a crack at it. So at least from my perspective, there are a couple of things that I hope to see over the next couple of years. One is that I think this idea of benchmarks and other standards representing consensus, we should be seeing more things like certification that represent more types of consensus. If benchmarking represents consensus around how to estimate and measure a thing, certification could end up representing agreement. A definition of what is good enough deserves some form of certification. I don’t know necessarily what that’s going to look like today, but I have to imagine that those sort of represent truces, the temporary agreements about this is good enough for my industry, this is good enough for my deployment, this is good enough for my use case.

So that’s what I’m hoping we start to see over the next two years. Anyone else want to add to that? Because, I mean, Chris, jump in, but we’ve seen some of these disclosures in the past, and people commit to environmental goals or DEI goals or other set of standards or disclosures. Stakeholder capitalism was a big deal, and now it’s more about shareholders. So I’d love to understand our perspective on how do we stay the course.

Chris Meserole

Yeah, I might distinguish a little bit between how do we future -proof these standards and then how do we kind of ensure that they’re implemented over time. And I think the way that we future -proof them is to some extent to go back to the point I was making earlier about process standards, right? The process is somewhat agnostic to the actual kind of, you know, AI system itself and the capabilities it has. If you have a good process for identifying risks, evaluating risks, that process can kind of be a bit future -proofed. The specific evals you run are probably going to have to be updated over time to account for the greater capabilities of models as they advance, right?

And I think… similar with some of the controls that might need to be kind of used to manage some of the risks if there’s certain thresholds or kind of if the evaluations kind of indicate a certain level of risk, right? So the subcomponents of it might need to be evaluated. The overarching framework hopefully can kind of have some legs behind it over time in terms of future -proofing it. So we must commit to a process. We can’t future -proof because we can’t predict the future, but the process is so important. Even a good example of this would be something like the, I think, 40 ,001 has come up a few times. Like there’s a certain class of AI that 40 ,001 is very kind of tailored to, but even that AI has changed over time.

But 40 ,001 is still a very good kind of standard for managing those kinds of risks for those kind of applications of AI across a broad array of machine learning algorithms. But the other point that I would make in terms of, you know, you alluded to some of the kind of implementation of standards over time and making sure that they have the same currency to them. And there, I think we can rely on some of the incentives and the need, again, for there to be collective action on this that we’ve talked about before. Some of the incentive to make sure that there’s a collective action problem is going to rest with policymakers, which is why you’ve seen some regulatory activity.

Even in areas where there’s not, to Juan C.’s point, there’s a clear market need for these standards to be developed and implemented over time because consumers want to see, you know, they want to trust that the, you know, whether it’s individual consumers or enterprise, they want to trust that the model is actually safe and secure to use. And so I don’t see kind of the standards, the importance of standards diminishing over time. In fact, if anything, as the capabilities advance, consumers and enterprises are going to be more and more interested in making sure that they

Bhushan Sethi

Yes, it’s going to be consumer -driven. Juan C., just from a regulatory perspective, any thoughts? Chris mentioned implementation. Which is the hard stuff of where lots of this stuff gets stuck. Any perspective on implementation or from your experience as a regulator to add here?

Lee Wan Sie

Implementation of standards? Yes. I mean, Chris put it very well, right? One, regulators could say, I expect you to comply with certain requirements and this is how you do it. And that’s where the standards set on how you do it. Or regulators can don’t provide certain requirements or certain expectations. And the market sets out these requirements and these expectations. If you do it, then we will buy your product, for example. So I think from an implementation point of view, I think there will be some momentum, either from the market or from regulations, to move standards. But I think where I think, back to your original question, what’s going to happen in two years, I hope we can actually move faster on standards in terms of the definitions of standards.

I think that would be super useful. We’re leading some work on testing, well, benchmarking and rate teaming, primarily methodology definition. But… Yeah. We hope that in the next one year that can be done and sorted and accepted within the ISO process. But the experience has shown us that it takes a while. So in the next few years, hopefully we will find a way in which we can move to standards faster.

Bhushan Sethi

So we need to move with speed from a regulatory perspective. Amanda is going to have the last word and then we’re going to go to questions. So please prepare them. Amanda?

Amanda Craig

I didn’t realize that. No, the one thing I wanted to add in terms of like a goal for where we can find ourselves two years from now is thinking about like a system of standards that are interoperable where we have a sort of modular approach, right, where across like general purpose technology and, for example, in different sort of deployment scenarios, different use cases, different sectors, we actually can get some efficiency from, you know, these standards are all going to need to continuously evolve and improve and we’re going to learn from the science. And we’re going to keep evolving the benchmarks and the kind of methodology around the evaluations. But we don’t want to like keep starting from scratch with every piece of that, you know, puzzle.

And so we need to figure out a way to actually ensure that. like where we are making progress on the evaluation science and how we are doing this in the context of like evaluating AI models or systems and then how we are evaluating AI and deployment in like critical sectors, for example, we actually have some synergy built into the standards ecosystem so that we are making kind of more dynamic progress across everything at the same time.

Bhushan Sethi

Yeah, so it needs to be interoperable and we can’t keep reinventing the wheel. So audience, questions? I’m going to collect questions, maybe three to five. So the gentleman at the front, the gentleman at the back, and then the lady with the hand up.

Audience

Hi there. Thanks for taking my question. Maybe I have a bit of a tricky question for you. You know, on the panel, obviously, we have a lot of commercial interests. My question is this. How do we know in your assurance program or whatever you’re proposing that it’s going to be done since it’s driven primarily by industry, how do we know that you’re not just going to create something that cheaply satisfies the industry in front of… of us versus what the public actually needs. And assuming you do have a program that you’re going to talk about, how does a government or external agency audit such a program, given the skill gap to create such a very sophisticated compliance program, how can world governments come?

Because I’ve been on a lot of panels this week. The fear, uncertainty, and doubt is not only just the policy gap. It’s actually the technical gap, the inability of world governments to audit properly whatever you have. Thank you.

Bhushan Sethi

Thank you. So keep the questions brief. Thank you for that. So that’s about, like, how do we make it real? How do we make it not performative? I’m going to collect two other questions, and then we’ll throw them to the panelists. So keep your hands raised. We have a gentleman at the back. And I think there was a lady or a gentleman with a tie. Yeah, hi.

Audience

So… As a recent computer science student, I’m interested in building AI for India. Specifically with such a distinguished panel, I thought I’d shoot my shot. I’m a little nervous, so I apologize about that. I want to talk specifically about language bias. Being in India, there are 22 official languages, and I’m constantly thinking in two to three different languages. And when I utilize tools, such amazing tools built by everybody here, I’m wondering how you guys would go about tackling language bias and building guardrails around that to ensure that, you know, a small model like a student like me is making does not go haywire. Yeah, great

Bhushan Sethi

question about language. Thank you, sir. And then, gentleman with a tie. Which doesn’t mean, like, more gentlemen wear ties, but, yes, please. Hi, Jules

Audience

Polonetsky at the Future of Privacy Forum and our AI Governance Center. The standards always say… seem to be an easier path when they are more technical than… and challenging social policy, and AI governance seems to capture the most broad potential collections of social policy. And given that there’s a lot of disagreement and some debate over whether one should even measure certain areas, do you imagine that we’re talking about minimum viable consensus with the broadest number of stakeholders, or is there a path to in some way address some issues that some stakeholders see as absolutely necessary and others don’t want on the table? Yep. All

Bhushan Sethi

right. Soundbite responses panel. Like how do we make it real? How do we deal with the skills gap? How do we deal with the MVP? Anyone? Go on, Jocelyn. On the

Joslyn Barnhart

performative question, I think now that standards have been referred to within actual regulation, I think to the extent that we want to use these standards as evidence of conformity with those particular regulations, that’s set up a lot of the work that we’re doing. that’s a kind of minimum bar at the very least, because I think if we make these things too high level, too abstract, or too essentially lowest common denominator, I don’t think regulators are going to look at those standards as evidence of conformity. So I think there is that kind of interlocking pressure created by the regulation itself for some sort of degree of quality. Thank you.

Bhushan Sethi

And Esther, do you want to comment on the language perspective and how you’re thinking about that at OpenAI? Thank you.

Esther Tetruashvily

Yes, we do a series of evaluations like MMLU for determining how well our models perform on a variety of languages. We also have a specific test actually in QA. There’s also a specific test in QA that we also kind of test our models on that has a variety of dialects within India. So I think the short answer is that this is an area where we need more participants. And I believe ML Commons is playing an active role in helping further our capacity building. And I think working with local ecosystems to help clean and collect good data so that we can do this appropriately. This is another area, right, just like we’ve been saying, where we need to work in partnership to figure out how do we both collect the type of information, how do we measure this stuff, how do we build the evaluations, and then how do we build an industry standard where all of the actors are kind of held to that standard.

And it’s going to have to be a collective effort. Yeah. Okay.

Etienne Chaponniere

Just to add a little bit on the question regarding the language. In the end, I don’t think there’s like a – there’s no silver bullet solution, right? There’s going to be a need to have this type of – Either safety test or safety prompt. which are required for different type of languages. And you’re not going to be able to address every single thing because there’s just a huge amount of diversity. I mean, take me. I’m French from cultural background. I speak English and think in French and English all the time. There’s weird stuff that I say that will not be captured by a model that’s only for American English, right? So there’s going to be a need for more than one language which are captured, and probably a lot of them, but this is where the community of basically everybody needs to come and say, hey, this is what I want to capture for my type of language.

What matters to make sure that there is scale and that it still remains efficient is that hopefully the tool and the software framework around it can be reused. And that’s really a big advantage for that. Thank you.

Bhushan Sethi

So in summary, and thank you, dear panelists, for the great discussion. So you heard today that standards are important. This is a fast -moving world. We’ve got to be designing for consumers, for business people. There’s a commitment. There’s a commitment here around measurement. It’s both art and science. We need to have the process that’s consistent. But across regulators, across standard -setters, around policymakers, and the business and the tech community, there’s a consistent understanding. So it’s going to be an emerging topic, which I know we’ll continue to discuss. Thank you, panelists, and thank you to the audience. Thank you. Thank you. Thank you. Thank you. Thank you.

Related ResourcesKnowledge base sources related to the discussion topics (13)
Factual NotesClaims verified against the Diplo knowledge base (4)
Confirmedhigh

“Standards solve a collective‑action problem by providing an open, credible process that levels the playing field”

The knowledge base notes that open standards act as a common technical language that can level the playing field for small companies and promote fairness, confirming the panel’s description of standards as a collective-action solution [S30] and highlights the role of standards in governance and risk-management coordination [S38].

Additional Contextmedium

“Kshitij Bathla said standards are tools that build consumer trust, assure quality, must be adaptable to Indian‑specific use‑cases while aligning with ISO”

Kshitij Bathla’s participation is recorded in the transcript (introductory remark) and the discussion references ISO-based frameworks such as ISO 42001 that provide a common set of requirements for national bodies, adding context to his emphasis on Indian-specific adaptation and ISO alignment [S71] and [S75].

Additional Contextmedium

“Open governance models (ML Commons, ISO, IEEE) enable smaller firms to adopt standards without building bespoke risk‑management systems”

Several knowledge-base entries describe how open, inclusive standards lower barriers for smaller actors, promote participation from diverse stakeholders, and are promoted through multistakeholder collaborations, supporting the panel’s point about open governance models [S30] and [S67] and the broader multistakeholder cooperation described in [S24].

Additional Contextlow

“Regulators are already referencing standards that have not yet been created, creating an urgent need for industry‑driven standardisation”

The knowledge base discusses how regulators rely on industry standards as part of AI governance and often look to standards processes to fill regulatory gaps, providing context for the claim that regulators cite yet-to-be-finalised standards, though it does not explicitly confirm the non-existence of those standards [S38] and [S79].

External Sources (81)
S1
Setting the Rules_ Global AI Standards for Growth and Governance — – Kshitij Bathla- Chris Meserole- Etienne Chaponniere- Rebecca Weiss- Bhushan Sethi – Kshitij Bathla- Chris Meserole- L…
S2
https://dig.watch/event/india-ai-impact-summit-2026/how-trust-and-safety-drive-innovation-and-sustainable-growth — I just have the image of the U.K. Information Commissioner doom -scrolling TikTok in my head now. Let’s do a quick round…
S3
How Trust and Safety Drive Innovation and Sustainable Growth — – Alexandra Reeve Givens- Amanda Craig – Denise Wong- Amanda Craig
S4
Setting the Rules_ Global AI Standards for Growth and Governance — -Etienne Chaponniere- Vice president of technical standards at Qualcomm
S5
https://dig.watch/event/india-ai-impact-summit-2026/setting-the-rules_-global-ai-standards-for-growth-and-governance — Just to add a little bit on the question regarding the language. In the end, I don’t think there’s like a – there’s no s…
S6
https://dig.watch/event/india-ai-impact-summit-2026/setting-the-rules_-global-ai-standards-for-growth-and-governance — And it’s going to have to be a collective effort. Yeah. Okay. Hi, everyone. My name is Esther Tetruashvily, and I’m the…
S7
Setting the Rules_ Global AI Standards for Growth and Governance — – Lee Wan Sie- Esther Tetruashvily- Chris Meserole – Rebecca Weiss- Esther Tetruashvily- Amanda Craig
S9
https://dig.watch/event/india-ai-impact-summit-2026/setting-the-rules_-global-ai-standards-for-growth-and-governance — And it doesn’t have to be the frontier model labs only. It could be app developers and so on. A way to differentiate the…
S10
S11
https://dig.watch/event/india-ai-impact-summit-2026/setting-the-rules_-global-ai-standards-for-growth-and-governance — And I think… similar with some of the controls that might need to be kind of used to manage some of the risks if there…
S12
Setting the Rules_ Global AI Standards for Growth and Governance — I’m Chris Meserole,. I’m the executive director of the Frontier Model Forum. Our mission is to advance Frontier AI safet…
S13
Ensuring Safe AI_ Monitoring Agents to Bridge the Global Assurance Gap — -Chris Meserole- CEO of FMF (organization not fully specified in transcript)
S14
Setting the Rules_ Global AI Standards for Growth and Governance — – Kshitij Bathla- Chris Meserole- Etienne Chaponniere- Rebecca Weiss- Bhushan Sethi
S15
AI-Powered Chips and Skills Shaping Indias Next-Gen Workforce — -Ashwini Vaishnaw- Role/Title: Honorable Minister (appears to be instrumental in India’s semiconductor industry developm…
S16
ElevenLabs Voice AI Session & NCRB/NPMFireside Chat — -Shailendra Pal Singh: Role/title not explicitly mentioned, but appears to be a co-presenter/expert on Bhashini translat…
S17
https://dig.watch/event/india-ai-impact-summit-2026/setting-the-rules_-global-ai-standards-for-growth-and-governance — Hello, everyone. I’m Kshitij Bathla from Bureau of Indian Standards, the NETS. National Standards Body of India, and her…
S18
Setting the Rules_ Global AI Standards for Growth and Governance — -Kshitij Bathla- Works at Bureau of Indian Standards (BIS), the National Standards Body of India, representing ISO ICJTC…
S19
WS #280 the DNS Trust Horizon Safeguarding Digital Identity — – **Audience** – Individual from Senegal named Yuv (role/title not specified)
S20
Building the Workforce_ AI for Viksit Bharat 2047 — -Audience- Role/Title: Professor Charu from Indian Institute of Public Administration (one identified audience member), …
S21
Nri Collaborative Session Navigating Global Cyber Threats Via Local Practices — – **Audience** – Dr. Nazar (specific role/title not clearly mentioned)
S22
WS #69 Beyond Tokenism Disability Inclusive Leadership in Ig — Astbrink highlights the complexity of implementing high-level global instruments at the national level. She emphasizes t…
S23
AI That Empowers Safety Growth and Social Inclusion in Action — And standards turn principles into action. They shape risk management, they clarify accountability, they guide human ove…
S24
International multistakeholder cooperation for AI standards | IGF 2023 WS #465 — Additionally, it provides e-learning materials to enhance understanding of AI standards. Moreover, the AI Standards Hub …
S25
U.S. AI Standards Shaping the Future of Trustworthy Artificial Intelligence — The discussion reveals extraordinary consensus among all speakers on the fundamental principles of AI agent standards de…
S26
Harmonizing High-Tech: The role of AI standards as an implementation tool — Conversations around standards are transparent and inclusive Experts take decisions by consensus Renowned for their lo…
S27
Standardisation – The Key to Unlock the Sustainable Development Goals (SDGs) — The SDGs tackle challenges of global proportions and it would be ill-advised to look for solutions that are not globally…
S28
Artificial Intelligence & Emerging Tech — Efforts should be made to avoid reinventing the wheel and use existing good/best practices Efforts to coordinate in the…
S29
Advancing Scientific AI with Safety Ethics and Responsibility — Thank you, Shyam. I think this is a very important question. And it’s also a topic that I’m really passionate about as w…
S30
Better governance for fairer digital markets: unlocking the innovation potential and leveling the playing field (UNCTAD) — Access to open markets through regulation is highlighted as beneficial for small messaging companies. This provides oppo…
S31
Day 0 Event #171 Legalization of data governance — Wolfgang Kleinwächter: Okay, thank you. Thank you very much and thank you for the invitation and thank you all the pri…
S32
The role of standards in shaping a safe and sustainable AI-driven future — Onoe acknowledged the rise of a novel AI innovation ecosystem and the indispensable role of standards in extending this …
S33
Artificial intelligence — Despite their technical nature – or rather because of that – standards have an important role to play in bridging techno…
S34
AI as critical infrastructure for continuity in public services — Thank you very much. Standards are a very important pillar of building trust. Another is inclusive governance. Changatai…
S35
Internet standards and human rights | IGF 2023 WS #460 — Ignacio Castro:Thank you. My name is Ignacio Castro, and I’m a lecturer in Queen Mary University of London, and I also c…
S36
High Level Dialogue: Strengthening the Resilience of Telecommunication Submarine Cables — Very high consensus with strong implications for effective policy coordination. The alignment suggests that the ITU’s In…
S37
Closing Session  — Sustained collaboration between governments, industry, and other stakeholders is essential for translating recommendatio…
S38
Searching for Standards: The Global Competition to Govern AI | IGF 2023 — Collaboration with industry was emphasized as crucial, and various arguments and evidence were presented throughout the …
S39
Keynote-Julie Sweet — She stresses that human leaders, not automated loops, must decide how AI tools are deployed responsibly, and that global…
S40
WS #103 Aligning strategies, protecting critical infrastructure — International cooperation and alignment of policies/standards is crucial
S41
Global Standards for a Sustainable Digital Future — This comment challenges the traditional static nature of standards development and proposes a paradigm shift toward dyna…
S42
Strengthen Digital Governance and International Cooperation to Build an Inclusive Digital Future — The forum revealed both the promise and complexity of international cooperation on digital governance. The strong consen…
S43
Setting the Rules_ Global AI Standards for Growth and Governance — The discussion revealed relatively low levels of fundamental disagreement among panelists, with most tensions arising ar…
S44
Open Forum #30 High Level Review of AI Governance Including the Discussion — High level of consensus with significant implications for AI governance development. The alignment suggests that despite…
S45
Global AI Policy Framework: International Cooperation and Historical Perspectives — High level of consensus on fundamental principles and approaches, with differences mainly in emphasis and specific imple…
S46
Main Session | Policy Network on Artificial Intelligence — Panelists debated the feasibility of a global AI governance regime, acknowledging the challenges of multilateralism but …
S47
Interdisciplinary approaches — AI-related issues are being discussed in various international spaces. In addition to the EU, OECD, and UNESCO, organisa…
S48
A Digital Future for All (afternoon sessions) — AI governance requires a multi-stakeholder approach due to the diverse nature of opportunities, risks, and inclusivity c…
S49
From principles to practice: Governing advanced AI in action — Chris emphasizes the importance of coordinating globally to standardize frontier AI risk management frameworks. He notes…
S50
Policymaker’s Guide to International AI Safety Coordination — Moderate disagreement with significant implications – while speakers share common concerns about AI safety, their differ…
S51
The geopolitics of digital standards: China’s role in standard-setting organisations — But if standardisation processes become overly politicised, this could slow them down. It could also mean that discussio…
S52
Navigating the Digital Future: Standards-led Digital Economy (BSI) — In conclusion, voluntary standards have a positive impact on globally diverse organizations, promoting economic efficien…
S53
U.S. AI Standards Shaping the Future of Trustworthy Artificial Intelligence — These key comments transformed what could have been a dry technical discussion into a compelling narrative about the str…
S54
The role of standards in shaping a safe and sustainable AI-driven future — Seizo Onoe:Thank you very much. Good morning, everyone, and very warm welcome to you all. Our discussions at this summit…
S55
Can (generative) AI be compatible with Data Protection? | IGF 2023 #24 — Furthermore, the analysis explores the role of regulation in the AI landscape. It suggests that regulation should not on…
S56
International multistakeholder cooperation for AI standards | IGF 2023 WS #465 — Context is highlighted as a crucial element for effective engagement in standards development. Australia’s experts have …
S57
Open Forum #34 How Do Technical Standards Shape Connectivity and Inclusion — Both audience members criticized the panel for discussing technical standards without including actual technical standar…
S58
The role of standards in shaping a safe and sustainable AI-driven future — Onoe acknowledged the rise of a novel AI innovation ecosystem and the indispensable role of standards in extending this …
S59
Setting the Rules_ Global AI Standards for Growth and Governance — I think it’s worth backing up from this thing. One of the original questions was, what are standards for? Is Chris’s min…
S60
AI as critical infrastructure for continuity in public services — “Trust also can influence economic confidence and cross -border collaboration.”[54]. “Standards are a very important pil…
S61
International multistakeholder cooperation for AI standards | IGF 2023 WS #465 — Standards are voluntary codes of best practice that companies adhere to. They assure quality, safety, environmental targ…
S62
Internet standards and human rights | IGF 2023 WS #460 — Ignacio Castro:Thank you. My name is Ignacio Castro, and I’m a lecturer in Queen Mary University of London, and I also c…
S63
WS #438 Digital Dilemmaai Ethical Foresight Vs Regulatory Roulette — von Knebel Moritz: Yeah, thank you and thanks for having me. People have often asked this question, what are the regulat…
S64
Closing Session  — Sustained collaboration between governments, industry, and other stakeholders is essential for translating recommendatio…
S65
Keynote-Julie Sweet — She stresses that human leaders, not automated loops, must decide how AI tools are deployed responsibly, and that global…
S66
WS #103 Aligning strategies, protecting critical infrastructure — International cooperation and alignment of policies/standards is crucial
S67
International Standards: A Commitment to Inclusivity — Charlyne Restivo:Ladies and gentlemen, distinguished guests, good afternoon, and welcome to this WSIS High-Level Dialogu…
S68
Resilient infrastructure for a sustainable world — Benjamin Frisch offered CERN’s perspective on open collaboration, explaining how creating open ecosystems around technol…
S69
Global Standards for a Sustainable Digital Future — ### Dynamic Standards for Rapidly Evolving Technologies Dimitrios Kalogeropoulos, an expert in AI applications in healt…
S70
YouthLead: Inclusive digital future for all — Melissa Michelle Munoz Suro: When I was 25, I found myself standing in a room full of policymakers, developers, designer…
S71
Democratizing AI: Open foundations and shared resources for global impact — Bernard Maissen: Yes, thank you. Hello, everybody, dear panelists. Nina, thank you for giving me the floor. In the globa…
S72
High-Level Session 3: Exploring Transparency and Explainability in AI: An Ethical Imperative — 3. Global collaboration: Li Junhua stressed the importance of cooperation among all stakeholders. His Excellency Dr. Ab…
S73
High-level AI Standards panel — ## Challenges and Future Considerations 3. **Include**: Engaging diverse stakeholders beyond traditional technical comm…
S74
WS #189 AI Regulation Unveiled: Global Pioneering for a Safer World — Auke Pals: Thank you very much, Lisa. So I hope at this point, the EU AI Act could steer us in the right direction a…
S75
Aligning AI Governance Across the Tech Stack ITI C-Suite Panel — It doesn’t mean that countries can’t have their own perspectives or sovereign outlooks, but there is sort of a… a move…
S76
Bridging the AI innovation gap — This comment provides a profound reframing of technical standards from bureaucratic requirements to tools of global equi…
S77
Importance of Professional standards for AI development and testing — Don Gotterbarn: Thank you, Stephen. The previous assertion to Stephen’s that says essentially, because there’s differenc…
S78
Google and Microsoft impress investors with AI growth — Microsoft Corp. and Google owner Alphabet Inc.impressedinvestors surpassing Wall Street expectations with robust quarter…
S79
Closing the Governance Gaps: New Paradigms for a Safer DNS — Although regulation in the DNS industry is inevitable, it should aim to avoid fragmented jurisdictional approaches. If t…
S80
AI and Human Connection: Navigating Trust and Reality in a Fragmented World — Current regulation approaches are inadequate and lag behind technological development Legal and regulatory | Economic …
S81
[WebDebate #12 summary] Standardisation: Practical solutions for strained negotiations, or an arena for realpolitik? — However, it is important to note that the main goal of a standard is to benefit the actors it applies to. For example, s…
Speakers Analysis
Detailed breakdown of each speaker’s arguments and positions
R
Rebecca Weiss
2 arguments205 words per minute679 words197 seconds
Argument 1
Benchmarking methodology as core of standards (Rebecca Weiss)
EXPLANATION
Rebecca explains that a technical AI standard must first define a clear measurement methodology and then provide the technical artifacts that let engineers embed this methodology into their development pipelines. This two‑part approach ensures that standards are not just abstract rules but actionable tools for consistent evaluation.
EVIDENCE
She states that ML Commons wants to “define the methodology for measurement and … create the technical artifacts that allow for engineers to integrate this methodology into their development life cycle” [13-14] and later describes a benchmark as consisting of a taxonomy, dataset, and evaluator system together with a measurement methodology and reference implementations [252-261].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Rebecca Weiss’s description of benchmarking components (taxonomy, dataset, evaluator) and the need for a clear measurement methodology is corroborated by S1, which outlines these three essential elements.
MAJOR DISCUSSION POINT
Definition and Purpose of AI Standards
AGREED WITH
Chris Meserole, Amanda Craig, Bhushan Sethi
Argument 2
Benchmark defined by taxonomy, dataset, evaluator; methodology essential (Rebecca Weiss)
EXPLANATION
Rebecca details the components that make up a benchmark: a well‑structured taxonomy, a representative data set, and an evaluation system, all tied together by a rigorous measurement methodology. These elements allow the benchmark to be reproducible and scalable across diverse AI deployments.
EVIDENCE
She outlines that “the definition of a benchmark … is a taxonomy, a data set, and an evaluator system” and that the methodology and reference builds enable engineers to scale the approach [252-254]. She also notes the challenge of estimating uncertainty and providing probabilistic risk estimates under defined assumptions [255-260].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S1 details the same definition of a benchmark—taxonomy, data set, evaluator system—tied together by a rigorous methodology, supporting Weiss’s claim.
MAJOR DISCUSSION POINT
Measurement and Benchmarking
A
Amanda Craig
4 arguments180 words per minute984 words327 seconds
Argument 1
Translating high‑level norms into practice aligns with policy (Amanda Craig)
EXPLANATION
Amanda argues that the difficulty lies in turning broad AI governance norms into concrete, actionable practices across the AI value chain. She stresses that standards are needed to bridge the gap between high‑level expectations and day‑to‑day risk‑management responsibilities of developers, deployers, and users.
EVIDENCE
She notes that “we want to know how AI providers are managing risk, but we are in the early days of defining really what that means in practice” and that this translation is essential for aligning with policy expectations [153-160].
MAJOR DISCUSSION POINT
Standards vs Regulation/Policy
Argument 2
Internal responsible AI standard aligns stakeholders; external standards provide common language (Amanda Craig)
EXPLANATION
Amanda describes Microsoft’s internal Responsible AI Standard, which aligns product, engineering, and sales teams around what “good” looks like. She adds that external standards are needed to give the broader ecosystem a shared language and expectations.
EVIDENCE
She explains that Microsoft defines a “responsible AI standard that applies to all of our internal … product groups, our engineering function, our sales function” to align internal stakeholders, and calls for partnership with industry and governments to define external standards [42-46].
MAJOR DISCUSSION POINT
Trust and Consumer Confidence
AGREED WITH
Bhushan Sethi, Esther Tetruashvily, Kshitij Bathla, Lee Wan Sie, Joslyn Barnhart, Chris Meserole
Argument 3
Standards needed to assess progress, uncertainty levels across sectors (Amanda Craig)
EXPLANATION
Amanda points out that without common standards it is hard to gauge whether the AI field has moved beyond its nascent stage or to what degree uncertainty is acceptable in different sectors. She calls for standardized ways to measure progress and determine when AI systems are reliable enough for deployment.
EVIDENCE
She asks “how do we know if we have made sufficient progress?” and argues that “we need to standardize how we assess progress, uncertainty levels, and when we can rely on these systems” [274-277].
MAJOR DISCUSSION POINT
Measurement and Benchmarking
AGREED WITH
Rebecca Weiss, Chris Meserole, Bhushan Sethi
Argument 4
Interoperable, modular standards avoid reinventing the wheel (Amanda Craig)
EXPLANATION
Amanda envisions a future where standards are modular and interoperable, allowing different sectors and use‑cases to reuse common components rather than building new ones from scratch. This approach would accelerate progress and keep standards up‑to‑date with evolving science.
EVIDENCE
She describes a “system of standards that are interoperable where we have a sort of modular approach” and stresses the need to avoid “starting from scratch with every piece of that puzzle” while evolving benchmarks and methodology [388-392].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need for interoperable, modular standards is reinforced by S25, which stresses open interoperability and avoiding vendor lock‑in, and by S28, which calls for avoiding reinventing the wheel and reusing existing best practices.
MAJOR DISCUSSION POINT
Implementation, Future‑Proofing, Outlook
AGREED WITH
Chris Meserole, Lee Wan Sie, Bhushan Sethi
E
Etienne Chaponniere
3 arguments194 words per minute1066 words328 seconds
Argument 1
AI safety standards trail products, differ from telecom compliance (Etienne Chaponniere)
EXPLANATION
Etienne contrasts the telecom world, where products cannot be shipped without compliance, with AI, where safety standards typically appear after products are already on the market. He emphasizes that AI standards must become widely available and easy for engineering teams to adopt.
EVIDENCE
He notes that “you cannot ship a product unless you comply to a standard” in telecom, whereas “in the world of AI standards, it’s a bit different… safety standards typically trail the products” [20-24].
MAJOR DISCUSSION POINT
Definition and Purpose of AI Standards
Argument 2
Open governance models ensure small firms can comply (Etienne Chaponniere)
EXPLANATION
Etienne argues that for standards to be inclusive, they must be open and governed in a way that allows participation from smaller companies that lack resources to develop their own standards. Open models like ML Commons and ISO enable this inclusivity.
EVIDENCE
He states that “there needs to be an opportunity for everyone to participate” and cites open governance models such as ML Commons, ISO, and Sentinelic as mechanisms for inclusive participation [179-182].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S24 highlights inclusive access to AI standards and broad stakeholder participation, while S30 discusses how small companies benefit from open, democratic standards, aligning with Etienne’s point.
MAJOR DISCUSSION POINT
Inclusivity and Global Cooperation
AGREED WITH
Rebecca Weiss, Lee Wan Sie, Chris Meserole
Argument 3
No silver bullet; multiple language safety tests required (Etienne Chaponniere)
EXPLANATION
Etienne acknowledges that language bias cannot be solved by a single solution; instead, a variety of safety tests and prompts must be created for different languages and dialects. He stresses community involvement to define language‑specific requirements while keeping tools reusable.
EVIDENCE
He says “there’s no silver bullet solution” and that “there’s going to be a need for more than one language” and that the community must decide what to capture for each language, while keeping the tooling efficient and reusable [451-460].
MAJOR DISCUSSION POINT
Language Bias and Technical Specifics
AGREED WITH
Esther Tetruashvily, Audience
E
Esther Tetruashvily
3 arguments180 words per minute1072 words355 seconds
Argument 1
Standards translate risk management into trust language, ISO certification (Esther Tetruashvily)
EXPLANATION
Esther explains that standards help convert OpenAI’s internal risk‑management practices into a language that customers can understand, building trust. She highlights OpenAI’s ISO 42001 certification as a concrete signal of compliance and reliability.
EVIDENCE
She says standards “translate some of our practices for risk management into the language of risk management for customers” and notes that OpenAI is “certified in ISO 42001” which signals trust to the market [56-59][134-136].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S23 explains how standards turn high‑level risk‑management principles into actionable tools that build trust with customers, supporting the claim about translating risk practices and the role of ISO certification.
MAJOR DISCUSSION POINT
Trust and Consumer Confidence
AGREED WITH
Bhushan Sethi, Amanda Craig, Kshitij Bathla, Lee Wan Sie, Joslyn Barnhart, Chris Meserole
Argument 2
OpenAI’s safety hub, model cards, ISO 42001 certification as measurement tools (Esther Tetruashvily)
EXPLANATION
Esther describes concrete measurement artifacts OpenAI uses: model cards detailing performance, a publicly updated safety hub, and ISO 42001 certification. These tools provide transparent evidence of safety and reliability for users and regulators.
EVIDENCE
She mentions “model cards performance on a variety of metrics” and a “safety hub that gets updated regularly” as well as the ISO 42001 certification that signals adherence to industry best practices [133-139].
MAJOR DISCUSSION POINT
Measurement and Benchmarking
Argument 3
MMLU and Indian dialect tests illustrate language evaluation efforts (Esther Tetruashvily)
EXPLANATION
Esther notes that OpenAI evaluates multilingual capabilities using the MMLU benchmark and specific tests covering Indian dialects, demonstrating an active approach to language bias assessment. She calls for broader participation to improve these evaluations.
EVIDENCE
She states “we do a series of evaluations like MMLU for determining how well our models perform on a variety of languages” and that they also test “a specific test in QA that has a variety of dialects within India” [441-444].
MAJOR DISCUSSION POINT
Language Bias and Technical Specifics
AGREED WITH
Etienne Chaponniere, Audience
L
Lee Wan Sie
4 arguments171 words per minute917 words320 seconds
Argument 1
Global norms define “good” for AI (Lee Wan Sie)
EXPLANATION
Lee describes standards as a way to set global norms that define what “good” looks like in AI governance, aligning expectations across jurisdictions. She emphasizes that these norms are technical, not merely checklist items.
EVIDENCE
She says standards mean “setting norms” and that this “means alignment globally on what good looks like” and that it involves “common methodologies and processes” [28-30].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S27 argues that international standards provide globally coherent metrics that define what “good” looks like in AI, echoing Lee’s statement about global norms.
MAJOR DISCUSSION POINT
Definition and Purpose of AI Standards
Argument 2
Standards useful without regulation; certification as market signal (Lee Wan Sie)
EXPLANATION
Lee argues that even in the absence of formal regulation, standards serve a valuable role by providing a market signal of quality and compliance. Certification, such as ISO 42001, allows organizations to differentiate themselves and demonstrate that they meet a recognized level of risk mitigation.
EVIDENCE
She notes that “even when there’s no regulations, I think the standards still are useful” and cites OpenAI’s ISO 42001 certification as an example of a market differentiator, stating that companies can “demonstrate that I have actually implemented something that’s good enough” [215-224].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S23 notes that standards give companies concrete tools to demonstrate compliance and act as market differentiators, supporting Lee’s view of certification as a market signal.
MAJOR DISCUSSION POINT
Standards vs Regulation/Policy
AGREED WITH
Chris Meserole
DISAGREED WITH
Chris Meserole, Joslyn Barnhart
Argument 3
Interconnected standards bodies require faster coordination (Lee Wan Sie)
EXPLANATION
Lee points out that many standards organizations (ISO, ML Commons, IEEE) are interlinked, and that faster coordination among them is needed to produce timely global standards. She mentions ongoing work to accelerate testing and benchmarking within the ISO process.
EVIDENCE
She lists the interconnected bodies – ISO, ML Commons, IEEE – and says “we hope that in the next one year that can be done and sorted and accepted within the ISO process” while acknowledging the typical slowness of standard development [206-212][376-382].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S1 calls for faster movement on testing and benchmarking methodologies, and S25 emphasizes the need for accelerated coordination among ISO, ML Commons, IEEE, matching Lee’s call for quicker coordination.
MAJOR DISCUSSION POINT
Inclusivity and Global Cooperation
AGREED WITH
Etienne Chaponniere, Rebecca Weiss, Chris Meserole
Argument 4
Need faster standard definition; market or regulator pressure drives adoption (Lee Wan Sie)
EXPLANATION
Lee stresses that both market demand and regulatory expectations can accelerate the creation and adoption of standards. She expresses optimism that within a year progress can be made on testing methodologies and that momentum will increase.
EVIDENCE
She says “there will be some momentum, either from the market or from regulations, to move standards” and that they hope to complete work on testing and benchmarking within a year and get it accepted in ISO [376-382].
MAJOR DISCUSSION POINT
Implementation, Future‑Proofing, Outlook
AGREED WITH
Chris Meserole, Amanda Craig, Bhushan Sethi
J
Joslyn Barnhart
3 arguments188 words per minute459 words146 seconds
Argument 1
Regulation cites non‑existent standards, creating compliance need (Joslyn Barnhart)
EXPLANATION
Joslyn observes that current regulations often refer to standards that have not yet been developed, forcing companies to anticipate or create those standards to achieve compliance. This creates pressure for organizations like Google DeepMind to prioritize standard‑setting activities.
EVIDENCE
She notes that “regulation has gone ahead and jumped to… we have regulated and essentially made reference to standards that do not yet exist” and that this makes standard development an “utmost priority” for Google DeepMind [49-51].
MAJOR DISCUSSION POINT
Standards vs Regulation/Policy
Argument 2
Standards raise safety floor, prevent race to the bottom (Joslyn Barnhart)
EXPLANATION
Joslyn argues that establishing common safety standards lifts the minimum level of safety across the industry, discouraging a race to the bottom where firms might cut corners. Collective incentives drive firms to adopt higher safety baselines.
EVIDENCE
She says “the worst thing for adoption would be a safety incident” and that there is a “collective incentive as an industry to make sure that we raise the floor to avoid that” [145-147].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S23 notes that standards raise the baseline of safety across the industry and help avoid a race to the bottom, supporting Barnhart’s argument.
MAJOR DISCUSSION POINT
Trust and Consumer Confidence
AGREED WITH
Bhushan Sethi, Amanda Craig, Esther Tetruashvily, Kshitij Bathla, Lee Wan Sie, Chris Meserole
Argument 3
Performance of standards as evidence of conformity (Joslyn Barnhart)
EXPLANATION
In response to audience concerns, Joslyn explains that when standards are referenced in regulation, they become a minimum bar that regulators can accept as evidence of conformity, ensuring that standards are not merely abstract but have practical regulatory weight.
EVIDENCE
She states that “if we make these things too high level… regulators are not going to look at those standards as evidence of conformity” and that standards create “interlocking pressure” from regulation [436-438].
MAJOR DISCUSSION POINT
Industry‑driven standards risk performativity; auditability challenge
AGREED WITH
Rebecca Weiss, Bhushan Sethi, Chris Meserole
C
Chris Meserole
3 arguments204 words per minute1311 words385 seconds
Argument 1
Standards give legitimacy and fill policy gaps (Chris Meserole)
EXPLANATION
Chris emphasizes that formal standard‑setting bodies provide legitimacy, openness, and credibility that pure industry or government efforts lack, thereby bridging policy gaps and supporting collective action on AI risk.
EVIDENCE
He notes that “standard-setting bodies are open” and that they bring “legitimacy and credibility” which are missing when standards are set only by industry or government [112-114].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S26 highlights that standard‑setting bodies bring legitimacy, openness, and credibility, directly supporting Meserole’s claim.
MAJOR DISCUSSION POINT
Standards vs Regulation/Policy
AGREED WITH
Etienne Chaponniere, Rebecca Weiss, Lee Wan Sie
Argument 2
Process standards need scientific benchmarks for risk evaluation (Chris Meserole)
EXPLANATION
Chris distinguishes high‑level process standards (identifying, evaluating, mitigating risks) from the scientific benchmarks needed to actually measure those risks. He argues that both layers are essential for a coherent risk‑management framework.
EVIDENCE
He describes the process of “identifying risks, evaluating risks, putting in place mitigations” as needing standardization, and then adds that “once we have agreed on what the risk is… we need scientific benchmarks” [295-298].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S1 outlines the benchmark methodology (taxonomy, dataset, evaluator) as the scientific foundation needed to evaluate AI risks, aligning with Meserole’s argument.
MAJOR DISCUSSION POINT
Measurement and Benchmarking
AGREED WITH
Rebecca Weiss, Amanda Craig, Bhushan Sethi
Argument 3
Process standards future‑proof; evaluations must evolve with model capabilities (Chris Meserole)
EXPLANATION
Chris contends that while the overarching risk‑management process can remain stable over time, the specific evaluation methods must be updated as AI models become more capable. This ensures standards stay relevant without needing complete redesign.
EVIDENCE
He states that “the process is somewhat agnostic” and that “specific evals you run are probably going to have to be updated over time to account for the greater capabilities of models” [348-351].
MAJOR DISCUSSION POINT
Implementation, Future‑Proofing, Outlook
AGREED WITH
Amanda Craig, Lee Wan Sie, Bhushan Sethi
B
Bhushan Sethi
3 arguments110 words per minute1735 words943 seconds
Argument 1
Standards demystify AI, distinguish from regulation (Bhushan Sethi)
EXPLANATION
Bhushan highlights the need to clarify what AI standards are, separating them from broader regulatory or legislative frameworks. He points out the confusion that exists among stakeholders about these concepts.
EVIDENCE
He says “we need to demystify what we mean by standard setting” and later notes the “confusion between standards, regulation, legislation” [4][67-68].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S23 describes how standards translate principles into concrete practices and clarify accountability, helping demystify AI standards versus regulation.
MAJOR DISCUSSION POINT
Definition and Purpose of AI Standards
Argument 2
Trust and verification are central to AI adoption (Bhushan Sethi)
EXPLANATION
Bhushan stresses that for AI to be widely adopted, there must be trustworthy, verifiable processes for reporting, disclosure, and credibility, moving beyond superficial check‑boxes.
EVIDENCE
He asks “How do we report? How do we disclose? How do we make it credible?” emphasizing the need for non-subjective verification [92-95].
MAJOR DISCUSSION POINT
Trust and Consumer Confidence
AGREED WITH
Amanda Craig, Esther Tetruashvily, Kshitij Bathla, Lee Wan Sie, Joslyn Barnhart, Chris Meserole
Argument 3
Two‑year goal: certification as consensus on “good enough” (Bhushan Sethi)
EXPLANATION
Bhushan envisions that within the next two years, the industry will see certifications that embody a consensus on what constitutes “good enough” for AI systems, providing a clear benchmark for compliance and trust.
EVIDENCE
He states his hope to see “certification that represent more types of consensus” and that “definition of what is good enough deserves some form of certification” over the next two years [337-340].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S5 discusses the concept of “good enough” as a consensus defined by standards, directly relating to Bhushan’s two‑year certification goal.
MAJOR DISCUSSION POINT
Implementation, Future‑Proofing, Outlook
AGREED WITH
Rebecca Weiss, Chris Meserole, Joslyn Barnhart
K
Kshitij Bathla
3 arguments149 words per minute526 words210 seconds
Argument 1
Standards verify AI claims, enable consumer trust (Kshitij Bathla)
EXPLANATION
Kshitij describes standards as tools that build consumer trust by ensuring that AI products meet quality expectations, thereby facilitating industry adoption and consumer confidence.
EVIDENCE
He says standards “are the tools which enables consumers’ trust in whatever ecosystem for which they are developed” and also help industry ensure quality [62-63].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S23 states that standards provide tools for risk management and help build consumer trust, supporting Bathla’s claim.
MAJOR DISCUSSION POINT
Definition and Purpose of AI Standards
AGREED WITH
Bhushan Sethi, Amanda Craig, Esther Tetruashvily, Lee Wan Sie, Joslyn Barnhart, Chris Meserole
Argument 2
Standards enable consumer trust and industry quality (Kshitij Bathla)
EXPLANATION
He reiterates that standards serve as a mechanism for both consumers to trust AI systems and for industries to maintain consistent quality across products.
EVIDENCE
He repeats that standards “enable consumer trust” and “ensure the quality and the consumer trust” [62-63].
MAJOR DISCUSSION POINT
Trust and Consumer Confidence
Argument 3
Global standards must be adaptable to local contexts (Kshitij Bathla)
EXPLANATION
Kshitij explains that while standards should be globally consistent, they must also be flexible enough to address India‑specific risks and use‑cases, such as those highlighted in the Manav mission.
EVIDENCE
He references the “Manav mission” as human-centric, notes that India’s governance guidelines provide a framework, and stresses the need to adapt global standards to local conditions while also developing India-specific guidance [201-208][212-218].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S24 emphasizes inclusive participation and adaptation to diverse contexts, while S30 notes the importance of small‑firm and local considerations, aligning with Bathla’s point.
MAJOR DISCUSSION POINT
Inclusivity and Global Cooperation
A
Audience
3 arguments159 words per minute387 words145 seconds
Argument 1
Industry‑driven standards risk performativity; auditability challenge (Audience)
EXPLANATION
An audience member questions whether industry‑led standards might become superficial, serving industry interests rather than public needs, and raises concerns about how governments can audit such programs given skill gaps.
EVIDENCE
The audience asks “How do we know … you’re not just creating something that cheaply satisfies the industry … how does a government or external agency audit such a program, given the skill gap?” [398-405].
MAJOR DISCUSSION POINT
Standards vs Regulation/Policy
Argument 2
Balancing minimum viable consensus with diverse stakeholder demands (Audience)
EXPLANATION
Another audience participant asks whether standards should aim for a minimal viable consensus that includes the broadest set of stakeholders, or whether they should attempt to satisfy all stakeholder demands, even when they conflict.
EVIDENCE
The audience asks “do you imagine that we’re talking about minimum viable consensus with the broadest number of stakeholders, or is there a path to address issues that some stakeholders see as absolutely necessary and others don’t?” [428-429].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
S24 stresses the importance of inclusive, multistakeholder participation in standards development, providing context for the discussion on minimum viable consensus.
MAJOR DISCUSSION POINT
Inclusivity and Global Cooperation
Argument 3
Audience query on handling multilingual bias in AI development (Audience)
EXPLANATION
A student from India raises a question about how to address language bias in AI models given India’s linguistic diversity, seeking guidance on building guardrails for multilingual development.
EVIDENCE
The audience member states “there are 22 official languages… how do you go about tackling language bias and building guardrails…” [418-420].
MAJOR DISCUSSION POINT
Language Bias and Technical Specifics
AGREED WITH
Esther Tetruashvily, Etienne Chaponniere
Agreements
Agreement Points
Standards are essential to build trust, credibility and consumer confidence in AI systems
Speakers: Bhushan Sethi, Amanda Craig, Esther Tetruashvily, Kshitij Bathla, Lee Wan Sie, Joslyn Barnhart, Chris Meserole
Trust and verification are central to AI adoption (Bhushan Sethi) Internal responsible AI standard aligns stakeholders; external standards provide common language (Amanda Craig) Standards translate risk management into trust language, ISO certification (Esther Tetruashvily) Standards verify AI claims, enable consumer trust (Kshitij Bathla) Standards useful without regulation; certification as market signal (Lee Wan Sie) Standards raise safety floor, prevent race to the bottom (Joslyn Barnhart) Standards give legitimacy and fill policy gaps (Chris Meserole)
Multiple panelists emphasized that standards provide a common language, certification and measurable assurances that help consumers and enterprises trust AI products, even serving as market differentiators and raising the overall safety floor [92-95][42-46][56-59][62-63][215-224][145-147][112-114].
POLICY CONTEXT (KNOWLEDGE BASE)
This view aligns with the strategic framing of open AI standards as essential for trust and market confidence highlighted in the U.S. AI Standards discussion, which draws parallels to historic standardisation in the internet and automotive sectors [S53], and reflects the ITU’s emphasis on building a trustworthy AI ecosystem through extensive standard publication [S54].
Defining “good enough” through consensus‑based standards and certification
Speakers: Rebecca Weiss, Bhushan Sethi, Chris Meserole, Joslyn Barnhart
Benchmarking methodology as core of standards (Rebecca Weiss) Two‑year goal: certification as consensus on “good enough” (Bhushan Sethi) Standards give legitimacy and fill policy gaps (Chris Meserole) Performance of standards as evidence of conformity (Joslyn Barnhart)
Speakers agreed that standards must embody a consensus on what is “good enough”, with benchmarking defining that threshold and certification signalling it to the market and regulators [97-102][337-340][108-110][112-114][436-438].
POLICY CONTEXT (KNOWLEDGE BASE)
Consensus-based approaches to defining adequacy are echoed in the low-disagreement findings of the Global AI Standards panel, where the balance between industry leadership and broader stakeholder input was stressed [S43], and in the high-level consensus on methodology for AI governance [S44].
Open, inclusive, multistakeholder governance is needed for AI standards to be accessible to small firms
Speakers: Etienne Chaponniere, Rebecca Weiss, Lee Wan Sie, Chris Meserole
Open governance models ensure small firms can comply (Etienne Chaponniere) Benchmarking methodology … need more stakeholders (Rebecca Weiss) Interconnected standards bodies require faster coordination (Lee Wan Sie) Standards give legitimacy and fill policy gaps (Chris Meserole)
Panelists highlighted that standards must be developed through open processes that allow participation from diverse stakeholders, ensuring smaller companies can adopt them without prohibitive costs [179-182][99-102][206-212][376-382][112-114].
POLICY CONTEXT (KNOWLEDGE BASE)
Multistakeholder governance is repeatedly underscored as critical for inclusive AI standards, from the IGF’s call for diverse participation across sectors [S48] to the interdisciplinary coordination involving UNESCO, OECD and others [S47], and the emphasis on contextual engagement to empower smaller actors [S56].
Benchmarking methodology and clear measurement taxonomy are foundational for effective AI standards
Speakers: Rebecca Weiss, Chris Meserole, Amanda Craig, Bhushan Sethi
Benchmarking methodology as core of standards (Rebecca Weiss) Process standards need scientific benchmarks for risk evaluation (Chris Meserole) Standards needed to assess progress, uncertainty levels across sectors (Amanda Craig) Clarity of taxonomy, measurement needed (Bhushan Sethi)
All agreed that a rigorous benchmark-comprising taxonomy, dataset and evaluator-paired with a clear methodology is essential to quantify AI risk and progress [252-261][295-298][274-277][262-264].
POLICY CONTEXT (KNOWLEDGE BASE)
The need for robust benchmarking and taxonomy is highlighted in discussions about coordinating frontier AI risk-management frameworks, where clear evaluation methods are deemed essential for effective standardisation [S49] and for aligning global AI safety coordination efforts [S50].
Standards provide value even in the absence of formal regulation
Speakers: Lee Wan Sie, Chris Meserole
Standards useful without regulation; certification as market signal (Lee Wan Sie) Regulators offload to standards (Chris Meserole)
Both speakers argued that standards serve as market signals and can be leveraged by regulators to define compliance expectations, even when explicit regulations are lacking [215-224][194-196].
POLICY CONTEXT (KNOWLEDGE BASE)
Voluntary standards are recognised for delivering economic and innovation benefits without regulatory mandates, as noted in analyses of digital-economy standards [S52] and arguments for market-oriented regulatory approaches that still rely on standards for trust [S55].
Future‑proofing AI standards through modular, interoperable designs and evolving evaluation methods
Speakers: Chris Meserole, Amanda Craig, Lee Wan Sie, Bhushan Sethi
Process standards future‑proof; evaluations must evolve with model capabilities (Chris Meserole) Interoperable, modular standards avoid reinventing the wheel (Amanda Craig) Need faster standard definition; market or regulator pressure drives adoption (Lee Wan Sie) Two‑year outlook, certification etc. (Bhushan Sethi)
Panelists concurred that while high-level processes can remain stable, specific benchmarks must be updated as AI models advance, and modular standards can accelerate adoption across sectors [348-351][388-392][376-382][336-342].
POLICY CONTEXT (KNOWLEDGE BASE)
Future-proofing is advocated in calls for faster, modular standard development to keep pace with frontier AI, emphasizing interoperable designs and iterative evaluation [S49], and reflected in the broader consensus on adaptable governance mechanisms [S45].
Addressing language bias requires multilingual evaluation efforts and community involvement
Speakers: Esther Tetruashvily, Etienne Chaponniere, Audience
MMLU and Indian dialect tests illustrate language evaluation efforts (Esther Tetruashvily) No silver bullet; multiple language safety tests required (Etienne Chaponniere) Audience query on handling multilingual bias in AI development (Audience)
All three highlighted the challenge of multilingual bias, noting existing tests (MMLU, Indian dialects) and the need for multiple language-specific safety tests developed collaboratively [441-444][451-460][418-420].
Similar Viewpoints
Both see standards as a tool for regulators and markets to define and meet AI risk expectations even when formal regulations are not yet in place [215-224][194-196].
Speakers: Lee Wan Sie, Chris Meserole
Standards useful without regulation; certification as market signal (Lee Wan Sie) Regulators offload to standards (Chris Meserole)
Both stress the need for open, coordinated standards bodies that enable participation from smaller companies and accelerate standard development [179-182][376-382].
Speakers: Etienne Chaponniere, Lee Wan Sie
Open governance models ensure small firms can comply (Etienne Chaponniere) Interconnected standards bodies require faster coordination (Lee Wan Sie)
Both argue that scientific benchmarking is a necessary component of AI standards to enable reliable risk assessment [252-261][295-298].
Speakers: Rebecca Weiss, Chris Meserole
Benchmarking methodology as core of standards (Rebecca Weiss) Process standards need scientific benchmarks for risk evaluation (Chris Meserole)
Both view standards as a bridge translating internal risk practices into a common external language that builds trust with customers and regulators [42-46][56-59].
Speakers: Amanda Craig, Esther Tetruashvily
Internal responsible AI standard aligns stakeholders; external standards provide common language (Amanda Craig) Standards translate risk management into trust language, ISO certification (Esther Tetruashvily)
Unexpected Consensus
Broad agreement that standards are needed and valuable even though global AI regulation is still fragmented
Speakers: Lee Wan Sie, Chris Meserole, Rebecca Weiss, Bhushan Sethi
Standards useful without regulation; certification as market signal (Lee Wan Sie) Regulators offload to standards (Chris Meserole) Benchmarking methodology as core of standards (Rebecca Weiss) Trust and verification are central to AI adoption (Bhushan Sethi)
Despite the lack of coordinated global AI legislation, industry leaders and standard-setting advocates converged on the necessity of standards to fill policy gaps, provide market signals, and establish trust, which was not an obvious expectation at the start of the discussion [215-224][194-196][13-14][92-95].
POLICY CONTEXT (KNOWLEDGE BASE)
Multiple sources report high consensus on the necessity of standards despite fragmented regulation, including the Global AI Policy Framework’s emphasis on common principles [S45] and the IGF’s observation of strong alignment on AI governance direction [S44].
Overall Assessment

The panel displayed strong consensus that AI standards are critical for building trust, defining “good enough”, ensuring inclusivity, and providing measurable benchmarks, with agreement across industry, policy, and technical perspectives. There is also shared recognition of the need for open, multistakeholder processes and future‑proof, modular designs.

High consensus across most thematic areas, indicating a unified stance that standards will be a cornerstone for responsible AI deployment and that coordinated, inclusive efforts are essential for their success.

Differences
Different Viewpoints
Whether AI standards should be primarily driven by regulation or can be valuable and market‑driven in the absence of regulation
Speakers: Lee Wan Sie, Chris Meserole, Joslyn Barnhart
Standards useful without regulation; certification as market signal (Lee Wan Sie) Standards fill policy gaps; regulators may offload requirements to standards (Chris Meserole) Standards become evidence of conformity when referenced in regulation (Joslyn Barnhart)
Lee argues that standards remain useful even when no formal regulations exist, serving as a market differentiator through certification [215-224]. Chris contends that standards are needed to fill policy gaps, with regulators often delegating risk-management requirements to the standards process [106-110]. Joslyn points out that standards gain regulatory weight only when they are explicitly referenced in law, serving as a minimum bar for compliance [436-438]. These positions reflect a tension between viewing standards as independent market tools versus as extensions of regulatory frameworks.
POLICY CONTEXT (KNOWLEDGE BASE)
The tension between regulation-led and market-driven standards is documented in debates over a market-oriented regulatory model [S55] and in moderate disagreement about divergent safety approaches that could lead to fragmented policy responses [S50].
Desired speed and coordination of AI standard development
Speakers: Lee Wan Sie, Etienne Chaponniere, Chris Meserole
Need faster coordination among standards bodies and hope to finalize testing work within a year (Lee Wan Sie) Acknowledges that standard development is slow and takes time (Etienne Chaponniere) Emphasises that standards must be open and credible, which can lengthen the process (Chris Meserole)
Lee pushes for accelerated progress, stating a goal to complete testing and benchmarking work within a year and be accepted by ISO [376-382]. Etienne notes the reality that standard-setting “takes a while” and cannot be rushed [383-384]. Chris adds that openness and legitimacy of standard-setting bodies are essential, implying that thorough, inclusive processes may limit speed [112-114]. The speakers thus differ on how quickly standards should be produced versus the practical constraints of inclusive, credible development.
POLICY CONTEXT (KNOWLEDGE BASE)
Disagreement over pace is highlighted in the Global AI Standards panel, where participants noted differing views on how quickly standards should be produced [S43], and in calls for accelerated formal standards to match rapid AI advances [S49].
Risk of industry‑driven standards being performative versus their legitimacy and auditability
Speakers: Audience, Chris Meserole, Lee Wan Sie
Industry‑driven standards may be superficial; governments lack capacity to audit (Audience) Standard‑setting bodies provide legitimacy and openness missing from pure industry or government efforts (Chris Meserole) Certification offers a market signal of quality even without regulation (Lee Wan Sie)
An audience member questions whether standards created mainly by industry will be merely performative and how governments can audit them given skill gaps [398-405]. Chris counters that formal standard-setting bodies bring openness, legitimacy, and credibility that pure industry or government actions lack [112-114]. Lee reinforces that certification can serve as an independent market signal of compliance, even absent regulation [215-224]. This reflects a disagreement between external skepticism about performativity and internal confidence in the legitimacy of standards processes.
POLICY CONTEXT (KNOWLEDGE BASE)
Concerns about the legitimacy of industry-led standards appear in analyses of politicised standard-setting that may compromise technical merit [S51] and in observations of challenges around industry engagement and auditability [S52].
Unexpected Differences
Audience skepticism about the performative nature of industry‑driven standards versus panel confidence in their legitimacy
Speakers: Audience, Chris Meserole, Lee Wan Sie
Industry‑driven standards risk being superficial; governments lack audit capacity (Audience) Standard‑setting bodies provide legitimacy and openness missing from pure industry or government (Chris Meserole) Certification offers a market signal of quality even without regulation (Lee Wan Sie)
The audience’s concern that standards may be crafted to satisfy industry interests rather than public needs, and that governments lack the technical capacity to audit them, was not directly addressed by the panelists, who instead emphasized the inherent legitimacy of standard‑setting bodies and the value of certification as an independent quality signal. This gap between external critique and internal assurance was not anticipated in the earlier discussion.
POLICY CONTEXT (KNOWLEDGE BASE)
Direct audience criticism of a panel’s lack of technical practitioner representation, questioning the legitimacy of discussed standards, was recorded at the IGF workshop [S57].
Overall Assessment

The panel largely converged on the importance of AI standards for trust, risk management, and global cooperation, but diverged on the relationship between standards and regulation, the pace of standard development, and the risk of performative, industry‑driven processes. These disagreements highlight the need for clearer governance frameworks, faster yet inclusive standard‑setting mechanisms, and stronger auditability to ensure standards serve public interests.

Moderate – while there is broad consensus on goals, the differing views on implementation pathways and regulatory interplay suggest potential friction that could affect the timely and effective adoption of AI standards.

Partial Agreements
All speakers agree that building trust and ensuring verifiable, reliable AI systems is essential, but they differ on the mechanisms: Bhushan emphasizes reporting and disclosure frameworks; Amanda highlights internal corporate standards and external common language; Esther points to ISO certification and safety hubs; Joslyn focuses on industry‑wide safety baselines; Chris stresses high‑level process standards coupled with scientific benchmarks. This shows consensus on the goal of trust, with varied pathways to achieve it.
Speakers: Bhushan Sethi, Amanda Craig, Esther Tetruashvily, Joslyn Barnhart, Chris Meserole
Trust and verification are central to AI adoption (Bhushan Sethi) Internal responsible AI standard aligns stakeholders; external standards provide common language (Amanda Craig) Standards translate risk management into trust language; ISO certification signals reliability (Esther Tetruashvily) Standards raise safety floor, prevent race to the bottom (Joslyn Barnhart) Process standards needed for risk identification, evaluation, mitigation (Chris Meserole)
The speakers share the objective of creating inclusive, widely applicable AI standards, but differ on emphasis: Rebecca stresses a technical benchmark methodology and the need for diverse stakeholder input [96-101]; Etienne highlights open governance and participation of smaller companies as essential for accessibility [179-182]; Lee focuses on establishing global norms and market‑driven certification as a signal of quality [215-224]. Thus, they agree on inclusivity but propose different primary levers.
Speakers: Rebecca Weiss, Etienne Chaponniere, Lee Wan Sie
Benchmarking methodology as core of standards; need broad stakeholder consensus (Rebecca Weiss) Open governance models ensure small firms can comply (Etienne Chaponniere) Global norms define “good” for AI; standards useful even without regulation (Lee Wan Sie)
Takeaways
Key takeaways
AI standards are essential to translate high‑level norms into concrete, verifiable practices and to build consumer and enterprise trust. Benchmarking methodology—taxonomy, dataset, and evaluator—is the technical core of AI standards and enables measurement of uncertainty and risk. Standards differ from regulation but can complement it; they provide legitimacy, a common language, and a market signal even where no formal rules exist. Global cooperation and open‑governance models are needed so that standards are inclusive, scalable, and usable by both large and small firms. Process‑oriented standards (risk identification, evaluation, mitigation) are more future‑proof, while specific technical benchmarks must evolve with model capabilities. Certification (e.g., ISO 42001) serves as evidence of “good enough” compliance and can differentiate trustworthy products. Addressing language bias requires dedicated multilingual test suites and community participation; no single solution will cover all languages. A modular, interoperable standards ecosystem is needed to avoid reinventing the wheel as AI applications diversify across sectors.
Resolutions and action items
Panelists agreed to deepen collaboration with ML Commons to develop benchmark methodologies and reference implementations. Qualcomm (Etienne) will advocate for open, accessible standards that enable smaller companies to achieve compliance without building their own frameworks. OpenAI (Esther) will continue publishing its safety hub, model cards, and pursue ISO 42001 certification as a benchmark for trust. Microsoft (Amanda) will work with industry and civil‑society stakeholders to define measurable progress indicators for AI maturity. Singapore (Lee) will push forward ISO‑level testing and benchmarking work, aiming for a draft within the next year. India (Kshitij) will align national standards with global ISO/IEC SC42 outputs while identifying India‑specific use‑case guidance. All participants committed to advancing modular, interoperable standards that can be updated as model capabilities evolve.
Unresolved issues
How governments and external auditors can effectively verify industry‑driven standards given the technical skill gap. The extent to which standards should be prescriptive versus allowing flexibility for diverse stakeholder needs. Mechanisms for achieving consensus on contentious risk categories where some jurisdictions consider certain risks essential and others do not. Concrete timelines and processes for rapidly developing and ratifying new standards to keep pace with fast‑moving AI models. Specific approaches for comprehensive multilingual evaluation and mitigation of language bias beyond existing test suites.
Suggested compromises
Adopt standards as a minimum floor (baseline) while allowing higher‑level or sector‑specific standards to be layered on top. Use a modular standards architecture so that common core components are shared, and specialized extensions can address local or domain‑specific requirements. Combine market‑driven adoption incentives with regulatory expectations to ensure standards attain sufficient rigor without being overly burdensome. Maintain open governance and inclusive participation to balance the interests of large tech firms with those of smaller companies and civil‑society groups.
Thought Provoking Comments
Regulation has jumped ahead and referenced standards that do not yet exist. For places like Google DeepMind who have not invested heavily in the standard space in the past, this is now of utmost priority because we actually need this to assist with implementation and compliance.
Highlights a critical mismatch where policymakers are demanding compliance to standards that are still under development, creating urgency for the industry to engage in standard‑setting.
Shifted the conversation from abstract benefits of standards to a concrete pressure point, prompting other panelists to discuss how their organisations are accelerating standard‑development and underscoring the need for rapid, collaborative action.
Speaker: Joslyn Barnhart (Google DeepMind)
A big part of what standards are for is to try and solve this collective action problem… having a formal standard‑setting body is open, so there’s legitimacy and credibility you don’t have if it’s just industry or just government.
Frames standards as a mechanism to align diverse actors and overcome the classic collective‑action dilemma, emphasizing openness and legitimacy as essential qualities.
Provided a theoretical foundation that other speakers (e.g., Rebecca, Etienne) referenced when discussing inclusivity and openness, steering the dialogue toward the governance structure of standard bodies rather than just technical details.
Speaker: Chris Meserole (Frontier Model Forum)
What is good enough? A standard represents a consensus about what is good enough. The problem is who contributes to that consensus – it shouldn’t be exclusively an industry perspective; you need broader stakeholder representation, and there’s both a scientific element (statistical guarantees) and a political element.
Introduces the nuanced question of “good enough” and points out the dual scientific‑political nature of standards, challenging the panel to think beyond technical metrics.
Prompted deeper discussion on inclusivity (Etienne’s point about smaller companies) and on the need for multi‑stakeholder processes, influencing later remarks about openness and the role of regulators.
Speaker: Rebecca Weiss (ML Commons)
In the telecom world you cannot ship a product unless you comply to a standard because you need it for interoperability. In AI standards we’re talking more about safety standards, and those typically trail the products. The products are out there, and then they’re going to comply to standards at some point when the standards are available.
Draws a clear contrast between mature, mandatory standards in telecom and the nascent, reactive nature of AI safety standards, exposing a timing gap that affects adoption.
Set the stage for later comments about the need for faster standard development (Lee’s regulatory timing, Amanda’s modular approach) and highlighted why AI standards are currently “behind” the technology curve.
Speaker: Etienne Chaponniere (Qualcomm)
If there are no regulations, standards are still useful – they can be a way for organisations to differentiate themselves, demonstrate that they have implemented something that is ‘good enough’, and provide certification assurance.
Counters the assumption that standards only matter when mandated, positioning them as market‑driven signals of trust and quality.
Re‑oriented the discussion toward the commercial value of standards, leading to Amanda’s and Etienne’s remarks about standards as a competitive advantage and the need for open, accessible frameworks.
Speaker: Lee Wan Sie (Singapore Government)
Future‑proofing standards is best done at the process‑level – a good risk‑identification and mitigation process can stay relevant even as model capabilities evolve; the specific evaluations will need updating, but the overarching framework can endure.
Provides a strategic lens for designing standards that remain relevant amid rapid AI advances, separating stable process elements from mutable technical tests.
Guided the conversation toward long‑term planning, influencing Amanda’s modular‑interoperable vision and reinforcing the importance of separating process standards from benchmark specifics.
Speaker: Chris Meserole (Frontier Model Forum)
We need a system of interoperable, modular standards so we don’t reinvent the wheel for every new use‑case or sector; the standards ecosystem should have synergy and evolve together.
Advocates for a cohesive, reusable standards architecture that balances evolution with efficiency, addressing the earlier concern about speed and duplication.
Synthesised earlier points about speed, openness, and modularity, and set a concrete direction for future work, prompting agreement from other panelists about avoiding siloed efforts.
Speaker: Amanda Craig (Microsoft)
If we make standards too high‑level or lowest‑common‑denominator, regulators won’t accept them as evidence of conformity. There needs to be a minimum bar of quality, otherwise standards become performative.
Highlights the risk of “performative” standards and stresses the need for substantive, regulator‑acceptable criteria, adding a pragmatic constraint to the earlier optimism.
Re‑focused the dialogue on the balance between accessibility and rigor, influencing later remarks about certification, credibility, and the role of regulators in enforcing standards.
Speaker: Joslyn Barnhart (Google DeepMind)
Overall Assessment

The discussion was driven forward by a handful of pivotal remarks that moved it from a generic endorsement of standards to a concrete, problem‑oriented dialogue. Joslyn’s observation about regulators demanding non‑existent standards created urgency; Chris’s framing of standards as a collective‑action solution gave the conversation a governance backbone; Rebecca’s ‘good enough’ question forced the panel to confront the scientific‑political trade‑offs and stakeholder inclusion; Etienne’s telecom analogy exposed the timing mismatch between product rollout and safety standards; Lee’s point about market‑driven differentiation showed standards’ value even without regulation; Chris’s future‑proofing insight introduced a strategic design principle; Amanda’s call for modular, interoperable standards offered a practical roadmap; and Joslyn’s warning against performative standards reminded everyone of the need for rigor. Together, these comments shifted the tone from abstract optimism to a nuanced, action‑oriented plan, shaping the panel’s consensus around speed, openness, legitimacy, and the balance between accessibility and regulatory credibility.

Follow-up Questions
What are the top three priority areas where standards are needed today (e.g., testing methodologies, transparency/disclosure formats, incident reporting and monitoring)?
Identifies key domains where standardisation can create alignment and trust across AI deployments.
Speaker: Lee Wan Sie
How should AI organisations report, disclose, and make compliance credible without it becoming a subjective tick‑box exercise?
Seeks concrete guidance on credible, verifiable reporting mechanisms to ensure trust and avoid perfunctory compliance.
Speaker: Bhushan Sethi
What additional considerations should be brought in from a standard‑setting perspective before the industry view is formed?
Requests input on foundational standard‑setting issues that may be overlooked by industry practitioners.
Speaker: Bhushan Sethi (to Chris Meserole and Rebecca Weiss)
Is there a disconnect between the strong industry consensus on AI standards and the lack of coordinated global regulations, and how should the audience interpret this gap?
Addresses the tension between voluntary standard adoption and the absence of binding regulatory frameworks.
Speaker: Bhushan Sethi (wild‑card question)
How can governments or external agencies effectively audit industry‑driven AI assurance programs given the technical skill gap?
Raises concern about public oversight and the ability of regulators to verify compliance with industry‑created standards.
Speaker: Audience member (unnamed)
What approaches can be used to detect and mitigate language bias in AI models, especially for multilingual contexts like India’s 22 official languages, and how can guardrails be built for small‑scale developers?
Highlights the need for multilingual fairness, evaluation data, and accessible tooling for developers.
Speaker: Audience member (computer‑science student)
Should AI standards aim for a minimum‑viable consensus that includes the broadest stakeholder base, or can they also address issues that some stakeholders consider essential but others deem unnecessary?
Explores the scope and inclusivity of standards‑setting processes versus targeted, high‑bar requirements.
Speaker: Audience member (Polonetsky)
What is needed to develop a clear, verifiable taxonomy, reference datasets, and evaluator systems for AI benchmarking, and how can these be standardized across industries?
Identifies foundational research required to create scalable, trustworthy benchmarking infrastructure.
Speaker: Rebecca Weiss
How can the speed of standards development be increased within bodies like ISO to keep pace with rapid AI advances?
Calls for process improvements to reduce lag between technology emergence and standard availability.
Speaker: Lee Wan Sie
How can standards be designed to be interoperable and modular across different sectors, use‑cases, and deployment scenarios, avoiding reinventing the wheel?
Emphasises the need for a cohesive, reusable standards ecosystem that adapts to varied applications.
Speaker: Amanda Craig
What mechanisms can future‑proof AI standards so they remain relevant as model capabilities evolve?
Seeks strategies to ensure standards retain applicability despite rapid technological change.
Speaker: Chris Meserole
How can smaller companies be included in the standard‑setting process and benefit from open, accessible standards despite limited resources?
Addresses inclusivity and the need for open governance models that lower participation barriers.
Speaker: Etienne Chaponniere
How can global standards be adapted to local contexts (e.g., India‑specific risks and use‑cases) while maintaining overall consistency?
Highlights the challenge of balancing worldwide harmonisation with region‑specific requirements.
Speaker: Kshitij Bathla
How can a consensus on what constitutes ‘good enough’ be reached across sectors with differing risk tolerances and expectations?
Points to the difficulty of defining acceptable risk thresholds that satisfy diverse industries.
Speaker: Rebecca Weiss
What is the role and impact of certification schemes (e.g., ISO 42001) on market trust and adoption of AI systems?
Investigates how formal certification can signal compliance and build stakeholder confidence.
Speaker: Esther Tetruashvily

Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.