Towards a Safer South Launching the Global South AI Safety Research Network

20 Feb 2026 17:00h - 18:00h

Towards a Safer South Launching the Global South AI Safety Research Network

Session at a glance

Summary

This discussion centered on the launch of the Global South Network for Trustworthy AI at the India AI Impact Summit, addressing the critical need for inclusive AI safety infrastructure that represents the perspectives of Global South countries. Dr. Urvashi Aneja, founder of Digital Futures Lab, opened by highlighting how AI systems are rapidly being deployed in critical sectors across the Global South, where low institutional capacity and deep inequities create both immense opportunities and significant risks. She emphasized that Global South organizations remain underrepresented in global AI safety governance, despite being uniquely positioned to identify real-world deployment risks invisible to lab-based evaluations.


Mr. Abhishek Singh from India’s AI Mission stressed that while everyone agrees on the need for safe AI, the challenge lies in developing technical tools and benchmarks to address risks, particularly for multilingual contexts where most models are evaluated only in English. Ambassador Philip Thigo from Kenya pointed out the structural exclusion of Global South countries from safety conversations and emphasized that safety must extend beyond technology to include environmental harms, biases, and full lifecycle accountability. The panelists identified several critical gaps in current AI safety approaches, including the need to redefine safety according to social and cultural contexts, address linguistic nuances beyond simple translation, and understand emerging harms during actual usage.


Industry representatives acknowledged the challenge of scaling context-sensitive evaluations across thousands of languages and millions of cultural settings while maintaining sustainability. The discussion concluded with commitments from various organizations to contribute to multilingual benchmarking, incident reporting tools, and infrastructure investments, establishing a foundation for collaborative AI safety work that centers Global South perspectives and needs.


Keypoints

Major Discussion Points:

Launch of the Global South Network for Trustworthy AI: The primary focus was announcing this new network bringing together research institutions from Asia, Africa, and Latin America to evaluate real-world AI impacts and build trust mechanisms localized to different linguistic, cultural, and infrastructural contexts.


Context-sensitive AI safety challenges in the Global South: Speakers emphasized that current AI safety frameworks miss critical contextual factors like local languages, cultural norms, gender dynamics, and social inequalities. Examples included AI models failing to understand local expressions (like “waters have broken” translating to “thrown away water”) and agricultural tools with male voices potentially exacerbating gender-based violence.


Structural gaps in global AI governance: Multiple speakers highlighted that Global South countries are underrepresented in global AI safety infrastructure, with only Kenya mentioned as a member of international AI safety institutes. This creates a disconnect between where AI harms occur most and where safety decisions are made.


Need for multilingual and multicultural evaluation systems: The discussion emphasized developing benchmarks beyond English-language models, creating evaluation tools that capture societal risks specific to Global South contexts, and building sustainable, community-led assessment frameworks.


Concrete initiatives and commitments: The network outlined five flagship projects including multilingual AI benchmarks, gender safety taxonomy, procurement guidelines, evaluation methodologies, and healthcare AI assessments. Industry partners like Microsoft committed to infrastructure investments and data sharing.


Overall Purpose:

The discussion aimed to formally launch the Global South Network for Trustworthy AI and establish a collaborative framework for addressing AI safety challenges specific to developing countries, while advocating for greater inclusion of Global South perspectives in international AI governance.


Overall Tone:

The tone was collaborative and urgent throughout, with speakers expressing both excitement about the network’s potential and concern about the pressing need for action. There was a consistent emphasis on moving beyond theoretical discussions to practical implementation, with industry and government representatives showing strong commitment to supporting the initiative. The conversation maintained a professional yet passionate quality, reflecting the speakers’ shared belief in the critical importance of inclusive AI safety.


Speakers

Speakers from the provided list:


Dr. Urvashi Aneja – Founder and Director of Digital Futures Lab


Mr. Abhishek Singh – Leadership role in India AI Summit and India AI mission


Ambassador Philip Thigo – Special Envoy on Technology from the Republic of Kenya


Mr. Quintin Chou-Lambert – Chief of Office and AI Lead, UN Office for Digital and Emerging Technologies


Ms. Natasha Crampton – Vice President and Chief Responsible AI Officer at Microsoft


Dr. Rachel Sibande – Senior Program Officer AI for Africa at the Gates Foundation


Ms. Chenai Chair – Director of the Masakane African Language Hub


Mr. Amir Banifatemi – Chief Responsible AI Officer at Cognizant


Dr. Balaraman Ravindran – Head Center of Responsible AI at IIT Madras


Additional speakers:


None – all speakers mentioned in the transcript were included in the provided speakers names list.


Full session report

This discussion centered on the launch of the Global South Network for Trustworthy AI at the India AI Impact Summit, bringing together government officials, industry leaders, civil society representatives, and academics to address the underrepresentation of Global South perspectives in international AI safety governance.


Network Vision and Structure

Dr. Urvashi Aneja, founder and director of Digital Futures Lab, opened by highlighting the fundamental challenge: while AI systems are being rapidly deployed across critical sectors like healthcare, education, and government in the Global South, these regions remain underrepresented in global safety infrastructure. She noted the paradox that countries with the greatest potential to leverage AI for development are precisely those most excluded from safety governance.


Dr. Aneja outlined five flagship projects for the network: multilingual AI benchmarking (with the Collective Intelligence Project and CARIA), a gender safety taxonomy project (with GXD Hub and the Global Center for AI Governance), procurement guidelines development, evaluation methodology work (with ITS Rio), and health information systems evaluation. The founding partners include Digital Futures Lab, Sirai from IIT Madras, Global Center for AI Governance, ITS Rio, and International Innovation Corps.


Government Support and Policy Alignment

Mr. Abhishek Singh from India’s AI Mission provided governmental backing, emphasizing that while consensus exists around the need for safe and trusted AI, the challenge lies in developing technical tools and benchmarks to address identified risks. He highlighted a critical gap: most AI models are assessed using predominantly English-language benchmarks, despite countries like India having 22 official languages and numerous dialects.


Singh noted the network’s alignment with the New Delhi Frontier AI commitments, which secured agreements from major AI companies to share usage data and develop multilingual performance benchmarks.


Structural Representation Challenges

Ambassador Philip Thigo from Kenya described the network as “timely but also late” due to the structural exclusion of Global South countries from safety conversations. He observed that Kenya is the only Global South member of international AI safety institutes, illustrating the representation gap. Thigo argued that “the global north of artificial intelligence is two countries and a few companies,” suggesting even traditionally developed nations face exclusion from AI governance.


He expanded the definition of AI safety beyond technical considerations to include environmental harms, biases, misinformation, and full lifecycle accountability from “minds to models,” emphasizing that governance is fundamentally about power.


Contextual Safety and Real-World Examples

Dr. Rachel Sibande from the Gates Foundation provided compelling examples of how linguistic nuance affects AI safety. She explained how the phrase “waters have broken”—a critical medical emergency—translates literally as “thrown away water” from local languages, potentially causing AI systems to miss life-threatening situations. She emphasized that language support requires understanding lived experiences and cultural contexts, not merely translation capabilities.


Sibande also discussed her work in Malawi, referencing the country’s identity as the “warm heart of Africa,” and highlighted challenges in measuring emerging harms during AI usage, such as cognitive substitution or emotional dependency.


Ms. Chenai Chair from the Masakhane African Language Hub reinforced these concerns, noting that Africa has over 2,000 documented languages, with Masakhane working on only 50. She provided an example of how agricultural tools with male voices could potentially increase gender-based violence in contexts with high gender inequality, demonstrating how design choices can have profound social consequences.


Industry Perspectives and Implementation Challenges

Ms. Natasha Crampton from Microsoft articulated the central scaling challenge: how to extend thoughtful, community-led evaluation work across thousands of languages and cultural settings while maintaining sustainability. She emphasized the need for continuous rather than one-time evaluation systems and referenced Microsoft’s commitment to the New Delhi Frontier AI agreements.


Mr. Amir Banifatemi offered a more critical assessment, arguing that safety lacks clear definition and is not integrated into companies’ financial planning or cost structures. He observed that “there is no penalty of not being safe,” highlighting the need for regulatory frameworks that make safety a financial imperative.


Coordination and Network Proliferation

Dr. Balaraman Ravindran from IIT Madras raised important questions about coordination, noting that multiple AI safety networks are launching simultaneously—including initiatives in Africa, China, and through UN processes. He called for coordination rather than competition among networks and suggested focusing on problems requiring cross-border collaboration.


Multilateral Integration

Mr. Quintin Chou-Lambert from the UN Office for Digital and Emerging Technologies discussed the evolution of international AI discussions from the concentrated focus of Bletchley Park to broader participation in subsequent summits, with over 100 countries now engaging in discussions. He referenced the UN Global Dialogue on AI Governance and the independent international scientific panel on AI as potential integration points for the network’s work.


Technical Challenges and Event Context

The event faced some technical difficulties, with portions of Dr. Aneja’s presentation becoming repetitive due to technical issues. Time constraints affected the panel discussion format, and the session concluded with mentions of a photo opportunity for participants.


Conclusion

The launch represents an important step toward including Global South perspectives in AI safety governance. The network aims to address critical gaps in current evaluation systems while navigating challenges around scaling, coordination with other initiatives, and translating evaluation work into meaningful protections for citizens across the Global South. Success will depend on the network’s ability to maintain contextual sensitivity while developing scalable methodologies and effectively integrating with broader international governance processes.


Session transcript

Dr. Urvashi Aneja

Thank you. Thank you. Thank you. Thank you. Thank you. Thank you. Thank you. Good evening, everyone. My name is Urvishya Neja. I am the founder and director of Digital Futures Lab. And I am so excited to see all of you here and to have you all here for the launch of this network. So it’s a real pleasure to welcome you to the launch of the Global South Network for Trustworthy AI here at the India AI Impact Summit. On behalf of Digital Futures Lab and our other founding partners, Sirai from IIT Madras, the Global Center for AI Governance, ITS Rio, International Innovation Corps, thank you all for being here. And we’re especially grateful to Mr. Abhishek Singh and Ambassador Philip Tigo and Mr. Quintin Chow and to all our distinguished speakers and guests who are joining us today.

Across the Global South, AI systems are being rapidly deployed in critical social sectors such as healthcare, education, judiciary, and in government. And while the opportunities are immense, in many of these contexts, many of these contexts are also marked by low institutional capacity, deep societal inequities, popularization, and populations with low levels of literacy. So while the potential is immense, the risks and harms are also immense. And so it’s particularly important that we figure out ways to make AI safe and trustworthy in these contexts to ensure not only that we protect the populations and to ensure that we don’t exasperate existing harms, but also to ensure that we build the infrastructure for safe and inclusive AI adoption.

Unfortunately, Global South organizations, Global South communities, Global South states remain underrepresented in global safety and governance infrastructures. And many countries in the Global South are actually unlikely to even have in the near term their own safety or oversight institutes. And there’s a real risk, therefore, that the concerns and priorities of these countries, of these communities remain underrepresented in the global safety infrastructure. And precisely those countries that have the most potential or the most opportunity to leverage AI. Independent civil society organizations are uniquely positioned to address this gap. Their proximity to real -world deployment contexts enables them to surface risks that are invisible to lab -based evaluations or testing. The form of grounded evidence that civil society organizations can bring can inform global safety benchmarks, standard -setting processes, and risk assessments, providing corrective signals to technical and regulatory institutions.

The Global South Network for Trustworthy AI works to advance exactly these objectives – to evaluate the real -world impact of AI systems, to build the trust and oversight mechanisms localized to different linguistic, cultural, and infrastructural contexts, and to elevate Global South perspectives in global AI governance forums. It is particularly encouraging that this initiative also aligns closely with the recently announced New Delhi Frontier AI. The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI systems, and to develop a more robust and efficient system The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI systems, and to develop a more robust and efficient system the real -world impact of AI systems, and to develop a more robust and efficient system The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI.

impact of AI. The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world systems. The Global South Network for Trustworthy AI brings together some of the leading research institutions from across the Global South. We are joined by a community of organizations from Asia, from Africa, from Latin America, and the names of which you see displayed behind you. I also want to take this opportunity to highlight some of the key activities that we’re going to be doing as part of the network. I think one of the key things that we want to do as part of the network is to really build an independent evidence base to generate community -informed analysis of the societal, ethical, and distributional risks of AI systems across diverse contexts.

We also want to do real -world deployment assessment to conduct contextual and public evaluations of models and applications across diverse social contexts. We also want to push the field of evaluations, push the science of evaluations, where we say that benchmarks are very important, but benchmarks as they stand today do not necessarily capture all the societal risks that we see in the Global South. So how do we ensure that the evaluation work that we’re doing also captures some of those harms? In some sense, what we want to do with the network is field building. We want to bring together Global South civil society organizations to pool in their collective intelligence, to pool in their capacities, and to advocate together for the representation of Global South concerns on global governance forums.

So what we are trying to do here is field building within the Global South around AI safety and around building that trust infrastructure. And eventually what we hope that all of this amounts to is collective advocacy. We see an important role that the network will play in creating a connective tissue between the global governance architecture, between the global safety infrastructure, and what’s happening on the ground. We hope the network can provide that visibility to real -world impact. to technology companies who are designing tools, who are designing safety infrastructure, as well as to governments and international organizations who are building the architecture of global AI governance. So with that, I want to thank you all. Oh, wait, I have one more thing to share with all of you.

I’m not ready to thank you yet. I also want to showcase some of the projects that we’ll be doing in the coming year. Picking up on yesterday’s commitments, one of the things that we’ll be doing is building benchmarks for multilingual AI. This is with our network partners, the Collective Intelligence Project and CARIA, and we’re really excited to start this work. We’re also going to be doing work on gender and safety. This is with our partners at GXD Hub and the Global Center for AI Governance to build a taxonomy of gender harm so that we can start building a more robust incident reporting database when it comes to gender -related harms and really advance gender safety in digital spaces.

The third piece that we’re going to be working on this year is around procurement. All of the evaluation work that we do, all the benchmarks that we build, all of that has to eventually feed into public policy. And so we hope that some of this work can support procurement. And procurement, we think, is a really important lever for countries in the global south to shape markets for responsible innovation. I think we’ve all heard a lot about the kind of third way of AI governance that India brings to the global governance landscape. And procurement can be an important lever of making that third way a reality and setting the bar for what responsible innovation looks like.

Like I mentioned earlier, we also want to push on the science of evaluation. What does good evaluation look like? What are the kind of methodologies that we need? What are the kind of methodologies that reflect the concerns and the capacities of communities in the global south? So we’re very excited to be doing this work with ITS Rio, who’s also one of the founding partners, and specifically to implement and advance this discussion on evaluations. We’ll be looking at labor market impacts in the global south. and finally we’re going to be looking at evaluations of health information systems do the existing generative AI tools and large language models that we see do they deliver for clinicians do they deliver for doctors what more can they do to support the needs of healthcare professionals in the global south.

So those are the five kind of big flagship projects that we’re going to be launching within the coming year. We’re going to be very busy as you can see we have a lot that we’re going to try and get done and we’re really excited to be on this journey with all of you and would love to engage with all of you post the launch and see how we build this civil society and research infrastructure together. So with that I am delighted to welcome our keynote speakers first and I would like to give the floor to Mr. Abhishek Singh. Sir thank you for your continued support Thank you for the network and for your leadership on the India AI Summit.

Over to you, sir.

Mr. Abhishek Singh

Thank you, Urvashi. And first and foremost, I’d like to congratulate all the team, the network which has brought this together, this Global South Network for Trustworthy AI. With a few months back when we started discussing this concept with Urvashi, with Kalika, with my team, we felt that how do we go about it? Because safe and trusted AI is something that nobody disagrees with. Everybody says that whenever AI innovation is happening, but we must ensure that we must protect ourselves, we must kind of secure ourselves from the harms that can come from misuse of AI or from the risks that frontier AI poses. So yes, we did have the Yoshua Bengio’s report, the scientific panel report, which is part of all the impact summits, the Action Summit and the Bletchley Park Summit, in which it has kind of… identifies the risks that frontier AI model poses.

But what we do believe is that just identifying the risk is not sufficient. We need to think of how do we address those risks. And for addressing those risks, you need to first have the technical tool, the capacity to identify those risks. What are the benchmarks on which you will evaluate them? Some of which Roshi identified, like how do various models perform on multilingual benchmarks? Because very often, most models are evaluated on benchmarks which are predominantly in English language. But if you look at India, a diverse country, we have 22 official languages and multiple other dialects. How do we evaluate how a model performs on various domains in prompts given in those languages? We don’t have specific linguistic benchmarks.

The same applies to many countries of global south. So it felt that while limited expertise exists in some institutions where research is going on, like Serai is one of them, where Professor Balram Ravindran is leading it. There are many labs, of course. whether it’s Microsoft Research or whether other labs wherein such work is going on. The AI Security Institute in UK is doing some work in this direction. The OECD has been doing some work. But how do we ensure that we enable the access to such resources, such tools, such studies to the larger global majority? So with that, this whole concept of creating a global south network for trustworthy AI came in. And then we immediately had these conversations with all the key stakeholders, partners.

We got a lot of support from almost all stakeholders. And along with that, the conversation for the New Delhi Frontier AI commitments was also going on, which Kalika from my team was leading it. And luckily, we were able to announce it in which all models committed to those two commitments about sharing usage data as also multilingual performance benchmarks. So that was a huge achievement. And I feel that the launch of this. Global South Network for trustworthy AI is a further step in that direction. How do we enable compliance to those commitments? How do we ensure that how this data will be shared? How do we create tools for evaluating models in various languages? How do we build up capacity in all countries of the global south?

How do we share resources? How do we share knowledge across? So this is just the beginning and I feel that we support from all industry organizations, the frontier AI labs, the research organizations, governments across the world. This can really, really grow into a resource that can be a global utility. So I compliment all that team which is involved in doing that. The launch of the network is the first step. But how do we action it out? How do we make it functional? How do we ensure that we get necessary support from all stakeholders? Very often whenever we talk about trusted AI, whenever we talk about safe AI, some people think that we are trying to stifle innovation.

The objective is not that. We always say that while the primary objective is to ensure diffusion of AI, primary objective is to ensure that more and more users benefit from the usage of AI. But at the same time, we need to do that in a responsible manner. We need to do it in a safe manner. We do need to do it in a trustworthy manner to limit the harm that can be caused. So this Global South Network for Trustworthy AI which is being launched will work in that direction. It will be an institution that will support not only India but the entire Global South. And I am sure with just the presence of all the speakers who are present in this session, the strong commitment that all industry and all countries and all multilateral organizations are showing to this initiative, I am sure this will get further strengthened in the days to come.

There is a lot of work that Urvashi and team are taking. They are taking on their own. But we will be there to provide all necessary support for India AI mission and we will work towards ensuring that you get the same level of support from every. participating country which is here. So thank you once again and congratulations for this launch and look forward to working towards the objectives in the near future. Thank you.

Dr. Urvashi Aneja

MS. Thank you, sir, for your remarks and most importantly for your support. I think it means a lot to us to be working so closely with the India AI mission and we’re really excited to be able to deliver on this promise. It’s now my honor to invite Ambassador Philip Tigo, the Special Envoy on Technology from the Republic of Kenya, to share his reflections.

Ambassador Philip Thigo

AMBASSADOR PHILIP TIGO, ASSOCIATE OF TECHNOLOGY, KENYA, Thank you so much for this opportunity to share my reflections. And I noticed that this is really a women -led network, so again, congratulations, Ovashi and Rachel, for putting this together. I think before we celebrate the launch of the network, I think we must acknowledge that we are working with the right people and we are working with the right people. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here.

And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think because you must acknowledge the structural problem around the safety conversations and the infrastructure that has been cutting safety in the last three years. I think the global south has always been excluded from this conversation. I say this from a position of strength because Kenya is the only, Kenya I think, we’re the only member of the international network of AI safety institutes.

And so there’s a challenge there. And so I think that model that is not inclusive to a global majority, that in most cases bears the brunt and the impacts of AI, is not acceptable. And so this network, in my sense, is timely but also late. And so there’s almost an urgency that we need to work very closely in how we scale up what this network does. The second part, of course, is, as I mentioned, that a lot of the global majority countries that are there are not. They are the ones that not just bear the brunt of the models, but bear the adverse societal harms of the models. Kenya is one of the countries that uses one of the models.

and from the use cases we see that they use it for the wrong reasons. Emotional support or companionship, it’s not necessarily for anything meaningful or productivity. And so as the world advances, it therefore behoves us that we work with these frontier model companies to ensure that their models are safe beyond secure, but also are more trustworthy. The second part, of course, is that part of model evaluations assumes access. We now know that a lot of my colleagues who are doing model evaluations are doing it from an external point of view. So we need to be very clear that global majority countries, and by this when I say global majority countries, we also have a new global south in AI, because it’s just not the global majority.

We know in the global north of artificial intelligence is two countries and a few companies. So we must, beyond this, extend to also include other colleagues, whether it’s from Europe, Western Europe, or Latin America. Safety must also go beyond technology. towards socio -technical issues. We look at AI in the countries of Kenya from minds to models and so safety must also include environmental harms, biases, misinformation, disinformation but also harms to water, environment and so we need full lifecycle accountability. It’s good to evaluate the models but also it’s good to evaluate the footprints of the model quickly. There are four structural gaps that we see and this is why I love this network and the network I think one is yes you want global majority folks to evaluate the models but we have great teaming capacity gaps so I hope that this network will look at this.

Secondly I think is also issues of access to compute. We can’t have global majority researchers trying to evaluate models without necessarily having access to compute to do that. Third part of course has been mentioned by I think his left issues around linguistic and cultural mismatch so we need to do that the other part of course is benchmarking. as governance power. Also, benchmarks are not neutral. Sometimes I think I like to be honest because that’s what evaluation needs to do. And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means. Governance is about power. And we must deconcentrate that power even if it’s unintentional.

Finally, I think for me, evaluation is also about agency. And we must have a question of agency, a notion of agency around these models, but also including sovereign capability. As we know, a lot of your countries are trying to build sovereign models, but also sovereign capabilities across the track. What should this network deliver, in my view? And I’ll humbly make these quick suggestions. One, I think, yes, good to have the network, but can we have regional nodes for this? So that, because Africa… I speak for Africa, Africa is another country, it’s 54 countries, expanded to have nodes. Secondly, include multilingual benchmark data sets. Could be an interesting annual red teaming exercise. Could be potentially, why not publish a Global South AI Safety Report with an expansive definition of what safety is.

And I would be remiss if I don’t say how do we fit this into the multilateral process. We already have a global UN scientific panel on AI, and there’s a global dialogue on AI governance. I’m one of the champions for this, so hopefully we will get this in there. Finally, let’s close the accountability loop. How do all this ultimately matter for citizens? We can evaluate all we want, but if they don’t translate

Dr. Urvashi Aneja

Thank you, Ambassador, for highlighting the urgency of this work and also reframing the safety conversation for the Global South. And just to say we are planning to have regional hubs, and we do. And I think the point about how we engage with the multilateral system is very important, and we will have the Indian AC as part of our steering committee, and we hope we can work with the government of Kenya as well. And, of course, we have Professor Ravindran, who is part of the scientific council, so we will be relying on him as well. But thank you. Thank you for your remarks. And with that, I’d like to call our final keynote speaker for the day, who represents the UN Office for Digital and Emerging Technologies.

I’m pleased to invite Mr. Quenchen Chow Lambert, the Chief of Office and AI Lead, to deliver the next keynote. Thank you for your keynote address.

Mr. Quintin Chou-Lambert:

There is less, perhaps, infrastructure or energy connection to go around. So the concept of AI safety becomes less of a, or it kind of edges into this more contextual field, and that’s where this kind of low perspectives, field -tested examples can be very helpful to surface, which we’re missing. And I’d say the idea of AI standards as technical standards don’t solve that issue because a one -size -fits -all standard will not be contextually sensitive. So moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small -language models to context turns the issue of AI safety into a more fuzzy kind of discussion and one which really needs empirical evidence.

And I think the trends in the institutional discussions from Bletchley Park to Sears, Seoul, where there were also around 30 countries signing the declaration, to Paris, where you had 60 -plus, and now here. over 100 countries engaging. We now have the United Nations Global Dialogue on AI Governance, which will include a whole 193 member states informed by analysis from an independent international scientific panel on AI, which will look at the risks and also opportunities and impacts of AI. And so as the conversation in these summit settings and in the international level has widened and to include more countries and more people and covered more of humanity, the focus has, through the open source developments, been allowed to become much more focused of encompassing other perspectives.

And that’s why, to close and to echo Ambassador Thieger, these kinds of networks play a crucial role in connecting and bringing examples of the challenges that we face. Thank you very much. cases of threats from various sources to local people into discussions so that international discussions do not ignore or omit or discount the perspectives of the vast majority of people on the planet. Thank you.

Dr. Urvashi Aneja

Thank you, Mr. Chow, for those remarks. I’d now like to call our panelists onto the stage. Ms. Natasha Crampton, Vice President and Chief Responsible AI Officer at Microsoft. Dr. Rachel Tabande, Senior Program Officer AI for Africa at the Gates Foundation. Before you sit, we’re going to take one quick picture. Ms. Chennai Chair. I don’t see you. Oh, there you are. Yes, okay. Director of the Masakane African Language Hub. Mr. Amin Banefatami, Chief Responsible AI Officer. I’m cognizant. And last but certainly not least, Dr. Balaram Ravindran, Head Center of Responsible AI at IIT Madras. Yes, and can we get the keynote speakers as well? Thank you. As with all good things in life, we’re short on time.

But so let’s get started. Rachel, I’m going to start with you. Thank you. where according to you what according to you or where according to you do you feel like we still lack clarity on how safe and reliable AI systems are when they’re deployed in real world context in the global south

Dr. Rachel Sibande

thank you so couple of things maybe two three things number one is we need to redefine what is safe and what is harmful in as far as AI models or applications are concerned according to the social cultural context that they are deployed in and that means that having models or applications that are great at understanding the data or the patterns to generate content is not enough if they do not understand the social norms the gender dynamics the religious beliefs the political sensitivities or indeed even the humor the slang or the tone particularly now that voice is being used in the media a key channel for delivery of AI. So we need to redefine safety and harm in the context in which AI models are deployed.

So I think we’re missing that, but hopefully we get there. I think the second piece is around language. It’s not enough for a large language model to have strong translation capabilities. Language in itself is not just about vocabulary. It’s also about the lived meaning, the lived experiences. I come from a beautiful country called Malawi. It’s also called the warm heart of Africa. Now, if you’re deploying a model for pregnant mothers to access advisory messaging there, if the mother says their waters have broken, which clinically is a critical incident that should warrant that mother to be referred to a health facility, but if you translate that from the local language to English, which is where most of these large language models and applications have been benchmarked on, that will literally mean I have thrown away water.

So if the model is not trained to understand that context, then you will miss that flag. And then finally, I wanted to say that we also need to understand the harms that emerge as people use the AI models. Currently, I think much of the benchmarking is done on the content and predefined metrics. So final example, personally, I use my AI companion as my therapist. So it’s the one persona that knows a lot about my personality from all spheres, as a mother, as a career person, my finances, all of that. But at what point can we then be able to track whether I’m substituting my cognizance and cognitive capabilities with that AI model or application, or that I’m becoming overly emotionally dependent?

So I think there are those three areas that we’re missing, and hopefully we can get better at it. Thank you.

Dr. Urvashi Aneja

Thank you, Rachel, and thank you also for those powerful examples, because I think we’ve been saying some of this at almost a theoretical level, but I think those examples really bring home the gaps in terms of where the current safety conversation is. Chennai, from a civil society perspective, what do you feel companies or developers often miss about the safety implications of deploying AI systems in the global south?

Ms. Chenai Chair

That’s one thing they miss, the user experience. So on a more serious note, thank you, Vashi. So this is great to actually piggyback from what you said, and I was like, are we reading the same notes? So I think what really is missed when people are deploying some of these solutions is around the context in which they’re deploying the tool. And this is particularly looking at an example where on the African continent, there is high levels of gender inequality, a very youthful population with young people often unemployed, and also older people forgotten in actually the development of technologies. So I don’t know who we’re developing for, but sometimes we actually don’t consider that diversity and the inequalities that exist.

So you can find that sometimes when these tools are deployed, they actually further exacerbate a situation of inequality. And I’ll give you one example where perhaps an agricultural tool that has a voice system on it to provide farmers or women information on what to plant may actually have a male -sounding voice. And if in that context there’s high issues of gender -based violence or lack of trust, and the community members were not consulted in the design process, what it actually leads to is just exacerbating an already existing situation. And that is an example. That actually did happen when people were deploying Internet solutions for a community. Then secondly, also thinking about who gets left behind in deploying these solutions.

This is where language, as Rachel was mentioning, comes in. So on the African continent, we have over 2 ,000 languages that have been documented. Masakhane is only working on 50 of those African languages to build up quality data sets. So what you then find is when people are deploying technologies, even if they deploy them in something like Kiswahili, which now has a large number of data sets, people just don’t speak Kiswahili across East Africa. And particularly in Kenya, if you go to Nairobi, the Kiswahili spoken in Nairobi will be Shang. Then you go to, it’s not even Kiswahili, as I’m being corrected. And then if you go to the coast in Mombasa, it will be completely different.

So we have to actually take into account the context and nuance of what is being deployed. And then lastly, the way in which the sector, the technology is actually used, if deployment doesn’t take into account. the whole ecosystem of the end user, it can actually result in misuse. And I want to specifically say that there’s two forms of misuse here. There could be people who unintentionally actually carry out a problematic, harmful act online based on how they’re interacting with the technology. And we know that content, particularly if it’s in their own language, and we know that content moderation for the global majority is not sufficient. Or people are underpaid, as we’ve seen the cases that were coming out about content moderators in Kenya.

Then there’s actually intentional misuse. Now, this is where we find gender disinformation, the use of deepfakes to discredit people, particularly around election period. And now with increased open AI that people can actually just type something and get something back, we are seeing that high level of deployment without thinking about what is the after -end impact. To close it off, because I’m talking about AI as if it’s coming later. A10. when they were deployed. It was great, I can track my missing bag on a flight. They have now been put in women’s bags or children’s bags by people who they do not know and they track them. That’s already an act of surveillance that was, if people had been consulted, it might have been mitigated against.

Yes, I do want to know where my bag is, but I don’t want to be tracked unknowingly.

Dr. Urvashi Aneja

Thanks, Chennai, for that and also for pointing out, bringing the gender dimension on the table and highlighting the issues around what seems like useful technology, how quickly it can become surveillance technology. I’d like to now bring the industry perspective into this conversation. So, Natasha, maybe I can start with you. As you scale systems globally, what are some of the hardest constraints that you as a company face in ensuring context -sensitive safety?

Ms. Natasha Crampton

Well, thanks for that question, Arati, and congratulations to everyone on the establishment of the network. I think it’s a really important step forward. So when I think about Microsoft, I think about sort of Microsoft’s scale, and our mission is really to try and empower every person in every organization in the world to achieve more. And so one of the challenges I think that we face with scaling up our efforts here is how do we take the very deep, careful, thoughtful, community -led evaluation work that animated a project like Samishka, which the CAIA organization, as well as the Collective Intelligence Project and Microsoft Research worked on together, which really developed very context -aware evaluations that were appropriate for the use case.

And how do we take that thoughtful work and really scale it up? Because really we want to do that type of work for thousands of languages and probably millions of different cultural settings. And so I think we really need to think about this system of how we are going to build multilingual and multicultural evaluations that we can really run broadly. I think sometimes we think evaluations, we don’t sort of understand how sustainably they need to be run. As in you can’t just do it once before you release a product. You need to run the evaluations on an ongoing basis to understand how there might have been shifts. And so I really think for us we need to think about this system.

How are we going to build a sustainable, grounded, community -led system of scalable evaluation?

Dr. Urvashi Aneja

Thanks, Natasha. And I hope in some sense also the network can actually play at least part of that function in building that kind of coherence to the space of evaluation and helping us at least build a shared vocabulary and a shared set of methodologies together as organizations. Amir, what do you think needs to change, whether it’s internally within companies or externally in terms of the ecosystem that we’re operating in, to make such grounded evaluations, the kind that Natasha was talking about, become the standard practice for industry? Should they be the standard practice? And if so, how? How do we get there?

Mr. Amir Banifatemi

Thank you for that question. And first, congratulations. I’m happy to be also part of this network and support it. I think Natasha mentioned part of the foundational questions. And I think from a, I’m putting my hat off, cognizant chief responsible, we work with a lot of companies and governments into deploying. new scenarios. We call it systems or applications or anything else. The concept of safety, I was mentioned, is diffused. It’s not very clear what we’d call safety. So evaluating the underlying element that needs to be changed or to be addressed is not obvious. When we talk about models, models are not just one thing that you deploy. It goes into an application, there’s a system, infrastructure, there’s network access, there’s API connected data access.

All of them are contextually different. That was mentioned before. And then the problem, one of the problems is that, you didn’t ask me about the problem, but there’s a problem issue is that there’s a lack of imagination. People that are building systems have no awareness about the context in which those situations occur and how they occur and what’s the causes and what’s the likelihood of solution to happen. So absent of that, all this context which language is part of it, culture is part of it, is not captured. So without that, there is very little capability. to address that from a regulation or incentive perspective. Safety, on the other side, is not costed into financial systems and so forth.

There is no penalty of not being safe. So as long as there is no constraint to put safety as a cost structure, which strong mandate, companies will not pay attention or enough attention. So if it’s not part of the financial planning and the processes and so forth, it won’t happen. So there is a disconnect between what we do as enterprises to make sure that systems and platforms are properly built and deployed. There is a disconnect between the system in which they are deployed. At the same time, there is a talent inclusion that is missing. So the inclusion part is that all the talent that is building into those safety conversations are not the talent that are exposed to those issues.

So that absent voice is also a piece that needs to be addressed, not just from a skilling perspective, but also from an integration perspective. And finally, the infrastructure part. The infrastructure is not just systems and models and data, it’s also the tooling and the evaluation. And it was mentioned that evaluation has to be done differently, but if you don’t know what harm or safety means, evaluation’s gotta be different. There is probably an opportunity here to come up with a series of evaluation tools that are not only built for model design, but also built for system deployment. And if we go from pilot to scaling, what issues occur and what examples are happening and what incidents are deployed, and incident reporting is a huge opportunity here because it will capture, nested in reporting, some of the hidden element of the control issues, data access, regulation, absence, or anything else.

Finally, there is a latency issue in global north, and you mentioned probably correctly that there’s a lot of latency issues. There are only a handful of countries in the global north that probably the new slot is much bigger. there are institutional framework, you have basically the rule of law, you have civil society which is very active, you have legal framework that basically creates an accelerated feedback loop into all this incident safety in most of the global south countries these mechanisms don’t exist which delays the feedback loop and basically compounds the possible harm and everything else so there is probably an opportunity to figure out how we can accelerate the learning capabilities and the skills at which we capture knowledge and data to be tied with tools that probably need to be implemented and deployed either on an open source matter or a free access matter and build it with a contextual environment, the talent pool to make it together so the ownership of the global south, all these pieces are important so the network can actually incentivize those different pieces that could complete each other to really play a role into the global south understanding better where safety issues are, where harm can happen and what corrections can be made in the rhythm that needs to happen because rhythms are not exportable and what we do from one country to another is not.

And finally the network could probably help bring it together.

Dr. Urvashi Aneja

Thank you for laying that out and also just pointing out how all the kind of pieces link to each other and we can’t just kind of go at it at one level alone and to the importance of capacity across all those. Professor Ravindran, AI deployment is accelerating in the global south, in India, in many other countries as well. But at the same time we don’t see or so far we haven’t seen as much investment in the kind of safety and safety infrastructure. Would you agree? You’re actually asking an academic about investment? Sure, of course there’s not enough money. Why not and how do we change it?

Dr. Balaraman Ravindran

so I’m going to answer a different question sure perfect like a true academic I’m sorry I’ll connect it back to what you asked so there are a whole lot of initiatives that are getting announced at the summit and also things that I kind of discovered while having various conversations that there are multiple networks that are getting launched there are already in operation there is a network in Africa looking at capacity building there is a network in China apparently which none of us seem to have heard about that’s being launched on AI safety and capacity building and that is our network that’s getting launched and that is the UN initiative on building this network of capacity building institutes for the global south which we had a meeting in the morning as well about that so there is just too many of these initiatives that are getting launched.

And we have to figure out a way how we would coordinate operations among these initiatives as well. So I think that would be a great multiplier instead of everybody going out and saying, okay, let me see what small piece of the pie that I can get so that I can do these activities. And after that, if there is a lot more coordination. And if you remember our initial conversations about when we wanted to start this thing was about this would be like this one node in the global AC. I can’t even say global network of safety institutes anymore, can I? So they’re not even safety institutes. So AC institutes, whatever ACs, this should be like one node in the AC network which kind of represents unheard voices there because almost except for, as the ambassador was pointing out, except for Kenya.

So we really have, and of course India, I presume. We don’t have safety institutes in the global south, right, that can participate in the dialogue. So I mean, that kind of larger collaboration framework is something that we should enable so that, I mean, even if you say, we go to Gates, and then how many different people, how many different networks would Gates want to spend their money to? If that is one way we can say that there’s this whole operation that’s happening, then that would be a great way of harmonizing our efforts. I can turn it back to the question. Thank you.

Dr. Urvashi Aneja

No, I mean, I think you raised a really important issue of kind of harmonizing these efforts, and also that how this network can play a really important role in the larger kind of AC network. Luckily, the S remains the same, so we can still go with the acronym, I guess, on the safety network. We’re almost at time, so let’s just do one kind of quick rapid -fire round with all the panelists, and maybe Natasha, I can start with you. What is the one concrete step your institution, Microsoft, could take in the next year to strengthen AI safety in the global south?

Ms. Natasha Crampton

Well I’m looking forward to making good on the New Delhi Frontier AI commitments that Microsoft made which is going to help advance multilingual and multicultural evaluation work as well as share data that will help policy makers make or understand AI adoption within their countries and make the sort of choices and policy interventions that help bring that broader access so if I can be sneaky and kind of come as one thing. The second thing I’m really excited about is we’re making large infrastructure investments across the global south to the tune of 50 billion dollars by the end of this decade now that infrastructure as Amir and others on the panel have mentioned is essential to being able to building up this scaled system of sustainable evaluation so I’m looking forward to those investments too.

Dr. Urvashi Aneja

Thank you.

Dr. Balaraman Ravindran

is that a fire alarm or something?

Dr. Urvashi Aneja

No, no, no, they’re telling us that we have to wrap up I think.

Dr. Balaraman Ravindran

Okay, great, so wrapping up, so we have to get the work going, rolling, right, so talking about it is one thing, but actually starting to do this collaboration and getting this research efforts going, we’d love to reach out to partners across the globe, in fact, I’m part of the other UN network as well and we have been talking about looking at problems that would necessarily require cross -border collaboration, right, as supposed to, you know, problems that we would anyway solve in our geography, then just working with somebody else to solve it in two geographies, okay but if you can pick problems that will necessarily require people across borders to collaborate, I think that will certainly drive this and also will, you know, kind of put forth the importance of having the network itself, not just information sharing, but actually problem solving that can be done only across the network.

Dr. Urvashi Aneja

Thank you. Rachel, 30 seconds.

Dr. Rachel Sibande

30 seconds I think from the foundation side is to really institutionalize the evaluation of safety of AI solutions right at deployment because what we see now is that safety issues almost emerge post deployment thank you

Ms. Chenai Chair

so from the hub side we actually do have a benchmarking initiative that’s going on this year so this will be one contributing to the African benchmarking work and so that will be our output in contribution

Dr. Urvashi Aneja

amazing looking forward to that thank you Chennai and Amir last but not least

Mr. Amir Banifatemi

we’re working already on with our two labs one in Bangalore actually and one in San Francisco on safety evaluations mostly on incident reporting and we already made it culturally contextual so I hope that we are helpful to basically provide open source tools for evaluation to disseminate them and work with that work to basically make them accessible to the public available to all partners.

Dr. Urvashi Aneja

Thank you.

D

Dr. Urvashi Aneja

Speech speed

129 words per minute

Speech length

2383 words

Speech time

1102 seconds

Underrepresentation and low institutional capacity create heightened risks

Explanation

Aneja highlights that many Global South contexts suffer from weak institutions and deep inequities, which raises the risk that their safety concerns are overlooked in global AI safety frameworks.


Evidence

“And while the opportunities are immense, in many of these contexts, many of these contexts are also marked by low institutional capacity, deep societal inequities, popularization, and populations with low levels of literacy.” [1]. “And there’s a real risk, therefore, that the concerns and priorities of these countries, of these communities remain underrepresented in the global safety infrastructure.” [2]. “Unfortunately, Global South organizations, Global South communities, Global South states remain underrepresented in global safety and governance infrastructures.” [8].


Major discussion point

Challenges and Gaps in AI Safety for the Global South


Topics

Capacity development | Human rights and the ethical dimensions of the information society | Artificial intelligence | Closing all digital divides


Build an independent evidence base, field‑building, and collective advocacy

Explanation

Aneja proposes that the Global South Network should generate community‑informed evidence on AI risks, engage in field‑building activities, and use the findings for collective advocacy at global forums.


Evidence

“I think one of the key things that we want to do as part of the network is to really build an independent evidence base to generate community -informed analysis of the societal, ethical, and distributional risks of AI systems across diverse contexts.” [58]. “In some sense, what we want to do with the network is field building.” [59]. “And eventually what we hope that all of this amounts to is collective advocacy.” [62].


Major discussion point

Role and Design of the Global South Network for Trustworthy AI


Topics

Monitoring and measurement | Capacity development | Artificial intelligence


Embed safety considerations into public procurement

Explanation

Aneja argues that procurement policies can be leveraged to set safety standards for AI, ensuring responsible innovation and market shaping in the Global South.


Evidence

“And procurement can be an important lever of making that third way a reality and setting the bar for what responsible innovation looks like.” [159]. “And so we hope that some of this work can support procurement.” [160]. “And procurement, we think, is a really important lever for countries in the global south to shape markets for responsible innovation.” [161].


Major discussion point

Governance, Representation, and Power Dynamics


Topics

Governance | Financial mechanisms | Human rights and the ethical dimensions of the information society


Launch gender‑harm taxonomy and incident reporting database

Explanation

Aneja notes that the network will develop a taxonomy of gender‑related harms and a corresponding incident‑reporting database to improve gender safety in digital spaces.


Evidence

“This is with our partners at GXD Hub and the Global Center for AI Governance to build a taxonomy of gender harm so that we can start building a more robust incident reporting database when it comes to gender -related harms and really advance gender safety in digital spaces.” [110].


Major discussion point

Concrete Actions and Commitments


Topics

Human rights and the ethical dimensions of the information society | Artificial intelligence | Data governance


M

Mr. Abhishek Singh

Speech speed

181 words per minute

Speech length

892 words

Speech time

294 seconds

Existing benchmarks are English‑centric, overlooking multilingual realities

Explanation

Singh points out that most AI model evaluations rely on English‑language benchmarks, leaving a gap for languages spoken in the Global South.


Evidence

“Because very often, most models are evaluated on benchmarks which are predominantly in English language.” [16]. “We don’t have specific linguistic benchmarks.” [17].


Major discussion point

Challenges and Gaps in AI Safety for the Global South


Topics

Artificial intelligence | Closing all digital divides


Enable compliance with New Delhi commitments and share evaluation tools

Explanation

Singh stresses the need to help stakeholders meet the New Delhi Frontier AI commitments by providing tools and data for multilingual performance and usage‑data sharing.


Evidence

“How do we enable compliance to those commitments?” [69]. “And along with that, the conversation for the New Delhi Frontier AI commitments was also going on, which Kalika from my team was leading it.” [71]. “And luckily, we were able to announce it in which all models committed to those two commitments about sharing usage data as also multilingual performance benchmarks.” [113].


Major discussion point

Role and Design of the Global South Network for Trustworthy AI


Topics

Artificial intelligence | Data governance


Create multilingual benchmarks, red‑team exercises, and gender‑harm taxonomies

Explanation

Singh advocates for developing multilingual evaluation suites, conducting red‑team testing, and building taxonomies that capture gender‑related harms.


Evidence

“Some of which Roshi identified, like how do various models perform on multilingual benchmarks?” [22]. “How do we evaluate how a model performs on various domains in prompts given in those languages?” [112].


Major discussion point

Multilingual and Culturally Contextual Evaluation


Topics

Artificial intelligence | Closing all digital divides | Human rights and the ethical dimensions of the information society


New Delhi Frontier AI commitments on multilingual performance and usage‑data sharing

Explanation

Singh reiterates that the commitments require models to disclose multilingual performance metrics and share usage data to foster transparency.


Evidence

“And luckily, we were able to announce it in which all models committed to those two commitments about sharing usage data as also multilingual performance benchmarks.” [113].


Major discussion point

Concrete Actions and Commitments


Topics

Artificial intelligence | Data governance


A

Ambassador Philip Thigo

Speech speed

196 words per minute

Speech length

1016 words

Speech time

309 seconds

Global South voices are excluded from safety governance; power must be de‑concentrated

Explanation

Thigo emphasizes that structural problems have left the Global South out of safety conversations and calls for de‑concentrating power away from a few institutions.


Evidence

“And I think because you must acknowledge the structural problem around the safety conversations and the infrastructure that has been cutting safety in the last three years.” [12]. “I think the global south has always been excluded from this conversation.” [141]. “And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means.” [42]. “And we must deconcentrate that power even if it’s unintentional.” [134].


Major discussion point

Governance, Representation, and Power Dynamics


Topics

Artificial intelligence | Governance | Human rights and the ethical dimensions of the information society


Benchmark standards should not be defined by a handful of institutions

Explanation

Thigo warns against a concentration of benchmark authority and urges broader participation in defining safety metrics.


Evidence

“And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means.” [42].


Major discussion point

Governance, Representation, and Power Dynamics


Topics

Artificial intelligence | Governance


Establish regional nodes and integrate with multilateral AI governance processes

Explanation

Thigo proposes creating regional hubs for the network and linking its work to existing UN AI panels and global governance dialogues.


Evidence

“One, I think, yes, good to have the network, but can we have regional nodes for this?” [82]. “We already have a global UN scientific panel on AI, and there’s a global dialogue on AI governance.” [83]. “And I would be remiss if I don’t say how do we fit this into the multilateral process.” [91].


Major discussion point

Role and Design of the Global South Network for Trustworthy AI


Topics

Artificial intelligence | Internet governance | Governance


Global South researchers lack access to compute resources for model evaluation

Explanation

Thigo notes that without sufficient compute, researchers in the Global South cannot effectively evaluate AI models, limiting their participation.


Evidence

“We can’t have global majority researchers trying to evaluate models without necessarily having access to compute to do that.” [135].


Major discussion point

Infrastructure, Capacity, and Resource Constraints


Topics

Artificial intelligence | Capacity development


M

Ms. Chenai Chair

Speech speed

167 words per minute

Speech length

683 words

Speech time

244 seconds

Gender bias and language diversity are ignored, leading to misuse and surveillance

Explanation

Chair stresses that overlooking gender‑based violence and linguistic diversity can exacerbate harms and turn useful technology into surveillance tools.


Evidence

“And if in that context there’s high issues of gender -based violence or lack of trust, and the community members were not consulted in the design process, what it actually leads to is just exacerbating an already existing situation.” [13]. “And this is particularly looking at an example where on the African continent, there is high levels of gender inequality, a very youthful population with young people often unemployed, and also older people forgotten in actually the development of technologies.” [15]. “Thanks, Chennai, for that and also for pointing out, bringing the gender dimension on the table and highlighting the issues around what seems like useful technology, how quickly it can become surveillance technology.” [44]. “Now, this is where we find gender disinformation, the use of deepfakes to discredit people, particularly around election period.” [51].


Major discussion point

Challenges and Gaps in AI Safety for the Global South


Topics

Human rights and the ethical dimensions of the information society | Closing all digital divides | Building confidence and security in the use of ICTs


Africa’s 2,000+ languages require localized datasets beyond dominant lingua‑franca

Explanation

Chair highlights the massive linguistic diversity in Africa and the need for localized data sets, noting current efforts cover only a fraction of languages.


Evidence

“So on the African continent, we have over 2 ,000 languages that have been documented.” [123]. “Masakhane is only working on 50 of those African languages to build up quality data sets.” [124]. “So what you then find is when people are deploying technologies, even if they deploy them in something like Kiswahili, which now has a large number of data sets, people just don’t speak Kiswahili across East Africa.” [126].


Major discussion point

Multilingual and Culturally Contextual Evaluation


Topics

Closing all digital divides | Artificial intelligence


African benchmarking initiative to produce localized evaluation metrics

Explanation

Chair mentions an ongoing African benchmarking project that will generate metrics tailored to local languages and contexts.


Evidence

“so from the hub side we actually do have a benchmarking initiative that’s going on this year so this will be one contributing to the African benchmarking work and so that will be our output in contribution” [177].


Major discussion point

Concrete Actions and Commitments


Topics

Artificial intelligence | Closing all digital divides


M

Ms. Natasha Crampton

Speech speed

136 words per minute

Speech length

404 words

Speech time

177 seconds

Sustainable, ongoing evaluation infrastructure is needed, not one‑off tests

Explanation

Crampton argues that evaluations must be continuous and supported by large‑scale infrastructure investments to track model behavior over time.


Evidence

“As in you can’t just do it once before you release a product.” [106]. “You need to run the evaluations on an ongoing basis to understand how there might have been shifts.” [145]. “The second thing I’m really excited about is we’re making large infrastructure investments across the global south to the tune of 50 billion dollars by the end of this decade now that infrastructure as Amir and others on the panel have mentioned is essential to being able to building up this scaled system of sustainable evaluation so I’m looking forward to those investments too.” [146].


Major discussion point

Infrastructure, Capacity, and Resource Constraints


Topics

Monitoring and measurement | Capacity development | Financial mechanisms


Scale community‑led multilingual evaluation sustainably across thousands of languages

Explanation

Crampton emphasizes the ambition to create community‑driven evaluation pipelines that can handle thousands of languages and cultural settings.


Evidence

“Because really we want to do that type of work for thousands of languages and probably millions of different cultural settings.” [120]. “I think sometimes we think evaluations, we don’t sort of understand how sustainably they need to be run.” [131]. “I think sometimes we think evaluations, we don’t sort of understand how sustainably they need to be run.” [132].


Major discussion point

Multilingual and Culturally Contextual Evaluation


Topics

Closing all digital divides | Capacity development | Artificial intelligence


Microsoft’s $50 billion infrastructure investment aims to support scalable evaluation

Explanation

Crampton notes Microsoft’s commitment of $50 billion to build infrastructure that will enable large‑scale, sustainable AI evaluation in the Global South.


Evidence

“The second thing I’m really excited about is we’re making large infrastructure investments across the global south to the tune of 50 billion dollars by the end of this decade now that infrastructure as Amir and others on the panel have mentioned is essential to being able to building up this scaled system of sustainable evaluation so I’m looking forward to those investments too.” [146].


Major discussion point

Concrete Actions and Commitments


Topics

Financial mechanisms | Enabling environment for digital development | Artificial intelligence


M

Mr. Amir Banifatemi

Speech speed

169 words per minute

Speech length

844 words

Speech time

299 seconds

Lack of a clear safety definition and financial penalties hampers responsible deployment

Explanation

Banifatemi points out that safety is a diffused concept, not costed into financial planning, and there are no penalties for unsafe AI, reducing incentives for responsible behavior.


Evidence

“The concept of safety, I was mentioned, is diffused.” [9]. “So as long as there is no constraint to put safety as a cost structure, which strong mandate, companies will not pay attention or enough attention.” [55]. “Safety, on the other side, is not costed into financial systems and so forth.” [56]. “There is no penalty of not being safe.” [150].


Major discussion point

Challenges and Gaps in AI Safety for the Global South


Topics

Artificial intelligence | Financial mechanisms | Building confidence and security in the use of ICTs


Open‑source, culturally contextual evaluation tools and incident reporting pipelines

Explanation

Banifatemi describes ongoing work in Bangalore and San Francisco to produce culturally contextual safety evaluations and to release open‑source tools for broader public use.


Evidence

“we’re working already on with our two labs one in Bangalore actually and one in San Francisco on safety evaluations mostly on incident reporting and we already made it culturally contextual so I hope that we are helpful to basically provide open source tools for evaluation to disseminate them and work with that work to basically make them accessible to the public available to all partners.” [72].


Major discussion point

Concrete Actions and Commitments


Topics

Artificial intelligence | Data governance | Open source


Develop incident‑reporting mechanisms to create feedback loops for regulation

Explanation

Banifatemi stresses that without a clear safety definition, evaluation must change, and incident reporting can capture hidden harms to inform regulation.


Evidence

“And it was mentioned that evaluation has to be done differently, but if you don’t know what harm or safety means, evaluation’s gotta be different.” [40]. “And if we go from pilot to scaling, what issues occur and what examples are happening and what incidents are deployed, and incident reporting is a huge opportunity here because it will capture, nested in reporting, some of the hidden element of the control issues, data access, regulation, absence, or anything else.” [163].


Major discussion point

Governance, Representation, and Power Dynamics


Topics

Monitoring and measurement | Building confidence and security in the use of ICTs | Artificial intelligence


D

Dr. Rachel Sibande

Speech speed

144 words per minute

Speech length

454 words

Speech time

188 seconds

Safety definitions must reflect local cultural, gender, and religious norms

Explanation

Sibande argues that AI safety must be redefined to incorporate societal norms, gender dynamics, religious beliefs, and other contextual factors.


Evidence

“thank you so couple of things maybe two three things number one is we need to redefine what is safe and what is harmful in as far as AI models or applications are concerned according to the social cultural context that they are deployed in and that means that having models or applications that are great at understanding the data or the patterns to generate content is not enough if they do not understand the social norms the gender dynamics the religious beliefs the political sensitivities or indeed even the humor the slang or the tone particularly now that voice is being used in the media a key channel for delivery of AI.” [32]. “So we need to redefine safety and harm in the context in which AI models are deployed.” [34].


Major discussion point

Challenges and Gaps in AI Safety for the Global South


Topics

Human rights and the ethical dimensions of the information society | Artificial intelligence


Language nuance goes beyond translation; contextual meaning is critical

Explanation

Sibande notes that language involves lived meaning and cultural context, which simple translation cannot capture.


Evidence

“Language in itself is not just about vocabulary.” [23]. “It’s not enough for a large language model to have strong translation capabilities.” [27].


Major discussion point

Multilingual and Culturally Contextual Evaluation


Topics

Closing all digital divides | Artificial intelligence


D

Dr. Balaraman Ravindran

Speech speed

172 words per minute

Speech length

565 words

Speech time

196 seconds

Coordinate among overlapping initiatives to avoid duplication and amplify impact

Explanation

Ravindran calls for mechanisms to coordinate the many emerging AI safety and capacity‑building initiatives across the Global South.


Evidence

“And we have to figure out a way how we would coordinate operations among these initiatives as well.” [103]. “And after that, if there is a lot more coordination.” [104].


Major discussion point

Role and Design of the Global South Network for Trustworthy AI


Topics

Capacity development | Artificial intelligence


Absence of dedicated safety institutes in the Global South limits capacity building

Explanation

Ravindran points out that many Global South countries lack safety or oversight institutes, hindering participation in global AI safety dialogues.


Evidence

“We don’t have safety institutes in the global south, right, that can participate in the dialogue.” [137]. “And many countries in the Global South are actually unlikely to even have in the near term their own safety or oversight institutes.” [138].


Major discussion point

Infrastructure, Capacity, and Resource Constraints


Topics

Capacity development | Artificial intelligence


M

Mr. Quintin Chou‑Lambert

Speech speed

Default speed

Speech length

Default length

Speech time

Default duration

Connect grassroots empirical findings to global governance forums

Explanation

Chou‑Lambert stresses that moving from large, generic models to context‑specific, smaller language models requires empirical evidence that can inform global governance.


Evidence

“So moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small -language models to context turns the issue of AI safety into a more fuzzy kind of discussion and one which really needs empirical evidence.” [122].


Major discussion point

Role and Design of the Global South Network for Trustworthy AI


Topics

Artificial intelligence | Monitoring and measurement


M

Mr. Quintin Chou-Lambert

Speech speed

135 words per minute

Speech length

324 words

Speech time

143 seconds

Networks bridge local AI safety challenges to global governance

Explanation

Chou‑Lambert stresses that multi‑stakeholder networks are essential for surfacing grassroots threats and examples so that international AI safety discussions do not overlook the perspectives of the majority of people worldwide. By aggregating experiences from over a hundred countries, these networks expand the conversation beyond a narrow set of institutions.


Evidence

“And that’s why, to close and to echo Ambassador Thieger, these kinds of networks play a crucial role in connecting and bringing examples of the challenges that we face.” [3]. “cases of threats from various sources to local people into discussions so that international discussions do not ignore or omit or discount the perspectives of the vast majority of people on the planet.” [4]. “over 100 countries engaging.” [5]. “And I think the trends in the institutional discussions … where you had 60‑plus, and now here.” [6]. “And so as the conversation in these summit settings and in the international level has widened and to include more countries and more people … the focus has … become much more focused of encompassing other perspectives.” [8].


Major discussion point

Role and Design of the Global South Network for Trustworthy AI


Topics

Artificial intelligence | Governance | Capacity development


Context‑specific AI safety demands empirical, field‑tested evidence

Explanation

Moving from monolithic, expensive models to smaller, locally‑tailored language models makes AI safety a more ambiguous issue. Chou‑Lambert argues that this shift requires concrete, empirical evidence from on‑the‑ground deployments to inform safety assessments.


Evidence

“So moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small‑language models to context turns the issue of AI safety into a more fuzzy kind of discussion and one which really needs empirical evidence.” [9]. “So the concept of AI safety becomes less of a, or it kind of edges into this more contextual field, and that’s where this kind of low perspectives, field‑tested examples can be very helpful to surface, which we’re missing.” [10].


Major discussion point

Challenges and Gaps in AI Safety for the Global South


Topics

Artificial intelligence | Monitoring and measurement | Capacity development


One‑size‑fits‑all technical standards are inadequate for diverse contexts

Explanation

Chou‑Lambert cautions that universal AI technical standards cannot capture the nuanced safety needs of varied cultural and infrastructural settings, calling for standards that are adaptable to local realities.


Evidence

“And I’d say the idea of AI standards as technical standards don’t solve that issue because a one‑size‑fits‑all standard will not be contextually sensitive.” [12].


Major discussion point

Governance, Representation, and Power Dynamics


Topics

Artificial intelligence | Governance | Human rights and the ethical dimensions of the information society


UN Global Dialogue on AI Governance provides an inclusive, evidence‑based platform

Explanation

He highlights the establishment of a United Nations‑backed Global Dialogue on AI Governance that brings together all 193 member states, drawing on analysis from an independent scientific panel to assess AI risks and opportunities, thereby offering a venue for incorporating Global South evidence.


Evidence

“We now have the United Nations Global Dialogue on AI Governance, which will include a whole 193 member states informed by analysis from an independent international scientific panel on AI, which will look at the risks and also opportunities and impacts of AI.” [11].


Major discussion point

Governance, Representation, and Power Dynamics


Topics

Artificial intelligence | International cooperation | Governance


Infrastructure and energy constraints shape AI safety priorities in the Global South

Explanation

Chou‑Lambert points out that limited infrastructure and energy connectivity in many regions restrict the deployment of large, compute‑intensive models, making safety considerations inherently tied to local resource realities.


Evidence

“There is less, perhaps, infrastructure or energy connection to go around.” [7].


Major discussion point

Infrastructure, Capacity, and Resource Constraints


Topics

Capacity development | Environmental impacts | Artificial intelligence


Agreements

Agreement points

Global South underrepresentation in AI safety governance requires urgent action

Speakers

– Dr. Urvashi Aneja
– Ambassador Philip Thigo

Arguments

Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums


Network is timely but also overdue given exclusion of global majority from safety conversations


Summary

Both speakers emphasize that Global South countries and communities have been systematically excluded from AI safety governance structures, with Ambassador Thigo noting Kenya is the only Global South member of international AI safety institutes, while Dr. Aneja highlights that many Global South countries lack their own safety institutes


Topics

Artificial intelligence | Capacity development


AI safety must be contextually defined and culturally sensitive

Speakers

– Dr. Rachel Sibande
– Ms. Chenai Chair
– Ambassador Philip Thigo

Arguments

Safety and harm must be redefined according to social, cultural, and linguistic contexts where AI is deployed


Developers miss user experience and context, often exacerbating existing inequalities like gender-based violence


AI safety extends beyond technology to include environmental harms and full lifecycle accountability


Summary

All three speakers agree that current AI safety definitions are inadequate for Global South contexts and must account for local social norms, cultural nuances, gender dynamics, and broader socio-technical impacts including environmental considerations


Topics

Artificial intelligence | Human rights and the ethical dimensions of the information society | Closing all digital divides


Need for multilingual and multicultural evaluation systems

Speakers

– Mr. Abhishek Singh
– Ms. Natasha Crampton
– Ms. Chenai Chair

Arguments

Initiative aligns with New Delhi Frontier AI commitments and supports compliance with multilingual benchmarks


Challenge lies in scaling thoughtful, community-led evaluation work across thousands of languages and millions of cultural settings


Masakhane Hub will contribute African benchmarking work for 50 African languages


Summary

All speakers recognize the critical need for AI evaluation systems that work across diverse languages and cultures, with specific commitments to multilingual benchmarking and acknowledgment of the scale challenge involved


Topics

Artificial intelligence | Closing all digital divides


Capacity building and institutional strengthening are essential

Speakers

– Ambassador Philip Thigo
– Mr. Amir Banifatemi
– Dr. Balaraman Ravindran

Arguments

Global South countries lack capacity building, access to compute, and face linguistic/cultural mismatches in benchmarking


Absence of institutional frameworks in Global South delays feedback loops and compounds potential harms


Multiple networks are launching simultaneously requiring coordination to avoid fragmentation


Summary

All speakers identify capacity building as a fundamental challenge, noting gaps in technical capacity, institutional frameworks, and the need for coordinated efforts to build sustainable capabilities across the Global South


Topics

Capacity development | Artificial intelligence


Evaluation must be continuous and proactive rather than reactive

Speakers

– Ms. Natasha Crampton
– Dr. Rachel Sibande

Arguments

Need for sustainable evaluation systems that run continuously, not just once before product release


Foundation will institutionalize safety evaluation right at deployment rather than post-deployment


Summary

Both speakers emphasize moving from one-time or post-deployment safety assessments to continuous, proactive evaluation systems that monitor AI systems throughout their lifecycle


Topics

Artificial intelligence | Monitoring and measurement


Similar viewpoints

Both speakers see the network as a crucial bridge between local, contextual AI deployment experiences and global governance structures, ensuring that international AI governance reflects the perspectives and needs of the global majority

Speakers

– Dr. Urvashi Aneja
– Mr. Quintin Chou-Lambert

Arguments

Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums


Network should connect empirical evidence from field testing to international governance discussions


Topics

Artificial intelligence | Follow-up and review


Both speakers emphasize that AI safety and innovation are complementary rather than competing objectives, with concrete commitments to responsible AI development and significant infrastructure investments

Speakers

– Mr. Abhishek Singh
– Ms. Natasha Crampton

Arguments

Innovation focus should not stifle responsible AI development but ensure safe and trustworthy deployment


Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South


Topics

Artificial intelligence | The enabling environment for digital development | Financial mechanisms


Both speakers highlight structural power imbalances in AI governance, with Ambassador Thigo focusing on institutional power concentration and Mr. Banifatemi on economic incentive structures that fail to prioritize safety

Speakers

– Ambassador Philip Thigo
– Mr. Amir Banifatemi

Arguments

Benchmarks are not neutral and power concentration in few institutions must be addressed


Safety lacks clear definition and is not integrated into financial planning or cost structures of companies


Topics

Artificial intelligence | Human rights and the ethical dimensions of the information society


Unexpected consensus

Economic incentives for AI safety

Speakers

– Mr. Amir Banifatemi
– Ms. Natasha Crampton

Arguments

Safety lacks clear definition and is not integrated into financial planning or cost structures of companies


Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South


Explanation

Unexpected alignment between a consultant’s critique of corporate safety incentives and a corporate representative’s substantial financial commitments, suggesting recognition across industry that current economic models inadequately incentivize safety


Topics

Artificial intelligence | Financial mechanisms | The enabling environment for digital development


Need for coordination among multiple AI initiatives

Speakers

– Dr. Balaraman Ravindran
– Dr. Urvashi Aneja

Arguments

Multiple networks are launching simultaneously requiring coordination to avoid fragmentation


Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums


Explanation

Academic and network founder both recognize the proliferation of AI safety initiatives could lead to fragmentation rather than strengthened capacity, showing pragmatic consensus on the need for strategic coordination


Topics

Artificial intelligence | Capacity development


Overall assessment

Summary

Strong consensus emerged around five key areas: Global South underrepresentation in AI governance, need for contextually-sensitive safety definitions, importance of multilingual evaluation systems, capacity building requirements, and shift toward proactive evaluation approaches. Speakers consistently emphasized that current AI safety frameworks inadequately serve Global South contexts and populations.


Consensus level

High level of consensus with remarkable alignment across government, industry, civil society, and academic perspectives. This suggests the Global South Network for Trustworthy AI addresses widely recognized gaps in current AI governance structures. The consensus implies strong potential for collaborative action and suggests the network fills a critical institutional void in global AI governance.


Differences

Different viewpoints

Timeline and urgency of network establishment

Speakers

– Ambassador Philip Thigo
– Dr. Urvashi Aneja

Arguments

Network is timely but also overdue given exclusion of global majority from safety conversations


Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums


Summary

Ambassador Thigo emphasizes that while the network is timely, it is also ‘late’ and there’s an urgency to scale up quickly due to structural exclusion, while Dr. Aneja presents it as a timely response to current gaps without the same sense of overdue urgency


Topics

Artificial intelligence | Capacity development


Scope and definition of AI safety

Speakers

– Ambassador Philip Thigo
– Mr. Amir Banifatemi
– Dr. Rachel Sibande

Arguments

AI safety extends beyond technology to include environmental harms and full lifecycle accountability


Safety lacks clear definition and is not integrated into financial planning or cost structures of companies


Safety and harm must be redefined according to social, cultural, and linguistic contexts where AI is deployed


Summary

The speakers disagree on what constitutes AI safety – Ambassador Thigo advocates for a comprehensive approach including environmental impacts, Mr. Banifatemi focuses on the lack of clear definition and financial integration, while Dr. Sibande emphasizes cultural and contextual redefinition


Topics

Artificial intelligence | Environmental impacts | Human rights and the ethical dimensions of the information society


Approach to scaling evaluation systems

Speakers

– Ms. Natasha Crampton
– Dr. Balaraman Ravindran

Arguments

Challenge lies in scaling thoughtful, community-led evaluation work across thousands of languages and millions of cultural settings


Multiple networks are launching simultaneously requiring coordination to avoid fragmentation


Summary

Ms. Crampton focuses on the technical challenge of scaling individual evaluation systems, while Dr. Ravindran emphasizes the need for coordination among multiple competing networks to avoid fragmentation and resource competition


Topics

Artificial intelligence | Capacity development


Unexpected differences

Network coordination vs individual network development

Speakers

– Dr. Balaraman Ravindran
– Dr. Urvashi Aneja

Arguments

Multiple networks are launching simultaneously requiring coordination to avoid fragmentation


Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums


Explanation

Unexpectedly, Dr. Ravindran, who is part of the founding network, raises concerns about too many similar networks launching simultaneously and the need for coordination, which somewhat contradicts the celebratory launch tone of the event


Topics

Artificial intelligence | Capacity development


Corporate responsibility approach

Speakers

– Ms. Natasha Crampton
– Mr. Amir Banifatemi

Arguments

Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South


Safety lacks clear definition and is not integrated into financial planning or cost structures of companies


Explanation

While both represent industry perspectives, Crampton emphasizes Microsoft’s commitments and investments, while Banifatemi argues that companies fundamentally won’t prioritize safety without financial penalties, creating an unexpected tension between corporate commitment and systemic critique


Topics

Artificial intelligence | The enabling environment for digital development | Financial mechanisms


Overall assessment

Summary

The main areas of disagreement center around the scope and definition of AI safety, the urgency and timing of network establishment, approaches to scaling evaluation systems, and the balance between individual network development versus coordination among multiple initiatives


Disagreement level

Moderate disagreement with significant implications – while all speakers support the network’s goals, their different emphases on environmental impacts, power dynamics, technical scaling, and institutional coordination could lead to different priorities and resource allocation decisions that may fragment efforts or create competing approaches to Global South AI safety


Partial agreements

Partial agreements

All speakers agree that current evaluation approaches are insufficient and need to be more systematic and continuous, but they disagree on the primary solution – Crampton focuses on sustainable technical systems, Banifatemi on institutional frameworks, and Sibande on timing of deployment

Speakers

– Ms. Natasha Crampton
– Mr. Amir Banifatemi
– Dr. Rachel Sibande

Arguments

Need for sustainable evaluation systems that run continuously, not just once before product release


Absence of institutional frameworks in Global South delays feedback loops and compounds potential harms


Foundation will institutionalize safety evaluation right at deployment rather than post-deployment


Topics

Artificial intelligence | Capacity development | Monitoring and measurement


All agree that current AI systems fail to account for Global South contexts and can cause harm, but they emphasize different aspects – Thigo focuses on power dynamics in benchmarking, Chair on deployment context and inequality, and Sibande on linguistic and cultural understanding

Speakers

– Ambassador Philip Thigo
– Ms. Chenai Chair
– Dr. Rachel Sibande

Arguments

Benchmarks are not neutral and power concentration in few institutions must be addressed


Developers miss user experience and context, often exacerbating existing inequalities like gender-based violence


Language models need understanding of lived experiences, not just translation capabilities


Topics

Artificial intelligence | Human rights and the ethical dimensions of the information society | Closing all digital divides


Similar viewpoints

Both speakers see the network as a crucial bridge between local, contextual AI deployment experiences and global governance structures, ensuring that international AI governance reflects the perspectives and needs of the global majority

Speakers

– Dr. Urvashi Aneja
– Mr. Quintin Chou-Lambert

Arguments

Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums


Network should connect empirical evidence from field testing to international governance discussions


Topics

Artificial intelligence | Follow-up and review


Both speakers emphasize that AI safety and innovation are complementary rather than competing objectives, with concrete commitments to responsible AI development and significant infrastructure investments

Speakers

– Mr. Abhishek Singh
– Ms. Natasha Crampton

Arguments

Innovation focus should not stifle responsible AI development but ensure safe and trustworthy deployment


Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South


Topics

Artificial intelligence | The enabling environment for digital development | Financial mechanisms


Both speakers highlight structural power imbalances in AI governance, with Ambassador Thigo focusing on institutional power concentration and Mr. Banifatemi on economic incentive structures that fail to prioritize safety

Speakers

– Ambassador Philip Thigo
– Mr. Amir Banifatemi

Arguments

Benchmarks are not neutral and power concentration in few institutions must be addressed


Safety lacks clear definition and is not integrated into financial planning or cost structures of companies


Topics

Artificial intelligence | Human rights and the ethical dimensions of the information society


Takeaways

Key takeaways

The Global South Network for Trustworthy AI was successfully launched to address the critical underrepresentation of Global South perspectives in global AI safety infrastructure and governance forums


AI safety must be redefined contextually for Global South deployment, considering social, cultural, linguistic, and infrastructural differences rather than applying one-size-fits-all standards


Current AI evaluation methods are inadequate for Global South contexts, missing critical aspects like lived experiences in language, cultural nuances, gender dynamics, and post-deployment harm tracking


Industry faces significant challenges in scaling thoughtful, community-led evaluation work across thousands of languages and millions of cultural settings while maintaining sustainability


Safety is not adequately integrated into companies’ financial planning and cost structures, creating insufficient incentives for comprehensive safety measures


Multiple similar networks are launching simultaneously, requiring coordination to avoid fragmentation and maximize impact through harmonized efforts


The network will serve as crucial connective tissue between global governance architecture and real-world deployment contexts in the Global South


Resolutions and action items

Network will launch five flagship projects in the coming year: multilingual AI benchmarks, gender safety taxonomy, procurement guidelines, evaluation methodologies, and health information systems evaluation


Microsoft committed to fulfilling New Delhi Frontier AI commitments on multilingual benchmarks and usage data sharing, plus $50 billion infrastructure investment in Global South by end of decade


Gates Foundation will institutionalize safety evaluation right at AI deployment rather than waiting for post-deployment issues


Masakhane Hub will contribute to African benchmarking work covering 50 African languages


Cognizant will provide open source safety evaluation tools with cultural context through their Bangalore and San Francisco labs


Network will establish regional nodes/hubs to better serve diverse Global South contexts


India AI mission committed to provide ongoing support for network operations and objectives


Network will work to integrate findings into multilateral processes including UN Global Dialogue on AI Governance and scientific panel


Focus on cross-border collaborative problem-solving that requires international cooperation rather than parallel work in separate geographies


Unresolved issues

How to effectively coordinate with multiple other AI safety networks launching simultaneously to avoid duplication and fragmentation


Lack of clear, universally accepted definition of what constitutes ‘safety’ in AI systems across different contexts


How to create sustainable funding mechanisms for ongoing evaluation work rather than one-time assessments


Absence of institutional frameworks and rule of law in many Global South countries that delays feedback loops and compounds potential harms


How to address the fundamental talent inclusion problem where those building safety systems lack exposure to Global South deployment contexts


How to create financial incentives and cost structures that make safety a priority for companies deploying AI in Global South


How to scale community-led, contextual evaluation work to cover thousands of languages and millions of cultural settings


How to close the accountability loop so that evaluation work translates into meaningful protection for citizens


How to address the ‘new global south in AI’ that includes countries beyond traditional Global South due to AI concentration in just two countries and few companies


Suggested compromises

Expanding the definition of ‘Global South’ in AI context to include other regions like parts of Europe and Latin America that face similar exclusion from AI governance


Balancing innovation promotion with safety requirements – ensuring safety measures don’t stifle AI development while protecting populations from harm


Creating hybrid evaluation approaches that combine technical benchmarks with real-world, contextual assessments


Developing both model-level and system-level evaluation tools to address the full deployment context rather than just underlying AI models


Establishing open source and free access tools for safety evaluation while building local capacity and ownership in Global South countries


Creating incident reporting systems that can capture both intentional and unintentional misuse of AI technologies


Developing procurement guidelines as a policy lever that Global South countries can use to shape markets for responsible innovation


Thought provoking comments

We now know in the global north of artificial intelligence is two countries and a few companies. So we must, beyond this, extend to also include other colleagues, whether it’s from Europe, Western Europe, or Latin America.

Speaker

Ambassador Philip Thigo


Reason

This comment reframes the entire global AI landscape by suggesting that even the ‘global north’ in AI is extremely concentrated, essentially challenging the traditional north-south binary. It introduces the concept of a ‘new global south in AI’ that includes traditionally developed countries who are also excluded from AI development.


Impact

This fundamentally shifted the discussion from a simple global south vs. global north framework to a more nuanced understanding of AI power concentration. It broadened the scope of who should be included in the network and influenced subsequent speakers to think beyond traditional geographical boundaries.


Language in itself is not just about vocabulary. It’s also about the lived meaning, the lived experiences… if the mother says their waters have broken, which clinically is a critical incident that should warrant that mother to be referred to a health facility, but if you translate that from the local language to English… that will literally mean I have thrown away water.

Speaker

Dr. Rachel Sibande


Reason

This powerful example moves the discussion from abstract concepts of multilingual AI to concrete life-or-death scenarios. It demonstrates how current translation-based approaches to multilingual AI can fail catastrophically in critical contexts, revealing the inadequacy of surface-level language support.


Impact

This comment grounded the entire safety discussion in tangible, high-stakes examples. It influenced subsequent speakers to focus more on contextual understanding rather than just technical capabilities, and reinforced the urgency of the network’s mission with a compelling real-world scenario.


Safety, on the other side, is not costed into financial systems… There is no penalty of not being safe. So as long as there is no constraint to put safety as a cost structure, which strong mandate, companies will not pay attention or enough attention.

Speaker

Mr. Amir Banifatemi


Reason

This comment cuts to the heart of why AI safety remains inadequate by identifying the fundamental economic incentive problem. It shifts the discussion from technical solutions to systemic economic and regulatory issues that drive corporate behavior.


Impact

This observation reframed the conversation from focusing solely on technical evaluation methods to addressing the underlying economic structures that perpetuate unsafe AI deployment. It influenced the discussion toward considering regulatory and financial mechanisms as essential components of any safety framework.


There is just too many of these initiatives that are getting launched… we have to figure out a way how we would coordinate operations among these initiatives as well. So I think that would be a great multiplier instead of everybody going out and saying, okay, let me see what small piece of the pie that I can get.

Speaker

Dr. Balaraman Ravindran


Reason

This meta-observation about the proliferation of AI safety networks introduces a critical coordination challenge that could undermine the effectiveness of all initiatives. It’s insightful because it addresses the risk of fragmentation in a field that requires collective action.


Impact

This comment introduced a sobering reality check into the celebratory launch atmosphere, forcing participants to consider how their network would differentiate itself and coordinate with others. It shifted the discussion toward practical implementation challenges and the need for strategic positioning within a crowded landscape.


Benchmarks are not neutral. Sometimes I think I like to be honest because that’s what evaluation needs to do. And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means. Governance is about power.

Speaker

Ambassador Philip Thigo


Reason

This comment exposes the political nature of AI evaluation by highlighting how seemingly technical benchmarks embed power structures and value judgments. It challenges the notion that evaluation can be objective and reveals how current systems perpetuate exclusion.


Impact

This observation elevated the discussion from technical evaluation methods to questions of power, representation, and democratic participation in AI governance. It reinforced the political urgency of the network’s mission and influenced other speakers to consider the governance implications of their technical work.


Overall assessment

These key comments fundamentally transformed what could have been a routine network launch into a sophisticated analysis of AI governance challenges. Ambassador Thigo’s interventions consistently elevated the discussion by introducing power dynamics and structural critiques, while Dr. Sibande’s concrete examples grounded abstract concepts in life-or-death realities. Mr. Banifatemi’s economic analysis and Dr. Ravindran’s coordination concerns added practical urgency to the theoretical framework. Together, these comments created a multi-layered conversation that addressed technical, economic, political, and practical dimensions of AI safety in the Global South, establishing a robust intellectual foundation for the network’s future work.


Follow-up questions

How do we ensure that we get necessary support from all stakeholders to make the Global South Network for Trustworthy AI functional?

Speaker

Mr. Abhishek Singh


Explanation

This addresses the critical challenge of securing ongoing commitment and resources from industry, governments, and multilateral organizations to operationalize the network beyond its launch


How do we create tools for evaluating models in various languages and build up capacity in all countries of the global south?

Speaker

Mr. Abhishek Singh


Explanation

This highlights the need for practical evaluation tools and capacity building mechanisms that can work across diverse linguistic and cultural contexts in the Global South


How do we ensure compliance to the New Delhi Frontier AI commitments regarding sharing usage data and multilingual performance benchmarks?

Speaker

Mr. Abhishek Singh


Explanation

This addresses the implementation gap between policy commitments and actual practice in AI safety and evaluation


How do we fit this network into the multilateral process, including the UN scientific panel on AI and global dialogue on AI governance?

Speaker

Ambassador Philip Thigo


Explanation

This is crucial for ensuring the network’s work influences global AI governance structures and doesn’t operate in isolation


How do we close the accountability loop so that evaluation work ultimately matters for citizens?

Speaker

Ambassador Philip Thigo


Explanation

This addresses the fundamental question of translating technical evaluation work into tangible benefits and protections for people in the Global South


How do we take thoughtful, community-led evaluation work and scale it up for thousands of languages and millions of different cultural settings?

Speaker

Ms. Natasha Crampton


Explanation

This highlights the scalability challenge of conducting context-sensitive AI safety evaluations across the vast diversity of the Global South


How do we build a sustainable, grounded, community-led system of scalable evaluation that can be run on an ongoing basis?

Speaker

Ms. Natasha Crampton


Explanation

This addresses the need for continuous rather than one-time evaluation systems that can adapt to changing AI systems and contexts


How do we coordinate operations among multiple AI safety and capacity building networks being launched globally?

Speaker

Dr. Balaraman Ravindran


Explanation

This addresses the risk of fragmentation and duplication of efforts across various AI safety initiatives in the Global South


How can we pick problems that will necessarily require cross-border collaboration rather than just parallel work in different geographies?

Speaker

Dr. Balaraman Ravindran


Explanation

This focuses on identifying research challenges that can only be solved through genuine international collaboration, strengthening the network’s value proposition


How do we accelerate learning capabilities and feedback loops in Global South countries that lack institutional frameworks for AI safety?

Speaker

Mr. Amir Banifatemi


Explanation

This addresses the structural disadvantage many Global South countries face in developing rapid response mechanisms for AI safety issues


How do we institutionalize the evaluation of safety of AI solutions right at deployment rather than post-deployment?

Speaker

Dr. Rachel Sibande


Explanation

This addresses the timing gap in current safety practices where issues are only identified after systems are already causing harm


At what point can we track whether users are substituting their cognitive capabilities with AI models or becoming overly emotionally dependent?

Speaker

Dr. Rachel Sibande


Explanation

This raises important questions about measuring psychological and cognitive impacts of AI use that current benchmarking doesn’t capture


How do we ensure that only a handful of institutions don’t define what risks are measured, what harms are prioritized, and what safe performance means?

Speaker

Ambassador Philip Thigo


Explanation

This addresses power concentration in AI safety standard-setting and the need for more democratic and inclusive governance structures


Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.