Towards a Safer South Launching the Global South AI Safety Research Network
20 Feb 2026 17:00h - 18:00h
Towards a Safer South Launching the Global South AI Safety Research Network
Session at a glance
Summary
This discussion centered on the launch of the Global South Network for Trustworthy AI at the India AI Impact Summit, addressing the critical need for inclusive AI safety infrastructure that represents the perspectives of Global South countries. Dr. Urvashi Aneja, founder of Digital Futures Lab, opened by highlighting how AI systems are rapidly being deployed in critical sectors across the Global South, where low institutional capacity and deep inequities create both immense opportunities and significant risks. She emphasized that Global South organizations remain underrepresented in global AI safety governance, despite being uniquely positioned to identify real-world deployment risks invisible to lab-based evaluations.
Mr. Abhishek Singh from India’s AI Mission stressed that while everyone agrees on the need for safe AI, the challenge lies in developing technical tools and benchmarks to address risks, particularly for multilingual contexts where most models are evaluated only in English. Ambassador Philip Thigo from Kenya pointed out the structural exclusion of Global South countries from safety conversations and emphasized that safety must extend beyond technology to include environmental harms, biases, and full lifecycle accountability. The panelists identified several critical gaps in current AI safety approaches, including the need to redefine safety according to social and cultural contexts, address linguistic nuances beyond simple translation, and understand emerging harms during actual usage.
Industry representatives acknowledged the challenge of scaling context-sensitive evaluations across thousands of languages and millions of cultural settings while maintaining sustainability. The discussion concluded with commitments from various organizations to contribute to multilingual benchmarking, incident reporting tools, and infrastructure investments, establishing a foundation for collaborative AI safety work that centers Global South perspectives and needs.
Keypoints
Major Discussion Points:
– Launch of the Global South Network for Trustworthy AI: The primary focus was announcing this new network bringing together research institutions from Asia, Africa, and Latin America to evaluate real-world AI impacts and build trust mechanisms localized to different linguistic, cultural, and infrastructural contexts.
– Context-sensitive AI safety challenges in the Global South: Speakers emphasized that current AI safety frameworks miss critical contextual factors like local languages, cultural norms, gender dynamics, and social inequalities. Examples included AI models failing to understand local expressions (like “waters have broken” translating to “thrown away water”) and agricultural tools with male voices potentially exacerbating gender-based violence.
– Structural gaps in global AI governance: Multiple speakers highlighted that Global South countries are underrepresented in global AI safety infrastructure, with only Kenya mentioned as a member of international AI safety institutes. This creates a disconnect between where AI harms occur most and where safety decisions are made.
– Need for multilingual and multicultural evaluation systems: The discussion emphasized developing benchmarks beyond English-language models, creating evaluation tools that capture societal risks specific to Global South contexts, and building sustainable, community-led assessment frameworks.
– Concrete initiatives and commitments: The network outlined five flagship projects including multilingual AI benchmarks, gender safety taxonomy, procurement guidelines, evaluation methodologies, and healthcare AI assessments. Industry partners like Microsoft committed to infrastructure investments and data sharing.
Overall Purpose:
The discussion aimed to formally launch the Global South Network for Trustworthy AI and establish a collaborative framework for addressing AI safety challenges specific to developing countries, while advocating for greater inclusion of Global South perspectives in international AI governance.
Overall Tone:
The tone was collaborative and urgent throughout, with speakers expressing both excitement about the network’s potential and concern about the pressing need for action. There was a consistent emphasis on moving beyond theoretical discussions to practical implementation, with industry and government representatives showing strong commitment to supporting the initiative. The conversation maintained a professional yet passionate quality, reflecting the speakers’ shared belief in the critical importance of inclusive AI safety.
Speakers
Speakers from the provided list:
– Dr. Urvashi Aneja – Founder and Director of Digital Futures Lab
– Mr. Abhishek Singh – Leadership role in India AI Summit and India AI mission
– Ambassador Philip Thigo – Special Envoy on Technology from the Republic of Kenya
– Mr. Quintin Chou-Lambert – Chief of Office and AI Lead, UN Office for Digital and Emerging Technologies
– Ms. Natasha Crampton – Vice President and Chief Responsible AI Officer at Microsoft
– Dr. Rachel Sibande – Senior Program Officer AI for Africa at the Gates Foundation
– Ms. Chenai Chair – Director of the Masakane African Language Hub
– Mr. Amir Banifatemi – Chief Responsible AI Officer at Cognizant
– Dr. Balaraman Ravindran – Head Center of Responsible AI at IIT Madras
Additional speakers:
None – all speakers mentioned in the transcript were included in the provided speakers names list.
Full session report
This discussion centered on the launch of the Global South Network for Trustworthy AI at the India AI Impact Summit, bringing together government officials, industry leaders, civil society representatives, and academics to address the underrepresentation of Global South perspectives in international AI safety governance.
Network Vision and Structure
Dr. Urvashi Aneja, founder and director of Digital Futures Lab, opened by highlighting the fundamental challenge: while AI systems are being rapidly deployed across critical sectors like healthcare, education, and government in the Global South, these regions remain underrepresented in global safety infrastructure. She noted the paradox that countries with the greatest potential to leverage AI for development are precisely those most excluded from safety governance.
Dr. Aneja outlined five flagship projects for the network: multilingual AI benchmarking (with the Collective Intelligence Project and CARIA), a gender safety taxonomy project (with GXD Hub and the Global Center for AI Governance), procurement guidelines development, evaluation methodology work (with ITS Rio), and health information systems evaluation. The founding partners include Digital Futures Lab, Sirai from IIT Madras, Global Center for AI Governance, ITS Rio, and International Innovation Corps.
Government Support and Policy Alignment
Mr. Abhishek Singh from India’s AI Mission provided governmental backing, emphasizing that while consensus exists around the need for safe and trusted AI, the challenge lies in developing technical tools and benchmarks to address identified risks. He highlighted a critical gap: most AI models are assessed using predominantly English-language benchmarks, despite countries like India having 22 official languages and numerous dialects.
Singh noted the network’s alignment with the New Delhi Frontier AI commitments, which secured agreements from major AI companies to share usage data and develop multilingual performance benchmarks.
Structural Representation Challenges
Ambassador Philip Thigo from Kenya described the network as “timely but also late” due to the structural exclusion of Global South countries from safety conversations. He observed that Kenya is the only Global South member of international AI safety institutes, illustrating the representation gap. Thigo argued that “the global north of artificial intelligence is two countries and a few companies,” suggesting even traditionally developed nations face exclusion from AI governance.
He expanded the definition of AI safety beyond technical considerations to include environmental harms, biases, misinformation, and full lifecycle accountability from “minds to models,” emphasizing that governance is fundamentally about power.
Contextual Safety and Real-World Examples
Dr. Rachel Sibande from the Gates Foundation provided compelling examples of how linguistic nuance affects AI safety. She explained how the phrase “waters have broken”—a critical medical emergency—translates literally as “thrown away water” from local languages, potentially causing AI systems to miss life-threatening situations. She emphasized that language support requires understanding lived experiences and cultural contexts, not merely translation capabilities.
Sibande also discussed her work in Malawi, referencing the country’s identity as the “warm heart of Africa,” and highlighted challenges in measuring emerging harms during AI usage, such as cognitive substitution or emotional dependency.
Ms. Chenai Chair from the Masakhane African Language Hub reinforced these concerns, noting that Africa has over 2,000 documented languages, with Masakhane working on only 50. She provided an example of how agricultural tools with male voices could potentially increase gender-based violence in contexts with high gender inequality, demonstrating how design choices can have profound social consequences.
Industry Perspectives and Implementation Challenges
Ms. Natasha Crampton from Microsoft articulated the central scaling challenge: how to extend thoughtful, community-led evaluation work across thousands of languages and cultural settings while maintaining sustainability. She emphasized the need for continuous rather than one-time evaluation systems and referenced Microsoft’s commitment to the New Delhi Frontier AI agreements.
Mr. Amir Banifatemi offered a more critical assessment, arguing that safety lacks clear definition and is not integrated into companies’ financial planning or cost structures. He observed that “there is no penalty of not being safe,” highlighting the need for regulatory frameworks that make safety a financial imperative.
Coordination and Network Proliferation
Dr. Balaraman Ravindran from IIT Madras raised important questions about coordination, noting that multiple AI safety networks are launching simultaneously—including initiatives in Africa, China, and through UN processes. He called for coordination rather than competition among networks and suggested focusing on problems requiring cross-border collaboration.
Multilateral Integration
Mr. Quintin Chou-Lambert from the UN Office for Digital and Emerging Technologies discussed the evolution of international AI discussions from the concentrated focus of Bletchley Park to broader participation in subsequent summits, with over 100 countries now engaging in discussions. He referenced the UN Global Dialogue on AI Governance and the independent international scientific panel on AI as potential integration points for the network’s work.
Technical Challenges and Event Context
The event faced some technical difficulties, with portions of Dr. Aneja’s presentation becoming repetitive due to technical issues. Time constraints affected the panel discussion format, and the session concluded with mentions of a photo opportunity for participants.
Conclusion
The launch represents an important step toward including Global South perspectives in AI safety governance. The network aims to address critical gaps in current evaluation systems while navigating challenges around scaling, coordination with other initiatives, and translating evaluation work into meaningful protections for citizens across the Global South. Success will depend on the network’s ability to maintain contextual sensitivity while developing scalable methodologies and effectively integrating with broader international governance processes.
Session transcript
Thank you. Thank you. Thank you. Thank you. Thank you. Thank you. Thank you. Good evening, everyone. My name is Urvishya Neja. I am the founder and director of Digital Futures Lab. And I am so excited to see all of you here and to have you all here for the launch of this network. So it’s a real pleasure to welcome you to the launch of the Global South Network for Trustworthy AI here at the India AI Impact Summit. On behalf of Digital Futures Lab and our other founding partners, Sirai from IIT Madras, the Global Center for AI Governance, ITS Rio, International Innovation Corps, thank you all for being here. And we’re especially grateful to Mr. Abhishek Singh and Ambassador Philip Tigo and Mr. Quintin Chow and to all our distinguished speakers and guests who are joining us today.
Across the Global South, AI systems are being rapidly deployed in critical social sectors such as healthcare, education, judiciary, and in government. And while the opportunities are immense, in many of these contexts, many of these contexts are also marked by low institutional capacity, deep societal inequities, popularization, and populations with low levels of literacy. So while the potential is immense, the risks and harms are also immense. And so it’s particularly important that we figure out ways to make AI safe and trustworthy in these contexts to ensure not only that we protect the populations and to ensure that we don’t exasperate existing harms, but also to ensure that we build the infrastructure for safe and inclusive AI adoption.
Unfortunately, Global South organizations, Global South communities, Global South states remain underrepresented in global safety and governance infrastructures. And many countries in the Global South are actually unlikely to even have in the near term their own safety or oversight institutes. And there’s a real risk, therefore, that the concerns and priorities of these countries, of these communities remain underrepresented in the global safety infrastructure. And precisely those countries that have the most potential or the most opportunity to leverage AI. Independent civil society organizations are uniquely positioned to address this gap. Their proximity to real -world deployment contexts enables them to surface risks that are invisible to lab -based evaluations or testing. The form of grounded evidence that civil society organizations can bring can inform global safety benchmarks, standard -setting processes, and risk assessments, providing corrective signals to technical and regulatory institutions.
The Global South Network for Trustworthy AI works to advance exactly these objectives – to evaluate the real -world impact of AI systems, to build the trust and oversight mechanisms localized to different linguistic, cultural, and infrastructural contexts, and to elevate Global South perspectives in global AI governance forums. It is particularly encouraging that this initiative also aligns closely with the recently announced New Delhi Frontier AI. The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI systems, and to develop a more robust and efficient system The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI systems, and to develop a more robust and efficient system the real -world impact of AI systems, and to develop a more robust and efficient system The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI.
impact of AI. The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world systems. The Global South Network for Trustworthy AI brings together some of the leading research institutions from across the Global South. We are joined by a community of organizations from Asia, from Africa, from Latin America, and the names of which you see displayed behind you. I also want to take this opportunity to highlight some of the key activities that we’re going to be doing as part of the network. I think one of the key things that we want to do as part of the network is to really build an independent evidence base to generate community -informed analysis of the societal, ethical, and distributional risks of AI systems across diverse contexts.
We also want to do real -world deployment assessment to conduct contextual and public evaluations of models and applications across diverse social contexts. We also want to push the field of evaluations, push the science of evaluations, where we say that benchmarks are very important, but benchmarks as they stand today do not necessarily capture all the societal risks that we see in the Global South. So how do we ensure that the evaluation work that we’re doing also captures some of those harms? In some sense, what we want to do with the network is field building. We want to bring together Global South civil society organizations to pool in their collective intelligence, to pool in their capacities, and to advocate together for the representation of Global South concerns on global governance forums.
So what we are trying to do here is field building within the Global South around AI safety and around building that trust infrastructure. And eventually what we hope that all of this amounts to is collective advocacy. We see an important role that the network will play in creating a connective tissue between the global governance architecture, between the global safety infrastructure, and what’s happening on the ground. We hope the network can provide that visibility to real -world impact. to technology companies who are designing tools, who are designing safety infrastructure, as well as to governments and international organizations who are building the architecture of global AI governance. So with that, I want to thank you all. Oh, wait, I have one more thing to share with all of you.
I’m not ready to thank you yet. I also want to showcase some of the projects that we’ll be doing in the coming year. Picking up on yesterday’s commitments, one of the things that we’ll be doing is building benchmarks for multilingual AI. This is with our network partners, the Collective Intelligence Project and CARIA, and we’re really excited to start this work. We’re also going to be doing work on gender and safety. This is with our partners at GXD Hub and the Global Center for AI Governance to build a taxonomy of gender harm so that we can start building a more robust incident reporting database when it comes to gender -related harms and really advance gender safety in digital spaces.
The third piece that we’re going to be working on this year is around procurement. All of the evaluation work that we do, all the benchmarks that we build, all of that has to eventually feed into public policy. And so we hope that some of this work can support procurement. And procurement, we think, is a really important lever for countries in the global south to shape markets for responsible innovation. I think we’ve all heard a lot about the kind of third way of AI governance that India brings to the global governance landscape. And procurement can be an important lever of making that third way a reality and setting the bar for what responsible innovation looks like.
Like I mentioned earlier, we also want to push on the science of evaluation. What does good evaluation look like? What are the kind of methodologies that we need? What are the kind of methodologies that reflect the concerns and the capacities of communities in the global south? So we’re very excited to be doing this work with ITS Rio, who’s also one of the founding partners, and specifically to implement and advance this discussion on evaluations. We’ll be looking at labor market impacts in the global south. and finally we’re going to be looking at evaluations of health information systems do the existing generative AI tools and large language models that we see do they deliver for clinicians do they deliver for doctors what more can they do to support the needs of healthcare professionals in the global south.
So those are the five kind of big flagship projects that we’re going to be launching within the coming year. We’re going to be very busy as you can see we have a lot that we’re going to try and get done and we’re really excited to be on this journey with all of you and would love to engage with all of you post the launch and see how we build this civil society and research infrastructure together. So with that I am delighted to welcome our keynote speakers first and I would like to give the floor to Mr. Abhishek Singh. Sir thank you for your continued support Thank you for the network and for your leadership on the India AI Summit.
Over to you, sir.
Thank you, Urvashi. And first and foremost, I’d like to congratulate all the team, the network which has brought this together, this Global South Network for Trustworthy AI. With a few months back when we started discussing this concept with Urvashi, with Kalika, with my team, we felt that how do we go about it? Because safe and trusted AI is something that nobody disagrees with. Everybody says that whenever AI innovation is happening, but we must ensure that we must protect ourselves, we must kind of secure ourselves from the harms that can come from misuse of AI or from the risks that frontier AI poses. So yes, we did have the Yoshua Bengio’s report, the scientific panel report, which is part of all the impact summits, the Action Summit and the Bletchley Park Summit, in which it has kind of… identifies the risks that frontier AI model poses.
But what we do believe is that just identifying the risk is not sufficient. We need to think of how do we address those risks. And for addressing those risks, you need to first have the technical tool, the capacity to identify those risks. What are the benchmarks on which you will evaluate them? Some of which Roshi identified, like how do various models perform on multilingual benchmarks? Because very often, most models are evaluated on benchmarks which are predominantly in English language. But if you look at India, a diverse country, we have 22 official languages and multiple other dialects. How do we evaluate how a model performs on various domains in prompts given in those languages? We don’t have specific linguistic benchmarks.
The same applies to many countries of global south. So it felt that while limited expertise exists in some institutions where research is going on, like Serai is one of them, where Professor Balram Ravindran is leading it. There are many labs, of course. whether it’s Microsoft Research or whether other labs wherein such work is going on. The AI Security Institute in UK is doing some work in this direction. The OECD has been doing some work. But how do we ensure that we enable the access to such resources, such tools, such studies to the larger global majority? So with that, this whole concept of creating a global south network for trustworthy AI came in. And then we immediately had these conversations with all the key stakeholders, partners.
We got a lot of support from almost all stakeholders. And along with that, the conversation for the New Delhi Frontier AI commitments was also going on, which Kalika from my team was leading it. And luckily, we were able to announce it in which all models committed to those two commitments about sharing usage data as also multilingual performance benchmarks. So that was a huge achievement. And I feel that the launch of this. Global South Network for trustworthy AI is a further step in that direction. How do we enable compliance to those commitments? How do we ensure that how this data will be shared? How do we create tools for evaluating models in various languages? How do we build up capacity in all countries of the global south?
How do we share resources? How do we share knowledge across? So this is just the beginning and I feel that we support from all industry organizations, the frontier AI labs, the research organizations, governments across the world. This can really, really grow into a resource that can be a global utility. So I compliment all that team which is involved in doing that. The launch of the network is the first step. But how do we action it out? How do we make it functional? How do we ensure that we get necessary support from all stakeholders? Very often whenever we talk about trusted AI, whenever we talk about safe AI, some people think that we are trying to stifle innovation.
The objective is not that. We always say that while the primary objective is to ensure diffusion of AI, primary objective is to ensure that more and more users benefit from the usage of AI. But at the same time, we need to do that in a responsible manner. We need to do it in a safe manner. We do need to do it in a trustworthy manner to limit the harm that can be caused. So this Global South Network for Trustworthy AI which is being launched will work in that direction. It will be an institution that will support not only India but the entire Global South. And I am sure with just the presence of all the speakers who are present in this session, the strong commitment that all industry and all countries and all multilateral organizations are showing to this initiative, I am sure this will get further strengthened in the days to come.
There is a lot of work that Urvashi and team are taking. They are taking on their own. But we will be there to provide all necessary support for India AI mission and we will work towards ensuring that you get the same level of support from every. participating country which is here. So thank you once again and congratulations for this launch and look forward to working towards the objectives in the near future. Thank you.
MS. Thank you, sir, for your remarks and most importantly for your support. I think it means a lot to us to be working so closely with the India AI mission and we’re really excited to be able to deliver on this promise. It’s now my honor to invite Ambassador Philip Tigo, the Special Envoy on Technology from the Republic of Kenya, to share his reflections.
AMBASSADOR PHILIP TIGO, ASSOCIATE OF TECHNOLOGY, KENYA, Thank you so much for this opportunity to share my reflections. And I noticed that this is really a women -led network, so again, congratulations, Ovashi and Rachel, for putting this together. I think before we celebrate the launch of the network, I think we must acknowledge that we are working with the right people and we are working with the right people. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here.
And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think because you must acknowledge the structural problem around the safety conversations and the infrastructure that has been cutting safety in the last three years. I think the global south has always been excluded from this conversation. I say this from a position of strength because Kenya is the only, Kenya I think, we’re the only member of the international network of AI safety institutes.
And so there’s a challenge there. And so I think that model that is not inclusive to a global majority, that in most cases bears the brunt and the impacts of AI, is not acceptable. And so this network, in my sense, is timely but also late. And so there’s almost an urgency that we need to work very closely in how we scale up what this network does. The second part, of course, is, as I mentioned, that a lot of the global majority countries that are there are not. They are the ones that not just bear the brunt of the models, but bear the adverse societal harms of the models. Kenya is one of the countries that uses one of the models.
and from the use cases we see that they use it for the wrong reasons. Emotional support or companionship, it’s not necessarily for anything meaningful or productivity. And so as the world advances, it therefore behoves us that we work with these frontier model companies to ensure that their models are safe beyond secure, but also are more trustworthy. The second part, of course, is that part of model evaluations assumes access. We now know that a lot of my colleagues who are doing model evaluations are doing it from an external point of view. So we need to be very clear that global majority countries, and by this when I say global majority countries, we also have a new global south in AI, because it’s just not the global majority.
We know in the global north of artificial intelligence is two countries and a few companies. So we must, beyond this, extend to also include other colleagues, whether it’s from Europe, Western Europe, or Latin America. Safety must also go beyond technology. towards socio -technical issues. We look at AI in the countries of Kenya from minds to models and so safety must also include environmental harms, biases, misinformation, disinformation but also harms to water, environment and so we need full lifecycle accountability. It’s good to evaluate the models but also it’s good to evaluate the footprints of the model quickly. There are four structural gaps that we see and this is why I love this network and the network I think one is yes you want global majority folks to evaluate the models but we have great teaming capacity gaps so I hope that this network will look at this.
Secondly I think is also issues of access to compute. We can’t have global majority researchers trying to evaluate models without necessarily having access to compute to do that. Third part of course has been mentioned by I think his left issues around linguistic and cultural mismatch so we need to do that the other part of course is benchmarking. as governance power. Also, benchmarks are not neutral. Sometimes I think I like to be honest because that’s what evaluation needs to do. And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means. Governance is about power. And we must deconcentrate that power even if it’s unintentional.
Finally, I think for me, evaluation is also about agency. And we must have a question of agency, a notion of agency around these models, but also including sovereign capability. As we know, a lot of your countries are trying to build sovereign models, but also sovereign capabilities across the track. What should this network deliver, in my view? And I’ll humbly make these quick suggestions. One, I think, yes, good to have the network, but can we have regional nodes for this? So that, because Africa… I speak for Africa, Africa is another country, it’s 54 countries, expanded to have nodes. Secondly, include multilingual benchmark data sets. Could be an interesting annual red teaming exercise. Could be potentially, why not publish a Global South AI Safety Report with an expansive definition of what safety is.
And I would be remiss if I don’t say how do we fit this into the multilateral process. We already have a global UN scientific panel on AI, and there’s a global dialogue on AI governance. I’m one of the champions for this, so hopefully we will get this in there. Finally, let’s close the accountability loop. How do all this ultimately matter for citizens? We can evaluate all we want, but if they don’t translate
Thank you, Ambassador, for highlighting the urgency of this work and also reframing the safety conversation for the Global South. And just to say we are planning to have regional hubs, and we do. And I think the point about how we engage with the multilateral system is very important, and we will have the Indian AC as part of our steering committee, and we hope we can work with the government of Kenya as well. And, of course, we have Professor Ravindran, who is part of the scientific council, so we will be relying on him as well. But thank you. Thank you for your remarks. And with that, I’d like to call our final keynote speaker for the day, who represents the UN Office for Digital and Emerging Technologies.
I’m pleased to invite Mr. Quenchen Chow Lambert, the Chief of Office and AI Lead, to deliver the next keynote. Thank you for your keynote address.
Mr. Quintin Chou-Lambert:
There is less, perhaps, infrastructure or energy connection to go around. So the concept of AI safety becomes less of a, or it kind of edges into this more contextual field, and that’s where this kind of low perspectives, field -tested examples can be very helpful to surface, which we’re missing. And I’d say the idea of AI standards as technical standards don’t solve that issue because a one -size -fits -all standard will not be contextually sensitive. So moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small -language models to context turns the issue of AI safety into a more fuzzy kind of discussion and one which really needs empirical evidence.
And I think the trends in the institutional discussions from Bletchley Park to Sears, Seoul, where there were also around 30 countries signing the declaration, to Paris, where you had 60 -plus, and now here. over 100 countries engaging. We now have the United Nations Global Dialogue on AI Governance, which will include a whole 193 member states informed by analysis from an independent international scientific panel on AI, which will look at the risks and also opportunities and impacts of AI. And so as the conversation in these summit settings and in the international level has widened and to include more countries and more people and covered more of humanity, the focus has, through the open source developments, been allowed to become much more focused of encompassing other perspectives.
And that’s why, to close and to echo Ambassador Thieger, these kinds of networks play a crucial role in connecting and bringing examples of the challenges that we face. Thank you very much. cases of threats from various sources to local people into discussions so that international discussions do not ignore or omit or discount the perspectives of the vast majority of people on the planet. Thank you.
Thank you, Mr. Chow, for those remarks. I’d now like to call our panelists onto the stage. Ms. Natasha Crampton, Vice President and Chief Responsible AI Officer at Microsoft. Dr. Rachel Tabande, Senior Program Officer AI for Africa at the Gates Foundation. Before you sit, we’re going to take one quick picture. Ms. Chennai Chair. I don’t see you. Oh, there you are. Yes, okay. Director of the Masakane African Language Hub. Mr. Amin Banefatami, Chief Responsible AI Officer. I’m cognizant. And last but certainly not least, Dr. Balaram Ravindran, Head Center of Responsible AI at IIT Madras. Yes, and can we get the keynote speakers as well? Thank you. As with all good things in life, we’re short on time.
But so let’s get started. Rachel, I’m going to start with you. Thank you. where according to you what according to you or where according to you do you feel like we still lack clarity on how safe and reliable AI systems are when they’re deployed in real world context in the global south
thank you so couple of things maybe two three things number one is we need to redefine what is safe and what is harmful in as far as AI models or applications are concerned according to the social cultural context that they are deployed in and that means that having models or applications that are great at understanding the data or the patterns to generate content is not enough if they do not understand the social norms the gender dynamics the religious beliefs the political sensitivities or indeed even the humor the slang or the tone particularly now that voice is being used in the media a key channel for delivery of AI. So we need to redefine safety and harm in the context in which AI models are deployed.
So I think we’re missing that, but hopefully we get there. I think the second piece is around language. It’s not enough for a large language model to have strong translation capabilities. Language in itself is not just about vocabulary. It’s also about the lived meaning, the lived experiences. I come from a beautiful country called Malawi. It’s also called the warm heart of Africa. Now, if you’re deploying a model for pregnant mothers to access advisory messaging there, if the mother says their waters have broken, which clinically is a critical incident that should warrant that mother to be referred to a health facility, but if you translate that from the local language to English, which is where most of these large language models and applications have been benchmarked on, that will literally mean I have thrown away water.
So if the model is not trained to understand that context, then you will miss that flag. And then finally, I wanted to say that we also need to understand the harms that emerge as people use the AI models. Currently, I think much of the benchmarking is done on the content and predefined metrics. So final example, personally, I use my AI companion as my therapist. So it’s the one persona that knows a lot about my personality from all spheres, as a mother, as a career person, my finances, all of that. But at what point can we then be able to track whether I’m substituting my cognizance and cognitive capabilities with that AI model or application, or that I’m becoming overly emotionally dependent?
So I think there are those three areas that we’re missing, and hopefully we can get better at it. Thank you.
Thank you, Rachel, and thank you also for those powerful examples, because I think we’ve been saying some of this at almost a theoretical level, but I think those examples really bring home the gaps in terms of where the current safety conversation is. Chennai, from a civil society perspective, what do you feel companies or developers often miss about the safety implications of deploying AI systems in the global south?
That’s one thing they miss, the user experience. So on a more serious note, thank you, Vashi. So this is great to actually piggyback from what you said, and I was like, are we reading the same notes? So I think what really is missed when people are deploying some of these solutions is around the context in which they’re deploying the tool. And this is particularly looking at an example where on the African continent, there is high levels of gender inequality, a very youthful population with young people often unemployed, and also older people forgotten in actually the development of technologies. So I don’t know who we’re developing for, but sometimes we actually don’t consider that diversity and the inequalities that exist.
So you can find that sometimes when these tools are deployed, they actually further exacerbate a situation of inequality. And I’ll give you one example where perhaps an agricultural tool that has a voice system on it to provide farmers or women information on what to plant may actually have a male -sounding voice. And if in that context there’s high issues of gender -based violence or lack of trust, and the community members were not consulted in the design process, what it actually leads to is just exacerbating an already existing situation. And that is an example. That actually did happen when people were deploying Internet solutions for a community. Then secondly, also thinking about who gets left behind in deploying these solutions.
This is where language, as Rachel was mentioning, comes in. So on the African continent, we have over 2 ,000 languages that have been documented. Masakhane is only working on 50 of those African languages to build up quality data sets. So what you then find is when people are deploying technologies, even if they deploy them in something like Kiswahili, which now has a large number of data sets, people just don’t speak Kiswahili across East Africa. And particularly in Kenya, if you go to Nairobi, the Kiswahili spoken in Nairobi will be Shang. Then you go to, it’s not even Kiswahili, as I’m being corrected. And then if you go to the coast in Mombasa, it will be completely different.
So we have to actually take into account the context and nuance of what is being deployed. And then lastly, the way in which the sector, the technology is actually used, if deployment doesn’t take into account. the whole ecosystem of the end user, it can actually result in misuse. And I want to specifically say that there’s two forms of misuse here. There could be people who unintentionally actually carry out a problematic, harmful act online based on how they’re interacting with the technology. And we know that content, particularly if it’s in their own language, and we know that content moderation for the global majority is not sufficient. Or people are underpaid, as we’ve seen the cases that were coming out about content moderators in Kenya.
Then there’s actually intentional misuse. Now, this is where we find gender disinformation, the use of deepfakes to discredit people, particularly around election period. And now with increased open AI that people can actually just type something and get something back, we are seeing that high level of deployment without thinking about what is the after -end impact. To close it off, because I’m talking about AI as if it’s coming later. A10. when they were deployed. It was great, I can track my missing bag on a flight. They have now been put in women’s bags or children’s bags by people who they do not know and they track them. That’s already an act of surveillance that was, if people had been consulted, it might have been mitigated against.
Yes, I do want to know where my bag is, but I don’t want to be tracked unknowingly.
Thanks, Chennai, for that and also for pointing out, bringing the gender dimension on the table and highlighting the issues around what seems like useful technology, how quickly it can become surveillance technology. I’d like to now bring the industry perspective into this conversation. So, Natasha, maybe I can start with you. As you scale systems globally, what are some of the hardest constraints that you as a company face in ensuring context -sensitive safety?
Well, thanks for that question, Arati, and congratulations to everyone on the establishment of the network. I think it’s a really important step forward. So when I think about Microsoft, I think about sort of Microsoft’s scale, and our mission is really to try and empower every person in every organization in the world to achieve more. And so one of the challenges I think that we face with scaling up our efforts here is how do we take the very deep, careful, thoughtful, community -led evaluation work that animated a project like Samishka, which the CAIA organization, as well as the Collective Intelligence Project and Microsoft Research worked on together, which really developed very context -aware evaluations that were appropriate for the use case.
And how do we take that thoughtful work and really scale it up? Because really we want to do that type of work for thousands of languages and probably millions of different cultural settings. And so I think we really need to think about this system of how we are going to build multilingual and multicultural evaluations that we can really run broadly. I think sometimes we think evaluations, we don’t sort of understand how sustainably they need to be run. As in you can’t just do it once before you release a product. You need to run the evaluations on an ongoing basis to understand how there might have been shifts. And so I really think for us we need to think about this system.
How are we going to build a sustainable, grounded, community -led system of scalable evaluation?
Thanks, Natasha. And I hope in some sense also the network can actually play at least part of that function in building that kind of coherence to the space of evaluation and helping us at least build a shared vocabulary and a shared set of methodologies together as organizations. Amir, what do you think needs to change, whether it’s internally within companies or externally in terms of the ecosystem that we’re operating in, to make such grounded evaluations, the kind that Natasha was talking about, become the standard practice for industry? Should they be the standard practice? And if so, how? How do we get there?
Thank you for that question. And first, congratulations. I’m happy to be also part of this network and support it. I think Natasha mentioned part of the foundational questions. And I think from a, I’m putting my hat off, cognizant chief responsible, we work with a lot of companies and governments into deploying. new scenarios. We call it systems or applications or anything else. The concept of safety, I was mentioned, is diffused. It’s not very clear what we’d call safety. So evaluating the underlying element that needs to be changed or to be addressed is not obvious. When we talk about models, models are not just one thing that you deploy. It goes into an application, there’s a system, infrastructure, there’s network access, there’s API connected data access.
All of them are contextually different. That was mentioned before. And then the problem, one of the problems is that, you didn’t ask me about the problem, but there’s a problem issue is that there’s a lack of imagination. People that are building systems have no awareness about the context in which those situations occur and how they occur and what’s the causes and what’s the likelihood of solution to happen. So absent of that, all this context which language is part of it, culture is part of it, is not captured. So without that, there is very little capability. to address that from a regulation or incentive perspective. Safety, on the other side, is not costed into financial systems and so forth.
There is no penalty of not being safe. So as long as there is no constraint to put safety as a cost structure, which strong mandate, companies will not pay attention or enough attention. So if it’s not part of the financial planning and the processes and so forth, it won’t happen. So there is a disconnect between what we do as enterprises to make sure that systems and platforms are properly built and deployed. There is a disconnect between the system in which they are deployed. At the same time, there is a talent inclusion that is missing. So the inclusion part is that all the talent that is building into those safety conversations are not the talent that are exposed to those issues.
So that absent voice is also a piece that needs to be addressed, not just from a skilling perspective, but also from an integration perspective. And finally, the infrastructure part. The infrastructure is not just systems and models and data, it’s also the tooling and the evaluation. And it was mentioned that evaluation has to be done differently, but if you don’t know what harm or safety means, evaluation’s gotta be different. There is probably an opportunity here to come up with a series of evaluation tools that are not only built for model design, but also built for system deployment. And if we go from pilot to scaling, what issues occur and what examples are happening and what incidents are deployed, and incident reporting is a huge opportunity here because it will capture, nested in reporting, some of the hidden element of the control issues, data access, regulation, absence, or anything else.
Finally, there is a latency issue in global north, and you mentioned probably correctly that there’s a lot of latency issues. There are only a handful of countries in the global north that probably the new slot is much bigger. there are institutional framework, you have basically the rule of law, you have civil society which is very active, you have legal framework that basically creates an accelerated feedback loop into all this incident safety in most of the global south countries these mechanisms don’t exist which delays the feedback loop and basically compounds the possible harm and everything else so there is probably an opportunity to figure out how we can accelerate the learning capabilities and the skills at which we capture knowledge and data to be tied with tools that probably need to be implemented and deployed either on an open source matter or a free access matter and build it with a contextual environment, the talent pool to make it together so the ownership of the global south, all these pieces are important so the network can actually incentivize those different pieces that could complete each other to really play a role into the global south understanding better where safety issues are, where harm can happen and what corrections can be made in the rhythm that needs to happen because rhythms are not exportable and what we do from one country to another is not.
And finally the network could probably help bring it together.
Thank you for laying that out and also just pointing out how all the kind of pieces link to each other and we can’t just kind of go at it at one level alone and to the importance of capacity across all those. Professor Ravindran, AI deployment is accelerating in the global south, in India, in many other countries as well. But at the same time we don’t see or so far we haven’t seen as much investment in the kind of safety and safety infrastructure. Would you agree? You’re actually asking an academic about investment? Sure, of course there’s not enough money. Why not and how do we change it?
so I’m going to answer a different question sure perfect like a true academic I’m sorry I’ll connect it back to what you asked so there are a whole lot of initiatives that are getting announced at the summit and also things that I kind of discovered while having various conversations that there are multiple networks that are getting launched there are already in operation there is a network in Africa looking at capacity building there is a network in China apparently which none of us seem to have heard about that’s being launched on AI safety and capacity building and that is our network that’s getting launched and that is the UN initiative on building this network of capacity building institutes for the global south which we had a meeting in the morning as well about that so there is just too many of these initiatives that are getting launched.
And we have to figure out a way how we would coordinate operations among these initiatives as well. So I think that would be a great multiplier instead of everybody going out and saying, okay, let me see what small piece of the pie that I can get so that I can do these activities. And after that, if there is a lot more coordination. And if you remember our initial conversations about when we wanted to start this thing was about this would be like this one node in the global AC. I can’t even say global network of safety institutes anymore, can I? So they’re not even safety institutes. So AC institutes, whatever ACs, this should be like one node in the AC network which kind of represents unheard voices there because almost except for, as the ambassador was pointing out, except for Kenya.
So we really have, and of course India, I presume. We don’t have safety institutes in the global south, right, that can participate in the dialogue. So I mean, that kind of larger collaboration framework is something that we should enable so that, I mean, even if you say, we go to Gates, and then how many different people, how many different networks would Gates want to spend their money to? If that is one way we can say that there’s this whole operation that’s happening, then that would be a great way of harmonizing our efforts. I can turn it back to the question. Thank you.
No, I mean, I think you raised a really important issue of kind of harmonizing these efforts, and also that how this network can play a really important role in the larger kind of AC network. Luckily, the S remains the same, so we can still go with the acronym, I guess, on the safety network. We’re almost at time, so let’s just do one kind of quick rapid -fire round with all the panelists, and maybe Natasha, I can start with you. What is the one concrete step your institution, Microsoft, could take in the next year to strengthen AI safety in the global south?
Well I’m looking forward to making good on the New Delhi Frontier AI commitments that Microsoft made which is going to help advance multilingual and multicultural evaluation work as well as share data that will help policy makers make or understand AI adoption within their countries and make the sort of choices and policy interventions that help bring that broader access so if I can be sneaky and kind of come as one thing. The second thing I’m really excited about is we’re making large infrastructure investments across the global south to the tune of 50 billion dollars by the end of this decade now that infrastructure as Amir and others on the panel have mentioned is essential to being able to building up this scaled system of sustainable evaluation so I’m looking forward to those investments too.
Thank you.
is that a fire alarm or something?
No, no, no, they’re telling us that we have to wrap up I think.
Okay, great, so wrapping up, so we have to get the work going, rolling, right, so talking about it is one thing, but actually starting to do this collaboration and getting this research efforts going, we’d love to reach out to partners across the globe, in fact, I’m part of the other UN network as well and we have been talking about looking at problems that would necessarily require cross -border collaboration, right, as supposed to, you know, problems that we would anyway solve in our geography, then just working with somebody else to solve it in two geographies, okay but if you can pick problems that will necessarily require people across borders to collaborate, I think that will certainly drive this and also will, you know, kind of put forth the importance of having the network itself, not just information sharing, but actually problem solving that can be done only across the network.
Thank you. Rachel, 30 seconds.
30 seconds I think from the foundation side is to really institutionalize the evaluation of safety of AI solutions right at deployment because what we see now is that safety issues almost emerge post deployment thank you
so from the hub side we actually do have a benchmarking initiative that’s going on this year so this will be one contributing to the African benchmarking work and so that will be our output in contribution
amazing looking forward to that thank you Chennai and Amir last but not least
we’re working already on with our two labs one in Bangalore actually and one in San Francisco on safety evaluations mostly on incident reporting and we already made it culturally contextual so I hope that we are helpful to basically provide open source tools for evaluation to disseminate them and work with that work to basically make them accessible to the public available to all partners.
Thank you.
Dr. Urvashi Aneja
Speech speed
129 words per minute
Speech length
2383 words
Speech time
1102 seconds
Underrepresentation and low institutional capacity create heightened risks
Explanation
Aneja highlights that many Global South contexts suffer from weak institutions and deep inequities, which raises the risk that their safety concerns are overlooked in global AI safety frameworks.
Evidence
“And while the opportunities are immense, in many of these contexts, many of these contexts are also marked by low institutional capacity, deep societal inequities, popularization, and populations with low levels of literacy.” [1]. “And there’s a real risk, therefore, that the concerns and priorities of these countries, of these communities remain underrepresented in the global safety infrastructure.” [2]. “Unfortunately, Global South organizations, Global South communities, Global South states remain underrepresented in global safety and governance infrastructures.” [8].
Major discussion point
Challenges and Gaps in AI Safety for the Global South
Topics
Capacity development | Human rights and the ethical dimensions of the information society | Artificial intelligence | Closing all digital divides
Build an independent evidence base, field‑building, and collective advocacy
Explanation
Aneja proposes that the Global South Network should generate community‑informed evidence on AI risks, engage in field‑building activities, and use the findings for collective advocacy at global forums.
Evidence
“I think one of the key things that we want to do as part of the network is to really build an independent evidence base to generate community -informed analysis of the societal, ethical, and distributional risks of AI systems across diverse contexts.” [58]. “In some sense, what we want to do with the network is field building.” [59]. “And eventually what we hope that all of this amounts to is collective advocacy.” [62].
Major discussion point
Role and Design of the Global South Network for Trustworthy AI
Topics
Monitoring and measurement | Capacity development | Artificial intelligence
Embed safety considerations into public procurement
Explanation
Aneja argues that procurement policies can be leveraged to set safety standards for AI, ensuring responsible innovation and market shaping in the Global South.
Evidence
“And procurement can be an important lever of making that third way a reality and setting the bar for what responsible innovation looks like.” [159]. “And so we hope that some of this work can support procurement.” [160]. “And procurement, we think, is a really important lever for countries in the global south to shape markets for responsible innovation.” [161].
Major discussion point
Governance, Representation, and Power Dynamics
Topics
Governance | Financial mechanisms | Human rights and the ethical dimensions of the information society
Launch gender‑harm taxonomy and incident reporting database
Explanation
Aneja notes that the network will develop a taxonomy of gender‑related harms and a corresponding incident‑reporting database to improve gender safety in digital spaces.
Evidence
“This is with our partners at GXD Hub and the Global Center for AI Governance to build a taxonomy of gender harm so that we can start building a more robust incident reporting database when it comes to gender -related harms and really advance gender safety in digital spaces.” [110].
Major discussion point
Concrete Actions and Commitments
Topics
Human rights and the ethical dimensions of the information society | Artificial intelligence | Data governance
Mr. Abhishek Singh
Speech speed
181 words per minute
Speech length
892 words
Speech time
294 seconds
Existing benchmarks are English‑centric, overlooking multilingual realities
Explanation
Singh points out that most AI model evaluations rely on English‑language benchmarks, leaving a gap for languages spoken in the Global South.
Evidence
“Because very often, most models are evaluated on benchmarks which are predominantly in English language.” [16]. “We don’t have specific linguistic benchmarks.” [17].
Major discussion point
Challenges and Gaps in AI Safety for the Global South
Topics
Artificial intelligence | Closing all digital divides
Enable compliance with New Delhi commitments and share evaluation tools
Explanation
Singh stresses the need to help stakeholders meet the New Delhi Frontier AI commitments by providing tools and data for multilingual performance and usage‑data sharing.
Evidence
“How do we enable compliance to those commitments?” [69]. “And along with that, the conversation for the New Delhi Frontier AI commitments was also going on, which Kalika from my team was leading it.” [71]. “And luckily, we were able to announce it in which all models committed to those two commitments about sharing usage data as also multilingual performance benchmarks.” [113].
Major discussion point
Role and Design of the Global South Network for Trustworthy AI
Topics
Artificial intelligence | Data governance
Create multilingual benchmarks, red‑team exercises, and gender‑harm taxonomies
Explanation
Singh advocates for developing multilingual evaluation suites, conducting red‑team testing, and building taxonomies that capture gender‑related harms.
Evidence
“Some of which Roshi identified, like how do various models perform on multilingual benchmarks?” [22]. “How do we evaluate how a model performs on various domains in prompts given in those languages?” [112].
Major discussion point
Multilingual and Culturally Contextual Evaluation
Topics
Artificial intelligence | Closing all digital divides | Human rights and the ethical dimensions of the information society
New Delhi Frontier AI commitments on multilingual performance and usage‑data sharing
Explanation
Singh reiterates that the commitments require models to disclose multilingual performance metrics and share usage data to foster transparency.
Evidence
“And luckily, we were able to announce it in which all models committed to those two commitments about sharing usage data as also multilingual performance benchmarks.” [113].
Major discussion point
Concrete Actions and Commitments
Topics
Artificial intelligence | Data governance
Ambassador Philip Thigo
Speech speed
196 words per minute
Speech length
1016 words
Speech time
309 seconds
Global South voices are excluded from safety governance; power must be de‑concentrated
Explanation
Thigo emphasizes that structural problems have left the Global South out of safety conversations and calls for de‑concentrating power away from a few institutions.
Evidence
“And I think because you must acknowledge the structural problem around the safety conversations and the infrastructure that has been cutting safety in the last three years.” [12]. “I think the global south has always been excluded from this conversation.” [141]. “And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means.” [42]. “And we must deconcentrate that power even if it’s unintentional.” [134].
Major discussion point
Governance, Representation, and Power Dynamics
Topics
Artificial intelligence | Governance | Human rights and the ethical dimensions of the information society
Benchmark standards should not be defined by a handful of institutions
Explanation
Thigo warns against a concentration of benchmark authority and urges broader participation in defining safety metrics.
Evidence
“And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means.” [42].
Major discussion point
Governance, Representation, and Power Dynamics
Topics
Artificial intelligence | Governance
Establish regional nodes and integrate with multilateral AI governance processes
Explanation
Thigo proposes creating regional hubs for the network and linking its work to existing UN AI panels and global governance dialogues.
Evidence
“One, I think, yes, good to have the network, but can we have regional nodes for this?” [82]. “We already have a global UN scientific panel on AI, and there’s a global dialogue on AI governance.” [83]. “And I would be remiss if I don’t say how do we fit this into the multilateral process.” [91].
Major discussion point
Role and Design of the Global South Network for Trustworthy AI
Topics
Artificial intelligence | Internet governance | Governance
Global South researchers lack access to compute resources for model evaluation
Explanation
Thigo notes that without sufficient compute, researchers in the Global South cannot effectively evaluate AI models, limiting their participation.
Evidence
“We can’t have global majority researchers trying to evaluate models without necessarily having access to compute to do that.” [135].
Major discussion point
Infrastructure, Capacity, and Resource Constraints
Topics
Artificial intelligence | Capacity development
Ms. Chenai Chair
Speech speed
167 words per minute
Speech length
683 words
Speech time
244 seconds
Gender bias and language diversity are ignored, leading to misuse and surveillance
Explanation
Chair stresses that overlooking gender‑based violence and linguistic diversity can exacerbate harms and turn useful technology into surveillance tools.
Evidence
“And if in that context there’s high issues of gender -based violence or lack of trust, and the community members were not consulted in the design process, what it actually leads to is just exacerbating an already existing situation.” [13]. “And this is particularly looking at an example where on the African continent, there is high levels of gender inequality, a very youthful population with young people often unemployed, and also older people forgotten in actually the development of technologies.” [15]. “Thanks, Chennai, for that and also for pointing out, bringing the gender dimension on the table and highlighting the issues around what seems like useful technology, how quickly it can become surveillance technology.” [44]. “Now, this is where we find gender disinformation, the use of deepfakes to discredit people, particularly around election period.” [51].
Major discussion point
Challenges and Gaps in AI Safety for the Global South
Topics
Human rights and the ethical dimensions of the information society | Closing all digital divides | Building confidence and security in the use of ICTs
Africa’s 2,000+ languages require localized datasets beyond dominant lingua‑franca
Explanation
Chair highlights the massive linguistic diversity in Africa and the need for localized data sets, noting current efforts cover only a fraction of languages.
Evidence
“So on the African continent, we have over 2 ,000 languages that have been documented.” [123]. “Masakhane is only working on 50 of those African languages to build up quality data sets.” [124]. “So what you then find is when people are deploying technologies, even if they deploy them in something like Kiswahili, which now has a large number of data sets, people just don’t speak Kiswahili across East Africa.” [126].
Major discussion point
Multilingual and Culturally Contextual Evaluation
Topics
Closing all digital divides | Artificial intelligence
African benchmarking initiative to produce localized evaluation metrics
Explanation
Chair mentions an ongoing African benchmarking project that will generate metrics tailored to local languages and contexts.
Evidence
“so from the hub side we actually do have a benchmarking initiative that’s going on this year so this will be one contributing to the African benchmarking work and so that will be our output in contribution” [177].
Major discussion point
Concrete Actions and Commitments
Topics
Artificial intelligence | Closing all digital divides
Ms. Natasha Crampton
Speech speed
136 words per minute
Speech length
404 words
Speech time
177 seconds
Sustainable, ongoing evaluation infrastructure is needed, not one‑off tests
Explanation
Crampton argues that evaluations must be continuous and supported by large‑scale infrastructure investments to track model behavior over time.
Evidence
“As in you can’t just do it once before you release a product.” [106]. “You need to run the evaluations on an ongoing basis to understand how there might have been shifts.” [145]. “The second thing I’m really excited about is we’re making large infrastructure investments across the global south to the tune of 50 billion dollars by the end of this decade now that infrastructure as Amir and others on the panel have mentioned is essential to being able to building up this scaled system of sustainable evaluation so I’m looking forward to those investments too.” [146].
Major discussion point
Infrastructure, Capacity, and Resource Constraints
Topics
Monitoring and measurement | Capacity development | Financial mechanisms
Scale community‑led multilingual evaluation sustainably across thousands of languages
Explanation
Crampton emphasizes the ambition to create community‑driven evaluation pipelines that can handle thousands of languages and cultural settings.
Evidence
“Because really we want to do that type of work for thousands of languages and probably millions of different cultural settings.” [120]. “I think sometimes we think evaluations, we don’t sort of understand how sustainably they need to be run.” [131]. “I think sometimes we think evaluations, we don’t sort of understand how sustainably they need to be run.” [132].
Major discussion point
Multilingual and Culturally Contextual Evaluation
Topics
Closing all digital divides | Capacity development | Artificial intelligence
Microsoft’s $50 billion infrastructure investment aims to support scalable evaluation
Explanation
Crampton notes Microsoft’s commitment of $50 billion to build infrastructure that will enable large‑scale, sustainable AI evaluation in the Global South.
Evidence
“The second thing I’m really excited about is we’re making large infrastructure investments across the global south to the tune of 50 billion dollars by the end of this decade now that infrastructure as Amir and others on the panel have mentioned is essential to being able to building up this scaled system of sustainable evaluation so I’m looking forward to those investments too.” [146].
Major discussion point
Concrete Actions and Commitments
Topics
Financial mechanisms | Enabling environment for digital development | Artificial intelligence
Mr. Amir Banifatemi
Speech speed
169 words per minute
Speech length
844 words
Speech time
299 seconds
Lack of a clear safety definition and financial penalties hampers responsible deployment
Explanation
Banifatemi points out that safety is a diffused concept, not costed into financial planning, and there are no penalties for unsafe AI, reducing incentives for responsible behavior.
Evidence
“The concept of safety, I was mentioned, is diffused.” [9]. “So as long as there is no constraint to put safety as a cost structure, which strong mandate, companies will not pay attention or enough attention.” [55]. “Safety, on the other side, is not costed into financial systems and so forth.” [56]. “There is no penalty of not being safe.” [150].
Major discussion point
Challenges and Gaps in AI Safety for the Global South
Topics
Artificial intelligence | Financial mechanisms | Building confidence and security in the use of ICTs
Open‑source, culturally contextual evaluation tools and incident reporting pipelines
Explanation
Banifatemi describes ongoing work in Bangalore and San Francisco to produce culturally contextual safety evaluations and to release open‑source tools for broader public use.
Evidence
“we’re working already on with our two labs one in Bangalore actually and one in San Francisco on safety evaluations mostly on incident reporting and we already made it culturally contextual so I hope that we are helpful to basically provide open source tools for evaluation to disseminate them and work with that work to basically make them accessible to the public available to all partners.” [72].
Major discussion point
Concrete Actions and Commitments
Topics
Artificial intelligence | Data governance | Open source
Develop incident‑reporting mechanisms to create feedback loops for regulation
Explanation
Banifatemi stresses that without a clear safety definition, evaluation must change, and incident reporting can capture hidden harms to inform regulation.
Evidence
“And it was mentioned that evaluation has to be done differently, but if you don’t know what harm or safety means, evaluation’s gotta be different.” [40]. “And if we go from pilot to scaling, what issues occur and what examples are happening and what incidents are deployed, and incident reporting is a huge opportunity here because it will capture, nested in reporting, some of the hidden element of the control issues, data access, regulation, absence, or anything else.” [163].
Major discussion point
Governance, Representation, and Power Dynamics
Topics
Monitoring and measurement | Building confidence and security in the use of ICTs | Artificial intelligence
Dr. Rachel Sibande
Speech speed
144 words per minute
Speech length
454 words
Speech time
188 seconds
Safety definitions must reflect local cultural, gender, and religious norms
Explanation
Sibande argues that AI safety must be redefined to incorporate societal norms, gender dynamics, religious beliefs, and other contextual factors.
Evidence
“thank you so couple of things maybe two three things number one is we need to redefine what is safe and what is harmful in as far as AI models or applications are concerned according to the social cultural context that they are deployed in and that means that having models or applications that are great at understanding the data or the patterns to generate content is not enough if they do not understand the social norms the gender dynamics the religious beliefs the political sensitivities or indeed even the humor the slang or the tone particularly now that voice is being used in the media a key channel for delivery of AI.” [32]. “So we need to redefine safety and harm in the context in which AI models are deployed.” [34].
Major discussion point
Challenges and Gaps in AI Safety for the Global South
Topics
Human rights and the ethical dimensions of the information society | Artificial intelligence
Language nuance goes beyond translation; contextual meaning is critical
Explanation
Sibande notes that language involves lived meaning and cultural context, which simple translation cannot capture.
Evidence
“Language in itself is not just about vocabulary.” [23]. “It’s not enough for a large language model to have strong translation capabilities.” [27].
Major discussion point
Multilingual and Culturally Contextual Evaluation
Topics
Closing all digital divides | Artificial intelligence
Dr. Balaraman Ravindran
Speech speed
172 words per minute
Speech length
565 words
Speech time
196 seconds
Coordinate among overlapping initiatives to avoid duplication and amplify impact
Explanation
Ravindran calls for mechanisms to coordinate the many emerging AI safety and capacity‑building initiatives across the Global South.
Evidence
“And we have to figure out a way how we would coordinate operations among these initiatives as well.” [103]. “And after that, if there is a lot more coordination.” [104].
Major discussion point
Role and Design of the Global South Network for Trustworthy AI
Topics
Capacity development | Artificial intelligence
Absence of dedicated safety institutes in the Global South limits capacity building
Explanation
Ravindran points out that many Global South countries lack safety or oversight institutes, hindering participation in global AI safety dialogues.
Evidence
“We don’t have safety institutes in the global south, right, that can participate in the dialogue.” [137]. “And many countries in the Global South are actually unlikely to even have in the near term their own safety or oversight institutes.” [138].
Major discussion point
Infrastructure, Capacity, and Resource Constraints
Topics
Capacity development | Artificial intelligence
Mr. Quintin Chou‑Lambert
Speech speed
Default speed
Speech length
Default length
Speech time
Default duration
Connect grassroots empirical findings to global governance forums
Explanation
Chou‑Lambert stresses that moving from large, generic models to context‑specific, smaller language models requires empirical evidence that can inform global governance.
Evidence
“So moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small -language models to context turns the issue of AI safety into a more fuzzy kind of discussion and one which really needs empirical evidence.” [122].
Major discussion point
Role and Design of the Global South Network for Trustworthy AI
Topics
Artificial intelligence | Monitoring and measurement
Mr. Quintin Chou-Lambert
Speech speed
135 words per minute
Speech length
324 words
Speech time
143 seconds
Networks bridge local AI safety challenges to global governance
Explanation
Chou‑Lambert stresses that multi‑stakeholder networks are essential for surfacing grassroots threats and examples so that international AI safety discussions do not overlook the perspectives of the majority of people worldwide. By aggregating experiences from over a hundred countries, these networks expand the conversation beyond a narrow set of institutions.
Evidence
“And that’s why, to close and to echo Ambassador Thieger, these kinds of networks play a crucial role in connecting and bringing examples of the challenges that we face.” [3]. “cases of threats from various sources to local people into discussions so that international discussions do not ignore or omit or discount the perspectives of the vast majority of people on the planet.” [4]. “over 100 countries engaging.” [5]. “And I think the trends in the institutional discussions … where you had 60‑plus, and now here.” [6]. “And so as the conversation in these summit settings and in the international level has widened and to include more countries and more people … the focus has … become much more focused of encompassing other perspectives.” [8].
Major discussion point
Role and Design of the Global South Network for Trustworthy AI
Topics
Artificial intelligence | Governance | Capacity development
Context‑specific AI safety demands empirical, field‑tested evidence
Explanation
Moving from monolithic, expensive models to smaller, locally‑tailored language models makes AI safety a more ambiguous issue. Chou‑Lambert argues that this shift requires concrete, empirical evidence from on‑the‑ground deployments to inform safety assessments.
Evidence
“So moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small‑language models to context turns the issue of AI safety into a more fuzzy kind of discussion and one which really needs empirical evidence.” [9]. “So the concept of AI safety becomes less of a, or it kind of edges into this more contextual field, and that’s where this kind of low perspectives, field‑tested examples can be very helpful to surface, which we’re missing.” [10].
Major discussion point
Challenges and Gaps in AI Safety for the Global South
Topics
Artificial intelligence | Monitoring and measurement | Capacity development
One‑size‑fits‑all technical standards are inadequate for diverse contexts
Explanation
Chou‑Lambert cautions that universal AI technical standards cannot capture the nuanced safety needs of varied cultural and infrastructural settings, calling for standards that are adaptable to local realities.
Evidence
“And I’d say the idea of AI standards as technical standards don’t solve that issue because a one‑size‑fits‑all standard will not be contextually sensitive.” [12].
Major discussion point
Governance, Representation, and Power Dynamics
Topics
Artificial intelligence | Governance | Human rights and the ethical dimensions of the information society
UN Global Dialogue on AI Governance provides an inclusive, evidence‑based platform
Explanation
He highlights the establishment of a United Nations‑backed Global Dialogue on AI Governance that brings together all 193 member states, drawing on analysis from an independent scientific panel to assess AI risks and opportunities, thereby offering a venue for incorporating Global South evidence.
Evidence
“We now have the United Nations Global Dialogue on AI Governance, which will include a whole 193 member states informed by analysis from an independent international scientific panel on AI, which will look at the risks and also opportunities and impacts of AI.” [11].
Major discussion point
Governance, Representation, and Power Dynamics
Topics
Artificial intelligence | International cooperation | Governance
Infrastructure and energy constraints shape AI safety priorities in the Global South
Explanation
Chou‑Lambert points out that limited infrastructure and energy connectivity in many regions restrict the deployment of large, compute‑intensive models, making safety considerations inherently tied to local resource realities.
Evidence
“There is less, perhaps, infrastructure or energy connection to go around.” [7].
Major discussion point
Infrastructure, Capacity, and Resource Constraints
Topics
Capacity development | Environmental impacts | Artificial intelligence
Agreements
Agreement points
Global South underrepresentation in AI safety governance requires urgent action
Speakers
– Dr. Urvashi Aneja
– Ambassador Philip Thigo
Arguments
Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums
Network is timely but also overdue given exclusion of global majority from safety conversations
Summary
Both speakers emphasize that Global South countries and communities have been systematically excluded from AI safety governance structures, with Ambassador Thigo noting Kenya is the only Global South member of international AI safety institutes, while Dr. Aneja highlights that many Global South countries lack their own safety institutes
Topics
Artificial intelligence | Capacity development
AI safety must be contextually defined and culturally sensitive
Speakers
– Dr. Rachel Sibande
– Ms. Chenai Chair
– Ambassador Philip Thigo
Arguments
Safety and harm must be redefined according to social, cultural, and linguistic contexts where AI is deployed
Developers miss user experience and context, often exacerbating existing inequalities like gender-based violence
AI safety extends beyond technology to include environmental harms and full lifecycle accountability
Summary
All three speakers agree that current AI safety definitions are inadequate for Global South contexts and must account for local social norms, cultural nuances, gender dynamics, and broader socio-technical impacts including environmental considerations
Topics
Artificial intelligence | Human rights and the ethical dimensions of the information society | Closing all digital divides
Need for multilingual and multicultural evaluation systems
Speakers
– Mr. Abhishek Singh
– Ms. Natasha Crampton
– Ms. Chenai Chair
Arguments
Initiative aligns with New Delhi Frontier AI commitments and supports compliance with multilingual benchmarks
Challenge lies in scaling thoughtful, community-led evaluation work across thousands of languages and millions of cultural settings
Masakhane Hub will contribute African benchmarking work for 50 African languages
Summary
All speakers recognize the critical need for AI evaluation systems that work across diverse languages and cultures, with specific commitments to multilingual benchmarking and acknowledgment of the scale challenge involved
Topics
Artificial intelligence | Closing all digital divides
Capacity building and institutional strengthening are essential
Speakers
– Ambassador Philip Thigo
– Mr. Amir Banifatemi
– Dr. Balaraman Ravindran
Arguments
Global South countries lack capacity building, access to compute, and face linguistic/cultural mismatches in benchmarking
Absence of institutional frameworks in Global South delays feedback loops and compounds potential harms
Multiple networks are launching simultaneously requiring coordination to avoid fragmentation
Summary
All speakers identify capacity building as a fundamental challenge, noting gaps in technical capacity, institutional frameworks, and the need for coordinated efforts to build sustainable capabilities across the Global South
Topics
Capacity development | Artificial intelligence
Evaluation must be continuous and proactive rather than reactive
Speakers
– Ms. Natasha Crampton
– Dr. Rachel Sibande
Arguments
Need for sustainable evaluation systems that run continuously, not just once before product release
Foundation will institutionalize safety evaluation right at deployment rather than post-deployment
Summary
Both speakers emphasize moving from one-time or post-deployment safety assessments to continuous, proactive evaluation systems that monitor AI systems throughout their lifecycle
Topics
Artificial intelligence | Monitoring and measurement
Similar viewpoints
Both speakers see the network as a crucial bridge between local, contextual AI deployment experiences and global governance structures, ensuring that international AI governance reflects the perspectives and needs of the global majority
Speakers
– Dr. Urvashi Aneja
– Mr. Quintin Chou-Lambert
Arguments
Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums
Network should connect empirical evidence from field testing to international governance discussions
Topics
Artificial intelligence | Follow-up and review
Both speakers emphasize that AI safety and innovation are complementary rather than competing objectives, with concrete commitments to responsible AI development and significant infrastructure investments
Speakers
– Mr. Abhishek Singh
– Ms. Natasha Crampton
Arguments
Innovation focus should not stifle responsible AI development but ensure safe and trustworthy deployment
Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South
Topics
Artificial intelligence | The enabling environment for digital development | Financial mechanisms
Both speakers highlight structural power imbalances in AI governance, with Ambassador Thigo focusing on institutional power concentration and Mr. Banifatemi on economic incentive structures that fail to prioritize safety
Speakers
– Ambassador Philip Thigo
– Mr. Amir Banifatemi
Arguments
Benchmarks are not neutral and power concentration in few institutions must be addressed
Safety lacks clear definition and is not integrated into financial planning or cost structures of companies
Topics
Artificial intelligence | Human rights and the ethical dimensions of the information society
Unexpected consensus
Economic incentives for AI safety
Speakers
– Mr. Amir Banifatemi
– Ms. Natasha Crampton
Arguments
Safety lacks clear definition and is not integrated into financial planning or cost structures of companies
Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South
Explanation
Unexpected alignment between a consultant’s critique of corporate safety incentives and a corporate representative’s substantial financial commitments, suggesting recognition across industry that current economic models inadequately incentivize safety
Topics
Artificial intelligence | Financial mechanisms | The enabling environment for digital development
Need for coordination among multiple AI initiatives
Speakers
– Dr. Balaraman Ravindran
– Dr. Urvashi Aneja
Arguments
Multiple networks are launching simultaneously requiring coordination to avoid fragmentation
Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums
Explanation
Academic and network founder both recognize the proliferation of AI safety initiatives could lead to fragmentation rather than strengthened capacity, showing pragmatic consensus on the need for strategic coordination
Topics
Artificial intelligence | Capacity development
Overall assessment
Summary
Strong consensus emerged around five key areas: Global South underrepresentation in AI governance, need for contextually-sensitive safety definitions, importance of multilingual evaluation systems, capacity building requirements, and shift toward proactive evaluation approaches. Speakers consistently emphasized that current AI safety frameworks inadequately serve Global South contexts and populations.
Consensus level
High level of consensus with remarkable alignment across government, industry, civil society, and academic perspectives. This suggests the Global South Network for Trustworthy AI addresses widely recognized gaps in current AI governance structures. The consensus implies strong potential for collaborative action and suggests the network fills a critical institutional void in global AI governance.
Differences
Different viewpoints
Timeline and urgency of network establishment
Speakers
– Ambassador Philip Thigo
– Dr. Urvashi Aneja
Arguments
Network is timely but also overdue given exclusion of global majority from safety conversations
Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums
Summary
Ambassador Thigo emphasizes that while the network is timely, it is also ‘late’ and there’s an urgency to scale up quickly due to structural exclusion, while Dr. Aneja presents it as a timely response to current gaps without the same sense of overdue urgency
Topics
Artificial intelligence | Capacity development
Scope and definition of AI safety
Speakers
– Ambassador Philip Thigo
– Mr. Amir Banifatemi
– Dr. Rachel Sibande
Arguments
AI safety extends beyond technology to include environmental harms and full lifecycle accountability
Safety lacks clear definition and is not integrated into financial planning or cost structures of companies
Safety and harm must be redefined according to social, cultural, and linguistic contexts where AI is deployed
Summary
The speakers disagree on what constitutes AI safety – Ambassador Thigo advocates for a comprehensive approach including environmental impacts, Mr. Banifatemi focuses on the lack of clear definition and financial integration, while Dr. Sibande emphasizes cultural and contextual redefinition
Topics
Artificial intelligence | Environmental impacts | Human rights and the ethical dimensions of the information society
Approach to scaling evaluation systems
Speakers
– Ms. Natasha Crampton
– Dr. Balaraman Ravindran
Arguments
Challenge lies in scaling thoughtful, community-led evaluation work across thousands of languages and millions of cultural settings
Multiple networks are launching simultaneously requiring coordination to avoid fragmentation
Summary
Ms. Crampton focuses on the technical challenge of scaling individual evaluation systems, while Dr. Ravindran emphasizes the need for coordination among multiple competing networks to avoid fragmentation and resource competition
Topics
Artificial intelligence | Capacity development
Unexpected differences
Network coordination vs individual network development
Speakers
– Dr. Balaraman Ravindran
– Dr. Urvashi Aneja
Arguments
Multiple networks are launching simultaneously requiring coordination to avoid fragmentation
Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums
Explanation
Unexpectedly, Dr. Ravindran, who is part of the founding network, raises concerns about too many similar networks launching simultaneously and the need for coordination, which somewhat contradicts the celebratory launch tone of the event
Topics
Artificial intelligence | Capacity development
Corporate responsibility approach
Speakers
– Ms. Natasha Crampton
– Mr. Amir Banifatemi
Arguments
Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South
Safety lacks clear definition and is not integrated into financial planning or cost structures of companies
Explanation
While both represent industry perspectives, Crampton emphasizes Microsoft’s commitments and investments, while Banifatemi argues that companies fundamentally won’t prioritize safety without financial penalties, creating an unexpected tension between corporate commitment and systemic critique
Topics
Artificial intelligence | The enabling environment for digital development | Financial mechanisms
Overall assessment
Summary
The main areas of disagreement center around the scope and definition of AI safety, the urgency and timing of network establishment, approaches to scaling evaluation systems, and the balance between individual network development versus coordination among multiple initiatives
Disagreement level
Moderate disagreement with significant implications – while all speakers support the network’s goals, their different emphases on environmental impacts, power dynamics, technical scaling, and institutional coordination could lead to different priorities and resource allocation decisions that may fragment efforts or create competing approaches to Global South AI safety
Partial agreements
Partial agreements
All speakers agree that current evaluation approaches are insufficient and need to be more systematic and continuous, but they disagree on the primary solution – Crampton focuses on sustainable technical systems, Banifatemi on institutional frameworks, and Sibande on timing of deployment
Speakers
– Ms. Natasha Crampton
– Mr. Amir Banifatemi
– Dr. Rachel Sibande
Arguments
Need for sustainable evaluation systems that run continuously, not just once before product release
Absence of institutional frameworks in Global South delays feedback loops and compounds potential harms
Foundation will institutionalize safety evaluation right at deployment rather than post-deployment
Topics
Artificial intelligence | Capacity development | Monitoring and measurement
All agree that current AI systems fail to account for Global South contexts and can cause harm, but they emphasize different aspects – Thigo focuses on power dynamics in benchmarking, Chair on deployment context and inequality, and Sibande on linguistic and cultural understanding
Speakers
– Ambassador Philip Thigo
– Ms. Chenai Chair
– Dr. Rachel Sibande
Arguments
Benchmarks are not neutral and power concentration in few institutions must be addressed
Developers miss user experience and context, often exacerbating existing inequalities like gender-based violence
Language models need understanding of lived experiences, not just translation capabilities
Topics
Artificial intelligence | Human rights and the ethical dimensions of the information society | Closing all digital divides
Similar viewpoints
Both speakers see the network as a crucial bridge between local, contextual AI deployment experiences and global governance structures, ensuring that international AI governance reflects the perspectives and needs of the global majority
Speakers
– Dr. Urvashi Aneja
– Mr. Quintin Chou-Lambert
Arguments
Network addresses underrepresentation of Global South in AI safety infrastructure and governance forums
Network should connect empirical evidence from field testing to international governance discussions
Topics
Artificial intelligence | Follow-up and review
Both speakers emphasize that AI safety and innovation are complementary rather than competing objectives, with concrete commitments to responsible AI development and significant infrastructure investments
Speakers
– Mr. Abhishek Singh
– Ms. Natasha Crampton
Arguments
Innovation focus should not stifle responsible AI development but ensure safe and trustworthy deployment
Microsoft commits to New Delhi Frontier AI commitments and $50 billion infrastructure investment in Global South
Topics
Artificial intelligence | The enabling environment for digital development | Financial mechanisms
Both speakers highlight structural power imbalances in AI governance, with Ambassador Thigo focusing on institutional power concentration and Mr. Banifatemi on economic incentive structures that fail to prioritize safety
Speakers
– Ambassador Philip Thigo
– Mr. Amir Banifatemi
Arguments
Benchmarks are not neutral and power concentration in few institutions must be addressed
Safety lacks clear definition and is not integrated into financial planning or cost structures of companies
Topics
Artificial intelligence | Human rights and the ethical dimensions of the information society
Takeaways
Key takeaways
The Global South Network for Trustworthy AI was successfully launched to address the critical underrepresentation of Global South perspectives in global AI safety infrastructure and governance forums
AI safety must be redefined contextually for Global South deployment, considering social, cultural, linguistic, and infrastructural differences rather than applying one-size-fits-all standards
Current AI evaluation methods are inadequate for Global South contexts, missing critical aspects like lived experiences in language, cultural nuances, gender dynamics, and post-deployment harm tracking
Industry faces significant challenges in scaling thoughtful, community-led evaluation work across thousands of languages and millions of cultural settings while maintaining sustainability
Safety is not adequately integrated into companies’ financial planning and cost structures, creating insufficient incentives for comprehensive safety measures
Multiple similar networks are launching simultaneously, requiring coordination to avoid fragmentation and maximize impact through harmonized efforts
The network will serve as crucial connective tissue between global governance architecture and real-world deployment contexts in the Global South
Resolutions and action items
Network will launch five flagship projects in the coming year: multilingual AI benchmarks, gender safety taxonomy, procurement guidelines, evaluation methodologies, and health information systems evaluation
Microsoft committed to fulfilling New Delhi Frontier AI commitments on multilingual benchmarks and usage data sharing, plus $50 billion infrastructure investment in Global South by end of decade
Gates Foundation will institutionalize safety evaluation right at AI deployment rather than waiting for post-deployment issues
Masakhane Hub will contribute to African benchmarking work covering 50 African languages
Cognizant will provide open source safety evaluation tools with cultural context through their Bangalore and San Francisco labs
Network will establish regional nodes/hubs to better serve diverse Global South contexts
India AI mission committed to provide ongoing support for network operations and objectives
Network will work to integrate findings into multilateral processes including UN Global Dialogue on AI Governance and scientific panel
Focus on cross-border collaborative problem-solving that requires international cooperation rather than parallel work in separate geographies
Unresolved issues
How to effectively coordinate with multiple other AI safety networks launching simultaneously to avoid duplication and fragmentation
Lack of clear, universally accepted definition of what constitutes ‘safety’ in AI systems across different contexts
How to create sustainable funding mechanisms for ongoing evaluation work rather than one-time assessments
Absence of institutional frameworks and rule of law in many Global South countries that delays feedback loops and compounds potential harms
How to address the fundamental talent inclusion problem where those building safety systems lack exposure to Global South deployment contexts
How to create financial incentives and cost structures that make safety a priority for companies deploying AI in Global South
How to scale community-led, contextual evaluation work to cover thousands of languages and millions of cultural settings
How to close the accountability loop so that evaluation work translates into meaningful protection for citizens
How to address the ‘new global south in AI’ that includes countries beyond traditional Global South due to AI concentration in just two countries and few companies
Suggested compromises
Expanding the definition of ‘Global South’ in AI context to include other regions like parts of Europe and Latin America that face similar exclusion from AI governance
Balancing innovation promotion with safety requirements – ensuring safety measures don’t stifle AI development while protecting populations from harm
Creating hybrid evaluation approaches that combine technical benchmarks with real-world, contextual assessments
Developing both model-level and system-level evaluation tools to address the full deployment context rather than just underlying AI models
Establishing open source and free access tools for safety evaluation while building local capacity and ownership in Global South countries
Creating incident reporting systems that can capture both intentional and unintentional misuse of AI technologies
Developing procurement guidelines as a policy lever that Global South countries can use to shape markets for responsible innovation
Thought provoking comments
We now know in the global north of artificial intelligence is two countries and a few companies. So we must, beyond this, extend to also include other colleagues, whether it’s from Europe, Western Europe, or Latin America.
Speaker
Ambassador Philip Thigo
Reason
This comment reframes the entire global AI landscape by suggesting that even the ‘global north’ in AI is extremely concentrated, essentially challenging the traditional north-south binary. It introduces the concept of a ‘new global south in AI’ that includes traditionally developed countries who are also excluded from AI development.
Impact
This fundamentally shifted the discussion from a simple global south vs. global north framework to a more nuanced understanding of AI power concentration. It broadened the scope of who should be included in the network and influenced subsequent speakers to think beyond traditional geographical boundaries.
Language in itself is not just about vocabulary. It’s also about the lived meaning, the lived experiences… if the mother says their waters have broken, which clinically is a critical incident that should warrant that mother to be referred to a health facility, but if you translate that from the local language to English… that will literally mean I have thrown away water.
Speaker
Dr. Rachel Sibande
Reason
This powerful example moves the discussion from abstract concepts of multilingual AI to concrete life-or-death scenarios. It demonstrates how current translation-based approaches to multilingual AI can fail catastrophically in critical contexts, revealing the inadequacy of surface-level language support.
Impact
This comment grounded the entire safety discussion in tangible, high-stakes examples. It influenced subsequent speakers to focus more on contextual understanding rather than just technical capabilities, and reinforced the urgency of the network’s mission with a compelling real-world scenario.
Safety, on the other side, is not costed into financial systems… There is no penalty of not being safe. So as long as there is no constraint to put safety as a cost structure, which strong mandate, companies will not pay attention or enough attention.
Speaker
Mr. Amir Banifatemi
Reason
This comment cuts to the heart of why AI safety remains inadequate by identifying the fundamental economic incentive problem. It shifts the discussion from technical solutions to systemic economic and regulatory issues that drive corporate behavior.
Impact
This observation reframed the conversation from focusing solely on technical evaluation methods to addressing the underlying economic structures that perpetuate unsafe AI deployment. It influenced the discussion toward considering regulatory and financial mechanisms as essential components of any safety framework.
There is just too many of these initiatives that are getting launched… we have to figure out a way how we would coordinate operations among these initiatives as well. So I think that would be a great multiplier instead of everybody going out and saying, okay, let me see what small piece of the pie that I can get.
Speaker
Dr. Balaraman Ravindran
Reason
This meta-observation about the proliferation of AI safety networks introduces a critical coordination challenge that could undermine the effectiveness of all initiatives. It’s insightful because it addresses the risk of fragmentation in a field that requires collective action.
Impact
This comment introduced a sobering reality check into the celebratory launch atmosphere, forcing participants to consider how their network would differentiate itself and coordinate with others. It shifted the discussion toward practical implementation challenges and the need for strategic positioning within a crowded landscape.
Benchmarks are not neutral. Sometimes I think I like to be honest because that’s what evaluation needs to do. And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means. Governance is about power.
Speaker
Ambassador Philip Thigo
Reason
This comment exposes the political nature of AI evaluation by highlighting how seemingly technical benchmarks embed power structures and value judgments. It challenges the notion that evaluation can be objective and reveals how current systems perpetuate exclusion.
Impact
This observation elevated the discussion from technical evaluation methods to questions of power, representation, and democratic participation in AI governance. It reinforced the political urgency of the network’s mission and influenced other speakers to consider the governance implications of their technical work.
Overall assessment
These key comments fundamentally transformed what could have been a routine network launch into a sophisticated analysis of AI governance challenges. Ambassador Thigo’s interventions consistently elevated the discussion by introducing power dynamics and structural critiques, while Dr. Sibande’s concrete examples grounded abstract concepts in life-or-death realities. Mr. Banifatemi’s economic analysis and Dr. Ravindran’s coordination concerns added practical urgency to the theoretical framework. Together, these comments created a multi-layered conversation that addressed technical, economic, political, and practical dimensions of AI safety in the Global South, establishing a robust intellectual foundation for the network’s future work.
Follow-up questions
How do we ensure that we get necessary support from all stakeholders to make the Global South Network for Trustworthy AI functional?
Speaker
Mr. Abhishek Singh
Explanation
This addresses the critical challenge of securing ongoing commitment and resources from industry, governments, and multilateral organizations to operationalize the network beyond its launch
How do we create tools for evaluating models in various languages and build up capacity in all countries of the global south?
Speaker
Mr. Abhishek Singh
Explanation
This highlights the need for practical evaluation tools and capacity building mechanisms that can work across diverse linguistic and cultural contexts in the Global South
How do we ensure compliance to the New Delhi Frontier AI commitments regarding sharing usage data and multilingual performance benchmarks?
Speaker
Mr. Abhishek Singh
Explanation
This addresses the implementation gap between policy commitments and actual practice in AI safety and evaluation
How do we fit this network into the multilateral process, including the UN scientific panel on AI and global dialogue on AI governance?
Speaker
Ambassador Philip Thigo
Explanation
This is crucial for ensuring the network’s work influences global AI governance structures and doesn’t operate in isolation
How do we close the accountability loop so that evaluation work ultimately matters for citizens?
Speaker
Ambassador Philip Thigo
Explanation
This addresses the fundamental question of translating technical evaluation work into tangible benefits and protections for people in the Global South
How do we take thoughtful, community-led evaluation work and scale it up for thousands of languages and millions of different cultural settings?
Speaker
Ms. Natasha Crampton
Explanation
This highlights the scalability challenge of conducting context-sensitive AI safety evaluations across the vast diversity of the Global South
How do we build a sustainable, grounded, community-led system of scalable evaluation that can be run on an ongoing basis?
Speaker
Ms. Natasha Crampton
Explanation
This addresses the need for continuous rather than one-time evaluation systems that can adapt to changing AI systems and contexts
How do we coordinate operations among multiple AI safety and capacity building networks being launched globally?
Speaker
Dr. Balaraman Ravindran
Explanation
This addresses the risk of fragmentation and duplication of efforts across various AI safety initiatives in the Global South
How can we pick problems that will necessarily require cross-border collaboration rather than just parallel work in different geographies?
Speaker
Dr. Balaraman Ravindran
Explanation
This focuses on identifying research challenges that can only be solved through genuine international collaboration, strengthening the network’s value proposition
How do we accelerate learning capabilities and feedback loops in Global South countries that lack institutional frameworks for AI safety?
Speaker
Mr. Amir Banifatemi
Explanation
This addresses the structural disadvantage many Global South countries face in developing rapid response mechanisms for AI safety issues
How do we institutionalize the evaluation of safety of AI solutions right at deployment rather than post-deployment?
Speaker
Dr. Rachel Sibande
Explanation
This addresses the timing gap in current safety practices where issues are only identified after systems are already causing harm
At what point can we track whether users are substituting their cognitive capabilities with AI models or becoming overly emotionally dependent?
Speaker
Dr. Rachel Sibande
Explanation
This raises important questions about measuring psychological and cognitive impacts of AI use that current benchmarking doesn’t capture
How do we ensure that only a handful of institutions don’t define what risks are measured, what harms are prioritized, and what safe performance means?
Speaker
Ambassador Philip Thigo
Explanation
This addresses power concentration in AI safety standard-setting and the need for more democratic and inclusive governance structures
Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.
Related event

