Towards a Safer South Launching the Global South AI Safety Research Network

20 Feb 2026 17:00h - 18:00h

Towards a Safer South Launching the Global South AI Safety Research Network

Session at a glanceSummary, keypoints, and speakers overview

Summary

The event marked the launch of the Global South Network for Trustworthy AI, introduced by Dr. Urvashi Aneja at the India AI Impact Summit to address AI deployment challenges in the Global South [8-9]. She highlighted that AI is rapidly being used in critical sectors across the Global South but that low institutional capacity and deep inequities create significant risks, and that the region is under-represented in global safety and governance structures [11-18]. Independent civil-society organisations were presented as uniquely positioned to provide grounded evidence from real-world deployments that can inform global benchmarks and standards [19-22].


The network’s core activities will include building an independent evidence base, conducting contextual real-world assessments, and advancing evaluation science beyond existing benchmarks [29-33]. Specific flagship projects for the coming year were announced: multilingual AI benchmarks with the Collective Intelligence Project, a gender-harm taxonomy with GXD Hub, and work to link evaluation outcomes to public-policy procurement mechanisms, which are seen as a lever for responsible innovation in the Global South [43-52]. Mr. Abhishek Singh emphasized that safe and trusted AI is a universal goal but that current multilingual benchmarks are lacking, noting India’s 22 official languages as an example and praising the New Delhi Frontier AI commitments as a step toward shared data and evaluation tools [66-77][88-92]. He warned that without capacity-building and resource sharing, the Global South will remain excluded from shaping AI safety standards [84-87][98-104].


Ambassador Philip Thigo reinforced the urgency of inclusion, calling the network timely yet late, and proposed regional nodes, multilingual benchmark datasets, an annual red-team exercise, and a Global South AI Safety Report to integrate the network into multilateral processes such as the UN AI governance panel [138-146][170-178][179-182]. Dr. Rachel Sibande argued that safety definitions must be re-defined to reflect local cultural, gender, and linguistic contexts, illustrating how mistranslation of a pregnant mother’s warning could miss a critical health signal [216-227]. Ms. Chenai Chair added that gender-biased voice interfaces and the diversity of African languages can exacerbate existing inequalities and even turn benign technologies into surveillance tools [240-269].


Natasha Crampton from Microsoft described the challenge of scaling community-led, multilingual evaluations to thousands of languages and stressed the need for sustainable, ongoing assessment processes [276-284]. Amir Banifatemi pointed out that safety is poorly defined, lacks financial incentives, and suffers from talent and infrastructure gaps, proposing open-source evaluation tools and incident-reporting systems to close feedback loops, especially where latency and regulatory mechanisms are weak [296-311][312-322]. Balaraman Ravindran noted the proliferation of overlapping AI safety initiatives and called for coordinated effort through a single node in the global accountability network to avoid duplication and amplify impact [330-337][338-342].


The speakers agreed that the network will serve as connective tissue between global governance, technology developers, and on-the-ground stakeholders, aiming to make AI trustworthy and inclusive for the Global South [37-39][59-60].


Keypoints


Major discussion points


Urgent need for trustworthy AI in the Global South and the current under-representation of these regions in global safety governance.


The speakers note that AI is rapidly deployed in critical sectors across the Global South, but low institutional capacity and deep inequities create high risks, while the region remains “under-represented in global safety and governance infrastructures” and often lacks its own oversight institutes [11-14][15-18].


The Global South Network for Trustworthy AI as a civil-society-driven platform to generate real-world evidence, improve contextual evaluation, and advocate for inclusive governance.


The network aims to “build an independent evidence base,” conduct “real-world deployment assessment,” and push the “science of evaluations” beyond standard benchmarks, while also “field building” and providing “connective tissue” between global governance and on-the-ground realities [29-33][34-38][43-50].


Key structural challenges identified: multilingual and cultural mismatches, limited access to compute, concentration of benchmark-setting power, and gaps in talent and infrastructure.


Participants highlight the scarcity of “multilingual benchmarks” for the many languages spoken in the Global South, the “access to compute” problem for researchers, the fact that “benchmarks are not neutral” and are often defined by a handful of institutions, and the lack of “talent inclusion” and appropriate “infrastructure” for evaluation [73-78][158-162][165-166][304-311].


Planned flagship projects to address these gaps, including multilingual benchmark development, gender-harm taxonomy, procurement-lever strategies, and sector-specific evaluations (e.g., health information systems).


The network will work with partners on “benchmarks for multilingual AI,” build a “taxonomy of gender harm,” support “procurement” as a lever for responsible innovation, and evaluate “labor market impacts” and “health information systems” in the Global South [43-50][54-58][59-60].


Calls for coordinated, regional, and multilateral structures to amplify impact and avoid duplication of effort.


The Ambassador proposes “regional nodes” and a “Global South AI Safety Report,” while other speakers stress the need to “harmonize” the many emerging initiatives, integrate the network into the UN AI governance process, and create a shared “steering committee” that includes Indian and Kenyan representatives [171-179][184-186][330-342].


Overall purpose / goal of the discussion


The session was convened to launch the Global South Network for Trustworthy AI and to articulate its mission: creating a civil-society-led ecosystem that generates context-specific evidence, builds multilingual and culturally aware evaluation tools, and advocates for the inclusion of Global South perspectives in global AI safety standards and governance frameworks.


Overall tone and its evolution


– The opening remarks are enthusiastic and celebratory, thanking partners and expressing excitement about the launch [4-9][11-14].


– The conversation then shifts to a problem-focused, analytical tone, detailing systemic gaps, risks, and technical challenges [15-22][73-78][158-166].


– As the panel proceeds, the tone becomes collaborative and solution-oriented, highlighting concrete project plans, regional coordination ideas, and commitments from industry and multilateral actors [43-50][171-179][348-349].


– The closing moments retain a hopeful and forward-looking tone, emphasizing rapid action, partnership, and the urgency of turning discussion into tangible outcomes [353-358].


Overall, the discussion moves from celebration of the network’s inception, through a sober assessment of existing deficiencies, to a constructive agenda for collective action.


Speakers

Dr. Urvashi Aneja – Founder and Director of Digital Futures Lab; host and moderator of the session.


Mr. Abhishek Singh – Under-Secretary, Ministry of Electronics and Information Technology, Government of India [S7].


Ambassador Philip Thigo – Special Envoy on Technology, Republic of Kenya [S4].


Mr. Quintin Chou-Lambert – Chief of Office and AI Lead, UN Office for Digital and Emerging Technologies [S16].


Ms. Natasha Crampton – Vice President and Chief Responsible AI Officer, Microsoft [S17].


Dr. Rachel Sibande – Senior Program Officer, AI for Africa, Gates Foundation [S10].


Ms. Chenai Chair – Director, Masakane African Language Hub [S12].


Dr. Balaraman Ravindran – Professor, IIT Madras; Head, Center of Responsible AI, IIT Madras; member of the UN scientific panel on AI [S1][S2].


Mr. Amir Banifatemi – (Speaker; specific title not stated in the transcript).


Additional speakers:


None identified beyond the list above.


Full session reportComprehensive analysis and detailed insights

The session opened with Dr Urvashi Aneja welcoming participants to the India AI Impact Summit and formally launching the Global South Network for Trustworthy AI. She opened the panel by asking Dr Rachel Sibande where clarity is lacking about safe AI in the Global South [210-214]. Aneja highlighted that AI is being deployed rapidly across health, education, the judiciary, and government in the Global South, creating “immense” opportunities but also “immense” risks because many of these contexts suffer from low institutional capacity, deep societal inequities and low literacy levels [11-14]. She warned that the region is “under-represented in global safety and governance infrastructures” and that many countries lack their own oversight bodies, leaving local concerns at risk of being ignored [15-18].


Aneja then positioned independent civil-society organisations as uniquely suited to fill this gap, arguing that their proximity to real-world deployments enables them to surface risks invisible to laboratory testing [19-21]. The Network’s core mission is to build an independent evidence base, conduct contextual real-world assessments, and advance the science of evaluation because existing benchmarks do not capture all societal risks [29-33]. The network will give visibility to technology companies designing tools and safety infrastructure, as well as to governments and international organisations shaping global AI-governance architecture[70-73]. It will also act as connective tissue between the global governance architecture, the global safety infrastructure, and what’s happening on the ground [70-73].


Five flagship projects for the first year were announced:


* development of multilingual AI benchmarks in partnership with the Collective Intelligence Project and CARIA [43-45];


* creation of a taxonomy of gender-related harms with GXD Hub and the Global Centre for AI Governance to improve incident-reporting databases [46-47];


* work on procurement levers, linking evaluation outcomes to public-policy procurement to shape markets for responsible innovation [48-53];


* a labour-market impact study; and


* health-information-system evaluations to test whether large language models meet clinicians’ needs in the Global South [54-58][59-60].


Mr Abhishek Singh reinforced that safe and trustworthy AI is a universal goal, but identified a critical shortfall: most benchmarks are English-centric, ignoring the 22 official languages of India and the linguistic diversity of other Global South nations [73-77]. He praised the New Delhi Frontier AI commitments, which require model developers to share usage data and to publish multilingual performance benchmarks, and asked how compliance can be ensured and capacity built across the region [84-92][98-104]. Singh cautioned that without tools and benchmarks, merely identifying risks is insufficient [71-73].


Ambassador Philip Thigo echoed the urgency, noting that the Global South has been “systematically excluded” from safety conversations and that Kenya is currently the only member of the international AI-safety-institute network [138-141]. He enumerated four structural gaps-limited team-capacity, access to compute, linguistic and cultural mismatches, and the non-neutrality of benchmarks, which concentrate power in a few institutions [157-166]-and proposed establishing regional nodes (e.g., an African hub) [160-162], creating multilingual benchmark datasets, organising an annual red-team exercise, and publishing a Global South AI Safety Report to feed into multilateral processes such as the UN AI-governance panel [170-179][180-182].


Dr Rachel Sibande (Gates Foundation) argued that safety must be re-defined to reflect local cultural, gender, religious and linguistic norms. She illustrated the danger of mistranslation with a pregnant mother’s phrase “waters have broken”, which could be rendered as “I have thrown away water” and thus miss a critical health alert [216-227]. She called for community-informed analyses of societal, ethical and distributional risks [216-218][229-232].


Ms Chenai Chair, Director of the Masakhane African Language Hub, added that developers often overlook user experience and gender dynamics, citing a voice-enabled agricultural tool that used a male-sounding voice in a context of gender-based violence, thereby exacerbating existing inequalities [236-247]. She highlighted the vast linguistic diversity of Africa-over 2 000 documented languages, of which Masakhane currently supports only about 50-leading to mismatches when tools are deployed in local dialects [248-255]. She warned that benign technologies can quickly become surveillance tools when communities are not consulted, giving the example of luggage-tracking devices being misused [256-269].


From the industry side, Natasha Crampton (Microsoft) described the challenge of scaling community-led, multilingual evaluations. She noted that projects like Samishka, which combined civil-society insight with research, must be turned into sustainable, ongoing evaluation pipelines that can operate across thousands of languages and cultural settings [276-284]. She stressed that benchmarks cannot be a one-off activity; they need to be run continuously to capture shifts in model behaviour [281-284].


Amir Banifatemi (ITS Rio) pointed out that safety is poorly defined and rarely costed into financial planning, meaning firms lack incentives to prioritise it [296-311]. He identified gaps in compute access, talent inclusion, and system-wide evaluation tools, arguing that current assessments focus narrowly on model design and ignore the broader ecosystem of APIs, data pipelines and infrastructure [312-322]. He advocated for open-source incident-reporting tools that capture contextual harms and for mechanisms to accelerate feedback loops in the Global South, where institutional latency hampers rapid response [321-322].


Professor Balaraman Ravindran (IIT Madras) observed a proliferation of overlapping AI-safety initiatives-including networks in Africa, China and UN-led capacity-building programmes-creating a risk of duplication [330-342]. He urged the Global South Network to serve as a single node within the broader accountability network, coordinating efforts and harmonising activities to amplify impact [330-337][338-342].


Mr Quintin Chou-Lambert (UN Office for Digital and Emerging Technologies) warned that technical standards alone cannot ensure safety, as a one-size-fits-all approach fails to capture contextual nuances [191-194]. He argued that field-tested, low-resource examples are essential to surface challenges that large-scale models overlook, and that the Network can feed such empirical evidence into the UN Global Dialogue on AI Governance [195-199].


Rapid-fire commitments


– Microsoft pledged to honour the New Delhi Frontier AI commitments by sharing multilingual data and investing $50 billion by the end of this decade in Global South infrastructure to support scalable evaluation [348-349].


– The Gates Foundation committed to institutionalise safety evaluation at the point of deployment, ensuring issues are caught early [355].


– Masakhane announced a benchmarking initiative for African languages to be delivered within the year [356].


– Amir’s labs in Bangalore and San Francisco will release open-source, culturally contextual incident-reporting tools for public use [358].


Across the discussion, participants converged on the need for multilingual, culturally aware benchmarks (Aneja, Singh, Crampton, Thigo, Chenai Chair, Sibande) [29-33][34-38][43-45][73-77][174-175][216-227][236-247]; they agreed that civil-society insight and inclusive talent are essential for surfacing risks (Aneja, Thigo, Chenai Chair, Banifatemi, Sibande) [19-22][138-141][236-247][304-315][216-218]; and they recognised capacity-building, compute access and infrastructure investment as prerequisites (Singh, Thigo, Crampton, Banifatemi, Aneja) [71-73][158-160][276-284][296-311][34-38].


Key points of disagreement emerged around who should define benchmarks. Singh called for multilingual benchmarks to address risks [73-77], while Thigo warned that benchmarks are not neutral and should not be set by a handful of institutions [161-165]; Banifatemi added that evaluations must consider the whole system, not just model performance [319-320]. On incentives, Banifatemi argued that safety is not costed into financial planning, reducing corporate motivation [307-311], whereas Singh stressed that safety should complement, not stifle, innovation [105-108]; Aneja suggested using public procurement as a lever to drive responsible AI [50-53]. Regarding the scope of safety, Singh focused on technical risk identification [68-73], Thigo broadened it to include environmental, misinformation and lifecycle harms[154-156], Banifatemi emphasised system-wide evaluation [319-320], and Sibande called for a culturally grounded definition of harm [216-218].


Take-aways: (i) AI deployment in the Global South offers great promise but also risks amplifying existing social, gender, linguistic and environmental harms; (ii) identifying risks is insufficient without tools, benchmarks and capacity-building; (iii) the Global South is systematically under-represented in AI-safety governance, and the Network aims to provide field-tested evidence and act as a bridge to global policy forums; (iv) English-centric benchmarks must be replaced by multilingual, culturally aware ones; (v) capacity gaps-compute, talent, sustainable evaluation mechanisms-must be addressed; (vi) governance must be de-concentrated, ensuring benchmarks are not dictated by a few institutions and that safety is financially incentivised; (vii) coordination across overlapping initiatives is essential to avoid duplication and maximise impact [29-33][34-38][43-50][174-175][216-227][236-247][276-284][319-322][330-342].


Unresolved issues include: (a) a precise, universally accepted definition of “safety” and “harm” that captures diverse cultural contexts; (b) concrete mechanisms to cost safety into corporate financial planning or impose penalties for unsafe AI; (c) design of ongoing, scalable evaluation frameworks beyond one-off tests; (d) equitable access to high-performance compute for Global South researchers; (e) detailed pathways for the Network to integrate with UN AI-governance processes; (f) strategies to de-concentrate benchmark authority and ensure inclusive risk prioritisation; (g) methods to close the accountability loop so that technical evaluations translate into tangible citizen-level benefits [216-218][307-311][281-284][158-160][170-179][161-165][180-182].


Suggested compromises involve establishing regional nodes to balance rapid activation with local expertise, adopting an open-source, collaborative benchmarking framework that allows multiple institutions to contribute, leveraging the New Delhi Frontier AI commitments as a baseline while expanding multilingual evaluation work, combining top-down UN engagement with bottom-up civil-society evidence generation, and using pilot projects and incremental infrastructure investments (e.g., Microsoft’s $50 bn pledge) as stepping stones toward a sustainable, global evaluation ecosystem [348-349][170-176][S3].


Overall, the launch marked a decisive step toward a coordinated, inclusive AI-safety ecosystem for the Global South, with broad consensus on the need for multilingual, context-sensitive evaluation and capacity-building, alongside notable divergences on benchmark governance, incentive structures and the breadth of safety considerations that will shape the Network’s future trajectory.


Session transcriptComplete transcript of the session
Dr. Urvashi Aneja

Thank you. Thank you. Thank you. Thank you. Thank you. Thank you. Thank you. Good evening, everyone. My name is Urvishya Neja. I am the founder and director of Digital Futures Lab. And I am so excited to see all of you here and to have you all here for the launch of this network. So it’s a real pleasure to welcome you to the launch of the Global South Network for Trustworthy AI here at the India AI Impact Summit. On behalf of Digital Futures Lab and our other founding partners, Sirai from IIT Madras, the Global Center for AI Governance, ITS Rio, International Innovation Corps, thank you all for being here. And we’re especially grateful to Mr. Abhishek Singh and Ambassador Philip Tigo and Mr.

Quintin Chow and to all our distinguished speakers and guests who are joining us today. Across the Global South, AI systems are being rapidly deployed in critical social sectors such as healthcare, education, judiciary, and in government. And while the opportunities are immense, in many of these contexts, many of these contexts are also marked by low institutional capacity, deep societal inequities, popularization, and populations with low levels of literacy. So while the potential is immense, the risks and harms are also immense. And so it’s particularly important that we figure out ways to make AI safe and trustworthy in these contexts to ensure not only that we protect the populations and to ensure that we don’t exasperate existing harms, but also to ensure that we build the infrastructure for safe and inclusive AI adoption.

Unfortunately, Global South organizations, Global South communities, Global South states remain underrepresented in global safety and governance infrastructures. And many countries in the Global South are actually unlikely to even have in the near term their own safety or oversight institutes. And there’s a real risk, therefore, that the concerns and priorities of these countries, of these communities remain underrepresented in the global safety infrastructure. And precisely those countries that have the most potential or the most opportunity to leverage AI. Independent civil society organizations are uniquely positioned to address this gap. Their proximity to real -world deployment contexts enables them to surface risks that are invisible to lab -based evaluations or testing. The form of grounded evidence that civil society organizations can bring can inform global safety benchmarks, standard -setting processes, and risk assessments, providing corrective signals to technical and regulatory institutions.

The Global South Network for Trustworthy AI works to advance exactly these objectives – to evaluate the real -world impact of AI systems, to build the trust and oversight mechanisms localized to different linguistic, cultural, and infrastructural contexts, and to elevate Global South perspectives in global AI governance forums. It is particularly encouraging that this initiative also aligns closely with the recently announced New Delhi Frontier AI. The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI systems, and to develop a more robust and efficient system The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI systems, and to develop a more robust and efficient system the real -world impact of AI systems, and to develop a more robust and efficient system The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world impact of AI.

impact of AI. The Global South Network for Trustworthy AI works to enhance the ability to evaluate the real -world systems. The Global South Network for Trustworthy AI brings together some of the leading research institutions from across the Global South. We are joined by a community of organizations from Asia, from Africa, from Latin America, and the names of which you see displayed behind you. I also want to take this opportunity to highlight some of the key activities that we’re going to be doing as part of the network. I think one of the key things that we want to do as part of the network is to really build an independent evidence base to generate community -informed analysis of the societal, ethical, and distributional risks of AI systems across diverse contexts.

We also want to do real -world deployment assessment to conduct contextual and public evaluations of models and applications across diverse social contexts. We also want to push the field of evaluations, push the science of evaluations, where we say that benchmarks are very important, but benchmarks as they stand today do not necessarily capture all the societal risks that we see in the Global South. So how do we ensure that the evaluation work that we’re doing also captures some of those harms? In some sense, what we want to do with the network is field building. We want to bring together Global South civil society organizations to pool in their collective intelligence, to pool in their capacities, and to advocate together for the representation of Global South concerns on global governance forums.

So what we are trying to do here is field building within the Global South around AI safety and around building that trust infrastructure. And eventually what we hope that all of this amounts to is collective advocacy. We see an important role that the network will play in creating a connective tissue between the global governance architecture, between the global safety infrastructure, and what’s happening on the ground. We hope the network can provide that visibility to real -world impact. to technology companies who are designing tools, who are designing safety infrastructure, as well as to governments and international organizations who are building the architecture of global AI governance. So with that, I want to thank you all. Oh, wait, I have one more thing to share with all of you.

I’m not ready to thank you yet. I also want to showcase some of the projects that we’ll be doing in the coming year. Picking up on yesterday’s commitments, one of the things that we’ll be doing is building benchmarks for multilingual AI. This is with our network partners, the Collective Intelligence Project and CARIA, and we’re really excited to start this work. We’re also going to be doing work on gender and safety. This is with our partners at GXD Hub and the Global Center for AI Governance to build a taxonomy of gender harm so that we can start building a more robust incident reporting database when it comes to gender -related harms and really advance gender safety in digital spaces.

The third piece that we’re going to be working on this year is around procurement. All of the evaluation work that we do, all the benchmarks that we build, all of that has to eventually feed into public policy. And so we hope that some of this work can support procurement. And procurement, we think, is a really important lever for countries in the global south to shape markets for responsible innovation. I think we’ve all heard a lot about the kind of third way of AI governance that India brings to the global governance landscape. And procurement can be an important lever of making that third way a reality and setting the bar for what responsible innovation looks like.

Like I mentioned earlier, we also want to push on the science of evaluation. What does good evaluation look like? What are the kind of methodologies that we need? What are the kind of methodologies that reflect the concerns and the capacities of communities in the global south? So we’re very excited to be doing this work with ITS Rio, who’s also one of the founding partners, and specifically to implement and advance this discussion on evaluations. We’ll be looking at labor market impacts in the global south. and finally we’re going to be looking at evaluations of health information systems do the existing generative AI tools and large language models that we see do they deliver for clinicians do they deliver for doctors what more can they do to support the needs of healthcare professionals in the global south.

So those are the five kind of big flagship projects that we’re going to be launching within the coming year. We’re going to be very busy as you can see we have a lot that we’re going to try and get done and we’re really excited to be on this journey with all of you and would love to engage with all of you post the launch and see how we build this civil society and research infrastructure together. So with that I am delighted to welcome our keynote speakers first and I would like to give the floor to Mr. Abhishek Singh. Sir thank you for your continued support Thank you for the network and for your leadership on the India AI Summit.

Over to you, sir.

Mr. Abhishek Singh

Thank you, Urvashi. And first and foremost, I’d like to congratulate all the team, the network which has brought this together, this Global South Network for Trustworthy AI. With a few months back when we started discussing this concept with Urvashi, with Kalika, with my team, we felt that how do we go about it? Because safe and trusted AI is something that nobody disagrees with. Everybody says that whenever AI innovation is happening, but we must ensure that we must protect ourselves, we must kind of secure ourselves from the harms that can come from misuse of AI or from the risks that frontier AI poses. So yes, we did have the Yoshua Bengio’s report, the scientific panel report, which is part of all the impact summits, the Action Summit and the Bletchley Park Summit, in which it has kind of…

identifies the risks that frontier AI model poses. But what we do believe is that just identifying the risk is not sufficient. We need to think of how do we address those risks. And for addressing those risks, you need to first have the technical tool, the capacity to identify those risks. What are the benchmarks on which you will evaluate them? Some of which Roshi identified, like how do various models perform on multilingual benchmarks? Because very often, most models are evaluated on benchmarks which are predominantly in English language. But if you look at India, a diverse country, we have 22 official languages and multiple other dialects. How do we evaluate how a model performs on various domains in prompts given in those languages?

We don’t have specific linguistic benchmarks. The same applies to many countries of global south. So it felt that while limited expertise exists in some institutions where research is going on, like Serai is one of them, where Professor Balram Ravindran is leading it. There are many labs, of course. whether it’s Microsoft Research or whether other labs wherein such work is going on. The AI Security Institute in UK is doing some work in this direction. The OECD has been doing some work. But how do we ensure that we enable the access to such resources, such tools, such studies to the larger global majority? So with that, this whole concept of creating a global south network for trustworthy AI came in.

And then we immediately had these conversations with all the key stakeholders, partners. We got a lot of support from almost all stakeholders. And along with that, the conversation for the New Delhi Frontier AI commitments was also going on, which Kalika from my team was leading it. And luckily, we were able to announce it in which all models committed to those two commitments about sharing usage data as also multilingual performance benchmarks. So that was a huge achievement. And I feel that the launch of this. Global South Network for trustworthy AI is a further step in that direction. How do we enable compliance to those commitments? How do we ensure that how this data will be shared?

How do we create tools for evaluating models in various languages? How do we build up capacity in all countries of the global south? How do we share resources? How do we share knowledge across? So this is just the beginning and I feel that we support from all industry organizations, the frontier AI labs, the research organizations, governments across the world. This can really, really grow into a resource that can be a global utility. So I compliment all that team which is involved in doing that. The launch of the network is the first step. But how do we action it out? How do we make it functional? How do we ensure that we get necessary support from all stakeholders?

Very often whenever we talk about trusted AI, whenever we talk about safe AI, some people think that we are trying to stifle innovation. The objective is not that. We always say that while the primary objective is to ensure diffusion of AI, primary objective is to ensure that more and more users benefit from the usage of AI. But at the same time, we need to do that in a responsible manner. We need to do it in a safe manner. We do need to do it in a trustworthy manner to limit the harm that can be caused. So this Global South Network for Trustworthy AI which is being launched will work in that direction. It will be an institution that will support not only India but the entire Global South.

And I am sure with just the presence of all the speakers who are present in this session, the strong commitment that all industry and all countries and all multilateral organizations are showing to this initiative, I am sure this will get further strengthened in the days to come. There is a lot of work that Urvashi and team are taking. They are taking on their own. But we will be there to provide all necessary support for India AI mission and we will work towards ensuring that you get the same level of support from every. participating country which is here. So thank you once again and congratulations for this launch and look forward to working towards the objectives in the near future.

Thank you.

Dr. Urvashi Aneja

MS. Thank you, sir, for your remarks and most importantly for your support. I think it means a lot to us to be working so closely with the India AI mission and we’re really excited to be able to deliver on this promise. It’s now my honor to invite Ambassador Philip Tigo, the Special Envoy on Technology from the Republic of Kenya, to share his reflections.

Ambassador Philip Thigo

AMBASSADOR PHILIP TIGO, ASSOCIATE OF TECHNOLOGY, KENYA, Thank you so much for this opportunity to share my reflections. And I noticed that this is really a women -led network, so again, congratulations, Ovashi and Rachel, for putting this together. I think before we celebrate the launch of the network, I think we must acknowledge that we are working with the right people and we are working with the right people. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here.

And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think we have a lot of good people here. And I think because you must acknowledge the structural problem around the safety conversations and the infrastructure that has been cutting safety in the last three years. I think the global south has always been excluded from this conversation. I say this from a position of strength because Kenya is the only, Kenya I think, we’re the only member of the international network of AI safety institutes.

And so there’s a challenge there. And so I think that model that is not inclusive to a global majority, that in most cases bears the brunt and the impacts of AI, is not acceptable. And so this network, in my sense, is timely but also late. And so there’s almost an urgency that we need to work very closely in how we scale up what this network does. The second part, of course, is, as I mentioned, that a lot of the global majority countries that are there are not. They are the ones that not just bear the brunt of the models, but bear the adverse societal harms of the models. Kenya is one of the countries that uses one of the models.

and from the use cases we see that they use it for the wrong reasons. Emotional support or companionship, it’s not necessarily for anything meaningful or productivity. And so as the world advances, it therefore behoves us that we work with these frontier model companies to ensure that their models are safe beyond secure, but also are more trustworthy. The second part, of course, is that part of model evaluations assumes access. We now know that a lot of my colleagues who are doing model evaluations are doing it from an external point of view. So we need to be very clear that global majority countries, and by this when I say global majority countries, we also have a new global south in AI, because it’s just not the global majority.

We know in the global north of artificial intelligence is two countries and a few companies. So we must, beyond this, extend to also include other colleagues, whether it’s from Europe, Western Europe, or Latin America. Safety must also go beyond technology. towards socio -technical issues. We look at AI in the countries of Kenya from minds to models and so safety must also include environmental harms, biases, misinformation, disinformation but also harms to water, environment and so we need full lifecycle accountability. It’s good to evaluate the models but also it’s good to evaluate the footprints of the model quickly. There are four structural gaps that we see and this is why I love this network and the network I think one is yes you want global majority folks to evaluate the models but we have great teaming capacity gaps so I hope that this network will look at this.

Secondly I think is also issues of access to compute. We can’t have global majority researchers trying to evaluate models without necessarily having access to compute to do that. Third part of course has been mentioned by I think his left issues around linguistic and cultural mismatch so we need to do that the other part of course is benchmarking. as governance power. Also, benchmarks are not neutral. Sometimes I think I like to be honest because that’s what evaluation needs to do. And so we need, in most cases, to ensure that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means. Governance is about power. And we must deconcentrate that power even if it’s unintentional.

Finally, I think for me, evaluation is also about agency. And we must have a question of agency, a notion of agency around these models, but also including sovereign capability. As we know, a lot of your countries are trying to build sovereign models, but also sovereign capabilities across the track. What should this network deliver, in my view? And I’ll humbly make these quick suggestions. One, I think, yes, good to have the network, but can we have regional nodes for this? So that, because Africa… I speak for Africa, Africa is another country, it’s 54 countries, expanded to have nodes. Secondly, include multilingual benchmark data sets. Could be an interesting annual red teaming exercise. Could be potentially, why not publish a Global South AI Safety Report with an expansive definition of what safety is.

And I would be remiss if I don’t say how do we fit this into the multilateral process. We already have a global UN scientific panel on AI, and there’s a global dialogue on AI governance. I’m one of the champions for this, so hopefully we will get this in there. Finally, let’s close the accountability loop. How do all this ultimately matter for citizens? We can evaluate all we want, but if they don’t translate

Dr. Urvashi Aneja

Thank you, Ambassador, for highlighting the urgency of this work and also reframing the safety conversation for the Global South. And just to say we are planning to have regional hubs, and we do. And I think the point about how we engage with the multilateral system is very important, and we will have the Indian AC as part of our steering committee, and we hope we can work with the government of Kenya as well. And, of course, we have Professor Ravindran, who is part of the scientific council, so we will be relying on him as well. But thank you. Thank you for your remarks. And with that, I’d like to call our final keynote speaker for the day, who represents the UN Office for Digital and Emerging Technologies.

I’m pleased to invite Mr. Quenchen Chow Lambert, the Chief of Office and AI Lead, to deliver the next keynote. Thank you for your keynote address.

Mr. Quintin Chou-Lambert

There is less, perhaps, infrastructure or energy connection to go around. So the concept of AI safety becomes less of a, or it kind of edges into this more contextual field, and that’s where this kind of low perspectives, field -tested examples can be very helpful to surface, which we’re missing. And I’d say the idea of AI standards as technical standards don’t solve that issue because a one -size -fits -all standard will not be contextually sensitive. So moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small -language models to context turns the issue of AI safety into a more fuzzy kind of discussion and one which really needs empirical evidence.

And I think the trends in the institutional discussions from Bletchley Park to Sears, Seoul, where there were also around 30 countries signing the declaration, to Paris, where you had 60 -plus, and now here. over 100 countries engaging. We now have the United Nations Global Dialogue on AI Governance, which will include a whole 193 member states informed by analysis from an independent international scientific panel on AI, which will look at the risks and also opportunities and impacts of AI. And so as the conversation in these summit settings and in the international level has widened and to include more countries and more people and covered more of humanity, the focus has, through the open source developments, been allowed to become much more focused of encompassing other perspectives.

And that’s why, to close and to echo Ambassador Thieger, these kinds of networks play a crucial role in connecting and bringing examples of the challenges that we face. Thank you very much. cases of threats from various sources to local people into discussions so that international discussions do not ignore or omit or discount the perspectives of the vast majority of people on the planet. Thank you.

Dr. Urvashi Aneja

Thank you, Mr. Chow, for those remarks. I’d now like to call our panelists onto the stage. Ms. Natasha Crampton, Vice President and Chief Responsible AI Officer at Microsoft. Dr. Rachel Tabande, Senior Program Officer AI for Africa at the Gates Foundation. Before you sit, we’re going to take one quick picture. Ms. Chennai Chair. I don’t see you. Oh, there you are. Yes, okay. Director of the Masakane African Language Hub. Mr. Amin Banefatami, Chief Responsible AI Officer. I’m cognizant. And last but certainly not least, Dr. Balaram Ravindran, Head Center of Responsible AI at IIT Madras. Yes, and can we get the keynote speakers as well? Thank you. As with all good things in life, we’re short on time.

But so let’s get started. Rachel, I’m going to start with you. Thank you. where according to you what according to you or where according to you do you feel like we still lack clarity on how safe and reliable AI systems are when they’re deployed in real world context in the global south

Dr. Rachel Sibande

thank you so couple of things maybe two three things number one is we need to redefine what is safe and what is harmful in as far as AI models or applications are concerned according to the social cultural context that they are deployed in and that means that having models or applications that are great at understanding the data or the patterns to generate content is not enough if they do not understand the social norms the gender dynamics the religious beliefs the political sensitivities or indeed even the humor the slang or the tone particularly now that voice is being used in the media a key channel for delivery of AI. So we need to redefine safety and harm in the context in which AI models are deployed.

So I think we’re missing that, but hopefully we get there. I think the second piece is around language. It’s not enough for a large language model to have strong translation capabilities. Language in itself is not just about vocabulary. It’s also about the lived meaning, the lived experiences. I come from a beautiful country called Malawi. It’s also called the warm heart of Africa. Now, if you’re deploying a model for pregnant mothers to access advisory messaging there, if the mother says their waters have broken, which clinically is a critical incident that should warrant that mother to be referred to a health facility, but if you translate that from the local language to English, which is where most of these large language models and applications have been benchmarked on, that will literally mean I have thrown away water.

So if the model is not trained to understand that context, then you will miss that flag. And then finally, I wanted to say that we also need to understand the harms that emerge as people use the AI models. Currently, I think much of the benchmarking is done on the content and predefined metrics. So final example, personally, I use my AI companion as my therapist. So it’s the one persona that knows a lot about my personality from all spheres, as a mother, as a career person, my finances, all of that. But at what point can we then be able to track whether I’m substituting my cognizance and cognitive capabilities with that AI model or application, or that I’m becoming overly emotionally dependent?

So I think there are those three areas that we’re missing, and hopefully we can get better at it. Thank you.

Dr. Urvashi Aneja

Thank you, Rachel, and thank you also for those powerful examples, because I think we’ve been saying some of this at almost a theoretical level, but I think those examples really bring home the gaps in terms of where the current safety conversation is. Chennai, from a civil society perspective, what do you feel companies or developers often miss about the safety implications of deploying AI systems in the global south?

Ms. Chenai Chair

That’s one thing they miss, the user experience. So on a more serious note, thank you, Vashi. So this is great to actually piggyback from what you said, and I was like, are we reading the same notes? So I think what really is missed when people are deploying some of these solutions is around the context in which they’re deploying the tool. And this is particularly looking at an example where on the African continent, there is high levels of gender inequality, a very youthful population with young people often unemployed, and also older people forgotten in actually the development of technologies. So I don’t know who we’re developing for, but sometimes we actually don’t consider that diversity and the inequalities that exist.

So you can find that sometimes when these tools are deployed, they actually further exacerbate a situation of inequality. And I’ll give you one example where perhaps an agricultural tool that has a voice system on it to provide farmers or women information on what to plant may actually have a male -sounding voice. And if in that context there’s high issues of gender -based violence or lack of trust, and the community members were not consulted in the design process, what it actually leads to is just exacerbating an already existing situation. And that is an example. That actually did happen when people were deploying Internet solutions for a community. Then secondly, also thinking about who gets left behind in deploying these solutions.

This is where language, as Rachel was mentioning, comes in. So on the African continent, we have over 2 ,000 languages that have been documented. Masakhane is only working on 50 of those African languages to build up quality data sets. So what you then find is when people are deploying technologies, even if they deploy them in something like Kiswahili, which now has a large number of data sets, people just don’t speak Kiswahili across East Africa. And particularly in Kenya, if you go to Nairobi, the Kiswahili spoken in Nairobi will be Shang. Then you go to, it’s not even Kiswahili, as I’m being corrected. And then if you go to the coast in Mombasa, it will be completely different. So we have to actually take into account the context and nuance of what is being deployed.

And then lastly, the way in which the sector, the technology is actually used, if deployment doesn’t take into account. the whole ecosystem of the end user, it can actually result in misuse. And I want to specifically say that there’s two forms of misuse here. There could be people who unintentionally actually carry out a problematic, harmful act online based on how they’re interacting with the technology. And we know that content, particularly if it’s in their own language, and we know that content moderation for the global majority is not sufficient. Or people are underpaid, as we’ve seen the cases that were coming out about content moderators in Kenya. Then there’s actually intentional misuse. Now, this is where we find gender disinformation, the use of deepfakes to discredit people, particularly around election period.

And now with increased open AI that people can actually just type something and get something back, we are seeing that high level of deployment without thinking about what is the after -end impact. To close it off, because I’m talking about AI as if it’s coming later. A10. when they were deployed. It was great, I can track my missing bag on a flight. They have now been put in women’s bags or children’s bags by people who they do not know and they track them. That’s already an act of surveillance that was, if people had been consulted, it might have been mitigated against. Yes, I do want to know where my bag is, but I don’t want to be tracked unknowingly.

Dr. Urvashi Aneja

Thanks, Chennai, for that and also for pointing out, bringing the gender dimension on the table and highlighting the issues around what seems like useful technology, how quickly it can become surveillance technology. I’d like to now bring the industry perspective into this conversation. So, Natasha, maybe I can start with you. As you scale systems globally, what are some of the hardest constraints that you as a company face in ensuring context -sensitive safety?

Ms. Natasha Crampton

Well, thanks for that question, Arati, and congratulations to everyone on the establishment of the network. I think it’s a really important step forward. So when I think about Microsoft, I think about sort of Microsoft’s scale, and our mission is really to try and empower every person in every organization in the world to achieve more. And so one of the challenges I think that we face with scaling up our efforts here is how do we take the very deep, careful, thoughtful, community -led evaluation work that animated a project like Samishka, which the CAIA organization, as well as the Collective Intelligence Project and Microsoft Research worked on together, which really developed very context -aware evaluations that were appropriate for the use case.

And how do we take that thoughtful work and really scale it up? Because really we want to do that type of work for thousands of languages and probably millions of different cultural settings. And so I think we really need to think about this system of how we are going to build multilingual and multicultural evaluations that we can really run broadly. I think sometimes we think evaluations, we don’t sort of understand how sustainably they need to be run. As in you can’t just do it once before you release a product. You need to run the evaluations on an ongoing basis to understand how there might have been shifts. And so I really think for us we need to think about this system.

How are we going to build a sustainable, grounded, community -led system of scalable evaluation?

Dr. Urvashi Aneja

Thanks, Natasha. And I hope in some sense also the network can actually play at least part of that function in building that kind of coherence to the space of evaluation and helping us at least build a shared vocabulary and a shared set of methodologies together as organizations. Amir, what do you think needs to change, whether it’s internally within companies or externally in terms of the ecosystem that we’re operating in, to make such grounded evaluations, the kind that Natasha was talking about, become the standard practice for industry? Should they be the standard practice? And if so, how? How do we get there?

Mr. Amir Banifatemi

Thank you for that question. And first, congratulations. I’m happy to be also part of this network and support it. I think Natasha mentioned part of the foundational questions. And I think from a, I’m putting my hat off, cognizant chief responsible, we work with a lot of companies and governments into deploying. new scenarios. We call it systems or applications or anything else. The concept of safety, I was mentioned, is diffused. It’s not very clear what we’d call safety. So evaluating the underlying element that needs to be changed or to be addressed is not obvious. When we talk about models, models are not just one thing that you deploy. It goes into an application, there’s a system, infrastructure, there’s network access, there’s API connected data access.

All of them are contextually different. That was mentioned before. And then the problem, one of the problems is that, you didn’t ask me about the problem, but there’s a problem issue is that there’s a lack of imagination. People that are building systems have no awareness about the context in which those situations occur and how they occur and what’s the causes and what’s the likelihood of solution to happen. So absent of that, all this context which language is part of it, culture is part of it, is not captured. So without that, there is very little capability. to address that from a regulation or incentive perspective. Safety, on the other side, is not costed into financial systems and so forth.

There is no penalty of not being safe. So as long as there is no constraint to put safety as a cost structure, which strong mandate, companies will not pay attention or enough attention. So if it’s not part of the financial planning and the processes and so forth, it won’t happen. So there is a disconnect between what we do as enterprises to make sure that systems and platforms are properly built and deployed. There is a disconnect between the system in which they are deployed. At the same time, there is a talent inclusion that is missing. So the inclusion part is that all the talent that is building into those safety conversations are not the talent that are exposed to those issues.

So that absent voice is also a piece that needs to be addressed, not just from a skilling perspective, but also from an integration perspective. And finally, the infrastructure part. The infrastructure is not just systems and models and data, it’s also the tooling and the evaluation. And it was mentioned that evaluation has to be done differently, but if you don’t know what harm or safety means, evaluation’s gotta be different. There is probably an opportunity here to come up with a series of evaluation tools that are not only built for model design, but also built for system deployment. And if we go from pilot to scaling, what issues occur and what examples are happening and what incidents are deployed, and incident reporting is a huge opportunity here because it will capture, nested in reporting, some of the hidden element of the control issues, data access, regulation, absence, or anything else.

Finally, there is a latency issue in global north, and you mentioned probably correctly that there’s a lot of latency issues. There are only a handful of countries in the global north that probably the new slot is much bigger. there are institutional framework, you have basically the rule of law, you have civil society which is very active, you have legal framework that basically creates an accelerated feedback loop into all this incident safety in most of the global south countries these mechanisms don’t exist which delays the feedback loop and basically compounds the possible harm and everything else so there is probably an opportunity to figure out how we can accelerate the learning capabilities and the skills at which we capture knowledge and data to be tied with tools that probably need to be implemented and deployed either on an open source matter or a free access matter and build it with a contextual environment, the talent pool to make it together so the ownership of the global south, all these pieces are important so the network can actually incentivize those different pieces that could complete each other to really play a role into the global south understanding better where safety issues are, where harm can happen and what corrections can be made in the rhythm that needs to happen because rhythms are not exportable and what we do from one country to another is not.

And finally the network could probably help bring it together.

Dr. Urvashi Aneja

Thank you for laying that out and also just pointing out how all the kind of pieces link to each other and we can’t just kind of go at it at one level alone and to the importance of capacity across all those. Professor Ravindran, AI deployment is accelerating in the global south, in India, in many other countries as well. But at the same time we don’t see or so far we haven’t seen as much investment in the kind of safety and safety infrastructure. Would you agree? You’re actually asking an academic about investment? Sure, of course there’s not enough money. Why not and how do we change it?

Dr. Balaraman Ravindran

so I’m going to answer a different question sure perfect like a true academic I’m sorry I’ll connect it back to what you asked so there are a whole lot of initiatives that are getting announced at the summit and also things that I kind of discovered while having various conversations that there are multiple networks that are getting launched there are already in operation there is a network in Africa looking at capacity building there is a network in China apparently which none of us seem to have heard about that’s being launched on AI safety and capacity building and that is our network that’s getting launched and that is the UN initiative on building this network of capacity building institutes for the global south which we had a meeting in the morning as well about that so there is just too many of these initiatives that are getting launched.

And we have to figure out a way how we would coordinate operations among these initiatives as well. So I think that would be a great multiplier instead of everybody going out and saying, okay, let me see what small piece of the pie that I can get so that I can do these activities. And after that, if there is a lot more coordination. And if you remember our initial conversations about when we wanted to start this thing was about this would be like this one node in the global AC. I can’t even say global network of safety institutes anymore, can I? So they’re not even safety institutes. So AC institutes, whatever ACs, this should be like one node in the AC network which kind of represents unheard voices there because almost except for, as the ambassador was pointing out, except for Kenya.

So we really have, and of course India, I presume. We don’t have safety institutes in the global south, right, that can participate in the dialogue. So I mean, that kind of larger collaboration framework is something that we should enable so that, I mean, even if you say, we go to Gates, and then how many different people, how many different networks would Gates want to spend their money to? If that is one way we can say that there’s this whole operation that’s happening, then that would be a great way of harmonizing our efforts. I can turn it back to the question. Thank you.

Dr. Urvashi Aneja

No, I mean, I think you raised a really important issue of kind of harmonizing these efforts, and also that how this network can play a really important role in the larger kind of AC network. Luckily, the S remains the same, so we can still go with the acronym, I guess, on the safety network. We’re almost at time, so let’s just do one kind of quick rapid -fire round with all the panelists, and maybe Natasha, I can start with you. What is the one concrete step your institution, Microsoft, could take in the next year to strengthen AI safety in the global south?

Ms. Natasha Crampton

Well I’m looking forward to making good on the New Delhi Frontier AI commitments that Microsoft made which is going to help advance multilingual and multicultural evaluation work as well as share data that will help policy makers make or understand AI adoption within their countries and make the sort of choices and policy interventions that help bring that broader access so if I can be sneaky and kind of come as one thing. The second thing I’m really excited about is we’re making large infrastructure investments across the global south to the tune of 50 billion dollars by the end of this decade now that infrastructure as Amir and others on the panel have mentioned is essential to being able to building up this scaled system of sustainable evaluation so I’m looking forward to those investments too.

Dr. Urvashi Aneja

Thank you.

Dr. Balaraman Ravindran

is that a fire alarm or something?

Dr. Urvashi Aneja

No, no, no, they’re telling us that we have to wrap up I think.

Dr. Balaraman Ravindran

Okay, great, so wrapping up, so we have to get the work going, rolling, right, so talking about it is one thing, but actually starting to do this collaboration and getting this research efforts going, we’d love to reach out to partners across the globe, in fact, I’m part of the other UN network as well and we have been talking about looking at problems that would necessarily require cross -border collaboration, right, as supposed to, you know, problems that we would anyway solve in our geography, then just working with somebody else to solve it in two geographies, okay but if you can pick problems that will necessarily require people across borders to collaborate, I think that will certainly drive this and also will, you know, kind of put forth the importance of having the network itself, not just information sharing, but actually problem solving that can be done only across the network.

Dr. Urvashi Aneja

Thank you. Rachel, 30 seconds.

Dr. Rachel Sibande

30 seconds I think from the foundation side is to really institutionalize the evaluation of safety of AI solutions right at deployment because what we see now is that safety issues almost emerge post deployment thank you

Ms. Chenai Chair

so from the hub side we actually do have a benchmarking initiative that’s going on this year so this will be one contributing to the African benchmarking work and so that will be our output in contribution

Dr. Urvashi Aneja

amazing looking forward to that thank you Chennai and Amir last but not least

Mr. Amir Banifatemi

we’re working already on with our two labs one in Bangalore actually and one in San Francisco on safety evaluations mostly on incident reporting and we already made it culturally contextual so I hope that we are helpful to basically provide open source tools for evaluation to disseminate them and work with that work to basically make them accessible to the public available to all partners.

Dr. Urvashi Aneja

Thank you.

Related ResourcesKnowledge base sources related to the discussion topics (41)
Factual NotesClaims verified against the Diplo knowledge base (5)
Confirmedhigh

“Dr Urvashi Aneja formally launched the Global South Network for Trustworthy AI at the India AI Impact Summit”

The knowledge base lists Dr Urvashi Aneja as a participant in the launch of the Global South AI Safety Research Network, confirming the launch event and her role [S3].

Confirmedmedium

“Independent civil‑society organisations are uniquely suited to surface risks invisible to laboratory testing”

Civil-society organisations are described as able to bridge the gap between citizens and governments by conducting independent assessments and surfacing risks that may not appear in lab settings [S49].

Additional Contextmedium

“The network will act as connective tissue between the global governance architecture, the global safety infrastructure, and what’s happening on the ground”

The knowledge base notes that such networks play a crucial role in connecting local challenges to global discussions and that they bring unique regional expertise to facilitate sharing and capacity-building across the Global South [S16] and [S128].

Additional Contextmedium

“The Global South is under‑represented in global safety and governance infrastructures, with many countries lacking their own oversight bodies”

Discussion in the knowledge base highlights infrastructural barriers and a lack of assurance mechanisms for many Global South countries, underscoring under-representation in safety and governance frameworks [S34].

Confirmedhigh

“Ambassador Philip Thigo is involved in the network and echoed the urgency of addressing AI safety in the Global South”

Ambassador Philip Thigo is listed among the participants in the launch of the Global South AI Safety Research Network, confirming his involvement and support for the initiative [S3].

External Sources (132)
S1
Panel Discussion AI & Cybersecurity _ India AI Impact Summit — -Balaraman Ravindran- Professor at IIT Madras (India), member of the UN scientific panel
S2
Why science metters in global AI governance — -Balaraman Ravindran- Professor at IIT Madras, member of International Independent Scientific Panel
S3
Towards a Safer South Launching the Global South AI Safety Research Network — – Dr. Balaraman Ravindran- Dr. Urvashi Aneja
S4
Responsible AI for Shared Prosperity — -Philip Thigo- His Excellency Ambassador, Special Technology Envoy of the Government of Kenya
S5
Philip Thigo named Kenya’s special envoy for technology — Philip Thigo, the Executive Director for Africa at Thunderbird School of Global Management, has been appointed as the Sp…
S6
https://dig.watch/event/india-ai-impact-summit-2026/toward-collective-action_-roundtable-on-safe-trusted-ai — And to explore those questions, we’ve got an amazing panel that I’m honored to introduce. We’ve got Dr. Chinasa Okolo on…
S7
Open Forum #30 High Level Review of AI Governance Including the Discussion — – **Abhishek Singh** – Under-Secretary from the Indian Ministry of Electronics and Information Technology Abhishek Sing…
S8
GPAI: A Multistakeholder Initiative on Trustworthy AI | IGF 2023 Open Forum #111 — Abhishek Singh:I can take that, no worries. Thank you, Abhishek. The floor is yours. You can give your question. Yeah, t…
S9
Announcement of New Delhi Frontier AI Commitments — -Abhishek: Role/Title: Not specified (invited as distinguished leader of organization), Area of expertise: Not specified
S10
Towards a Safer South Launching the Global South AI Safety Research Network — – Dr. Rachel Sibande- Ms. Chenai Chair- Ambassador Philip Thigo – Ms. Natasha Crampton- Dr. Rachel Sibande
S11
Published by DiploFoundation (2011) — Malta: 4th Floor, Regional Building Regional Rd. Msida, MSD 2033, Malta Switzerland: Rue de Lausanne 56 CH-1202 Ge…
S12
Towards a Safer South Launching the Global South AI Safety Research Network — -Ms. Chenai Chair- Director of the Masakane African Language Hub
S13
Responsible AI for Shared Prosperity — -Chenai Chair- Director of the Mazakani African Languages Hub -Co-Moderator- Role/title not specified
S14
IGF to GDC- An Equitable Framework for Developing Countries | IGF 2023 Open Forum #46 — Moderator:based intergovernmental international organization dedicated to promoting and supporting the development of th…
S15
Towards a Safer South Launching the Global South AI Safety Research Network — – Dr. Urvashi Aneja- Mr. Quintin Chou-Lambert
S16
https://dig.watch/event/india-ai-impact-summit-2026/towards-a-safer-south-launching-the-global-south-ai-safety-research-network — I’m pleased to invite Mr. Quenchen Chow Lambert, the Chief of Office and AI Lead, to deliver the next keynote. Thank you…
S17
https://dig.watch/event/india-ai-impact-summit-2026/towards-a-safer-south-launching-the-global-south-ai-safety-research-network — Thank you, Mr. Chow, for those remarks. I’d now like to call our panelists onto the stage. Ms. Natasha Crampton, Vice Pr…
S18
Towards a Safer South Launching the Global South AI Safety Research Network — – Mr. Abhishek Singh- Ms. Natasha Crampton- Ms. Chenai Chair – Ms. Natasha Crampton- Dr. Rachel Sibande
S19
Multi-stakeholder Discussion on issues about Generative AI — Natasha Crampton:So, I’m Natasha Crankjian from Microsoft. I’m incredibly optimistic about AI’s potential to help us hav…
S20
Towards a Safer South Launching the Global South AI Safety Research Network — – Dr. Urvashi Aneja- Ambassador Philip Thigo
S21
Towards a Safer South Launching the Global South AI Safety Research Network — – Ambassador Philip Thigo- Mr. Amir Banifatemi
S22
https://dig.watch/event/india-ai-impact-summit-2026/secure-finance-risk-based-ai-policy-for-the-banking-sector — Compliance functions increasingly rely on automated pattern recognition, while adaptive cybersecurity models respond to …
S23
morning session — Aouad argues that without proper regulations and safeguards, the population could experience negative consequences. Furt…
S24
https://dig.watch/event/india-ai-impact-summit-2026/ai-safety-at-the-global-level-insights-from-digital-ministers-of — Thank you. Certainly, the reason that I continue to be involved with this is because… under Yoshua’s chairmanship of t…
S25
Global AI Policy Framework: International Cooperation and Historical Perspectives — Velasco explains the complementary roles of the UN’s two new AI governance mechanisms, with the scientific panel offerin…
S26
Leveraging the UN system to advance global AI Governance efforts — Gilbert Houngbo highlights the imperative role of the United Nations in spearheading global coordination efforts, thereb…
S27
Main Session 2: The governance of artificial intelligence — Importance of bringing voices from the global south and underrepresented communities to governance dialogues
S28
What is it about AI that we need to regulate? — Ensuring Better Representation of Developing and Least-Developed Countries in Global Digital GovernanceThe question of h…
S29
GC3B: Mainstreaming cyber resilience and development agenda | IGF 2023 Open Forum #72 — One of the main arguments put forward at the conference was the necessity for individuals and nations to be aware of the…
S30
Advancing Scientific AI with Safety Ethics and Responsibility — And I think, I think, So, just in terms of paradigm change that we are seeing and that you mentioned, is that there need…
S31
WS #123 Responsible AI in Security Governance Risks and Innovation — Jingjie He: So I think the inclusive engagement across stakeholders is essential for the effective global governance of …
S32
AI in Africa: Beyond the algorithm — ### The Systematic Exclusion of the Global South
S33
WS #82 A Global South perspective on AI governance — AUDIENCE: Ends up. We cannot hear. Rely on ISO 31,000 is what they see as the kind of framework for risk assessments…
S34
Ensuring Safe AI_ Monitoring Agents to Bridge the Global Assurance Gap — Develop multilingual evaluations and benchmarks that account for diverse language ecosystems
S35
How to ensure cultural and linguistic diversity in the digital and AI worlds? — Xianhong Hu:Thank you very much Mr. Ambassador. Good morning everyone. First of all please allow me, I’d like to be able…
S37
Shaping the Future AI Strategies for Jobs and Economic Development — Thank you. and how safety is governed under real constraints, how AI systems actually reach the people and states often …
S38
Inclusive AI For A Better World, Through Cross-Cultural And Multi-Generational Dialogue — Factors such as restricted access to computing resources and data further impede policy efficacy. Nevertheless, the cont…
S39
Panel Discussion Summary: AI Governance Implementation and Capacity Building in Government — The tone was pragmatic and solution-oriented throughout, with speakers acknowledging both challenges and opportunities i…
S40
Smart Regulation Rightsizing Governance for the AI Revolution — Bella Wilkinson from Chatham House provided a realistic assessment of the current geopolitical landscape, arguing that g…
S41
MASTERPLAN FLAGSHIP PROGRAMMES — | Outcomes | Objectives …
S42
MASTERPLAN FLAGSHIP PROGRAMMES — | Outcomes | Objectives …
S43
Trade regulations in the digital environment: Is there a gender component? (UNCTAD) — In conclusion, the analysis reinforces the potential of digitalisation and emerging technologies, such as artificial int…
S44
WS #479 Gender Mainstreaming in Digital Connectivity Strategies — Ivy Tuffuor Hoetu: Yes, thank you. And picking up from where Dr. landed, it’s true we need the metrics and we need the d…
S45
WS #231 Address Digital Funding Gaps in the Developing World — The conversation concluded with calls for greater coordination among stakeholders to avoid duplication of efforts and ma…
S46
Successes & challenges: cyber capacity building coordination | IGF 2023 — Furthermore, the discussion emphasizes the importance of coordinating with multiple stakeholders or through bilateral in…
S47
Resilient infrastructure for a sustainable world — Collaboration and Partnership Importance Ng explains that UNDRR depends on partnerships due to being a small organizati…
S48
Part 3: ‘Readiness across the spectrum: Countries’ — The EU strategy’s emphasis on the 2030 Digital Agenda aligns closely with the IMI’s Access pillar, providing a strong fo…
S49
WS #302 Upgrading Digital Governance at the Local Level — The discussion maintained a consistently professional and collaborative tone throughout. It began with formal introducti…
S50
Fireside Conversation: 02 — Timeline expectations, hype, and the AGI narrative
S51
Shaping the Future AI Strategies for Jobs and Economic Development — The discussion maintained an optimistic yet pragmatic tone throughout. While acknowledging significant challenges around…
S52
Advancing Scientific AI with Safety Ethics and Responsibility — High level of consensus with significant implications for AI governance policy. The agreement across speakers from diffe…
S53
WS #162 Overregulation: Balance Policy and Innovation in Technology — Regulation is necessary but should not stifle innovation
S54
WS #283 AI Agents: Ensuring Responsible Deployment — Balance needed between privacy protection and innovation Despite representing different sectors (industry, government, …
S55
Aligning AI Governance Across the Tech Stack ITI C-Suite Panel — High level of consensus with significant implications for AI governance policy. The agreement among industry leaders fro…
S56
Main Session | Policy Network on Artificial Intelligence — These key comments shaped the discussion by broadening its scope beyond technical and policy considerations to include e…
S57
Towards a Safer South Launching the Global South AI Safety Research Network — We know in the global north of artificial intelligence is two countries and a few companies. So we must, beyond this, ex…
S58
From Technical Safety to Societal Impact Rethinking AI Governanc — Virginia stresses that AI safety cannot be limited to technical robustness, accuracy or alignment. It must incorporate m…
S59
DiploNews – Issue 312 – 15 November 2016 — News headlines are featuring more and more cases of severe cyber incidents, some openly attributed to states and their i…
S60
The impact of big data on geopolitics, negotiations, and diplomacy — At a global level, data is addressed by a wide range of organisations. Within the World Trade Organization (WTO), data f…
S61
Tech Diplomacy: New Impulses for the Geneva Ecosystem? (Science Diplomacy Week) — Mr Jean-Yves Art, Senior Director, Strategic Partnerships, Microsoft: Companies within the tech sector have a responsibi…
S62
https://dig.watch/event/india-ai-impact-summit-2026/towards-a-safer-south-launching-the-global-south-ai-safety-research-network — Finally, I think for me, evaluation is also about agency. And we must have a question of agency, a notion of agency arou…
S63
Building the Next Wave of AI_ Responsible Frameworks & Standards — I think there is a significant role the governments, innovation hubs, academia, and startups have to play in developing …
S64
Who Watches the Watchers Building Trust in AI Governance — Thank you, Greg. And again, congratulations. Stephen was the publisher of the great report. And I think, first of all, I…
S65
Networking Session #74 Digital Innovations Forum- Solutions for the Offline People — Better coordination and collaboration among donors is needed to avoid duplication of efforts and maximize impact.
S66
Opening address of the co-chairs of the AI Governance Dialogue — The speakers demonstrate strong consensus on the fundamental principles of AI governance, including the need for inclusi…
S67
WS #335 Global Perspectives on Network Fees and Net Neutrality — Despite representing different stakeholder groups (regulator, private sector, civil society), these speakers all emphasi…
S68
Can we test for trust? The verification challenge in AI — **Anja Kaspersen** stressed the importance of bringing technical professional organizations into governance conversation…
S69
Panel Discussion Data Sovereignty India AI Impact Summit — “One, of course, is basically the policies need to evolve along with the infrastructure.”[37]. “As far as governments ar…
S70
Beyond universality: the meaningful connectivity imperative | IGF 2023 — However, concerns remain regarding device affordability and availability, the need for inclusivity in content and servic…
S71
AI That Empowers Safety Growth and Social Inclusion in Action — “investors should ask whether there is clear board level responsibility on AI risk whether executive incentives are alig…
S72
Ensuring Safe AI_ Monitoring Agents to Bridge the Global Assurance Gap — And how do we demonstrate that the risks have been managed well? And that is where the assurance ecosystem that Rebecca …
S73
How the EU’s GPAI Code Shapes Safe and Trustworthy AI Governance India AI Impact Summit 2026 — This comment fundamentally redirected the conversation from discussing what rules to impose on companies to how to creat…
S74
Towards a Safer South Launching the Global South AI Safety Research Network — “And while the opportunities are immense, in many of these contexts, many of these contexts are also marked by low insti…
S75
Ensuring Safe AI_ Monitoring Agents to Bridge the Global Assurance Gap — And this is almost like a test for me of kind of saying. These names of these institutions through this panel. But they …
S76
Developing capacities for bottom-up AI in the Global South: What role for the international community? — ## Areas of Different Emphasis and Debate ## Practical Applications and Examples ## Unresolved Questions and Future Di…
S77
WS #362 Incorporating Human Rights in AI Risk Management — Jhalak Mrignayani Kakkar: Thank you. Thanks, Min. I think there’s a lot of work happening globally on human rights due d…
S78
What is it about AI that we need to regulate? — Ensuring Better Representation of Developing and Least-Developed Countries in Global Digital GovernanceThe question of h…
S79
Shaping the Future AI Strategies for Jobs and Economic Development — Thank you. and how safety is governed under real constraints, how AI systems actually reach the people and states often …
S80
WS #100 Integrating the Global South in Global AI Governance — Overall, the panel emphasized that while challenges remain, there are promising avenues to increase meaningful inclusion…
S82
How Multilingual AI Bridges the Gap to Inclusive Access — “And I think on these three capabilities, we need to jointly increase, and whoever doesn’t have it should be able to eas…
S83
Smart Regulation Rightsizing Governance for the AI Revolution — Bella Wilkinson from Chatham House provided a realistic assessment of the current geopolitical landscape, arguing that g…
S84
WS #214 AI Readiness in Africa in a Shifting Geopolitical Landscape — Fundamental infrastructure challenges—including limited computing power, inadequate connectivity, and capacity gaps—requ…
S85
WS #462 Bridging the Compute Divide a Global Alliance for AI — This comment deepened the discussion by introducing the concept of compound disadvantages and helped other panelists rec…
S86
Benchmarking countries’ progress globally on closing the gender digital divide ( Women in Digital Transformation) — Data generation is essential to be able to address these gaps, especially gender gaps
S87
Future-Ready Education: Enhancing Accessibility & Building | IGF 2023 — In conclusion, the analysis underscores the need for equitable access to the internet to ensure inclusive and quality di…
S88
MASTERPLAN FLAGSHIP PROGRAMMES — | Outcomes | Objectives …
S89
WS #110 AI Innovation Responsible Development Ethical Imperatives — Guilherme Canela de Souza Godoi: Thank you very much. First and foremost, thank you so much for the invitation to be her…
S90
MASTERPLAN FLAGSHIP PROGRAMMES — | Outcomes | Objectives …
S91
WS #231 Address Digital Funding Gaps in the Developing World — The conversation concluded with calls for greater coordination among stakeholders to avoid duplication of efforts and ma…
S92
Resilient infrastructure for a sustainable world — Collaboration and Partnership Importance Ng explains that UNDRR depends on partnerships due to being a small organizati…
S93
BPF: CYBERSECURITY — By working together strategically, they can pool resources, expertise, and knowledge to better respond to and mitigate c…
S94
Agenda item 5: discussions on substantive issues contained in paragraph 1 of General Assembly resolution 75/240 (continued)/5/OEWG 2025 — Dominican Republic: Thank you, Chairman. The Dominican Republic reiterates its firm conviction that capacity building …
S95
Digital Cooperation and Empowerment: Insights and Best Practices for Strengthening Multistakeholder and Inclusive Participation — ### Regional and Multilingual Strategies Amrita Choudhury emphasised the need to “keep processes open and inclusive wit…
S96
Opening and introduction — The AU’s commitment to working with Member States in adopting the meeting’s recommendations was reaffirmed, alongside th…
S97
Closing remarks — The tone is consistently celebratory, optimistic, and forward-looking throughout the discussion. It maintains an enthusi…
S98
Launch / Award Event #159 Book Launch Netmundial+10 Statement in the 6 UN Languages — The tone was consistently celebratory, appreciative, and forward-looking throughout the session. Participants expressed …
S99
Scaling Innovation Building a Robust AI Startup Ecosystem — The tone was consistently celebratory, appreciative, and inspirational throughout. It began formally with the awards cer…
S101
Strengthening Corporate Accountability on Inclusive, Trustworthy, and Rights-based Approach to Ethical Digital Transformation — The discussion maintained a professional, collaborative tone throughout, with speakers demonstrating expertise while ack…
S102
WS #225 Bridging the Connectivity Gap for Excluded Communities — The discussion maintained a professional but increasingly urgent tone throughout. It began optimistically with solution-…
S103
AI and Digital Developments Forecast for 2026 — The tone begins as analytical and educational but becomes increasingly cautionary and urgent throughout the conversation…
S104
New Technologies and the Impact on Human Rights — This IGF session demonstrated a maturing of debates around technology and human rights, with stakeholders from different…
S105
Comprehensive Discussion Report: Governance Frameworks for Reducing Digital Divides in African and Francophone Contexts — The tone was pragmatic and solution-oriented, with speakers expressing both frustration with past failures and cautious …
S106
Tackling disinformation in electoral context — The tone of the discussion was largely collaborative and solution-oriented, with panelists sharing insights from differe…
S107
Panel 5 – Ensuring Digital Resilience: Linking Submarine Cables to Broader Resilience Goals — The tone was largely collaborative and solution-oriented. Panelists built on each other’s points and offered complementa…
S108
Panel 2 – Anticipating and Mitigating Risks Along the Global Subsea Network  — The discussion maintained a professional, collaborative tone throughout, with participants demonstrating technical exper…
S109
Panel 4 – Resilient Subsea Infrastructure for Underserved Regions  — The discussion maintained a professional, collaborative tone throughout, with panelists building on each other’s insight…
S110
High-Level Track Facilitators Summary and Certificates — The discussion maintained a consistently positive and celebratory tone throughout, characterized by gratitude, accomplis…
S111
Closing Ceremony — The overall tone was positive and forward-looking. Speakers expressed gratitude to the hosts and participants, emphasize…
S112
(Plenary segment) Summit of the Future – General Assembly, 4th plenary meeting, 79th session — The tone of the discussion was generally optimistic and forward-looking, with speakers emphasizing the need for urgent a…
S113
AI: Lifting All Boats / DAVOS 2025 — The tone was largely optimistic and solution-oriented, with speakers acknowledging challenges but focusing on opportunit…
S114
Friday Opening Ceremony: Summit of the Future Action Days — The overall tone was inspirational, hopeful and energetic. Speakers aimed to motivate and empower youth attendees while …
S115
Impact of the Rise of Generative AI on Developing Countries | IGF 2023 Town Hall #29 — Tomoyuki Naito:Ladies and gentlemen, good evening. I know this is today’s last session, that’s why not over 100 people c…
S116
Powering AI _ Global Leaders Session _ AI Impact Summit India Part 2 — And we’ll hear some of that in addition to global elements. A lot of that is also having a lot of innovation that will r…
S117
Keynote by Vivek Mahajan CTO Fujitsu India AI Impact Summit — “For the next session, we have a fireside chat between Mr. Vivek Kaneja, Executive Director, CDAT, Mr. Nitin Bajaj, Dire…
S118
Global South’s role in AI governance explored at IGF 2024 — The inclusion of the Global South, particularly theMENA region, in AI governance emerged as a key focus in a recentpanel…
S119
Main Session on Artificial Intelligence | IGF 2023 — There is inadequate representation from the Global South in these discussions
S120
Harnessing AI for Child Protection | IGF 2023 — Artificial Intelligence is giving a lot of opportunities in various fields such as education, law, etc.
S121
WS #205 Contextualising Fairness: AI Governance in Asia — Milton Mueller: Can you hear me? Am I on? Okay, thank you very much. Yeah, I am going to, yeah, first issue you a f…
S122
Planetary Limits of AI: Governance for Just Digitalisation? | IGF 2023 Open Forum #37 — The digital transformation increasingly contributes to greenhouse gas (GHG) emissions. For example, generative artificia…
S123
From India to the Global South_ Advancing Social Impact with AI — That itself is offensive because what are we trying to say? So I think those things will get blurred because opportuniti…
S124
Development of Cyber capacities in emerging economies | IGF 2023 Open Forum #6 — Central America is facing significant challenges in the field of cybersecurity. The region is underdeveloped in terms of…
S125
Open Forum #13 Bridging the Digital Divide Focus on the Global South — ICANN co-chair Tripti Sinha emphasized that the divide encompasses participation and inclusiveness beyond mere access, a…
S126
Main Topic 4: Transatlantic rift on Freedom of Expression — Yasur argues that civil society organizations are uniquely positioned to bridge the gap between platforms, governments, …
S127
Global network strengthens AI measurement and evaluation — Leaders around the worldhave committedto strengthening the scientific measurement and evaluation of AI following a recen…
S128
https://dig.watch/event/india-ai-impact-summit-2026/panel-discussion-ai-cybersecurity-_-india-ai-impact-summit — The network would bring unique expertise and perspectives from different regions of the world. This diversity would only…
S129
WS #98 Towards a global, risk-adaptive AI governance framework — Paloma Villa Mateos: Yeah, thank you. Thank you. Can you listen to me well? It’s okay? Okay, great. Well, thank you. W…
S130
AI Meets Cybersecurity Trust Governance & Global Security — “AI governance now faces very similar tensions.”[27]”AI may shape the balance of power, but it is the governance or AI t…
S131
WS #97 Interoperability of AI Governance: Scope and Mechanism — Mauricio Gibson: Thank you. Yeah, I mean, just building on what Chet was saying, I think, and what you were saying, Olg…
S132
Networking Session #50 AI and Environment: Sustainable Development | IGF 2023 — The role of international organisations, such as the OECD, is highlighted by one speaker in facilitating cooperation on …
Speakers Analysis
Detailed breakdown of each speaker’s arguments and positions
D
Dr. Urvashi Aneja
3 arguments129 words per minute2383 words1102 seconds
Argument 1
Risk of amplifying existing harms without adequate safeguards (Dr. Urvashi Aneja)
EXPLANATION
Dr. Aneja warns that while AI offers great opportunities in the Global South, the same contexts also feature low institutional capacity and deep inequities. Without proper safeguards, AI could worsen existing harms rather than alleviate them.
EVIDENCE
She notes that AI systems are being rapidly deployed in critical sectors such as healthcare, education, judiciary and government across the Global South, and that these contexts are marked by low institutional capacity, deep societal inequities, popularization and low literacy, which together mean that the risks and harms are immense and could exacerbate existing problems if not addressed [11-14].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need for proper regulations and safeguards to prevent negative consequences is highlighted in [S23], and systematic exclusion of the Global South from safety governance, which can exacerbate harms, is discussed in [S32].
MAJOR DISCUSSION POINT
Risk of amplifying existing harms without adequate safeguards
AGREED WITH
Mr. Abhishek Singh
Argument 2
Network will build independent evidence, contextual evaluations, and act as a bridge to global governance (Dr. Urvashi Aneja)
EXPLANATION
The Global South Network for Trustworthy AI will generate real‑world evidence, conduct contextual assessments, and connect local insights with global AI governance structures. This aims to ensure that safety standards reflect the linguistic, cultural and infrastructural realities of the Global South.
EVIDENCE
Dr. Aneja describes the network’s purpose to evaluate the real-world impact of AI systems, build trust and oversight mechanisms tailored to different contexts, and elevate Global South perspectives in global AI governance forums [22-23]. She also explains that the network will serve as connective tissue between global safety infrastructure and on-the-ground realities, providing visibility to technology companies, governments and international organisations [37-38].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
UN mechanisms that bridge scientific evidence with policy and promote inclusive governance are described in [S25] and [S26], while the importance of bringing Global South voices into AI governance is emphasized in [S27].
MAJOR DISCUSSION POINT
Network will build independent evidence, contextual evaluations, and act as a bridge to global governance
AGREED WITH
Dr. Balaraman Ravindran, Mr. Quintin Chou‑Lambert
Argument 3
Engagement with UN‑led AI governance processes and inclusion of Global South voices are essential (Dr. Urvashi Aneja)
EXPLANATION
Dr. Aneja emphasizes the importance of linking the network’s work with multilateral AI governance mechanisms, including UN‑led dialogues and the Indian AI Council. This ensures that Global South perspectives are represented in global policy making.
EVIDENCE
She thanks the ambassador for highlighting urgency and notes that the network will have regional hubs, will involve the Indian AI Council as part of its steering committee, and will collaborate with the Kenyan government and Professor Ravindran on the scientific council [183-186].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The role of UN-led dialogues and the need for inclusive participation of developing countries are outlined in [S25], [S26] and [S27].
MAJOR DISCUSSION POINT
Engagement with UN‑led AI governance processes and inclusion of Global South voices are essential
M
Mr. Abhishek Singh
4 arguments181 words per minute892 words294 seconds
Argument 1
Identifying risks is not enough; tools and benchmarks are needed to address them (Mr. Abhishek Singh)
EXPLANATION
Mr. Singh argues that merely recognizing AI risks does not solve the problem; concrete technical tools, capacity building and appropriate benchmarks are required to mitigate those risks, especially in multilingual contexts.
EVIDENCE
He references the Yoshua Bengio report and other scientific panel reports that identify frontier AI risks, then stresses the need for technical tools, capacity to identify risks, and benchmarks such as multilingual performance tests, noting the lack of specific linguistic benchmarks for India’s 22 official languages and many other Global South countries [68-73].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The New Delhi Frontier AI commitments call for strengthened multilingual and contextual evaluations, underscoring the need for concrete tools and benchmarks [S9]; broader discussions on decentralized checks and technical tool development are found in [S30] and [S3].
MAJOR DISCUSSION POINT
Identifying risks is not enough; tools and benchmarks are needed to address them
AGREED WITH
Ambassador Philip Thigo, Ms. Natasha Crampton, Mr. Amir Banifatemi, Dr. Urvashi Aneja
DISAGREED WITH
Ambassador Philip Thigo, Mr. Amir Banifatemi
Argument 2
Network enables compliance with New Delhi Frontier AI commitments and capacity‑building across countries (Mr. Abhishek Singh)
EXPLANATION
The launch of the network is presented as a mechanism to help implement the New Delhi Frontier AI commitments, which include sharing usage data and multilingual benchmark performance, while also building capacity throughout the Global South.
EVIDENCE
He explains that conversations with stakeholders led to the New Delhi Frontier AI commitments, where models agreed to share usage data and multilingual performance benchmarks, and that the network will support compliance, data sharing, tool creation and capacity-building across countries [87-95].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The announcement of the New Delhi Frontier AI commitments, which include multilingual evaluation requirements, directly supports this claim [S9].
MAJOR DISCUSSION POINT
Network enables compliance with New Delhi Frontier AI commitments and capacity‑building across countries
Argument 3
Current benchmarks are English‑centric; multilingual benchmarks are essential for accurate assessment (Mr. Abhishek Singh)
EXPLANATION
Mr. Singh points out that most AI models are evaluated on English‑only benchmarks, which fails to capture performance in the many languages spoken across the Global South, making multilingual benchmarks a necessity.
EVIDENCE
He notes that most models are evaluated on predominantly English benchmarks and highlights India’s 22 official languages and many dialects, stressing the need for evaluation on prompts in those languages because specific linguistic benchmarks are lacking [73-77].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The prevalence of English-only benchmarks and the lack of linguistic benchmarks are noted in [S3]; initiatives for multilingual evaluation are described in [S34] and cultural-linguistic diversity considerations in [S35].
MAJOR DISCUSSION POINT
Current benchmarks are English‑centric; multilingual benchmarks are essential for accurate assessment
AGREED WITH
Dr. Urvashi Aneja, Ms. Natasha Crampton, Ambassador Philip Thigo, Ms. Chenai Chair, Dr. Rachel Sibande
Argument 4
Responsible AI diffusion should not be framed as stifling innovation; safety must coexist with broad benefit
EXPLANATION
Singh emphasizes that the goal of trustworthy AI is to ensure that more users benefit from AI while maintaining responsibility, not to hinder technological progress. He stresses balancing diffusion with safety safeguards.
EVIDENCE
He notes that some people mistakenly think trusted AI aims to stifle innovation, but the primary objective is to expand AI benefits responsibly and safely for all users [105-108].
MAJOR DISCUSSION POINT
AI safety should complement, not hinder, innovation
A
Ambassador Philip Thigo
6 arguments196 words per minute1016 words309 seconds
Argument 1
Global South is systematically excluded from AI safety governance structures (Ambassador Philip Thigo)
EXPLANATION
The ambassador asserts that AI safety conversations and institutions have historically left out Global South nations, resulting in a governance model that does not reflect the majority of AI users and the harms they experience.
EVIDENCE
He states that the Global South has always been excluded from safety conversations, that Kenya is the only member of the international network of AI safety institutes, and that a model not inclusive of the global majority is unacceptable [138-141].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Systematic exclusion of the Global South from AI safety discussions is documented in [S32]; the need for Global South representation in governance is highlighted in [S27] and [S28].
MAJOR DISCUSSION POINT
Global South is systematically excluded from AI safety governance structures
AGREED WITH
Dr. Urvashi Aneja, Ms. Chenai Chair, Mr. Amir Banifatemi, Dr. Rachel Sibande
Argument 2
Proposes regional nodes, multilingual benchmark datasets, and an annual AI safety report (Ambassador Philip Thigo)
EXPLANATION
He suggests expanding the network with regional nodes across Africa’s 54 countries, creating multilingual benchmark datasets, organizing red‑teaming exercises, and publishing an annual Global South AI Safety Report to broaden participation and accountability.
EVIDENCE
He recommends regional nodes for Africa, the creation of multilingual benchmark datasets, an annual red-teaming exercise, and publishing a Global South AI Safety Report with an expansive definition of safety [170-176].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Calls for multilingual benchmark datasets and regional capacity building are echoed in [S34]; the importance of cultural and linguistic diversity is discussed in [S35]; considerations of smaller-footprint solutions for resource-constrained settings appear in [S36].
MAJOR DISCUSSION POINT
Proposes regional nodes, multilingual benchmark datasets, and an annual AI safety report
AGREED WITH
Dr. Urvashi Aneja, Mr. Abhishek Singh, Ms. Natasha Crampton, Ms. Chenai Chair, Dr. Rachel Sibande
Argument 3
Benchmark design reflects power; only a few institutions should not dictate risk definitions (Ambassador Philip Thigo)
EXPLANATION
The ambassador argues that benchmarks are not neutral; when a handful of institutions define risk metrics, they concentrate governance power, which can marginalize the Global South.
EVIDENCE
He notes that benchmarks are not neutral, that only a handful of institutions should not define what risks are measured, what harms are prioritized, and what safe performance means, emphasizing that governance is about power and must be de-concentrated [161-165].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The argument that benchmark authority should not be concentrated in a handful of institutions is made explicit in [S3]; concerns about centralized governance and the need for decentralized checks are raised in [S30].
MAJOR DISCUSSION POINT
Benchmark design reflects power; only a few institutions should not dictate risk definitions
DISAGREED WITH
Mr. Abhishek Singh, Mr. Amir Banifatemi
Argument 4
Limited access to compute resources hampers Global South researchers’ ability to evaluate models (Ambassador Philip Thigo)
EXPLANATION
He highlights that without sufficient compute capacity, researchers in the Global South cannot effectively evaluate AI models, creating a structural disadvantage.
EVIDENCE
He identifies two structural gaps: the lack of teaming capacity and the issue of access to compute, stating that global majority researchers cannot evaluate models without such resources [158-160].
MAJOR DISCUSSION POINT
Limited access to compute resources hampers Global South researchers’ ability to evaluate models
AGREED WITH
Mr. Abhishek Singh, Ms. Natasha Crampton, Mr. Amir Banifatemi, Dr. Urvashi Aneja
Argument 5
Benchmarks are not neutral; concentration of benchmark authority concentrates governance power (Ambassador Philip Thigo)
EXPLANATION
Reiterating the earlier point, he stresses that benchmark authority should be decentralized to avoid power imbalances in AI governance.
EVIDENCE
He again states that benchmarks are not neutral and that only a handful of institutions should not define risk metrics, underscoring the need to de-concentrate governance power [161-165].
MAJOR DISCUSSION POINT
Benchmarks are not neutral; concentration of benchmark authority concentrates governance power
Argument 6
AI safety must include socio‑technical and environmental dimensions, such as impacts on water and ecosystems
EXPLANATION
The ambassador argues that safety considerations should go beyond algorithmic performance to cover broader societal and ecological harms, ensuring full lifecycle accountability for AI systems deployed in the Global South.
EVIDENCE
He states that safety must also address environmental harms, biases, misinformation, and specific harms to water and the environment, calling for comprehensive lifecycle accountability [154-156].
MAJOR DISCUSSION POINT
Broaden AI safety to socio‑technical and environmental impacts
D
Dr. Rachel Sibande
4 arguments144 words per minute454 words188 seconds
Argument 1
Safety must be re‑defined to reflect local cultural, gender, religious, and linguistic norms (Dr. Rachel Sibande)
EXPLANATION
Dr. Sibande argues that safety definitions need to incorporate the social, cultural, gender, religious and linguistic contexts of deployment, because models that ignore these factors can cause harm.
EVIDENCE
She explains that safety and harm must be redefined according to the social-cultural context, noting that models must understand gender dynamics, religious beliefs, political sensitivities, humor, slang and tone, especially as voice interfaces become common [216-218].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The necessity of incorporating cultural, linguistic, and gender dimensions into AI safety is emphasized in [S35]; multilingual evaluation frameworks that respect local contexts are described in [S34].
MAJOR DISCUSSION POINT
Safety must be re‑defined to reflect local cultural, gender, religious, and linguistic norms
AGREED WITH
Dr. Urvashi Aneja, Ambassador Philip Thigo, Ms. Chenai Chair, Mr. Amir Banifatemi
DISAGREED WITH
Mr. Abhishek Singh, Ambassador Philip Thigo, Mr. Amir Banifatemi
Argument 2
Language translation must capture lived meaning; mis‑translations can cause critical safety failures (Dr. Rachel Sibande)
EXPLANATION
She illustrates that literal translation can miss crucial contextual meanings, leading to dangerous misinterpretations in health‑related applications.
EVIDENCE
Using an example from Malawi, she describes how a phrase indicating that a pregnant woman’s water has broken could be mistranslated as “I have thrown away water,” causing a critical safety flag to be missed if the model does not understand the lived meaning [224-227].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The importance of culturally aware translation for safety-critical applications is highlighted in [S35] and reinforced by multilingual benchmark initiatives in [S34].
MAJOR DISCUSSION POINT
Language translation must capture lived meaning; mis‑translations can cause critical safety failures
AGREED WITH
Dr. Urvashi Aneja, Mr. Abhishek Singh, Ms. Natasha Crampton, Ambassador Philip Thigo, Ms. Chenai Chair
Argument 3
The Gates Foundation will institutionalize safety evaluation at the point of deployment to catch issues early (Dr. Rachel Sibande)
EXPLANATION
She states that the Gates Foundation plans to embed safety evaluation into the deployment stage of AI solutions, ensuring that harms are identified before they spread.
EVIDENCE
She says that from the foundation side they will “really institutionalize the evaluation of safety of AI solutions right at deployment because we see now that safety issues almost emerge post deployment” [355-356].
MAJOR DISCUSSION POINT
The Gates Foundation will institutionalize safety evaluation at the point of deployment to catch issues early
Argument 4
AI companionship can create psychological dependence, raising new safety concerns
EXPLANATION
Rachel highlights that users may become emotionally reliant on AI companions, potentially substituting their own cognitive abilities and decision‑making with the system’s guidance. This form of harm extends beyond technical errors to mental‑health impacts.
EVIDENCE
She describes her personal use of an AI companion as a therapist and questions at what point the user might be overly emotionally dependent, indicating a need to track such psychological effects [230-232].
MAJOR DISCUSSION POINT
Psychological dependence on AI companions as a safety risk
M
Ms. Chenai Chair
4 arguments167 words per minute683 words244 seconds
Argument 1
Companies often overlook user experience, gender dynamics, and language diversity, leading to unintended harms (Ms. Chenai Chair)
EXPLANATION
Ms. Chair highlights that AI deployments frequently ignore the diverse contexts of users, especially gender and language nuances, which can exacerbate existing inequalities and create new harms.
EVIDENCE
She provides examples such as gender-biased voice interfaces in agricultural tools, the multitude of African languages versus limited language support, and misuse scenarios like surveillance of bags, illustrating how lack of contextual design leads to gender-based violence, misinformation, and privacy violations [236-267].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The gap between compliance-focused risk assessment and human-rights-focused evaluation, especially regarding gender and language, is discussed in [S33]; broader calls for cultural and linguistic inclusivity appear in [S35].
MAJOR DISCUSSION POINT
Companies often overlook user experience, gender dynamics, and language diversity, leading to unintended harms
AGREED WITH
Dr. Urvashi Aneja, Ambassador Philip Thigo, Mr. Amir Banifatemi, Dr. Rachel Sibande
Argument 2
Africa’s linguistic diversity (2,000+ languages) is largely ignored in AI deployments (Ms. Chenai Chair)
EXPLANATION
She points out that while Africa has over two thousand documented languages, AI projects typically support only a tiny fraction, resulting in mismatched language support and ineffective tools.
EVIDENCE
She notes that Masakhane works on only 50 African languages out of more than 2,000 documented, and that even widely spoken languages like Kiswahili have regional variations that are not accounted for in deployments [249-255].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The massive linguistic diversity of Africa and its under-representation in AI systems is noted in [S35]; efforts to develop multilingual benchmarks for African languages are described in [S34].
MAJOR DISCUSSION POINT
Africa’s linguistic diversity (2,000+ languages) is largely ignored in AI deployments
AGREED WITH
Dr. Urvashi Aneja, Mr. Abhishek Singh, Ms. Natasha Crampton, Ambassador Philip Thigo, Dr. Rachel Sibande
Argument 3
Masakhane African Language Hub will deliver a benchmarking initiative for African languages this year (Ms. Chenai Chair)
EXPLANATION
The hub commits to producing a benchmark for African languages, addressing the gap identified earlier in the discussion.
EVIDENCE
She states that the hub has a benchmarking initiative underway for the year, which will contribute to African language benchmarking efforts [356].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The development of multilingual benchmark initiatives for African languages is highlighted in [S34].
MAJOR DISCUSSION POINT
Masakhane African Language Hub will deliver a benchmarking initiative for African languages this year
Argument 4
AI technologies can be repurposed for covert surveillance, violating privacy when deployed without community consent
EXPLANATION
She warns that AI‑enabled tracking devices, originally marketed for benign uses, can be inserted into personal belongings and used for surveillance without users’ knowledge. Lack of consultation amplifies privacy risks.
EVIDENCE
She recounts that AI-enabled trackers were placed in women’s and children’s bags, creating an act of surveillance that could have been mitigated if communities had been consulted before deployment [266-269].
MAJOR DISCUSSION POINT
Unconsented AI‑driven surveillance threatens privacy
M
Mr. Quintin Chou‑Lambert
2 arguments0 words per minute0 words1 seconds
Argument 1
Provides field‑tested examples to inform standards and connects local realities with international policy (Mr. Quintin Chou‑Lambert)
EXPLANATION
He argues that real‑world, context‑specific evidence from the Global South is essential to shape AI standards and ensure that international policy reflects on‑the‑ground challenges.
EVIDENCE
He explains that low-infrastructure contexts make AI safety a fuzzy discussion that needs empirical evidence, and that networks like this help bring local challenges into global dialogues, preventing them from being ignored [191-199].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The importance of bringing Global South evidence into UN-level AI governance is discussed in [S27]; bridging local insights with global policy is a core aim of the UN mechanisms described in [S25].
MAJOR DISCUSSION POINT
Provides field‑tested examples to inform standards and connects local realities with international policy
AGREED WITH
Dr. Balaraman Ravindran, Dr. Urvashi Aneja
Argument 2
One‑size‑fits‑all technical standards fail to capture contextual risks; empirical field evidence is crucial (Mr. Quintin Chou‑Lambert)
EXPLANATION
He stresses that universal technical standards cannot address the varied contexts of the Global South, and that field‑tested, empirical data is needed to create appropriate safety measures.
EVIDENCE
He notes that one-size-fits-all standards will not be contextually sensitive and that moving from large, expensive models to small, language-specific models turns safety into a fuzzy discussion that requires empirical evidence [193-195].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The critique of centralized benchmark authority and the call for context-sensitive standards are made in [S3]; the need for decentralized oversight is reinforced in [S30].
MAJOR DISCUSSION POINT
One‑size‑fits‑all technical standards fail to capture contextual risks; empirical field evidence is crucial
D
Dr. Balaraman Ravindran
3 arguments172 words per minute565 words196 seconds
Argument 1
Calls for coordination among overlapping initiatives to avoid duplication and increase impact (Dr. Balaraman Ravindran)
EXPLANATION
He highlights the proliferation of AI safety and capacity‑building networks and urges a coordinated framework to maximise resources and avoid fragmented efforts.
EVIDENCE
He mentions multiple initiatives across Africa, China, and UN-led capacity-building networks, calling for coordination to prevent each actor from working on a small piece of the puzzle and to create a multiplier effect [330-342].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The need for coordinated global AI governance and avoiding fragmented efforts is emphasized in [S25] and [S26]; inclusion of Global South perspectives is highlighted in [S27].
MAJOR DISCUSSION POINT
Calls for coordination among overlapping initiatives to avoid duplication and increase impact
AGREED WITH
Dr. Urvashi Aneja, Mr. Quintin Chou‑Lambert
Argument 2
The network will prioritize cross‑border problem solving that cannot be addressed by single‑country efforts (Dr. Balaraman Ravindran)
EXPLANATION
He proposes that the network focus on challenges requiring collaboration across borders, thereby demonstrating the added value of a shared platform beyond national initiatives.
EVIDENCE
He states that the network should target problems that necessarily require cross-border collaboration, rather than issues solvable within a single geography, and that this approach will drive the network’s relevance [353-357].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
UN-led collaborative frameworks that enable cross-border AI safety work are described in [S25].
MAJOR DISCUSSION POINT
The network will prioritize cross‑border problem solving that cannot be addressed by single‑country efforts
Argument 3
Existing AI safety institutes are scarce in the Global South; AI Centers of Excellence (AC) should serve as safety nodes
EXPLANATION
Ravindran observes that most Global South countries do not have dedicated AI safety institutes, and proposes that AI Centers of Excellence become the functional nodes for safety work and representation in global governance. This reframes the network’s architecture toward existing academic and research hubs.
EVIDENCE
He remarks that the network cannot be called a global safety institute network because such institutes are largely absent; instead, AC institutes should act as nodes that give voice to unheard communities, especially in Kenya and India [335-338].
MAJOR DISCUSSION POINT
Use AI Centers of Excellence as safety nodes in the Global South
M
Ms. Natasha Crampton
3 arguments136 words per minute404 words177 seconds
Argument 1
Scaling community‑led, multilingual evaluations requires sustainable systems and infrastructure (Ms. Natasha Crampton)
EXPLANATION
Ms. Crampton explains that while community‑led, context‑aware evaluations exist, scaling them to thousands of languages and cultural settings demands a sustainable, ongoing evaluation framework and robust infrastructure.
EVIDENCE
She references the Samishka project as an example of community-led evaluation, and stresses the need to build a system that can run multilingual and multicultural evaluations at scale, continuously, not just as a one-off test [277-284].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Sustainable multilingual evaluation infrastructures are discussed in [S34]; the need for culturally aware evaluation at scale is highlighted in [S35].
MAJOR DISCUSSION POINT
Scaling community‑led, multilingual evaluations requires sustainable systems and infrastructure
AGREED WITH
Dr. Urvashi Aneja, Mr. Abhishek Singh, Ambassador Philip Thigo, Ms. Chenai Chair, Dr. Rachel Sibande
Argument 2
Microsoft will honor New Delhi Frontier AI commitments, share multilingual data, and invest $50 bn in Global South infrastructure (Ms. Natasha Crampton)
EXPLANATION
She commits Microsoft to fulfilling the New Delhi Frontier AI pledges, providing multilingual benchmark data, and allocating substantial investment to build digital infrastructure across the Global South.
EVIDENCE
She states that Microsoft will help advance multilingual and multicultural evaluation work, share data to aid policy makers, and that Microsoft is making large infrastructure investments totaling $50 billion by the end of the decade [348-349].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The New Delhi Frontier AI commitments, which include multilingual data sharing, are outlined in [S9]; large-scale infrastructure investment aligns with UN development partnership goals noted in [S26].
MAJOR DISCUSSION POINT
Microsoft will honor New Delhi Frontier AI commitments, share multilingual data, and invest $50 bn in Global South infrastructure
Argument 3
Sustainable, ongoing evaluation mechanisms are needed rather than one‑off tests (Ms. Natasha Crampton)
EXPLANATION
She emphasizes that evaluations must be continuous to capture shifts over time, rather than being performed only once before product release.
EVIDENCE
She notes that evaluations need to be run on an ongoing basis to understand shifts, highlighting that one-off tests are insufficient for sustained safety assurance [281-284].
MAJOR DISCUSSION POINT
Sustainable, ongoing evaluation mechanisms are needed rather than one‑off tests
M
Mr. Amir Banifatemi
6 arguments169 words per minute844 words299 seconds
Argument 1
Lack of imagination and inclusion of local talent leads to blind spots in safety design (Mr. Amir Banifatemi)
EXPLANATION
He argues that system designers often lack awareness of local contexts and fail to involve talent from the regions affected, resulting in safety assessments that miss cultural, linguistic and contextual risks.
EVIDENCE
He points out that people building systems have no awareness of the contexts in which they operate, that language and culture are not captured, and that talent inclusion is missing, both in terms of skilling and integration, leading to blind spots [304-315].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The omission of local expertise and its impact on safety assessments is discussed in [S33]; cultural and linguistic blind spots are further emphasized in [S35].
MAJOR DISCUSSION POINT
Lack of imagination and inclusion of local talent leads to blind spots in safety design
AGREED WITH
Dr. Urvashi Aneja, Ambassador Philip Thigo, Ms. Chenai Chair, Dr. Rachel Sibande
Argument 2
Safety is not currently costed into financial planning, reducing incentives for firms to invest in it (Mr. Amir Banifatemi)
EXPLANATION
He notes that without financial penalties or budgeting for safety, companies lack motivation to prioritize safety measures in their product development cycles.
EVIDENCE
He explains that safety is not costed into financial systems, there is no penalty for being unsafe, and without a financial mandate, companies will not allocate resources to safety [307-311].
MAJOR DISCUSSION POINT
Safety is not currently costed into financial planning, reducing incentives for firms to invest in it
Argument 3
Absence of penalties for unsafe AI means companies lack financial motivation to prioritize safety (Mr. Amir Banifatemi)
EXPLANATION
He reiterates that without regulatory or financial penalties, firms have little incentive to invest in safety, perpetuating risk.
EVIDENCE
He again stresses that there is no penalty for unsafe AI, so companies lack financial motivation to prioritize safety [307-311].
MAJOR DISCUSSION POINT
Absence of penalties for unsafe AI means companies lack financial motivation to prioritize safety
Argument 4
Open‑source, culturally contextual incident‑reporting tools will be released to broaden evaluation access (Mr. Amir Banifatemi)
EXPLANATION
He announces that his labs in Bangalore and San Francisco are developing open‑source, culturally aware incident‑reporting tools to make evaluation resources publicly available.
EVIDENCE
He says the labs are working on safety evaluations, incident reporting, making tools culturally contextual and open-source, aiming to disseminate them to all partners and the public [358].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Open-source multilingual evaluation tools and community-driven incident reporting are highlighted in [S34].
MAJOR DISCUSSION POINT
Open‑source, culturally contextual incident‑reporting tools will be released to broaden evaluation access
Argument 5
Latency and institutional framework gaps in the Global South delay feedback loops, compounding AI‑related harms
EXPLANATION
Amir points out that the Global South lacks the rapid feedback mechanisms present in many Global North countries, leading to slower detection and mitigation of AI safety incidents. This latency, combined with weaker rule‑of‑law institutions, means harms can grow unchecked.
EVIDENCE
He notes that only a handful of Global North countries have robust legal frameworks and active civil societies that create accelerated feedback loops for incident safety, whereas most Global South nations lack these mechanisms, causing delayed responses and amplified harms [321-322].
MAJOR DISCUSSION POINT
Latency and institutional gaps delay AI safety feedback loops
Argument 6
Evaluation tools must address whole system deployment, not just model design
EXPLANATION
Amir argues that current evaluation approaches focus narrowly on model performance, overlooking the broader system context—including APIs, data pipelines, and infrastructure—that influences safety outcomes. He calls for tools that evaluate the entire deployment ecosystem.
EVIDENCE
He suggests creating a series of evaluation tools that are built for system deployment as well as model design, and highlights incident-reporting mechanisms that can capture hidden control issues, data access problems, and regulatory gaps [319-320].
MAJOR DISCUSSION POINT
Need for system‑wide AI safety evaluation tools
M
Mr. Quintin Chou-Lambert
3 arguments135 words per minute324 words143 seconds
Argument 1
Infrastructure and energy constraints in the Global South shape AI safety requirements
EXPLANATION
He points out that limited infrastructure and unreliable energy supplies in many Global South contexts mean that AI safety measures cannot be designed as if resources were abundant. Safety solutions must be adapted to these material constraints to be effective.
EVIDENCE
He notes that there is “less, perhaps, infrastructure or energy connection to go around” and that because of this, AI safety “edges into this more contextual field” where field-tested examples are needed to surface missing considerations [191-192].
MAJOR DISCUSSION POINT
Infrastructure limitations affect AI safety design
Argument 2
The move from large, expensive models to smaller, language‑specific models makes AI safety a fuzzy, context‑dependent problem
EXPLANATION
He explains that scaling down from massive, costly AI models to tailored, small‑language models changes the safety landscape, requiring new empirical evidence and evaluation methods that account for linguistic and cultural nuances.
EVIDENCE
He describes the transition as “moving from this kind of scaling a small, a very concentrated, highly expensive model across a massive user base to more tailored, small-language models” which turns AI safety into a “more fuzzy discussion” that “needs empirical evidence” [194].
MAJOR DISCUSSION POINT
Shift to small language models creates new safety challenges
Argument 3
The expanding UN Global Dialogue on AI Governance provides a venue to embed field‑tested evidence, but must ensure local threats are not ignored; open‑source developments can help broaden participation
EXPLANATION
He highlights that the United Nations Global Dialogue now includes all 193 member states and is informed by an independent scientific panel, offering an opportunity to integrate real‑world, ground‑level insights. However, he warns that without deliberate inclusion of local perspectives, international discussions may overlook threats faced by communities, and suggests open‑source tools as a way to bring those perspectives into the dialogue.
EVIDENCE
He references the trend of increasing participation from 30 to over 100 countries and the establishment of the UN Global Dialogue on AI Governance involving all member states and an independent scientific panel [195-197]. He then stresses that networks like this are crucial to “connect and bring examples of the challenges that we face” so that “international discussions do not ignore or omit or discount the perspectives of the vast majority of people on the planet” [198-199].
MAJOR DISCUSSION POINT
Leverage UN AI governance platform and open‑source to include local evidence
Agreements
Agreement Points
All participants stress the urgent need for multilingual, culturally‑aware benchmarks and evaluation frameworks to assess AI systems in the Global South.
Speakers: Dr. Urvashi Aneja, Mr. Abhishek Singh, Ms. Natasha Crampton, Ambassador Philip Thigo, Ms. Chenai Chair, Dr. Rachel Sibande
Network will build independent evidence, contextual evaluations, and act as a bridge to global governance (Dr. Urvashi Aneja) Current benchmarks are English‑centric; multilingual benchmarks are essential for accurate assessment (Mr. Abhishek Singh) Scaling community‑led, multilingual evaluations requires sustainable systems and infrastructure (Ms. Natasha Crampton) Proposes regional nodes, multilingual benchmark datasets, and an annual AI safety report (Ambassador Philip Thigo) Africa’s linguistic diversity (2,000+ languages) is largely ignored in AI deployments (Ms. Chenai Chair) Language translation must capture lived meaning; mis‑translations can cause critical safety failures (Dr. Rachel Sibande)
Speakers repeatedly highlighted that existing English-only benchmarks miss performance in the many languages spoken across the Global South, and that building multilingual, context-sensitive benchmarks and evaluation tools is essential for trustworthy AI deployment. Dr. Aneja announced multilingual benchmark work [44-45], Singh warned about English-centric tests [73-77], Crampton pledged to advance multilingual evaluation under the New Delhi commitments [348-349], the Ambassador called for multilingual datasets and regional nodes [174-175], the Chenai Chair underscored the gap of 2,000 African languages [249-255], and Dr. Sibande gave a concrete mistranslation example [224-227].
POLICY CONTEXT (KNOWLEDGE BASE)
This aligns with the launch of the Global South AI Safety Research Network, which highlighted multilingual benchmark datasets and regional nodes as priority actions [S57]. It also reflects broader calls for inclusive connectivity and culturally relevant AI evaluation noted in the IGF discussion on meaningful connectivity [S70] and Microsoft’s tech-diplomacy emphasis on multilingual resources [S61].
Broad consensus that local context, civil‑society insight, and inclusive talent are indispensable for identifying and mitigating AI risks.
Speakers: Dr. Urvashi Aneja, Ambassador Philip Thigo, Ms. Chenai Chair, Mr. Amir Banifatemi, Dr. Rachel Sibande
Independent civil society organizations are uniquely positioned to address this gap (Dr. Urvashi Aneja) Global South is systematically excluded from AI safety governance structures (Ambassador Philip Thigo) Companies often overlook user experience, gender dynamics, and language diversity, leading to unintended harms (Ms. Chenai Chair) Lack of imagination and inclusion of local talent leads to blind spots in safety design (Mr. Amir Banifatemi) Safety must be re‑defined to reflect local cultural, gender, religious, and linguistic norms (Dr. Rachel Sibande)
All speakers agreed that AI safety cannot be designed in a vacuum; it must draw on civil-society, regional expertise, and culturally aware perspectives. Dr. Aneja emphasized civil-society’s proximity to real-world deployments [19-21], the Ambassador warned of systemic exclusion and the need for regional nodes [138-141][170-176], the Chenai Chair highlighted missed gender and language nuances [236-247], Amir pointed to the absence of local talent and imagination [304-315], and Dr. Sibande called for redefining safety based on social-cultural context [216-218].
POLICY CONTEXT (KNOWLEDGE BASE)
The inclusive multi-stakeholder approach endorsed by the AI Governance Dialogue co-chairs mirrors this view [S66], and the UN-linked network’s emphasis on socio-technical issues underscores the role of civil-society and local expertise [S57][S58].
All speakers recognize the critical need for capacity‑building, compute resources, and infrastructure investments to enable effective AI safety work in the Global South.
Speakers: Mr. Abhishek Singh, Ambassador Philip Thigo, Ms. Natasha Crampton, Mr. Amir Banifatemi, Dr. Urvashi Aneja
Identifying risks is not enough; tools and benchmarks are needed to address them (Mr. Abhishek Singh) Limited access to compute resources hampers Global South researchers’ ability to evaluate models (Ambassador Philip Thigo) Microsoft will honor New Delhi Frontier AI commitments and invest $50 bn in Global South infrastructure (Ms. Natasha Crampton) Latency and institutional framework gaps in the Global South delay feedback loops, compounding AI‑related harms (Mr. Amir Banifatemi) Network will act as connective tissue between global governance and on‑the‑ground realities, building capacity (Dr. Urvashi Aneja)
Participants stressed that without technical capacity, compute power, and financial investment, safety initiatives cannot succeed. Singh called for tools and benchmarks [71-73][84-95], the Ambassador highlighted compute gaps [158-160], Crampton announced a $50 bn infrastructure pledge [348-349], Amir described latency and institutional gaps [321-322], and Dr. Aneja positioned the network as a capacity-building bridge [34-36][37-38].
POLICY CONTEXT (KNOWLEDGE BASE)
The EU 2030 Digital Agenda’s Access pillar and the IMI’s infrastructure focus provide a policy backdrop for such capacity-building [S48]; similarly, AI strategy discussions for jobs and economic development highlighted infrastructure and skills gaps as key challenges [S51]. The Global South network also calls for these investments [S57].
Consensus that AI safety should complement, not hinder, innovation and that responsible diffusion of AI must be balanced with safeguards.
Speakers: Mr. Abhishek Singh, Dr. Urvashi Aneja
Responsible AI diffusion should not be framed as stifling innovation (Mr. Abhishek Singh) Risk of amplifying existing harms without adequate safeguards (Dr. Urvashi Aneja)
Both speakers agreed that AI’s benefits must be realized while ensuring safety, rejecting the notion that safety measures impede progress. Singh explicitly said safety should not stifle innovation [105-108], while Dr. Aneja warned that unchecked AI could amplify harms [11-14].
POLICY CONTEXT (KNOWLEDGE BASE)
This balance is echoed in the Overregulation panel stressing that regulation must not stifle innovation [S53] and the AI Agents session calling for responsible deployment alongside privacy protection [S54]. Industry consensus on safe yet innovative AI standards further supports this view [S55][S73].
All participants see value in coordinated, cross‑border collaboration among the many emerging AI safety initiatives to avoid duplication and increase impact.
Speakers: Dr. Balaraman Ravindran, Dr. Urvashi Aneja, Mr. Quintin Chou‑Lambert
Calls for coordination among overlapping initiatives to avoid duplication and increase impact (Dr. Balaraman Ravindran) Network will build independent evidence, contextual evaluations, and act as a bridge to global governance (Dr. Urvashi Aneja) Provides field‑tested examples to inform standards and connects local realities with international policy (Mr. Quintin Chou‑Lambert)
Speakers highlighted the proliferation of AI safety networks and the need for a coordinated framework. Dr. Ravindran listed multiple overlapping initiatives and urged coordination [330-342], Dr. Aneja described the network as connective tissue between global and local actors [37-38], and Mr. Quintin emphasized the role of such networks in feeding field evidence into global standards [198-199].
POLICY CONTEXT (KNOWLEDGE BASE)
Coordination to avoid duplication was a key recommendation from the Digital Innovations Forum networking session [S65], and the AI Governance Dialogue highlighted the need for international cooperation and structured expert engagement [S66]. The Global South AI Safety Research Network also stresses cross-border collaboration [S57].
Similar Viewpoints
All three see the network (or their organizations) as a platform to develop and disseminate sustainable, community‑driven evaluation tools that can be scaled globally, emphasizing open‑source and infrastructural support [34-36][277-284][358].
Speakers: Dr. Urvashi Aneja, Ms. Natasha Crampton, Mr. Amir Banifatemi
Network will build independent evidence, contextual evaluations, and act as a bridge to global governance (Dr. Urvashi Aneja) Scaling community‑led, multilingual evaluations requires sustainable systems and infrastructure (Ms. Natasha Crampton) Open‑source, culturally contextual incident‑reporting tools will be released to broaden evaluation access (Mr. Amir Banifatemi)
Both highlight structural gaps – the former in governance representation, the latter in technical tooling – that prevent the Global South from effectively managing AI risks [138-141][71-73].
Speakers: Ambassador Philip Thigo, Mr. Abhishek Singh
Global South is systematically excluded from AI safety governance structures (Ambassador Philip Thigo) Identifying risks is not enough; tools and benchmarks are needed to address them (Mr. Abhishek Singh)
Both point to a systemic blind‑spot in design processes caused by insufficient inclusion of diverse user perspectives and local expertise [236-247][304-315].
Speakers: Ms. Chenai Chair, Mr. Amir Banifatemi
Companies often overlook user experience, gender dynamics, and language diversity, leading to unintended harms (Ms. Chenai Chair) Lack of imagination and inclusion of local talent leads to blind spots in safety design (Mr. Amir Banifatemi)
Unexpected Consensus
Industry (Microsoft) and diplomatic representatives (Kenyan Ambassador) both advocate for regional nodes and multilingual benchmark datasets.
Speakers: Ms. Natasha Crampton, Ambassador Philip Thigo
Scaling community‑led, multilingual evaluations requires sustainable systems and infrastructure (Ms. Natasha Crampton) Proposes regional nodes, multilingual benchmark datasets, and an annual AI safety report (Ambassador Philip Thigo)
It is notable that a corporate leader and a government envoy converge on the need for decentralized regional structures and multilingual data resources, indicating cross-sector alignment on capacity-building mechanisms [348-349][174-175].
POLICY CONTEXT (KNOWLEDGE BASE)
Microsoft’s tech-diplomacy statements describe its role in fostering regional AI hubs and multilingual resources [S61], while the Global South AI Safety Research Network explicitly calls for regional nodes and multilingual benchmarks, a position echoed by diplomatic actors [S57].
Both the UN‑linked network lead (Dr. Urvashi Aneja) and the private sector (Microsoft) stress the importance of ongoing, continuous evaluation rather than one‑off testing.
Speakers: Dr. Urvashi Aneja, Ms. Natasha Crampton
Network will build independent evidence, contextual evaluations, and act as a bridge to global governance (Dr. Urvashi Aneja) Sustainable, ongoing evaluation mechanisms are needed rather than one‑off tests (Ms. Natasha Crampton)
While the UN-focused speaker emphasizes building a continuous evidence base, the Microsoft executive explicitly calls for sustained evaluation processes, showing an unexpected alignment on the need for long-term monitoring [34-36][281-284].
POLICY CONTEXT (KNOWLEDGE BASE)
Continuous verification was highlighted by Anja Kaspersen as essential for trustworthy AI governance [S68], and the Global South network emphasizes ongoing evaluation as a core principle [S57]. Microsoft also advocates for iterative assessment in its responsible AI frameworks [S63].
Overall Assessment

The discussion reveals strong convergence around three core pillars: (1) the creation of multilingual, culturally‑aware benchmarks and evaluation tools; (2) the centrality of local civil‑society insight, inclusive talent, and contextual understanding; and (3) the necessity of capacity‑building, compute access, and financial investment to operationalise safety measures. Participants from academia, civil society, industry, and diplomacy all endorse these themes, indicating a shared vision for a coordinated, inclusive AI safety ecosystem in the Global South.

High consensus – most speakers, across sectors, articulate overlapping priorities, suggesting that future collaborative actions (regional nodes, open‑source tools, coordinated networks) have broad stakeholder buy‑in and are likely to shape policy and implementation agendas.

Differences
Different Viewpoints
Who should define and control AI safety benchmarks
Speakers: Mr. Abhishek Singh, Ambassador Philip Thigo, Mr. Amir Banifatemi
Identifying risks is not enough; tools and benchmarks are needed to address them (Mr. Abhishek Singh) Benchmark design reflects power; only a few institutions should not dictate risk definitions (Ambassador Philip Thigo) Evaluation tools must address whole system deployment, not just model design (Mr. Amir Banifatemi)
Singh calls for multilingual benchmarks to evaluate models in many languages [73-77]. The Ambassador warns that benchmarks are not neutral and should not be set by a handful of institutions, emphasizing power concentration [161-165]. Amir adds that current evaluations focus narrowly on models and lack system-wide tools, calling for broader evaluation frameworks [319-320] and noting a lack of imagination about local contexts [304-315]. These positions diverge on who should design benchmarks and how inclusive they must be.
POLICY CONTEXT (KNOWLEDGE BASE)
The debate over benchmark governance is reflected in discussions about “who watches the watchers,” where technical professional bodies like IEEE are urged to participate in AI governance and benchmark definition [S68][S64]. Calls for government and multi-stakeholder involvement in standards development also provide context [S63].
How to create incentives for AI safety compliance
Speakers: Mr. Amir Banifatemi, Mr. Abhishek Singh, Dr. Urvashi Aneja, Ms. Natasha Crampton
Safety is not currently costed into financial planning, reducing incentives for firms (Mr. Amir Banifatemi) Responsible AI diffusion should not be framed as stifling innovation (Mr. Abhishek Singh) Procurement is a lever for countries in the Global South to shape markets for responsible innovation (Dr. Urvashi Aneja) Microsoft will invest $50 bn in Global South infrastructure to support evaluation and policy work (Ms. Natasha Crampton)
Amir argues that without financial penalties or budgeting for safety, companies lack motivation to invest in safety [307-311]. Singh stresses that safety must coexist with innovation and should not hinder it [105-108]. Aneja proposes using public procurement as a policy lever to drive safe AI adoption [50-53]. Natasha commits large infrastructure investment to enable evaluation and policy support [348-349]. The speakers agree safety is needed but disagree on whether market-based procurement, regulatory penalties, or private investment are the primary driver.
POLICY CONTEXT (KNOWLEDGE BASE)
Incentive structures were addressed in the AI safety investment panel, recommending board-level responsibility and alignment of executive compensation with long-term risk mitigation [S71]. Industry-wide standards and systemic conditions that enable responsible behavior rather than punitive regulation further inform this discussion [S55][S73].
Breadth of AI safety considerations (technical vs socio‑technical and environmental)
Speakers: Mr. Abhishek Singh, Ambassador Philip Thigo, Mr. Amir Banifatemi, Dr. Rachel Sibande
Identifying risks is not enough; need technical tools and benchmarks (Mr. Abhishek Singh) AI safety must also include socio‑technical issues, environmental harms, water, etc. (Ambassador Philip Thigo) Evaluation tools must address whole system deployment, not just model design (Mr. Amir Banifatemi) Safety must be re‑defined to reflect local cultural, gender, religious, and linguistic norms (Dr. Rachel Sibande)
Singh focuses on technical risk identification and multilingual benchmarks as primary safety measures [68-73]. The Ambassador expands safety to cover environmental impacts, misinformation, and full lifecycle accountability [154-156]. Amir stresses that safety evaluation should consider the entire deployment ecosystem, not just model performance [319-320]. Rachel argues that safety definitions need to incorporate cultural and linguistic contexts [216-218]. These viewpoints differ on how broadly safety should be defined and which dimensions are essential.
POLICY CONTEXT (KNOWLEDGE BASE)
The Main Session on AI Policy Network expanded the conversation to ethical, environmental, and societal dimensions, urging a holistic view of AI safety [S56]. Virginia’s remarks on multidisciplinary governance further stress the need to go beyond purely technical metrics [S58].
Unexpected Differences
Optimism about the network’s timeliness versus perception of it being late
Speakers: Dr. Urvashi Aneja, Ambassador Philip Thigo
Network will provide visibility to real‑world impact and connect stakeholders (Dr. Urvashi Aneja) Network is timely but also late; urgency needed to scale up work (Ambassador Philip Thigo)
Aneja highlights the network’s role as connective tissue between global safety infrastructure and on-the-ground realities [37-38], expressing confidence in its impact. The Ambassador, however, characterizes the initiative as “timely but also late” and stresses an urgent need to scale up quickly [142-144], indicating a more cautious view of the network’s readiness.
Role of regulation versus voluntary industry action in driving safety
Speakers: Mr. Abhishek Singh, Mr. Amir Banifatemi
Responsible AI diffusion should not be framed as stifling innovation (Mr. Abhishek Singh) Safety is not costed into financial planning; lack of penalties reduces firm motivation (Mr. Amir Banifatemi)
Singh argues that safety measures should complement innovation and not be seen as restrictive [105-108]. Amir counters that without regulatory penalties or financial mandates, companies will deprioritize safety altogether [307-311], suggesting a need for stronger enforcement rather than purely voluntary action.
POLICY CONTEXT (KNOWLEDGE BASE)
Panels on overregulation and responsible deployment argue for a balanced approach where regulation sets guardrails but does not impede innovation, complemented by voluntary industry measures [S53][S54]. The EU GPAI Code discussion highlighted systemic conditions that enable companies to act responsibly without heavy-handed rules [S73], and data-sovereignty talks emphasized co-accountability partnerships between governments and industry [S69].
Overall Assessment

The discussion reveals broad consensus on the necessity of a Global South network for trustworthy AI, but significant divergences arise around benchmark governance, incentive structures, and the scope of safety. While speakers align on goals—building evidence, fostering capacity, and ensuring inclusive governance—their preferred pathways (regional nodes, corporate investment, open‑source tools, procurement levers, or regulatory penalties) differ markedly. These disagreements highlight challenges in harmonising technical standards, financing mechanisms, and interdisciplinary safety definitions across diverse stakeholders.

Moderate to high: The core objective is shared, yet the lack of agreement on implementation strategies and the breadth of safety considerations could impede coordinated action unless reconciled. The implications are that without a unified approach to benchmarks, incentives, and scope, the network may face fragmentation, slower adoption of standards, and uneven protection for vulnerable populations.

Partial Agreements
All speakers concur that a coordinated Global South network is essential for trustworthy AI, but they differ on the primary mechanism: Aneja emphasizes evidence generation and governance linkage [22-23]; Singh stresses compliance and capacity‑building [87-95]; the Ambassador calls for regional nodes and reporting structures [170-176]; Natasha focuses on corporate commitments and infrastructure investment [348-349]; Amir proposes open‑source tooling and incident reporting [358].
Speakers: Dr. Urvashi Aneja, Mr. Abhishek Singh, Ambassador Philip Thigo, Ms. Natasha Crampton, Mr. Amir Banifatemi
Network will build independent evidence, contextual evaluations, and act as a bridge to global governance (Dr. Urvashi Aneja) Network enables compliance with New Delhi Frontier AI commitments and capacity‑building (Mr. Abhishek Singh) Proposes regional nodes, multilingual benchmark datasets, and an annual AI safety report (Ambassador Philip Thigo) Microsoft will honor New Delhi Frontier AI commitments, share multilingual data, and invest $50 bn in Global South infrastructure (Ms. Natasha Crampton) Open‑source, culturally contextual incident‑reporting tools will be released to broaden evaluation access (Mr. Amir Banifatemi)
Takeaways
Key takeaways
AI deployment in the Global South presents huge opportunities but also significant risks of amplifying existing social, gender, linguistic, and environmental harms. Identifying risks is insufficient; concrete tools, benchmarks, and capacity‑building are needed to evaluate and mitigate those risks. The Global South is systematically under‑represented in AI safety governance; a dedicated network can provide independent, field‑tested evidence and act as a bridge to global policy forums. Current evaluation benchmarks are English‑centric and power‑concentrated; multilingual, culturally aware benchmarks are essential for trustworthy AI in diverse contexts. Capacity gaps—including limited compute resources, talent inclusion, and sustainable evaluation mechanisms—must be addressed to enable effective safety work. Governance mechanisms must be de‑centralised; benchmarks and standards should not be defined by a handful of institutions, and safety should be financially incentivised. Collaboration across overlapping initiatives (UN, regional networks, industry, civil society) is critical to avoid duplication and maximise impact.
Resolutions and action items
Launch of the Global South Network for Trustworthy AI as a coordinating platform for civil‑society, research, and policy actors. Establish regional nodes (e.g., African node) to decentralise activities and increase local relevance. Develop multilingual benchmark datasets and conduct an annual Global South AI Safety Report. Align network activities with the New Delhi Frontier AI commitments, including sharing usage data and multilingual performance benchmarks. Microsoft to honor its Frontier AI commitments, share multilingual data, and invest $50 bn in infrastructure across the Global South by 2030. The Gates Foundation will institutionalise safety evaluation at the point of deployment to capture issues early. Masakhane African Language Hub will deliver a benchmarking initiative for African languages within the year. Open‑source, culturally contextual incident‑reporting tools will be created and made publicly available (led by Amir’s labs). Coordinate with existing UN AI governance processes (UN Global Dialogue on AI Governance, scientific panel) to ensure Global South voices are included. Facilitate cross‑border problem‑solving projects that require collaboration beyond single‑country efforts.
Unresolved issues
Precise definition of ‘safety’ and ‘harm’ that reflects varied cultural, gender, religious, and linguistic contexts remains open. Mechanisms to financially cost safety into corporate planning and to impose penalties for unsafe AI have not been established. Sustainable, ongoing evaluation frameworks (beyond one‑off tests) need concrete design and funding models. How to ensure equitable access to compute resources for Global South researchers is still undetermined. Details on how the network will integrate with and influence UN‑led AI governance structures are pending. Strategies for de‑concentrating benchmark authority and preventing power imbalances in standard‑setting are not fully resolved. Methods to close the accountability loop so that citizen‑level impacts are directly addressed were discussed but not finalized.
Suggested compromises
Create regional nodes to balance the need for rapid network activation with the requirement for local contextual expertise. Adopt a shared, open‑source benchmarking framework that allows multiple institutions to contribute, mitigating concentration of power. Leverage existing commitments (New Delhi Frontier AI) as a common baseline while expanding them through collaborative, multilingual evaluation work. Combine top‑down UN engagement with bottom‑up civil‑society evidence generation to satisfy both global governance and local relevance. Use pilot projects and incremental infrastructure investments (e.g., Microsoft’s $50 bn) as stepping stones toward broader, sustainable evaluation systems.
Thought Provoking Comments
Across the Global South, AI systems are being rapidly deployed in critical social sectors … while the potential is immense, the risks and harms are also immense. It is particularly important that we figure out ways to make AI safe and trustworthy in these contexts to ensure we protect populations and build infrastructure for safe and inclusive AI adoption.
Sets the foundational problem statement, highlighting the paradox of high opportunity versus low institutional capacity, and frames the need for a dedicated network.
Established the urgency of the discussion, prompting subsequent speakers to propose concrete mechanisms (benchmarks, regional hubs, civil‑society involvement) to address the identified gap.
Speaker: Dr. Urvashi Aneja
Identifying the risk is not sufficient. We need to think of how do we address those risks. For that we need technical tools, benchmarks, especially multilingual benchmarks, because most models are evaluated only in English.
Moves the conversation from risk identification to actionable solutions, emphasizing multilingual evaluation as a concrete technical need.
Shifted the dialogue toward practical steps (benchmarks, capacity building) and reinforced the network’s relevance; later speakers referenced multilingual benchmarks as a priority.
Speaker: Mr. Abhishek Singh
We are the only member of the international network of AI safety institutes from the Global South. The model that is not inclusive to a global majority is not acceptable. There are four structural gaps: teaming capacity, access to compute, linguistic and cultural mismatch, and the non‑neutrality of benchmarks.
Highlights systemic inequities and enumerates specific structural gaps, challenging the audience to consider power dynamics and resource asymmetries.
Created a turning point by broadening the scope from technical benchmarks to governance, power, and agency; later participants (e.g., Amir, Rachel) expanded on cultural safety and sovereign capability.
Speaker: Ambassador Philip Thigo
We need to redefine what is safe and what is harmful according to the social‑cultural context … language is not just vocabulary, it’s lived meaning. Example: a pregnant mother saying ‘waters have broken’ could be mistranslated and the model would miss a critical health flag.
Provides vivid, real‑world examples that illustrate how current evaluation metrics miss contextual harms, especially in health and language nuances.
Deepened the conversation about the limits of existing benchmarks and spurred others (Chenai Chair, Natasha) to discuss user experience, gendered voice, and the need for context‑aware evaluation.
Speaker: Dr. Rachel Sibande
What often gets missed is the user experience and the diversity of the community. A voice‑enabled agricultural tool with a male‑sounding voice can exacerbate gender‑based violence if the community wasn’t consulted.
Links technical design choices (voice gender) to social harms, illustrating unintended consequences of poorly contextualized AI deployments.
Shifted the tone toward concrete design pitfalls and reinforced the call for participatory design; prompted further discussion on surveillance and misuse.
Speaker: Ms. Chenai Chair
The challenge is scaling community‑led, context‑aware evaluations sustainably. You can’t do a one‑off test; you need an ongoing system that can run at scale across thousands of languages and cultural settings.
Identifies the scalability and sustainability problem of evaluation, moving the conversation from theory to operational feasibility.
Guided the panel toward discussing infrastructure investments and the need for systematic, repeatable evaluation pipelines; later echoed in Amir’s remarks about tooling and incident reporting.
Speaker: Ms. Natasha Crampton
Safety is not costed into financial systems; there is no penalty for being unsafe. Without financial incentives or regulatory mandates, companies will not prioritize safety.
Points out a fundamental economic barrier to safety, challenging the assumption that good intentions alone will drive responsible AI.
Introduced a new dimension—economic incentives—into the debate, prompting later suggestions about integrating safety into budgeting and policy (e.g., procurement lever mentioned by Urvashi).
Speaker: Mr. Amir Banifatemi
There are too many parallel initiatives; we need coordination and a harmonized framework so that efforts are not duplicated and resources can be pooled.
Raises the meta‑issue of ecosystem fragmentation, urging strategic alignment across networks and funders.
Steered the conversation toward collaboration mechanisms, influencing the rapid‑fire round where participants mentioned concrete joint actions (benchmarking, incident reporting, UN coordination).
Speaker: Dr. Balaraman Ravindran
AI standards as technical standards don’t solve the issue because a one‑size‑fits‑all standard will not be contextually sensitive. We need empirical evidence from low‑resource, field‑tested examples.
Challenges the reliance on universal technical standards and underscores the necessity of context‑specific evidence.
Reinforced earlier points about cultural and linguistic mismatch, supporting the call for regional nodes and localized benchmarks.
Speaker: Mr. Quintin Chou‑Lambert
Overall Assessment

The discussion was shaped by a series of pivotal remarks that moved the conversation from a high‑level problem statement to concrete, multidimensional solutions. Dr. Aneja’s opening framed the urgency, while Mr. Singh introduced actionable benchmarks. Ambassador Thigo’s enumeration of structural gaps broadened the lens to include power and resource inequities, prompting participants to surface real‑world examples (Rachel, Chenai) that illustrated cultural and gendered harms. Technical scalability concerns (Natasha) and economic incentives (Amir) added layers of operational complexity, and Dr. Ravindran’s call for coordination highlighted the risk of fragmented efforts. Together, these comments redirected the dialogue toward actionable, context‑aware, and collaborative pathways, culminating in a rapid‑fire round where each participant committed to concrete steps. The key comments thus acted as turning points that deepened analysis, shifted perspectives, and forged a shared agenda for the Global South Network.

Follow-up Questions
How can we ensure that evaluation work captures the societal, ethical, and distributional harms specific to Global South contexts?
She highlighted the need for evaluations to go beyond technical metrics and reflect real‑world risks in low‑capacity, inequitable settings.
Speaker: Dr. Urvashi Aneja
How can we enable compliance with the New Delhi Frontier AI commitments, particularly regarding data sharing and multilingual performance benchmarks?
He asked how to operationalize the commitments made by AI developers to share usage data and benchmark multilingual performance.
Speaker: Mr. Abhishek Singh
What tools and processes are needed to evaluate AI models in diverse languages and build capacity across Global South countries?
He emphasized the lack of linguistic benchmarks and the need for capacity‑building to assess models in many local languages.
Speaker: Mr. Abhishek Singh
What steps are required to make the Global South Network for Trustworthy AI functional and to secure necessary support from all stakeholders?
He raised concerns about moving from launch to actionable, sustainable operations with stakeholder buy‑in.
Speaker: Mr. Abhishek Singh
How can we prevent benchmarks from being neutral or dominated by a handful of institutions, ensuring diverse risk priorities and deconcentrating power?
He warned that benchmark design can embed power imbalances and called for broader, inclusive definition of risks.
Speaker: Ambassador Philip Thigo
Should the network establish regional nodes (e.g., across Africa) and how might that be organized?
He suggested creating sub‑regional hubs to better address the diversity of contexts within the Global South.
Speaker: Ambassador Philip Thigo
How can multilingual benchmark datasets be developed and an annual red‑team exercise be instituted for the Global South?
He proposed concrete mechanisms—datasets and red‑teamings—to continuously test models in local languages.
Speaker: Ambassador Philip Thigo
Can a Global South AI Safety Report be published that adopts an expansive definition of safety?
He recommended producing a regular report to synthesize findings and set a broader safety agenda.
Speaker: Ambassador Philip Thigo
How should the network’s work be integrated into multilateral processes such as the UN AI governance panels?
He asked how the network can feed its evidence into existing global governance structures.
Speaker: Ambassador Philip Thigo
How do we close the accountability loop so that evaluations translate into tangible benefits for citizens?
He highlighted the risk that technical assessments may not reach end‑users without clear pathways to impact.
Speaker: Ambassador Philip Thigo
What sustainable, community‑led system can be built to scale multilingual and multicultural evaluations globally?
She identified the challenge of turning deep, context‑aware pilots into a repeatable, large‑scale evaluation infrastructure.
Speaker: Ms. Natasha Crampton
What internal and external ecosystem changes are needed for grounded evaluations to become standard practice in industry?
He asked how companies and the broader ecosystem must evolve for context‑sensitive safety assessments to be routine.
Speaker: Mr. Amir Banifatemi
How can safety be incorporated into financial planning and create penalties or incentives for unsafe AI?
He noted that without financial stakes, companies lack motivation to prioritize safety.
Speaker: Mr. Amir Banifatemi
How can talent inclusion be improved so that diverse voices are part of safety conversations and tool development?
He pointed out the current lack of representation of people who understand local contexts in safety work.
Speaker: Mr. Amir Banifatemi
How can compute access gaps for Global South researchers evaluating models be addressed?
He identified limited access to high‑performance compute as a structural barrier to evaluation.
Speaker: Ambassador Philip Thigo
How can the many AI safety capacity‑building initiatives and networks be coordinated and harmonized globally?
He observed a proliferation of overlapping efforts and called for a coordinated framework.
Speaker: Prof. Balaraman Ravindran
What cross‑border problems require collaboration across geographies, and how can the network prioritize them?
He suggested focusing on issues that cannot be solved within a single country to drive genuine collaboration.
Speaker: Prof. Balaraman Ravindran
How can learning capabilities be accelerated, incident reporting be improved, and open‑source safety evaluation tools be disseminated?
He highlighted opportunities to build tooling and reporting mechanisms that capture contextual harms quickly.
Speaker: Mr. Amir Banifatemi
How can the evaluation of AI safety be institutionalized at the point of deployment rather than after harms emerge?
She emphasized the need for safety checks to be built into deployment workflows.
Speaker: Dr. Rachel Sibande
How can African language benchmarking initiatives be expanded beyond the current limited set of languages?
She noted that Masakhane covers only ~50 of 2,000 documented African languages, leaving many unserved.
Speaker: Ms. Chenai Chair
How can we prevent AI tools from becoming surveillance technologies without informed consent?
She gave examples of tracking devices being misused, underscoring the need for consent‑driven design.
Speaker: Ms. Chenai Chair
How can the environmental footprints of AI models be evaluated and accounted for throughout their lifecycle?
He called for full‑lifecycle accountability, including water and environmental impacts, in safety assessments.
Speaker: Ambassador Philip Thigo

Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.