AI for Good Technology That Empowers People

20 Feb 2026 10:00h - 11:00h

AI for Good Technology That Empowers People

Session at a glanceSummary, keypoints, and speakers overview

Summary

The session opened with Speaker 1 introducing Fred Werner, Chief of Strategic Engagement at the ITU, to give opening remarks [1-3]. Fred framed AI as potentially the last human invention and argued that ensuring AI serves “good” requires proactive governance, citing a conversation with AI-safety expert Roman Jampolsky [4-12]. He outlined the evolution of the ITU’s AI for Good initiative from its 2017 launch-initially hype-driven-to today’s focus on generative AI, autonomous agents, robotics, brain-computer interfaces and space computing, noting a “zero-click” world where agents act on our behalf [13-22]. Emphasising the breadth of applications, Fred listed sectors such as affordable healthcare, education, food security and disaster response as key use-cases for AI for Good [23-24]. He described AI for Good as a year-long movement built on three pillars-solutions, skills and standards-and highlighted over 400 AI standards in development, including work on future networks and AI-native architectures relevant to the session’s edge-AI theme [55-71].


Brijesh Lal then presented IT Delhi’s edge-AI research, focusing on haptic feedback, split-control architectures and intent-based signal conversion to enable low-latency, safety-critical applications [87-112]. He also reported collaborative technical reports from TSDSI on dynamic AI models for V2X, security, digital twins and AI-native 6G RAN, underscoring the importance of global standard forums such as ITU-R IMT 2030 and CGPP for edge development [115-133].


Ranjitha Prasad followed with a technical overview of federated learning, explaining how data explosion and sub-10 ms latency requirements in 6G networks drive edge-centric training to preserve privacy and reduce bandwidth use [136-152]. She illustrated two use-cases-traffic-prediction during a football event and V2X road-condition sharing-showing how federated models run at the edge while only metadata is sent to the cloud, and noted her role leading the Intellicom lab’s collaboration with IT Delhi [153-166].


The panel, moderated by Fred, began with Mala Kumar describing XR-assisted medical emergency care using public 5G and private 5G for on-premise HCI, and expressed a desire to open-source these solutions through AI for Good platforms [172-188]. Alagan Mahalingam recounted deploying edge AI for small-scale farmers in Portugal and Sri Lanka, adapting hardware such as Raspberry Pi to deliver AI services where connectivity is limited, and emphasized model quantisation to fit edge constraints [200-236]. Sakshi Gupta of Qualcomm highlighted the shift of inference to devices-from smartphones to cars and IoT-pointing to on-device large models and Qualcomm’s “Tech for Good” program that supports startups like India’s Raksa Health with edge AI health assistants [246-286].


Ambassador Egriselda Lopez summarized that “HAI” means placing AI close to people and services, improving speed, cost and privacy in low-connectivity settings, and called for continued cooperation to avoid fragmented approaches [313-317]. Ambassador Reintam Saar outlined the upcoming UN Global Dialogue on AI Governance, stressing inclusive, outcome-oriented discussions that will build on the practical insights shared during the panel, thereby linking standards work with real-world impact [338-350].


Keypoints


Major discussion points


AI for Good – its evolution, goals and operating pillars – Fred introduced the AI for Good programme, noting its start in 2017, the shift from hype to concrete solutions, the rise of generative AI and autonomous agents, and its mission to “unlock AI’s potential to serve humanity” [4-13][14-21][24-28]. He later clarified that AI for Good is a year-long ecosystem built around three pillars - solutions, skills and standards – with hackathons, machine-learning challenges and a standards portfolio of over 400 AI standards [55-60][68-70].


Edge AI as a catalyst for the Global South – Multiple speakers highlighted why edge computing is essential for low-latency, privacy-preserving and context-aware AI. Brijesh described the convergence of communication, compute and control and the need for strong edge capabilities for haptics and V2X [90-99]. Ranjitha explained federated learning as a way to keep data at the edge, improving privacy, latency and bandwidth for telecom use-cases [136-144]. Alagan gave concrete examples such as AI-enabled soil-sensing and farmer advisory systems that were adapted to offline villages in Sri Lanka using edge devices [200-215]. Mala showcased XR-assisted medical emergency care that leverages private and public 5G edge networks [172-182]. Sakshi described Qualcomm’s on-device AI, edge-cloud and automotive deployments, and highlighted startup pilots like India’s Raksa Health that run AI locally [246-286].


Standards, collaboration and capacity-building across UN agencies and industry – Fred emphasized the role of the ITU and its 50 UN sister agencies in co-creating AI standards and standards-related work on future networks, AI-native RAN and edge AI [68-70][29-31]. He later noted that the “standards work … will emerge … to make this work at scale” [292-296]. Panelists referenced existing standardisation frameworks (e.g., ITU-IMT 2030, CGPP, M2M) and the need for inclusive, interoperable specifications [130-133][338-345].


Human-centred AI governance and the upcoming Global AI Governance Dialogue – The Ambassador of El Salvador stressed that AI must stay “close to people, services, communities” and protect privacy while delivering speed and cost benefits [309-317]. She called for people to remain at the centre, for closing the digital divide, and for avoiding fragmented approaches [322-330]. Ambassador Reintam outlined the mandate of the first UN Global AI Governance Dialogue, stressing inclusivity, capacity-building and actionable outcomes [338-345]. Throughout, speakers linked these governance goals back to the AI for Good ethos of putting humanity first [45-48].


Overall purpose / goal of the discussion


The session was convened to showcase how the AI for Good programme is mobilising research, standards-development and multi-stakeholder collaboration to harness edge AI for solving concrete societal challenges-especially in the Global South-while embedding those efforts within a human-centred governance framework that the UN will advance through its upcoming Global AI Governance Dialogue.


Overall tone and its evolution


– The opening remarks were formal and optimistic, framing AI for Good as a visionary movement.


– As technical speakers took the floor, the tone shifted to informative and pragmatic, focusing on specific edge-AI use-cases, challenges, and engineering solutions.


– During the panel, the conversation became collaborative and solution-oriented, with participants sharing real-world deployments and emphasizing standards and open-source sharing.


– The closing remarks adopted a hopeful and inclusive tone, stressing human-centred values, the need for global cooperation, and the promise of forthcoming governance dialogues. Throughout, the tone remained constructive, moving from high-level vision to concrete action and back to a unifying call for collective responsibility.


Speakers

Fred Werner – Chief of Strategic Engagement Department, ITU; moderator and panelist; expertise in AI for Good, AI standards, edge AI. [S18][S16]


Speaker 1 – Host/moderator of the session; role not specified.


Brijesh Lal – Professor; former Bharti School Chairman; researcher focusing on edge AI, haptics, and global-south initiatives. [S8]


Mala Kumar – Technologist at the Center of Excellence Wired and Wireless Technologies, Art Park; former post-doctoral researcher at Technical University Berlin; visiting researcher at UC Davis and TU Berlin; works on AI-enabled XR applications.


Alagan Mahalingam – Founder, CEO, and Chief Software Architect of RootCode; ICT Entrepreneur of the Year 2021; Young Entrepreneur of the Year 2024; Envoy for Estonia e-residency. [S1]


Sakshi Gupta – Global Government Affairs lead for Qualcomm; tech-policy professional focusing on AI, emerging technologies, market research and stakeholder engagement.


Ambassador Egriselda Lopez – Her Excellency, Ambassador, Permanent Representative of the Republic of El Salvador to the United Nations Office and other International Organizations in Geneva.


Ambassador Reintam Saar – Co-chair of the UN Global Dialogue on AI Governance; responsible for organizing the dialogue and producing summary reports. [S2]


Ranjitha Prasad – PhD researcher specializing in causal inference, survival analysis, Bayesian neural networks, and federated learning; Principal Investigator of the Intellicom Lab at IIIT Delhi. [S11]


Additional speakers:


Vijay Singh – Mentioned in the introduction; no role or title provided.


Vishnu ji – Referenced as a host/introducer in the transcript; no specific role or title provided.


Full session reportComprehensive analysis and detailed insights

The session opened with Speaker 1 thanking the audience and quickly introducing Fred Werner, Chief of Strategic Engagement at the ITU, who delivered the opening remarks [1-3]. Fred began with a provocative question – “What if the last thing that humans ever invent is invention itself?” – and recounted a conversation with AI-safety expert Roman Jampolsky about whether AI should be “for good” or “for good, forever” [4-12]. He used this dialogue to stress that, if AI becomes humanity’s final invention, it must be deliberately guided toward beneficial outcomes.


Fred traced the evolution of the ITU’s AI for Good programme. Launched in 2017, the initiative moved from an early focus on hype and “fear, promise and hype” [15-16] to a concrete effort that now embraces generative AI, autonomous agents, robotics, brain-computer interfaces and even space-based computing [18-23]. He described a “zero-click” world where agents act without explicit prompts [20-21] and highlighted the breadth of societal challenges AI can address, from affordable healthcare to food security and disaster response [24-25]. Fred underscored that AI for Good cannot succeed in isolation; it relies on partnerships with more than 50 UN sister agencies that contribute expertise, drive standards work and foster cooperation on AI governance [28-31].


Fred noted that AI for Good is a year-long movement and global community, not just an annual summit, organised around three pillars-solutions, skills and standards-each supporting concrete activities such as machine-learning challenges, the AI Skills Coalition sandbox, and over 400 emerging AI standards for future networks and AI-native architectures [55-60][61-63][63-66][68-71].


Brijesh Lal from IT Delhi then presented his research on edge-AI, beginning with the convergence of communication, compute and control that makes edge capability essential for safety-critical applications such as haptics [90-95]. He argued that latency-sensitive haptic feedback cannot tolerate errors, so strong edge processing is required [96-99]. Brijesh described a “split-control” architecture that moves substantial processing from the cloud to the edge, and an “intent-based” signal conversion that abstracts raw pressure data into higher-level commands [106-111]. He also highlighted collaborative technical reports from TSDSI on dynamic AI models for V2X, security-enhanced digital twins and AI-native 6G RAN, and pointed to standard-setting bodies such as ITU-R IMT 2030, ITU-T CGPP and M2M as crucial forums for global edge development [115-133][130-133].


Ranjitha Prasad, Principal Investigator of the Intellicom Lab at IIT Delhi and collaborator with the ITU-Delhi team, provided a technical overview of federated learning (FL) as an enabler of edge-centric AI [??-??]. She linked the exponential growth of mobile data traffic and the sub-10 ms latency requirements of 6G services to the need for privacy-preserving, distributed intelligence [136-144]. FL brings the code to the data, allowing training to occur locally while only model updates are shared with the cloud, thereby reducing bandwidth and safeguarding user privacy [139-144]. Ranjitha illustrated two concrete use cases: traffic-prediction during a football-match event, where edge base stations aggregate local traffic before a MEC controller optimises routing [147-152]; and V2X road-condition sharing, where each vehicle communicates with a local edge server before contributing to a global model [153-166].


The panel, moderated by Fred, opened with Mala describing XR-assisted medical emergency care. Using public 5G, first-responders equipped with XR glasses and IoT wearables receive real-time vitals overlaid on video, enabling remote medical experts to guide CPR or AED deployment [172-180]. A separate private 5G deployment supports on-premise XR tours for Industry 5.0 applications [182-186]. Mala expressed a desire to open-source these solutions through the AI for Good sandbox so that the international community can test, fine-tune and scale them [187-188].


Alagan Mahalingam then shared real-world edge-AI deployments. He recounted a farmer-advisory system built for Portugal that combines soil-sensing hardware, mobile-app image analysis and AI models to advise small-scale growers [204-207]. When the solution was trialled in a remote Sri Lankan village with unreliable connectivity, the team introduced a Raspberry Pi edge node running lightweight models (e.g., GemR) to retain functionality offline [208-214]. He also described a “tuk-tuk data-center” concept-edge compute mounted on a mobile vehicle to serve rural villages-illustrating creative deployment ideas [??-??]. This experience reinforced his view that edge is indispensable not only in the Global South but also in well-connected regions when users move out of coverage, as illustrated by a remote-patient-monitoring service in the United States [224-227]. Alagan stressed a “task-first” design philosophy: start from the specific problem, then distil or quantise models to fit edge constraints, avoiding unnecessary large-language-model deployments [230-236].


Sakshi Gupta of Qualcomm discussed why AGI is important for the Global South, highlighting key considerations-latency, security, privacy, personalization, cost and power-that affect both AGI and edge-AI deployments. She noted that modern smartphones already run on-device models with up to ten billion parameters, enabling AI use even in flight mode [262-266], and that similar capabilities are emerging in cars, IoT devices and smart glasses [267-270]. When Fred asked for concrete evaluation metrics for edge AI, Sakshi responded by outlining these considerations but did not propose specific metrics [??]. She also described Qualcomm’s “Tech for Good” programme, which mentors startups such as India’s Raksa Health that have built on-device AI health assistants capable of offline symptom checking and prescription lookup [284-286].


Fred linked these examples back to standards work, observing that the proliferation of edge AI creates a need for new specifications on hardware availability, connectivity quality, privacy safeguards and data handling [240-246]. He warned that without appropriate standards, scaling these solutions will be difficult, yet he also acknowledged the urgency of fast-track deployments that address immediate needs [292-296].


Ambassador Egriselda Lopez, based in New York, framed the discussion in human-centred terms, defining “HAI” as AI placed close to people, services and communities, which improves speed, reduces cost and enhances privacy in low-connectivity settings [313-318]. She reiterated three policy messages: keep people at the centre of AI development, provide decisive support to close the digital divide, and avoid fragmented national approaches by fostering cooperation [322-330].


Ambassador Reintam Saar outlined the forthcoming UN Global AI Governance Dialogue. He explained that the dialogue will bring together governments and multi-stakeholder groups to exchange best practices, focus on practical outcomes, align with existing UN processes and avoid duplication [338-345]. Capacity-building, trust, transparency and a human-rights grounding were identified as core principles, and the dialogue will draw on the “wisdom” of participants to produce a roadmap for future action [346-350].


Across the session, participants reached strong consensus on several points. All agreed that edge AI is essential for delivering low-latency, privacy-preserving services in both underserved and well-connected environments [91-95][202-215][260-266]; that task-driven, lightweight models are preferable to large generic LLMs for edge deployment [230-236][262-264]; that privacy preservation justifies moving processing to the edge [139-144][274-279][313-318]; and that AI for Good must be pursued through inclusive, multi-stakeholder governance [26-27][338-345][55-60].


However, the discussion also revealed disagreements. Alagan argued that edge solutions should rely on quantised, task-specific models and avoid large LLMs [230-236], whereas Sakshi highlighted that current smartphones already host 10-billion-parameter models on-device [262-264], illustrating a tension between perceived resource constraints and actual hardware capabilities. Sakshi responded to Fred’s request for concrete evaluation metrics by outlining key considerations but did not provide specific metrics [??]; this request remained unaddressed by other speakers who focused on use-cases and standards [68-70][230-236]. Finally, Fred emphasised the importance of standards development for AI-native networks [68-70], while Alagan’s fast-track, task-first deployment approach implied that waiting for formal standards could delay impact [230-236].


Key take-aways


(i) AI for Good’s overarching goal is to unlock AI’s potential for humanity through a continuous ecosystem of solutions, skills and standards [55-60][S1];


(ii) edge AI is a critical enabler because the convergence of communication, compute and control permits safety-critical and context-specific applications, especially in the Global South [90-99][200-215];


(iii) practical edge-AI use cases demonstrated include XR-assisted emergency care, farmer advisory systems, remote patient monitoring, traffic-prediction and V2X services [172-188][200-207][224-227][147-152];


(iv) task-driven, lightweight models-achieved via quantisation, pruning or distillation-are preferred over large foundation models for edge deployments [230-236];


(v) federated learning provides a privacy-preserving pathway to edge training while reducing bandwidth and latency [139-144][145-152];


(vi) the ITU is advancing standards for AI-native networks, future 5G/6G architectures and multimodal QoE, laying the groundwork for scalable edge AI [68-70][71];


(vii) human-centred AI (HAI) that brings intelligence close to people improves speed, cost and privacy [313-318];


(viii) the upcoming UN Global AI Governance Dialogue will focus on inclusive, outcome-oriented discussions, capacity-building and alignment with existing UN processes [338-345].


Unresolved issues


– Defining concrete metrics and benchmarks for evaluating edge-AI deployments (e.g., latency thresholds, hardware availability, privacy safeguards).


– Ensuring interoperability of heterogeneous edge devices and haptic interfaces across manufacturers.


– Establishing sustainable funding and business models for scaling edge solutions in under-connected regions.


– Clarifying the balance between on-device training versus cloud-based training in federated learning, especially for large-scale models.


– Creating mechanisms for coordinated data sharing and knowledge transfer among UN agencies, national governments and private-sector partners [Agreements][Disagreements].


Suggested compromises


Adopt a task-centric approach that designs lightweight, purpose-built models for edge deployment; combine cloud-based heavy training with edge-based inference and federated updates to preserve privacy; leverage open-source repositories within the AI for Good sandbox to enable community testing and rapid iteration; pilot regional edge solutions (e.g., in India, Sri Lanka, Portugal) that can be adapted elsewhere; and align standards development with existing UN frameworks to avoid duplication while running in parallel with fast-track pilots [Suggested compromises].


In closing, the session reaffirmed the collective commitment to advance edge AI as a means of realising the AI for Good vision, to develop inclusive standards and governance structures, and to continue collaborative research and capacity-building activities throughout the year. The speakers thanked the participants and invited them to contribute further to the forthcoming Global AI Governance Dialogue in July 2024 [338-350].


Session transcriptComplete transcript of the session
Speaker 1

Thank you. Thank you very much. We have very little time, so I want to first of all introduce Fred. Fred Werner is the Chief of Strategic Engagement Department at ITU Welcome Fred to give the opening remarks

Fred Werner

Hello Let me start with a question What if the last thing that humans ever invent is invention itself Now what do I mean by this? We had, if you’re familiar with Roman Jampolsky He’s a leading AI safety expert And I met him in New York at the UNGA last fall And he said, Fred, what is AI for good? I said, well, what do you mean? He said, well, is it for good or for good? Well, what do you mean? And he said, well, for good as in beneficial, as in good Or as in for good, forever I said, hmm, good point And he said, what if AI is the last thing that humans ever invent?

Now, you might agree or disagree with that statement, but it’s not hard to imagine a future where most future inventions will either be invented by an AI or with the help of an AI. And if that is the case, then I think we do need to make sure that AI, if it’s going to be for good, is indeed for good. So my name’s Fred Werner from the ITU. It’s the United Nations Specialized Agency for Digital Technologies, and we’re also the organizers of AI for Good with 50 -plus UN sister agencies. Now, AI for Good was created in 2017, and if you think about that, that’s basically an eternity in terms of AI years, looking at how fast it’s been developing.

And back then, it was really all about the fear and the promise and the hype of AI. Most solutions existed in fancy PowerPoint slides, but there wasn’t a whole lot of substance. But that changed rather quickly. In 2023, we saw the advent of generative AI. Last year, the unofficial theme of the summit was the rise of the AI agents. And now we’re looking at a world where you’re basically entering a zero -click world where agents are not waiting for our prompts. They’re actually acting on our behalf. And in addition, you have the physical embodiment of AI in the form of robotics, embodied AI, brain -computer interfaces, and we’re even looking at space AI computing now. Now, so I think we’re safe to say there’s no shortage of high -potential AI use cases that can be used to help solve global challenges.

Anything from affordable healthcare to education for all, food security, disaster response, the use cases are definitely there. So what is the goal of AI for Good? Well, simply put, it’s to unlock AI’s potential to serve humanity. And how do we do this? Well, first of all, we can’t do this alone. Nobody can. That’s why we have AI. We have 50 UN sister agencies as partners of AI for Good, contributing knowledge, sharing expertise, helping to drive our standards work. building cooperation around AI governance and we’re very privileged to have here the two co -chairs and facilitators of the UN AI Global Dialogue who will be doing the closing remarks. Now, I could talk about AI for Good for days but to save us some time, I just want to show you a little video so you can actually see AI for Good in action from our last summit.

If we could please play the video. I have a joke that I always say for these occasions. AI is easy, AV is difficult. Actually, we don’t need to see the video. Oh, ah. Is it going to happen? Yes. But now we need sound. Since there’s no sound, that’s lovely, Geneva. Ah, that’s good. We are more than the AI generation. We are the generation that is determined, ladies and gentlemen, determined to shape AI for good. So no matter how fast technology moves, let us never stop putting AI at the service of all people and our planet. If you want an AI literate society, meaning resilient and ready for the future, we need to integrate these new tools into schools, curricula.

Let’s build a future where AI advances progress for all humanity. A shared digital future that is again inclusive, equitable, prosperous and sustainable for all. It is no coincidence that this era of profound innovation has prompted many to reflect on what it means to be human and on humanity’s role in the world. AI must help bring us closer, not to divide us apart. That’s one of the foundational promises of AI for good. We all now have, I think, a much greater level of awareness around AI, and we all need to shift into that as fast as possible because this technology is moving so fast. Ladies and gentlemen, this was a real… fast -track operation that we did, which we call the International AI Standards Exchange Database.

in your domain or industry that require this type of trigger. And we have just started the last step right from the general division. Let’s go! I think it’s fair to say that AI for Good is indeed more than a summit. It’s a movement, it’s a global community, and it would be nothing without you, the participants. 3, 2, 1! Thanks for watching. I’m not sure who that last guy was. Now, I think one of our… I think people often misunderstand that AI for Good, it’s known as a summit that takes place each year in Geneva. But it’s actually a year -long activity. We have online events almost every day of the week, all year long. And we’re organized around three pillars.

Solutions, skills, and standards. And if you look at the solutions pillar, we have machine learning challenges, we have startup pitching competitions, all types of activities to identify real practical applications of AI that you can use here and today. And on the topic of Edge AI, we had a build -a -thon on Edge AI just a few weeks ago here in India. And we also had machine learning challenges on tiny ML, tiny machine learning devices. And when we’re looking at skills, we launched the AI Skills Coalition. And a big piece of that is going to be creating basically machine learning environment sandboxes where we can do training and mentoring for governments to upskill their constituencies on the use of AI using the data from our machine learning challenges.

So it’s not hypothetical. It’s using real data for real solutions. And the last piece, of course, the bread and butter of ITU, is standards. And we have over 400 AI standards published or in development covering a whole suite of topics. But more specifically related to the session, we have a standards work on future networks, basically 5G, 6G and beyond, and a pre -standardization effort on AI native networks. So basically, these are examples of AI for good in action. And the theme of this session is actually edge AI in action in the global south. And I’m very much looking forward to the discussion. And thank you for your time and attention.

Speaker 1

Thank you so much, Fred. Now, we have the keynotes coming. Thank you. First of all, let me call Professor Lal. Brijesh is my great friend as well as colleague. He was the Bharti School Chairman, but also right now he is currently looking at edge AI research. Our touch points with ITU are many, where he has hosted AI for Good Challenges, WTSA Hackathon. He was a judge, as well as Kaleidoscope. He is very active. Thank you very much, Vijay Singh, for coming, and over to you.

Brijesh Lal

So it’s been a while. Thank you, Vishnuji, for having me. I’ve been participating in these AI for Good activities, so there’s been a lot happening, not just these talks that you have, but also something on the ground. Hackathon is an example of that, with participation from all over the globe. So today I’m going to talk about some of the work that’s happening here at IT Delhi, where we’re trying to leverage the edge. And the other thing that I’m going to run through very quickly, is TSI and its role in edge. So because we’re focusing here on accelerating development across the global south, so I’m going to pick up those two examples today. Right. So what we’re trying to say is that you have lots and lots of edge agents that will now act simultaneously and in coordination.

So the reason why edge is becoming more and more important is this converge of communication, compute and control. And this convergence is now quite real. And because this convergence is real, it is enabled at least in today’s technology only by a strong edge control specifically for tasks in the area of haptics. As I will show in the next slide, require you to not miss or make mistakes because some of them are catastrophic. And for that reason, a strong development in the area of edge is important. The other reason why looking at edge is important from the perspective of Global South is that. While it might not be easy to have foundation models that solve all the problems of the world, at least.

to an extent context has become increasingly important in modern times. People want to provide solutions which are very very specific to the task at hand and context can be best leveraged or used if there is a strong edge capability that is present. So in that light it is important that the global south focuses on building its strength in the area of edge. This slide here talks about some of the work that we are doing with respect to haptics. Haptics as you know is this sense of touch primarily consists of two aspects. One is kinesthetics which is the pressure that we feel and the second is tactile or texture which is the quality of surface that we you know the fine grained texture of the surface that we are able to measure using our skin.

So the thing with this kind of a modality is that while it seems to be almost abstract it is quite pervasive. It is all around us the temperature. including you know the hardness the softness or the way people meet each other greet each other you know all of that is very very important we just don’t you know it’s not overt but it’s important nonetheless so we sort of take it for granted however it is very very important and therefore it needs to be looked at a little carefully now the challenge with haptics is that while as we move from speech to video people did talk about bandwidth and they did talk about latency and there were quality of experience measures that evolved with haptics it goes to the next level because if you have unsynced and delayed haptic inputs or feedback then it becomes quite confounding and it confuses the person and it sometimes can be quite disconcerting so for this reason it is extremely important that the haptics data that you receive is accurate and received on time.

So for this it becomes extremely important that there is a strong capability that is present at the edge. Now here at IIT we are trying to implement it using two ways. One is what we term as split control where we have tried to move from having solutions deployed only in the cloud and the endpoint. We try to put in significant amount of capability on the edge itself. The other aspect that we are looking at carefully is trying to convert signals which are haptic in form to signals which give you the intent rather than actual measurements of pressure as what haptics is to machines. So these two things are primarily handled at the edge. The first one is quite clear.

Let me just say a few words about the second. So when we talk about intent in today’s world whenever you look at a haptic solution it is sort of locked in right from the operator to the end point where you have some kind of manipulation, dexterous manipulation of the environment around the device. However it’s very very hard for devices of different different manufacturers to interoperate and this happens because it is very very tightly coupled to all the signals that are generated and the form factor of the devices it’s not as simple as pick up any camera and the image that you get you can show it on any display so for that reason the idea is to convert those signals into intent send the intent to the other side and the edge on the other side makes sense of the intent and converts into into a signal that the other or the far point can then use to do whatever works needed so these are the two things that we sort of look at with with reasonable amount of interest at iit delhi and we continue to contribute to standards primarily in the area of msc and quality of experience where multi -modality is involved right now this is the edge foundation network.

I’ll skip this in interest of time because I do have a couple of slides that I want to walk you through because there’s some work that’s also being done by TSDSI which is our SDO here in India and they have in conjunction with ITU doing quite a lot of interesting work which is edge centric. So let me talk about few of those. So there are a few technical reports that have come out of late. There’s a stock of dynamic AI ML models for self -sustainable V2X applications. So V2X applications is being looked at carefully. There’s also work in the area of security aspects and advanced and AI enhanced passive digital twinning initiatives. So we have some technical reports that have happened in this area.

There’s also developing of standards work that’s happening. There’s architectural support for tactile applications that I just spoke about. There’s talk of 6G AI architecture for RAN and also AI native scalable reference. Architectures. I think maybe we’ll talk about quality of experience in the next slide but that’s another thing we’re looking at. We’re also carrying out technical studies in all of these areas in interest of time. You’ll have the slides you can go through them when you find the time. This is the other thing that they wanted me to bring to light to this audience. Just a couple of minutes. So the global standard forums that are of interest to the audience here people who look at edge carefully.

There’s ITUR IMT 2030 framework for included ubiquitous intelligence for overarching design and then there’s ITUT related standards CGPP standards and of course the M2M. So all of these standards are of interest to the audience here and people trying to do research in this area and besides this TSTSI has been trying to be inclusive by holding these flagship conferences annual ones so that more and more people get insight into what is happening. With that I’ll close because we’re really short of time here. Vishnu ji back to you

Speaker 1

Thank you Bajeshji Thank you for bringing out the Indian research in the topic and bringing out the 8GI framework also it’s very less time let me invite Ranjitha Ranjitha obtained her PhD from IIC her current research involves causal inference, survival analysis and Bayesian neural networks over to you Ranjitha

Ranjitha Prasad

Yeah so something that he also missed, I actually do work in federated learning and many other learning paradigms so let me just start, so mine is going to be a technical talk where I’ll tell you the motivation for using federated learning, especially the role of federated learning in telecom networks and why really are people discussing about this the motivation is of course data explosion there’s exponential growth of in mobile data data traffic and you have all these diverse services that are there in 6G EMBB, URLC, I’m sure this audience is well aware of this then there are bottlenecks in these legacy networks which actually motivated moving more towards edge centric architectures. The goal is of course I think this is something very important that most of the standards are looking at.

Predictive zero touch automation closed loop wireless control and this loop closure latency requirement about less than 10 milliseconds for mission critical optimization. And this is exactly where federated learning comes in as a key enabler of privacy preserving and distributed intelligence. So all of this is captured in the AI native network concept where now AI is no longer a peripheral layer but it’s actually coming into the RAN. So this is enabled by what is called as this ORAN alliance particularly the RIC or the RAN intelligence controller and this is how the whole sub 10 millisecond latency requirement is fulfilled. Something that is not very clear here is So why do you really require edge intelligence, right? So to make it even faster and achieve the sub 10 milliseconds, you actually have to bring in inference and training to the edge rather than taking data to the cloud, right?

So that’s where the whole paradigm shifted and this argument about edge intelligence or edge native intelligence came in. And especially something called as MEC or multi -axis edge computing also was introduced. So this brought in a huge architectural change. That is, now we have the core network talking to RAN and then RAN talking to the UEs. And this is where the whole, you know, the UEs basically now have the intelligence along with the MEC controllers. So federated learning. Upon all these things, one very important aspect. that’s how we relate to AI for good is that of privacy right so think of the use case of traffic prediction where there is you know there’s a need for loads and loads of data but you know this data consists of raw user logs location history or I mean if you share it with the with the centralized controller it’s just privacy violation so the solution is to now bring code to the data and not take data to the code right so that’s the that’s where federated learning comes in the intelligence now or the training happens at the edge and only certain metadata is given to the cloud so what is this what is its implication in telecom so there’s impact on privacy that’s exactly where it’s supposed to make the impact and then of course there’s impact on latency and bandwidth so personalization of AI models is possible in real time large -scale training can still happen in the core network but the personalization of smaller models for you know for the edge and there’s impact on bandwidth because i no longer need to send data to the server and of course there’s a huge impact on architecture because you saw there that it becomes a hierarchical style of an architecture where core network is at the top and user ues are at the bottom okay so i just wanted to introduce quickly introduce two use cases in fact left it’s in fact a use case from uh from france it was this is for um a traffic prediction in fact uh predicting certain traffic spikes when they had a football match and uh this scenario is where you need to allocate dynamically uh resources for this particular stadium event so here what’s happening is each of these ues or base stations are picking up uh the traffic in their local area sharing it with the core network sharing it with a mec controller and then the core network is able to say you know how to really route the traffic so that you know there’s less congestion the other one is v2x so this is again for uh sharing road conditions or you know accident information and other things it’s very easy to see why uh fl may be useful here each cars can talk to its own edge server and then go to the cloud server where the global model is trained so this this sort of envisages how federated learning has become a very important technology so last but not the least i come from i have i’m the pi of the lab which is called as intellicom lab at triple it delhi uh we have a collaboration with it delhi uh for this entire work

Speaker 1

Thank you, Ranjitha, for the excellent talk. We had an introduction at least for federated running and also the framework that architecture that she explained is really interesting. Last time when ITU colleagues were here, we had visited the lab. If you haven’t done that, please talk to her. It’s a very exciting research which they do. And we also have great collaboration with BAPI and colleagues in IIT Delhi. Thank you, Ranjitha, for coming. we have a panel now we have approximately 20 minutes maybe for the panel let’s kick off the panel can I invite Fred to moderate the panel and can I invite the panelists Mala Alagan and Sakshi to please take the seats Fred to kick off, thank you very much over to you Fred

Fred Werner

thank you so I’m looking forward to this panel where we can aim to demystify Edge AI a little bit and explore the practical use cases and AI strategies but first I’ll introduce the panel so the first panelist her name is Mala she has a full name but she personally asked me to just call her Mala and I wish all panelists would do that it’s much easier that way so Mala is currently a technologist at the Center of Excellence Wired and Wireless Technologies at Art Park sorry, Art Park so Mala is currently a technologist Prior to this, she was a postdoctoral research at the Teikian Group at Technical University Berlin. She’s also involved in 6G initiatives such as AI RAN for efficient resource allocation and millimeter wave communications.

And she also has been a visiting researcher at UC Davis and TU Berlin. Welcome. Our next panelist is Alagan Mahalingam, founder, CEO, chief software architect of RootCode. Alagan is the founder of RootCode, and in his early 20s, he worked as a researcher at international research organizations such as the Geoinformatics Center at the Asian Institute of Technology, Thailand, also the University of Tokyo, Japan, where he worked on satellite communications and solar panel optimization algorithms. Alagan. Mahalingam was also awarded the special title of ICT Entrepreneur of the Year at the National ICT Awards in 2021. and also the Young Entrepreneur of the Year in 2024. And he’s also the envoy for the government of Estonia e -residency. So I see a lot of Estonia connections here today.

Last but not least, we have Shaxi Gupta. So she’s the Global Government Affairs responsible for Qualcomm. She’s a tech policy professional in AI and emerging technology policy analysis, market research, and stakeholder engagement. So if we could have a please warm welcome for the panel. So first question is for Mala. As an AI -enabled XR applications, and they’re split between private 5G and on -premise public 5G, could you please give us some examples of XR applications in different scenarios, and what are the trends? And what are the trade -offs in scalability, security, and security?

Mala Kumar

They get the immersive experience in their own preferred regional languages. And one other application that we have done is the XR -assisted medical emergency care. Here the focus is on the, to provide timely medical response to the patient who was suffering with a cardiac arrest and so on. And an SOS alert would be sent from the, by the bystanders from the life circuit exact to the first responders. And the medical experts and the ambulance through 5G connectivity. Once the first responder gets the alert, he arrives at the scene with XR glasses and IOT wearables. and also the AED kit. So, and while giving the CPR, the IOT patient vitals would be displayed, augmented onto the real -time video.

And the real -time video would also be sent to the medical expert. And the medical expert will guide whether to continue the CPR or it could be the AED and so on. So, the timely response will save multiple lives. So, in this case, we have used public 5G network. But for the XR -assisted facility tour, we have used private 5G network. So, the private 5G network is mainly to have on -premise HCI applications. And this would bring the core next to, the data generation. And then we can also… do real -time decision -making for industry 5 .0 applications. And going forward, we would like to have some of our applications to be in the open source and have it in the best place, like ITUs, AI for good, right?

So then the international community can access this open source AI models and they can fine -tune the models and they can do the rigorous testing before it is bringing it to a real -world deployment. That is what I’m looking forward for this.

Fred Werner

Yeah, thanks so much, Mala. And I think this really is a good example of AI for good in action. And I think, to your point, these solutions don’t happen by magic. There’s a lot of difficult problems. There’s a lot of problems to solve. And by putting these solutions in the AI for good, good sandbox that might lead to future standards which could make them replicable and then you could have that adoption at scale. So I’ll just go to the next panelist, Alagan. Given your rich experience in developing AI solutions for partners in different geographies, can you please give us some examples of edge AI deployment in real world scenarios, their impacts, the nuances you see in AI strategies on edge AI in the different regions?

Because from your bio you’ve been involved in many different parts of the world. Thanks.

Alagan Mahalingam

I started RootCode 11 years back because I was in love with building AI solutions as a college student and then now 11 years later the technology that we have built is used by more than 92 million people across 27 countries including many European governments like the government of Estonia, Portugal and many others. We chose to build edge AI in many cases. One, the obvious one, to bring technology to under -connected spaces and also to increase speed in many cases and sovereignty. And the most interesting project that we have done recently, let me tell that story, a couple of years back, Portugal realized that their farmers, especially the small -scale farmers, didn’t get enough access to advisory and intelligence to grow their crops.

And things have been changing because climate change and unpredictability in growing crops, a lot of people were leaving farming. And so we built a solution from a hardware, a software product and also an AI model. The hardware goes into the soil, so you understand the soil nutrition and you take pictures with the mobile app and we can process the pictures to understand is there a problem with the plant, right? And we built and it worked out fantastically well. And then I tried to bring that to Sri Lanka. I grew up in Sri Lanka and to date a big part of our development team is in Colombo, in Sri Lanka, more than 120 people. And so we went into one of these villages in the middle of the mountains of Sri Lanka, Nubar earlier and I was super fascinated and when we tried to deploy this, we realized they don’t have reliable connectivity in some corners of the villages and our solution was worthless.

And that’s where we started bringing in Edge. So we brought in a new version, we had a Raspberry Pi and we started testing models like GemR, and also we did our own convolutional networks like 2D things to figure out like where do you optimize? You don’t want to use LLM for everything, right? And by the end of it, we managed to bring the same value that the software gave to connected users. to people who didn’t even have internet in some part of Sri Lanka. And that reminded me how much edge is needed, especially in the global south. And yesterday I was at a dinner talking to some of the development finance colleagues from DZ. And somebody was talking about why don’t we put computes on the wheels in a tuk -tuk?

So imagine like we can’t process too many things on a small device of Raspberry Pi. What if you get a tuk -tuk coming to your village every other day or once a week with a data center built in, with Wi -Fi LAN, so farmers can connect and do the processing. Smaller banks, smaller institutions can do. And I was like, yeah. So this week has been super fascinating. And sometimes when we think about edge, we think it’s needed only in places that are not really connected, like rural parts. We have built this, we have built a beautiful solution that’s used in America. If you think America is well connected, you should take a road trip. When you go out of the city, you realize some parts are very disconnected.

And we built, for one of our clients, we built a solution that helps rural patients who are at high risk with remote patient monitoring. And then, yeah, EDGE works all around the world, not just in the South. If I, when I think about all my learnings, because there are so many learnings building EDGE for multiple geographies, multiple customers, multiple communities. If I were to single out, I would single out the fact that when you are trying to do something in the EDGE, you shouldn’t try to think of the model and go find a solution. But instead think of the task and then work backwards on how do you build and distill or fine tune a smaller model.

And that runs on the EDGE because in the EDGE, you can’t do everything, right? if you are building an AI assistant for farmers, you don’t want the AI to be able to tell why two of the famous CEOs didn’t want to hold hands. I mean, that doesn’t matter. You want it to answer about plants and agriculture. So the heavier the model is, it becomes impossible to deploy. So we work on multiple technologies to one, quantize or prune the models in a way that creates a smaller version that does exactly what’s supposed to happen. And I think the global south needs to grow with this AI transformation of the world because infrastructure takes decades, but the next few years is going to change the way we live.

And that’s why we are here. So I’m excited

Fred Werner

Yeah, thanks a lot. And I think what you’re saying here has almost been the theme of this week where you don’t need the biggest AI or the biggest large language model. And I think if you look at the example of India, where they’ve managed to enable… billions to have a digital ID to enable financial inclusion, financial payments with the public interest at heart and with relatively low -tech solutions, you can indeed bring AI to the edge in cases that make a lot of sense. So thanks for that. Sapsky, question for you. In your experience with Europe, the intersection of technology, innovation, and AI strategies, what do you think are the metrics to evaluate the usage of edge AI such as availability and capability of hardware at the edge and also the connectivity, privacy, and data issues that you see in your line of work?

Sakshi Gupta

Thank you, Fred. And let me start by saying it’s an absolute pleasure to be sitting with fellow panelists and speakers who have preceded me who are deploying edge AI and are doing research on edge AI. At Qualcomm, we are very focused on edge AI and think that that’s going to be the future of how not just Global South, but globally, we’re going to be using AI. So we… Um… And I really relate to what you said about the way to think about deployment of AI is actually to think backwards about what is the use case that you’re trying to solve. And then you think about what is the best architecture that you want to use.

Is it just cloud? Is it on -prem? Is it an edge cloud? Or is it on -device AI? So we have to think about it from a distributed architecture point of view when we think about the use cases that we have here in the Global South. And I do want to mention one important distinction here, which was touched upon earlier also, is that when we think about AI, there’s a training part of it, and then there’s an inferencing part of it. Inferencing is where you’re thinking processing is actually happening. So while training can continue to happen on the cloud, a lot of it, a lot of the inferencing, as we’re seeing, is moving towards the edge.

Now, in terms of availability, if I want to talk about it, I think we’re increasing. We’re increasingly seeing, and Qualcomm is deploying this at the edge of… So, you know, AI being available at the edge, not from, you know, the very basic thing that we all use every day is your smartphones. That we have on -device capabilities coming onto smartphones with 10 billion parameters models already running on -device. So that means that you do not need to be connected. If you’re in flight mode, you do not need to be connected on internet and you can still use AI. So that’s amazing in my point of view. We also have it coming to actually cars. So Qualcomm has developed that technology where you can now use HCI onto the, or it’s actually in development.

We have demos at the Qualcomm booth, which I’ll come to later, but which you can, so HCI is coming to the cars as well. And it’s increasingly coming to IoT devices and your smart glasses as well. So in terms of availability, I think we are seeing increasingly that it’s coming to all types of devices. That are connected to internet now. Now, why is AGI relevant? And some of my panelists have already touched on it. I think latency, security, privacy, personalization, low cost, low power are all very important factors for why AGI becomes important for Global South. We may not have access to as much power. We may not have access to as much water as needed.

But with AGI, we don’t have to worry about that. Apart from that, I do want to touch on one thing. That is Qualcomm, one of the things that we have is a program called Tech for Good, wherein we partner and work with startups and small businesses around the world. We invest in them. We mentor them. They use Qualcomm hardware to develop solutions at the edge. In fact, I do want to encourage that in Hall 4 at our Qualcomm booth, we have some of these startups who are displaying this technology. One of the examples is actually from India. It’s called Raksa Health. They’ve actually built an on -device AI healthcare assistant where it’s for doctors and patients both, where the doctors can take down symptoms and provide solutions for their patients and for patients to actually look up their prescriptions and be able to access all their records offline and ask questions about it.

So, yeah, I think that’s how we’re seeing the transition happen. Thank you.

Fred Werner

Thank you. Yeah, some amazing use cases. And I think this week’s coming out of Davos where the narrative was all about go, go, go, the insatiable demands for energy. There’s talk of putting data centers in space. But I think this panel also brings things a bit down to earth where, you know, you can have AI on the edge, and, of course, there’s a lot of things to solve there. when it comes to connectivity, when it comes to data compute. I think there’ll be a lot of standards development work that needs to emerge from this to make this work at scale. But I think your use cases and the way you’re approaching the problem, especially starting from the what are you trying to solve and work backwards from that, I think is very refreshing compared to all the headlines we’ve been seeing lately.

And I don’t see it either or. I see it as a big piece, a complementary piece of the puzzle. So with that, I really want to thank the panel and if we could have a round of applause for them. Thank you.

Speaker 1

Thank you very much, Fred, for running the tight panel. Now we are coming to the closing. Thank you, panelists, insightful remarks. Yes? Okay. Can I ask a quick group photo of the panelists, please? Panelists? Yes? Thank you. Thank you very much. Thank you very much. Now we are coming to the closing. There are excellent closing remarks coming. Can I please request Her Excellency Miss Lopez, Ambassador, Permanent Representative, Permanent Mission of the Republic of El Salvador to United Nations Office and other international organizations in Geneva to please give her closing remarks.

Ambassador Egriselda Lopez

I’m actually based in New York. Thank you. Well, good afternoon. I know that I don’t have much time, but I had just to say that this discussion was very enlightening. Thank you so much for sharing everything what you’re doing on the ground. And I guess that it was very clear to me that HAI means simply using an AI closer to where things happen. That means closer to people, closer to services, communities, rather than deepening only faraway systems. So amazing what you’re already doing. So this can be important for development because it can work better in places with limited connectivity, as we were hearing, and it can help with speed. And it can help with speed, cost, and privacy, since not everything has to be sent everywhere.

So I guess that I had to begin also with something. I am also the co -chair of the Global Dialogue on AI Governance. This is going to happen in July this year, and it’s going to be the first dialogue of its kind. So trying to also bring together what we have been hearing from member states and also other stakeholders in these months, I can tell you three specific things connecting with what we just heard today. First, that people must remain at the center. And we have heard with all these examples. And I guess that a common message that we have been hearing also in this week is that AI should be developed and used in a way that protects but also helps people.

Second, closing the gap is not a slogan. We are hearing this a lot. It requires decisive support. And I was very pleased, for instance, saying that you’ve been trying to replicate in some countries what it has in others, for instance. And I think that’s a really important thing. This information sharing, this is critical if we’re talking about closing the gaps. And the third message, the final one, is that we should avoid a world of disconnected approaches. And this also is aligned with what I was just saying, that cooperation across different national but also regional approaches, it will help us to reduce fragmentation. So, with that, I just have to tell you that we’re very looking forward to see some of you in Geneva in July, so we can hear and learn more about what AI is.

So, it’s my pleasure to give the floor to my distinguished co -chair, Ambassador Reintam Saar. He’s going to explain to you very, very shortly what the global dialogue on AI governance is. And this is really important work that we are putting a lot of effort to it. Thank you so much again for the invitation.

Ambassador Reintam Saar

Thank you. yes hello hello everyone frankly i really feel humbled among real experts not to say i feel helpless so please allow me then to do a little bit of awareness raising when it comes to the first global dialogue on ai governance and maybe this way i’ll try to fit into a discussion that we’ve heard here today so three points on my side first about tasking so the tasking was to put together a distinctive identifiable un global dialogue with all the elements that are prescribed in the mandate so bringing governments and stakeholders together to exchange best practices and of course to focus on cooperation and to execute it back to back with itu ai for good summit in july in geneva produce co -chair summary.

So this is what we are going to do. So, so far we’ve engaged with member states, with stakeholders, multi -stakeholders, and from member states we’ve kind of covered three different approaches, I would say a little bit. Risks versus opportunities, state -centric approach versus multi -stakeholder approach, closing AI divide versus free market innovation. But we also were able to pick up three convergences. Practical outcomes preferred over endless theoretical discussions, alignment with existing UN processes, avoiding duplication, clear timeline formats and thematic focus to produce actionable insights. And the unified element, I would say, in these discussions is that the dialogue needs to be inclusive. and capacity building was absolutely a crucial element that is, of course, one of the most important things to a global self.

So from multi -stakeholders, what we’ve heard, the key words, so to say, were trust, transparency, no duplication, interoperability, equal access and participation for everyone, rooting the dialogue in human rights and to be of a practical value and innovative in form. So what we are going to do, we will guide the discussions, but we will not predetermine the outcome. It’s for member states, it’s for you, for stakeholders. And, of course, we will engage also with international scientific panel that was also established through the same resolution. We will rely on member states and your wisdom, we would need to collect this wisdom somehow. and this is something that we are going to do because we would need this wisdom so that the dialogue would be really inclusive we would come up on certain point with a road map to Geneva where you would see building blocks towards dialogue and whatever opportunities to engage into dialogue and of course I very much hope that all these fantastic ideas and frankly I mean chapeau to the panel because you are already making or changing life on the ground and it’s absolutely fantastic we really need this also to inform our dialogue and so that the dialogue would be also result oriented on the ground.

Thank you very much.

Speaker 1

Thank you Monsieur Excellencies. At this point I would like to call Fred to give out the mementos if you don’t mind Fred please. and can we have the momentos for Brijeshji thank you very much Ranjitha Ranjitha please Moala can I request Nodal officer to please felicitate Fred yeah we have an event with him yeah yeah yeah yeah yeah yeah yeah yeah yeah yeah thank you very much for attending the session session is closed thank you thank you Thank you. Thank you.

Related ResourcesKnowledge base sources related to the discussion topics (36)
Factual NotesClaims verified against the Diplo knowledge base (5)
Confirmedhigh

“Fred Werner, Chief of Strategic Engagement at the ITU, delivered the opening remarks of the session.”

The knowledge base records Frederic Werner, Head of Strategic Engagement at the ITU, delivering opening remarks at the AI Policy Summit [S102].

Confirmedhigh

“Fred began with the question “What if the last thing that humans ever invent is invention itself?” and referenced a conversation with AI‑safety expert Roman Jampolsky.”

The transcript excerpt in the knowledge base contains the same opening question and mentions Roman Yampolsky as the AI safety expert [S2].

!
Correctionmedium

“AI for Good relies on partnerships with more than 50 UN sister agencies.”

The knowledge base states that 47 UN organizations are collaborating on the AI for Good initiative, not “more than 50” [S19].

Additional Contextmedium

“AI for Good works together with many UN agencies that contribute expertise and drive standards work.”

The AI Governance Dialogue notes ITU’s AI for Good programme partners with numerous UN agencies on standards and governance, confirming the collaborative nature of the effort [S21].

Additional Contextlow

“AI for Good is organised around three pillars—solutions, skills and standards—supporting activities such as machine‑learning challenges, the AI Skills Coalition sandbox, and over 400 emerging AI standards.”

A related source describes AI for Good’s three pillars (including comprehensive skills development and inclusive governance) and highlights its focus on standards and collaborative projects [S23].

External Sources (114)
S1
AI for Good Technology That Empowers People — -Alagan Mahalingam- Founder, CEO, and Chief Software Architect of RootCode; ICT Entrepreneur of the Year (2021), Young E…
S2
https://dig.watch/event/india-ai-impact-summit-2026/ai-for-good-technology-that-empowers-people-2 — So, it’s my pleasure to give the floor to my distinguished co -chair, Ambassador Reintam Saar. He’s going to explain to …
S3
S4
https://dig.watch/event/india-ai-impact-summit-2026/ai-for-good-technology-that-empowers-people-2 — Alagan was also awarded the special title of ICT Entrepreneur of the Year at the National ICT Awards in 2021. Alagan was…
S5
WS #53 Promoting Children’s Rights and Inclusion in the Digital Age — Audience: The lady’s name is Sasha. Okay, Sasha, could you unmute yourself and ask the question? Hello, can you hear …
S6
How AI Is Transforming Indias Workforce for Global Competitivene — -Sangeeta Gupta- Panel moderator (role/title not specified in transcript) -Srikrishna Ramakarthikeyan- (Role/title not …
S7
AI for Good Technology That Empowers People — – Alagan Mahalingam- Mala Kumar- Professor Brijesh Lall – Alagan Mahalingam- Professor Brijesh Lall
S8
https://dig.watch/event/india-ai-impact-summit-2026/ai-for-good-technology-that-empowers-people-2 — Thank you so much, Fred. Now, we have the keynotes coming. Thank you. First of all, let me call Professor Lal. Brijesh i…
S9
AI for Good Technology That Empowers People — – Frederick Werner- Professor Brijesh Lall- Alagan Mahalingam- Egriselda López- Qualcomm Member – Ranjitha Prasad- Qual…
S10
[Online Event] Cables, Novels and Nobels: The Journey of Diplomacy and Literature  — Amr Aljowaily:Great, we can learn from fiction then and the writers can do also their research. So thank you very much. …
S11
AI for Good Technology That Empowers People — -Ranjitha Prasad- PhD from ISE, researcher specializing in causal inference, survival analysis, Bayesian neural networks…
S12
Keynote-Martin Schroeter — -Speaker 1: Role/Title: Not specified, Area of expertise: Not specified (appears to be an event moderator or host introd…
S13
Responsible AI for Children Safe Playful and Empowering Learning — -Speaker 1: Role/title not specified – appears to be a student or child participant in educational videos/demonstrations…
S14
Building Trusted AI at Scale Cities Startups & Digital Sovereignty – Keynote Vijay Shekar Sharma Paytm — -Speaker 1: Role/Title: Not mentioned, Area of expertise: Not mentioned (appears to be an event host or moderator introd…
S15
Driving Social Good with AI_ Evaluation and Open Source at Scale — – Mala Kumar- Audience – Mala Kumar- Tarunima Prabhakar- Ashwani Sharma – Sanket Verma- Mala Kumar Mala Kumar strongl…
S16
AI for Good Technology That Empowers People — Now, you might agree or disagree with that statement, but it’s not hard to imagine a future where most future inventions…
S17
https://dig.watch/event/india-ai-impact-summit-2026/indias-ai-leap-policy-to-practice-with-aip2 — As Dr. said, this is really the earthquake of AI and we are at the epicenter. And as you can see, after five days, we ar…
S18
https://dig.watch/event/india-ai-impact-summit-2026/ai-for-good-technology-that-empowers-people-2 — Thank you. Thank you very much. We have very little time, so I want to first of all introduce Fred. Fred Werner is the C…
S19
AI for Good Impact Initiative — Werner Vogels:Hello Geneva. I’m Dr. Werner Vogels. As a technologist, I’m constantly inspired by young businesses and or…
S20
Indias AI Leap Policy to Practice with AIP2 — As Dr. said, this is really the earthquake of AI and we are at the epicenter. And as you can see, after five days, we ar…
S21
The role of standards in shaping an AI-driven future — ## ITU’s Current Initiatives – Role/Title: Not explicitly mentioned, but appears to be in a leadership position at ITU …
S22
The role of standards in shaping a safe and sustainable AI-driven future — Seizo Onoe:Thank you very much. Good morning, everyone, and very warm welcome to you all. Our discussions at this summit…
S23
We are the AI Generation — In her conclusion, Martin articulated that the fundamental question should not be “who can build the most powerful model…
S24
Designing Indias Digital Future AI at the Core 6G at the Edge — But delivering intelligence is not about LLMs or training LLMs. It is about delivering this entire ecosystem to the last…
S25
Edge computing growth boosted by Duos and Accu-Tech partnership — Duos Technologies Group, through its subsidiary Duos Edge AI,has entered a strategic partnershipwith Accu-Tech to expand…
S26
‘The elephant in the AI room’: Does more computing power really bring more useful AI? — Let’s think back to our daily routines before the AI era. Did we need a full set of Encyclopedia Britannica open on our …
S27
Scaling Trusted AI_ How France and India Are Building Industrial & Innovation Bridges — Again, I’m sure you’ll find, I’d be happy to talk about any of these for much longer, but we only have a short time. The…
S28
WS #208 Democratising Access to AI with Open Source LLMs — Daniele Turra: Thank you for the question. I think the answer is that technology licensing itself cannot really alone…
S29
The strategic shift toward open-source AI — The release of DeepSeek’s open-source reasoning model in January 2025, followed by the Trump administration’s July endor…
S30
AI Impact Summit 2026: Global Ministerial Discussions on Inclusive AI Development — “Standards, skills, and solutions”[104]. “First, standards”[106]. “Second, skills”[183]. “And third, solutions”[182]. “T…
S31
WS #82 A Global South perspective on AI governance — Lufuno argues that the Global South needs to move beyond being merely a subject of discussion in AI governance. She emph…
S32
Searching for Standards: The Global Competition to Govern AI | IGF 2023 — In conclusion, the UNESCO recommendation on AI ethics provides crucial guidance for global AI governance. By grounding A…
S33
International multistakeholder cooperation for AI standards | IGF 2023 WS #465 — Aurelie Jacquet :Thank you. So, following on Wansi’s point, I think what’s important to know is it’s actually good to se…
S34
AI Governance Dialogue: Steering the future of AI — Doreen Bogdan Martin: Thank you. And we now have a chance together to reflect on AI governance with someone who has a un…
S35
AI Governance Dialogue: Presidential address — – H.E. Mr. Alar Karis Human rights | Legal and regulatory | Development Importance of global cooperation and coordinat…
S36
AI governance efforts centre on human rights — At theInternet Governance Forum 2025in Lillestrøm, Norway, a keysessionspotlighted the launch of the Freedom Online Coal…
S37
Opening of the session/OEWG 2025 — El Salvador stresses the importance of developing governance frameworks and ethical guidelines for artificial intelligen…
S38
Opening — The overall tone was formal yet optimistic. Speakers acknowledged the serious challenges posed by rapid technological ch…
S39
Opening address of the co-chairs of the AI Governance Dialogue — The tone is consistently formal, diplomatic, and optimistic throughout. It maintains a ceremonial quality appropriate fo…
S40
AI, Data Governance, and Innovation for Development — Martha Omoekpen Alade: So, Martha Omoekpen Alade has actually helped us to answer one of the questions, right? Okay, but…
S41
Scaling Innovation Building a Robust AI Startup Ecosystem — The tone was consistently celebratory, appreciative, and inspirational throughout. It began formally with the awards cer…
S42
HETEROGENEOUS COMPUTE FOR DEMOCRATIZING ACCESS TO AI — The disagreement level is moderate but significant, particularly around philosophical approaches to security and the opt…
S43
WS #82 A Global South perspective on AI governance — Lufuno T Tshikalange: Okay, thank you. Hopefully I will finish what I’m saying this time around. In terms of the risk…
S44
Catalyzing Global Investment in AI for Health_ WHO Strategic Roundtable — The discussion highlighted how AI tools are often forced into clinical practice at inconvenient points in healthcare wor…
S45
WSIS Action Line C10: Ethics in AI: Shaping a Human-Centred Future in the Digital Age — Low level of fundamental disagreement with moderate differences in implementation strategies. The speakers largely agree…
S46
AI for Good Technology That Empowers People — Collaboration and Multi‑Stakeholder Engagement
S47
WS #110 AI Innovation Responsible Development Ethical Imperatives — Three core policy dimensions must be addressed: inclusive development, global governance alignment, and multi-stakeholde…
S48
A bottom-up approach: IG processes and multistakeholderism | IGF 2023 Open Forum #23 — In conclusion, the analysis highlights the importance of multi-stakeholder engagement in policy processes, with specific…
S49
AI for Bharat’s Health_ Addressing a Billion Clinical Realities — “I’m asking a suggestion from you, so like what model should, like someone who’s creating such solution for voice and tr…
S50
Democratizing AI Building Trustworthy Systems for Everyone — “I think open source is going to be in my mind a critical aspect of it”[32]. “Sustainability also requires these kinds o…
S51
The Expanding Universe of Generative Models — Aidan Gomez argues that enhancing the efficiency of large language models relies on increasing compute power. He asserts…
S52
State of Play: AI Governance / DAVOS 2025 — Arthur Mensch: I would say I think we can we can split responsibilities in between industries and governance. The firs…
S53
Industrial sectors push private 5G momentum — Private 5Gis often dismissedas too complex or narrow, yet analysts argue it carries strong potential for mission-critica…
S54
Designing Indias Digital Future AI at the Core 6G at the Edge — Questions about network API monetization and the practical implementation of distributed edge computing also highlighted…
S55
Resilient infrastructure for a sustainable world — An unexpected consensus emerged around the tension between the time needed for proper standards development and the rapi…
S56
Indias AI Leap Policy to Practice with AIP2 — Fred acknowledges that standards development is traditionally slow but emphasizes the dual nature of AI requiring carefu…
S57
Global Standards for a Sustainable Digital Future — Instead of developing standards that dictate specific actions, the focus should be on creating descriptive standards tha…
S58
AI Impact Summit 2026: Global Ministerial Discussions on Inclusive AI Development — In the global south. the timing and the location are equally important. As AI technology has continued to advance so has…
S59
Policy Network on Artificial Intelligence | IGF 2023 — Sarayu Natarajan:Thank you for that question. I think it’s a broad and difficult one, and I’ll try my best to do as much…
S60
Edge computing growth boosted by Duos and Accu-Tech partnership — Duos Technologies Group, through its subsidiary Duos Edge AI,has entered a strategic partnershipwith Accu-Tech to expand…
S61
Edge AI gains momentum in Europe’s innovation strategy — Europe is accelerating efforts tobuild digital sovereigntythrough high-performance technologies that do not increase pow…
S62
Searching for Standards: The Global Competition to Govern AI | IGF 2023 — Audience:Thank you so much, Dr. Ali Mahmood. I’m from Pakistan. I’m heading a provincial government entity that is invol…
S63
Open Forum #38 Harnessing AI innovation while respecting privacy rights — Audience: Hello. Thank you so much for the presentation. I’m from Nanting Youth Development Service Centre, and in my…
S64
Open Forum: A Primer on AI — Privacy protection is another important aspect discussed in the analysis. It is noted that AI training often involves th…
S65
From Human Potential to Global Impact_ Qualcomm’s AI for All Workshop — The privacy implications are equally significant. By processing personal data locally, edge AI addresses growing concern…
S66
AI for Good Technology That Empowers People — The AI for Good initiative, launched in 2017, has evolved from a concept-focused summit addressing the “fear, promise, a…
S67
AI Impact Summit 2026: Global Ministerial Discussions on Inclusive AI Development — “Standards, skills, and solutions”[104]. “First, standards”[106]. “Second, skills”[183]. “And third, solutions”[182]. “T…
S68
We are the AI Generation — In her conclusion, Martin articulated that the fundamental question should not be “who can build the most powerful model…
S69
Ministerial Roundtable — Strategy built around four pillars: Governance and Ethics with Clear Regulatory Standards and Human Oversight
S70
WS #100 Integrating the Global South in Global AI Governance — AUDIENCE: Can I add to this? Yeah, please. Okay. So I’m just going to be brief and quick on this. I think there is no …
S71
Building Scalable AI Through Global South Partnerships — Artificial intelligence | Information and communication technologies for development | Social and economic development
S72
Designing Indias Digital Future AI at the Core 6G at the Edge — This disagreement is unexpected because both speakers are discussing the same technological trend (edge computing) but a…
S73
The role of standards in shaping a safe and sustainable AI-driven future — He further expounded on the collaborative essence of standardisation work, which relies on mutual trust, understanding, …
S74
The role of standards in shaping an AI-driven future — He positioned this approach as leveraging ITU’s 160 years of experience and its global community’s commitment to collabo…
S75
Setting the Rules_ Global AI Standards for Growth and Governance — Key areas of convergence included the importance of process-oriented standards that can adapt to evolving capabilities, …
S76
Harmonizing High-Tech: The role of AI standards as an implementation tool — Bilel Jamoussi:Fantastic. Thank you, Philippe, for that excellent introduction of the three organizations. Between ISO, …
S77
International multistakeholder cooperation for AI standards | IGF 2023 WS #465 — Ashley Casovan:Yes, thank you so much, and thanks for having us here to present about the work that we’re doing, and als…
S78
Day 0 Event #172 Major challenges and gaps in intelligent society governance — Poncelet Ileleji: Thank you very much. Mr. Ilericic, can you hear me? Yes, can you hear me? Can you hear me? Okay. …
S79
AI governance efforts centre on human rights — At theInternet Governance Forum 2025in Lillestrøm, Norway, a keysessionspotlighted the launch of the Freedom Online Coal…
S80
AI Governance Dialogue: Steering the future of AI — Doreen Bogdan Martin: Thank you. And we now have a chance together to reflect on AI governance with someone who has a un…
S81
AI Governance Dialogue: Presidential address — Ettore Balestrero: On behalf of His Holiness Pope Leo XIV, I would like to extend his cordial greetings to all participa…
S82
First round of informal consultations with member states, observers and stakeholders (2024) — Advocating for a human-centred design of the GDC, the speaker argued that ensuring data privacy and broadening technolog…
S83
Opening — The overall tone was formal yet optimistic. Speakers acknowledged the serious challenges posed by rapid technological ch…
S84
From Innovation to Impact_ Bringing AI to the Public — These key comments transformed what could have been a typical AI hype discussion into a nuanced exploration of civilizat…
S85
Announcement of New Delhi Frontier AI Commitments — Opening remarks and framing of the event
S86
Keynote by Uday Shankar Vice Chairman_JioStar India — The tone is consistently optimistic and visionary throughout, beginning with congratulatory remarks and maintaining an i…
S87
Opening address of the co-chairs of the AI Governance Dialogue — The tone is consistently formal, diplomatic, and optimistic throughout. It maintains a ceremonial quality appropriate fo…
S88
Comprehensive Report: European Approaches to AI Regulation and Governance — The discussion maintained a professional, collaborative tone throughout. Both speakers demonstrated mutual respect and a…
S89
Shaping the Future AI Strategies for Jobs and Economic Development — The discussion maintained an optimistic yet pragmatic tone throughout. While acknowledging significant challenges around…
S90
AI as critical infrastructure for continuity in public services — The discussion maintained a collaborative and constructive tone throughout, with participants building on each other’s p…
S91
Indias AI Leap Policy to Practice with AIP2 — The discussion maintained a constructive and collaborative tone throughout, with speakers building on each other’s point…
S92
Business Engagement Session: Sustainable Leadership in the Digital Age – Shaping the Future of Business — The discussion maintained a consistently collaborative and optimistic tone throughout. It began with academic framing bu…
S93
AI for Social Good Using Technology to Create Real-World Impact — The tone was consistently optimistic and collaborative throughout, with speakers demonstrating genuine enthusiasm for AI…
S94
Open Forum #66 the Ecosystem for Digital Cooperation in Development — – **From dialogue to action**: A recurring theme was the need to move beyond policy discussions to concrete implementati…
S95
Panel 5 – Ensuring Digital Resilience: Linking Submarine Cables to Broader Resilience Goals — The tone was largely collaborative and solution-oriented. Panelists built on each other’s points and offered complementa…
S96
Building Population-Scale Digital Public Infrastructure for AI — The tone is optimistic and collaborative throughout, with speakers sharing concrete examples of successful implementatio…
S97
Parliamentary Closing Closing Remarks and Key Messages From the Parliamentary Track — The discussion maintained a collaborative and constructive tone throughout, characterized by diplomatic language and mut…
S98
Closing remarks – Charting the path forward — The tone throughout was consistently formal, diplomatic, and optimistic. It maintained a collaborative and forward-looki…
S99
World Economic Forum 2025 Annual Meeting Opening Ceremony: Summary — The tone began earnestly optimistic about dialogue and cooperation, with leaders acknowledging criticisms of elite gathe…
S100
Ensuring Safe AI_ Monitoring Agents to Bridge the Global Assurance Gap — The tone was collaborative and solution-oriented throughout, with participants acknowledging both the urgency and comple…
S101
Strengthen Digital Governance and International Cooperation to Build an Inclusive Digital Future — The discussion maintained a consistently collaborative and optimistic tone throughout, with speakers emphasizing partner…
S102
AI Policy Summit Opening Remarks: Discussion Report — The AI Policy Summit at ETH Zurich, part of the inaugural Zurich AI Festival, opened with remarks from Fabian Streiff re…
S103
Bridging the Digital Divide: Achieving Universal and Meaningful Connectivity (ITU) — Practitioners committed to advancing the digital agenda and connectivity in their respective roles shared their experien…
S104
Ethical limits of rapidly advancing AI debated at Doha forum — Doha Debates, an initiative of Qatar Foundation, hosted a town hall examining the ethical, political, and social implica…
S105
AI for Democracy_ Reimagining Governance in the Age of Intelligence — Hello. Good evening, ladies and gentlemen. It is a pleasure to be back here in India. As a fellow citizen of the Global …
S106
AI for Good – food and agriculture — **Doreen Bogdan Martin, ITU Secretary General** participated in announcing the youth robotics challenge and highlighted …
S107
Comprehensive Discussion Report: AI’s Transformative Potential for Global Economic Growth — These key comments fundamentally shaped the discussion by systematically addressing and reframing common concerns about …
S108
Toward Collective Action_ Roundtable on Safe & Trusted AI — So I think, you know, I don’t want to speak to what an individual person wants, but I think that what we all want is emp…
S109
Open Forum #53 AI for Sustainable Development Country Insights and Strategies — Oluwaseun Adepoju: Thank you so much, and thank you to the panellists who have spoken before me. I think they’ve raised …
S110
Enhancing rather than replacing humanity with AI — Right now, amid valid concerns about displacement, manipulation, and loss of human agency, there are also real examples …
S111
Keynote-Alexandr Wang — He highlighted innovative applications by Indian organizations addressing societal challenges. iSTEM has developed voice…
S112
AI and the future of work: Global forum highlights risks, promise, and urgent choices — At the20th Internet Governance Forum held in Lillestrøm, Norway, global leaders, industry experts, and creatives gathere…
S113
How AI agents are quietly rebuilding the foundations of the global economy  — AI agents have rapidly moved from niche research concepts to one of the most discussed technology topics of 2025. Search…
S114
AI for food systems — The initiative builds upon proven collaborative frameworks, particularly the ITU-FAO focus group on AI and IoT for digit…
Speakers Analysis
Detailed breakdown of each speaker’s arguments and positions
F
Fred Werner
3 arguments147 words per minute2013 words817 seconds
Argument 1
AI for Good purpose – unlocking AI’s potential for humanity
EXPLANATION
Fred defines the core mission of AI for Good as making AI serve humanity by unlocking its potential for beneficial applications. He frames this as the central purpose guiding the initiative’s activities.
EVIDENCE
He states, “So what is the goal of AI for Good? Well, simply put, it’s to unlock AI’s potential to serve humanity.” [26-27]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The AI for Good initiative’s mission to unlock AI for humanity is emphasized in the summit report, which frames AI for Good as ensuring AI serves people and is truly ‘for good’ [S1].
MAJOR DISCUSSION POINT
AI for Good purpose – unlocking AI’s potential for humanity
AGREED WITH
Ambassador Reintam Saar, Speaker 1
Argument 2
AI for Good pillars – solutions, skills, standards as a year‑long movement
EXPLANATION
Fred explains that AI for Good is organized around three continuous pillars—solutions, skills, and standards—delivered through daily online events throughout the year. This structure ensures ongoing engagement beyond the annual summit.
EVIDENCE
He describes, “We have online events almost every day of the week, all year long… organized around three pillars. Solutions, skills, and standards.” [55-60]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The evolution of AI for Good into a year-round movement organized around solutions, skills and standards is described in the AI for Good Technology report [S1] and reinforced by the summary of its three-pillar model [S23].
MAJOR DISCUSSION POINT
AI for Good pillars – solutions, skills, standards as a year‑long movement
Argument 3
Standards development – AI native network standards, future 5G/6G/AI‑centric architectures under ITU
EXPLANATION
Fred highlights ITU’s work on developing standards for future networks, including 5G, 6G, and AI‑native architectures, to support the safe and interoperable deployment of edge AI. These standards are part of the broader AI for Good agenda.
EVIDENCE
He notes, “We have a standards work on future networks, basically 5G, 6G and beyond, and a pre-standardization effort on AI native networks.” [68-70]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
ITU’s standardisation work on AI-centric future networks, including 5G, 6G and AI-native architectures, is outlined in the ITU standards briefings [S21] and the follow-up on safe AI-driven futures [S22].
MAJOR DISCUSSION POINT
Standards development – AI native network standards, future 5G/6G/AI‑centric architectures under ITU
DISAGREED WITH
Alagan Mahalingam
S
Speaker 1
1 argument71 words per minute487 words405 seconds
Argument 1
AI for Good structure – three pillars and continuous activities throughout the year
EXPLANATION
Speaker 1 emphasizes that AI for Good is not just an annual summit but a year‑long programme built around three pillars, ensuring continuous engagement and impact. This framing underscores the ongoing nature of the initiative.
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The AI for Good report notes that the programme now runs continuously across the year, built around the three pillars of solutions, skills and standards [S1].
MAJOR DISCUSSION POINT
AI for Good structure – three pillars and continuous activities throughout the year
B
Brijesh Lal
2 arguments167 words per minute1277 words458 seconds
Argument 1
Edge convergence – communication, compute, and control make edge essential for safety‑critical tasks like haptics
EXPLANATION
Brijesh argues that the convergence of communication, compute, and control at the edge is crucial for safety‑critical applications such as haptics, where timing errors can be catastrophic. This convergence enables reliable, low‑latency processing close to the user.
EVIDENCE
He explains, “the reason why edge is becoming more and more important is this converge of communication, compute and control… enabled… for tasks in the area of haptics… require you to not miss or make mistakes because some of them are catastrophic.” [91-95]
MAJOR DISCUSSION POINT
Edge convergence – communication, compute, and control make edge essential for safety‑critical tasks like haptics
Argument 2
Context‑specific solutions – strong edge capability is needed to deliver locally relevant AI in the Global South
EXPLANATION
Brijesh stresses that delivering AI solutions tailored to local contexts in the Global South requires robust edge capabilities, as foundation models alone cannot address highly specific tasks. Edge computing enables the fine‑grained, context‑aware AI needed for diverse local challenges.
EVIDENCE
He says, “While it might not be easy to have foundation models… context has become increasingly important… People want solutions very specific to the task… can be best leveraged if there is a strong edge capability… important that the global south focuses on building its strength in the area of edge.” [96-99]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The Sri Lanka case study shows that without edge capability the solution failed, highlighting the need for strong edge infrastructure for context-specific AI in the Global South [S1]; a related discussion on delivering intelligence to the last mile reinforces this point [S24].
MAJOR DISCUSSION POINT
Context‑specific solutions – strong edge capability is needed to deliver locally relevant AI in the Global South
A
Alagan Mahalingam
2 arguments161 words per minute826 words307 seconds
Argument 1
Edge for connectivity – brings AI to under‑connected regions, improves speed and data sovereignty
EXPLANATION
Alagan describes how edge AI is deployed to reach under‑connected areas, improving response speed and preserving data sovereignty. He gives examples from Portugal’s farmer advisory system and its adaptation for Sri Lanka’s connectivity‑limited villages.
EVIDENCE
He recounts, “We chose to build edge AI… to bring technology to under-connected spaces and also to increase speed… In Portugal we built a hardware-software-AI solution for small-scale farmers… When we tried to deploy it in Sri Lanka, lack of reliable connectivity made the solution worthless, so we added edge with Raspberry Pi and local models, restoring functionality even without internet.” [202-215]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The deployment of edge AI in Portugal and its adaptation for connectivity-limited villages in Sri Lanka demonstrates how edge improves speed and preserves data sovereignty [S1].
MAJOR DISCUSSION POINT
Edge for connectivity – brings AI to under‑connected regions, improves speed and data sovereignty
Argument 2
Task‑driven models – lightweight, task‑specific AI models are required; large LLMs are unnecessary for edge
EXPLANATION
Alagan argues that edge deployments should start from the specific task and then create compact, task‑focused models rather than trying to run large language models. Model quantization, pruning, and distillation are essential to fit AI onto constrained edge devices.
EVIDENCE
He notes, “you shouldn’t try to think of the model and go find a solution. Instead think of the task and work backwards… you can’t do everything… we work on multiple technologies to quantize or prune the models in a way that creates a smaller version that does exactly what’s supposed to happen.” [230-236]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Critiques of relying on large LLMs and the push for task-specific, quantised models are discussed in the analysis of AI model efficiency [S26]; the emphasis on building a last-mile ecosystem rather than large models is echoed in the edge-centric strategy paper [S24].
MAJOR DISCUSSION POINT
Task‑driven models – lightweight, task‑specific AI models are required; large LLMs are unnecessary for edge
DISAGREED WITH
Sakshi Gupta
M
Mala Kumar
2 arguments108 words per minute301 words166 seconds
Argument 1
XR medical emergency – public 5G enables first‑responders with XR glasses and IoT wearables; private 5G supports on‑premise industrial tours
EXPLANATION
Mala presents a use case where public 5G connects first responders equipped with XR glasses and IoT wearables to provide real‑time vitals and remote expert guidance during cardiac emergencies. She also contrasts this with private 5G used for on‑premise industrial tours, highlighting different deployment scenarios.
EVIDENCE
She describes, “XR-assisted medical emergency… an SOS alert is sent… first responder arrives with XR glasses, IoT wearables and an AED kit. Patient vitals are overlaid on real-time video and sent to a medical expert who guides the response… This uses public 5G. For XR-assisted facility tours we use private 5G for on-premise HCI applications.” [172-186]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
A description of an XR-assisted medical emergency using public 5G, XR glasses and IoT wearables matches the scenario outlined in the supplemental case note [S2].
MAJOR DISCUSSION POINT
XR medical emergency – public 5G enables first‑responders with XR glasses and IoT wearables; private 5G supports on‑premise industrial tours
Argument 2
Open‑source AI – sharing open‑source models through AI for Good accelerates testing and deployment
EXPLANATION
Mala advocates for publishing AI models as open‑source within the AI for Good ecosystem, enabling the global community to test, fine‑tune, and validate them before real‑world deployment. This approach promotes transparency and collaborative improvement.
EVIDENCE
She says, “we would like to have some of our applications in the open source and have it in the best place, like ITU’s AI for Good… so the international community can access this open-source AI models, fine-tune them and do rigorous testing before bringing them to real-world deployment.” [187-189]
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The role of open-source AI in accelerating social-good projects is highlighted in the ‘Driving Social Good with AI’ report and the discussion on open-source LLMs for collaborative development [S15][S28].
MAJOR DISCUSSION POINT
Open‑source AI – sharing open‑source models through AI for Good accelerates testing and deployment
S
Sakshi Gupta
4 arguments172 words per minute671 words233 seconds
Argument 1
Device‑level edge AI – smartphones, cars, IoT devices now run on‑device inference, enabling offline AI use
EXPLANATION
Sakshi outlines how on‑device AI inference is now feasible on smartphones, cars, and IoT gadgets, allowing AI functionality without continuous internet connectivity. This shift supports privacy and resilience in diverse environments.
EVIDENCE
She notes, “AI being available at the edge… on-device capabilities coming onto smartphones with 10 billion-parameter models already running on-device… also coming to cars, IoT devices, smart glasses… you do not need to be connected; you can still use AI in flight mode.” [260-270]
MAJOR DISCUSSION POINT
Device‑level edge AI – smartphones, cars, IoT devices now run on‑device inference, enabling offline AI use
Argument 2
Edge benefits – reduced latency, enhanced security, privacy, low cost, and energy efficiency for Global South applications
EXPLANATION
Sakshi emphasizes that edge AI delivers lower latency, stronger security and privacy, lower cost, and reduced power consumption, which are critical for applications in the Global South where resources may be limited.
EVIDENCE
She states, “latency, security, privacy, personalization, low cost, low power are all very important factors for why AGI becomes important for Global South… impact on privacy, latency, bandwidth, architecture.” [274-279]
MAJOR DISCUSSION POINT
Edge benefits – reduced latency, enhanced security, privacy, low cost, and energy efficiency for Global South applications
Argument 3
Evaluation metrics – availability of edge hardware, connectivity quality, privacy safeguards, and data handling are key measures
EXPLANATION
Sakshi identifies the key metrics for assessing edge AI deployments: hardware availability, connectivity quality, privacy protections, and data‑handling practices. These metrics guide effective and responsible edge implementations.
EVIDENCE
She remarks, “what are the metrics to evaluate the usage of edge AI such as availability and capability of hardware at the edge and also the connectivity, privacy, and data issues?” [245-247]
MAJOR DISCUSSION POINT
Evaluation metrics – availability of edge hardware, connectivity quality, privacy safeguards, and data handling are key measures
DISAGREED WITH
Other panelists (implicitly Fred Werner, Alagan Mahalingam)
Argument 4
Bandwidth & latency reduction – FL aligns with edge inference, lowering bandwidth needs and improving real‑time performance
EXPLANATION
Sakshi explains that federated learning, by keeping training data at the edge and only sharing model updates, reduces bandwidth consumption and latency, complementing edge inference for real‑time AI services.
EVIDENCE
She observes, “FL … impact on latency and bandwidth … personalization of AI models is possible in real time… large-scale training can still happen in the core network but personalization of smaller models for the edge… reduces bandwidth because data no longer needs to be sent to the server.” [258-263]
MAJOR DISCUSSION POINT
Bandwidth & latency reduction – FL aligns with edge inference, lowering bandwidth needs and improving real‑time performance
R
Ranjitha Prasad
2 arguments168 words per minute842 words300 seconds
Argument 1
Privacy‑preserving intelligence – federated learning keeps raw user data at the edge while sharing only model updates
EXPLANATION
Ranjitha describes federated learning as a privacy‑preserving approach where raw user data never leaves the edge device; only aggregated model updates are transmitted to the cloud, protecting user privacy.
EVIDENCE
She explains, “privacy … bring code to the data, not take data to the code … federated learning … training happens at the edge, only certain metadata is given to the cloud.” [139-144]
MAJOR DISCUSSION POINT
Privacy‑preserving intelligence – federated learning keeps raw user data at the edge while sharing only model updates
Argument 2
FL use cases – traffic‑prediction for events and V2X road‑condition sharing illustrate FL’s practical impact
EXPLANATION
Ranjitha provides concrete examples where federated learning improves traffic prediction for large events and enables vehicle‑to‑everything (V2X) sharing of road conditions, demonstrating its real‑world utility.
EVIDENCE
She details, “traffic-prediction for a football match… each base station shares local traffic with the core network and MEC controller, which then routes traffic to reduce congestion. Another use case is V2X where each car talks to its edge server, then to a cloud server for a global model, enabling sharing of road-condition and accident information.” [146-152]
MAJOR DISCUSSION POINT
FL use cases – traffic‑prediction for events and V2X road‑condition sharing illustrate FL’s practical impact
A
Ambassador Egriselda Lopez
1 argument151 words per minute457 words180 seconds
Argument 1
Human‑centric AI – AI should be close to people and services, improving speed, cost, and privacy, especially where connectivity is limited
EXPLANATION
Ambassador Lopez stresses that AI must be deployed close to users and services (HAI) to enhance speed, reduce costs, and protect privacy, particularly in regions with limited connectivity. This human‑centric approach aligns AI with development goals.
EVIDENCE
She says, “HAI means simply using an AI closer to where things happen… closer to people, services, communities… can work better in places with limited connectivity… help with speed, cost, and privacy.” [313-318]
MAJOR DISCUSSION POINT
Human‑centric AI – AI should be close to people and services, improving speed, cost, and privacy, especially where connectivity is limited
A
Ambassador Reintam Saar
1 argument123 words per minute478 words231 seconds
Argument 1
Inclusive governance – the Global AI Governance Dialogue seeks practical, non‑duplicative outcomes, capacity building, and multi‑stakeholder trust
EXPLANATION
Ambassador Saar outlines the objectives of the Global AI Governance Dialogue: to produce actionable, non‑redundant outcomes, foster capacity building, and ensure inclusive, multi‑stakeholder participation built on trust and transparency.
EVIDENCE
He explains, “the dialogue needs to be inclusive… practical outcomes preferred over endless theoretical discussions… alignment with existing UN processes… avoid duplication… capacity building… trust, transparency, interoperability, equal participation.” [340-346]
MAJOR DISCUSSION POINT
Inclusive governance – the Global AI Governance Dialogue seeks practical, non‑duplicative outcomes, capacity building, and multi‑stakeholder trust
AGREED WITH
Fred Werner, Speaker 1
Agreements
Agreement Points
Edge AI is essential for delivering AI services in low‑connectivity and underserved contexts, improving speed, cost and privacy.
Speakers: Brijesh Lal, Alagan Mahalingam, Sakshi Gupta, Ambassador Egriselda Lopez
Edge convergence — communication, compute, and control make edge essential for safety‑critical tasks like haptics Context‑specific solutions — strong edge capability is needed to deliver locally relevant AI in the Global South Edge for connectivity — brings AI to under‑connected regions, improves speed and data sovereignty Edge benefits — reduced latency, enhanced security, privacy, low cost, and energy efficiency for Global South applications Human‑centric AI — AI should be close to people and services, improving speed, cost, and privacy, especially where connectivity is limited
All four speakers stress that placing AI close to the user (at the edge) is crucial for safety-critical, context-specific, and under-connected scenarios, delivering faster, cheaper and more private services. Brijesh highlights the convergence of communication, compute and control for haptics and the need for local context [91-95][96-99]; Alagan describes edge deployments that restore functionality in connectivity-limited villages and improve data sovereignty [202-215]; Sakshi points to latency, security, privacy and cost benefits for the Global South [274-279]; Lopez frames this as a human-centric approach that works better where connectivity is scarce [313-318].
POLICY CONTEXT (KNOWLEDGE BASE)
Recognized in policy discussions on democratizing AI through heterogeneous compute to balance centralized and edge resources, and highlighted in recent edge-computing partnerships targeting underserved areas and European digital sovereignty initiatives [S42][S60][S61].
Edge deployments should be driven by the specific task and use lightweight, task‑specific models rather than large generic LLMs.
Speakers: Alagan Mahalingam, Sakshi Gupta, Ranjitha Prasad
Task‑driven models — lightweight, task‑specific AI models are required; large LLMs are unnecessary for edge Device‑level edge AI — smartphones, cars, IoT devices now run on‑device inference, enabling offline AI use Privacy‑preserving intelligence — federated learning keeps raw user data at the edge while sharing only model updates
Alagan argues that edge solutions must start from the task and be distilled into small models via quantisation or pruning [230-236]; Sakshi notes that on-device inference now runs large models on smartphones and other devices, enabling offline use [260-266]; Ranjitha adds that federated learning keeps data on the edge and only shares model updates, supporting lightweight, task-specific intelligence [258-263]. All three converge on the need for compact, purpose-built models at the edge.
POLICY CONTEXT (KNOWLEDGE BASE)
Policy briefs and industry dialogues stress the need for task-specific, low-parameter models for edge scenarios, emphasizing open-source and energy-efficient designs [S49][S50].
Privacy preservation is a key driver for moving AI processing to the edge.
Speakers: Ranjitha Prasad, Sakshi Gupta, Ambassador Egriselda Lopez
Privacy‑preserving intelligence — federated learning keeps raw user data at the edge while sharing only model updates Edge benefits — reduced latency, enhanced security, privacy, low cost, and energy efficiency for Global South applications Human‑centric AI — AI should be close to people and services, improving speed, cost, and privacy, especially where connectivity is limited
Ranjitha describes federated learning as a privacy-preserving approach that keeps raw data on devices [139-144]; Sakshi highlights privacy as one of the primary benefits of edge AI, alongside latency and cost reductions [274-279]; Lopez reinforces that locating AI close to people improves privacy in low-connectivity settings [315-318].
POLICY CONTEXT (KNOWLEDGE BASE)
Privacy is foregrounded in AI governance frameworks, with calls for local processing to protect data sovereignty and compliance with emerging privacy standards [S42][S52][S63][S64][S65].
AI for Good’s overarching purpose is to serve humanity and must be pursued through inclusive, collaborative governance and multi‑stakeholder engagement.
Speakers: Fred Werner, Ambassador Reintam Saar, Speaker 1
AI for Good purpose – unlocking AI’s potential for humanity Inclusive governance – the Global AI Governance Dialogue seeks practical, non‑duplicative outcomes, capacity building, and multi‑stakeholder trust
Fred defines AI for Good as unlocking AI’s potential to serve humanity [26-27]; Ambassador Saar outlines the need for inclusive, capacity-building governance that avoids duplication and produces actionable outcomes [340-346]; Speaker 1 reinforces the year-long, three-pillar structure that underpins continuous collaboration (solutions, skills, standards) [55-60]. Together they present a unified vision of AI for Good as a human-centric, inclusive initiative.
POLICY CONTEXT (KNOWLEDGE BASE)
AI for Good principles are anchored in multi-stakeholder and inclusive governance models endorsed by UN-IGF, WHO, and AI for Good initiatives, emphasizing collaboration across sectors [S43][S45][S46][S47][S48].
Similar Viewpoints
Both emphasize that edge computing is the technical enabler that makes AI usable in environments with limited or unreliable connectivity, whether for safety‑critical haptic control or for agricultural advisory services in remote villages [91-95][202-215].
Speakers: Brijesh Lal, Alagan Mahalingam
Edge convergence — communication, compute, and control make edge essential for safety‑critical tasks like haptics Edge for connectivity — brings AI to under‑connected regions, improves speed and data sovereignty
Both link the privacy advantages of edge/federated learning to broader development benefits, arguing that keeping data local reduces bandwidth, improves latency and safeguards user privacy [274-279][139-144].
Speakers: Sakshi Gupta, Ranjitha Prasad
Edge benefits — reduced latency, enhanced security, privacy, low cost, and energy efficiency for Global South applications Privacy‑preserving intelligence — federated learning keeps raw user data at the edge while sharing only model updates
Both frame AI’s ultimate goal as serving people directly – Fred through the AI for Good mission, Lopez through the concept of Human‑Centric AI that brings technology nearer to users [26-27][313-318].
Speakers: Fred Werner, Ambassador Egriselda Lopez
AI for Good purpose – unlocking AI’s potential for humanity Human‑centric AI – AI should be close to people and services, improving speed, cost, and privacy, especially where connectivity is limited
Unexpected Consensus
Edge AI is equally critical in highly connected regions as in the Global South.
Speakers: Alagan Mahalingam, Sakshi Gupta, Brijesh Lal
Edge for connectivity — brings AI to under‑connected regions, improves speed and data sovereignty Device‑level edge AI — smartphones, cars, IoT devices now run on‑device inference, enabling offline AI use Edge convergence — communication, compute, and control make edge essential for safety‑critical tasks like haptics
Alagan notes that even in well-connected countries like the United States, edge AI is needed when traveling outside urban coverage [224-227]; Sakshi points out that on-device AI works everywhere, including in flight mode [260-266]; Brijesh stresses edge for safety-critical haptics regardless of overall network quality [91-95]. This convergence shows an unexpected consensus that edge is not only a solution for the Global South but a universal requirement.
POLICY CONTEXT (KNOWLEDGE BASE)
Statements from Global South perspectives and European strategies underline that edge AI benefits both low-connectivity regions and highly connected economies, reinforcing its universal relevance [S43][S58][S61].
Overall Assessment

The discussion shows strong convergence around four core ideas: (1) edge AI is indispensable for delivering inclusive, low‑latency, privacy‑preserving services in both underserved and well‑connected environments; (2) edge solutions must be task‑driven and lightweight, avoiding reliance on massive LLMs; (3) privacy is a primary justification for moving AI processing to the edge; (4) the AI for Good initiative should be pursued through inclusive, multi‑stakeholder governance that keeps humanity at the centre.

High consensus – the majority of speakers independently arrived at the same conclusions about the role of edge AI, privacy, and inclusive governance, indicating a solid shared understanding that will likely shape future standards, deployments and policy frameworks.

Differences
Different Viewpoints
Size and type of AI models suitable for edge deployment
Speakers: Alagan Mahalingam, Sakshi Gupta
Task‑driven models – lightweight, task‑specific AI models are required; large LLMs are unnecessary for edge Device‑level edge AI – smartphones, cars, IoT devices now run on‑device inference, 10 billion‑parameter models already running on‑device
Alagan argues that edge deployments should start from the task and use quantised, pruned, or distilled lightweight models rather than large language models, stating “you shouldn’t try to think of the model and go find a solution… you can’t do everything… we work on multiple technologies to quantize or prune the models” [230-236]. Sakshi, however, points out that today on-device inference can already handle very large models, noting “10 billion-parameter models already running on-device” on smartphones and that AI is becoming available on cars and IoT devices [262-264]. The two positions clash over whether edge AI can realistically accommodate very large models or must stay limited to small, task-specific ones.
POLICY CONTEXT (KNOWLEDGE BASE)
Debates on model size reference calls for simpler, lower-parameter models versus the push for larger generative models, reflecting tensions in standard-setting and resource constraints [S50][S51].
What metrics should be used to evaluate edge AI deployments
Speakers: Sakshi Gupta, Other panelists (implicitly Fred Werner, Alagan Mahalingam)
Evaluation metrics – availability of edge hardware, connectivity quality, privacy safeguards, and data handling are key measures No concrete metric framework was presented by other speakers, who focused on use‑cases, standards or technical solutions without specifying measurement criteria
Sakshi explicitly asks for a set of evaluation metrics for edge AI, listing “availability and capability of hardware at the edge, connectivity, privacy, and data issues” [245-247]. The remaining speakers (e.g., Fred’s discussion of standards [68-70] and Alagan’s focus on task-driven model design [230-236]) do not address measurement, leading to a disagreement on whether metric definition is a priority for the discussion.
Emphasis on standards development versus rapid, ad‑hoc edge deployments
Speakers: Fred Werner, Alagan Mahalingam
Standards development – AI native network standards, future 5G/6G/AI‑centric architectures under ITU Task‑driven, fast‑track edge deployments – focus on building lightweight models and hardware solutions without waiting for formal standards
Fred highlights the importance of standards, noting “We have a standards work on future networks… 5G, 6G and beyond, and a pre-standardization effort on AI native networks” [68-70]. Alagan stresses a pragmatic, task-first approach, saying “you shouldn’t try to think of the model… work backwards… we work on multiple technologies to quantize or prune the models” [230-236], implying that waiting for standards could delay impact. This creates a tension between a standards-centric roadmap and a fast-track deployment mindset.
POLICY CONTEXT (KNOWLEDGE BASE)
A recurring theme in standards discussions is the trade-off between timely deployment and thorough standards development, noted in multiple policy forums and reports [S42][S55][S56][S57][S62].
Unexpected Differences
Feasibility of very large on‑device models versus edge resource constraints
Speakers: Alagan Mahalingam, Sakshi Gupta
Task‑driven models – lightweight, task‑specific AI models are required; large LLMs are unnecessary for edge Device‑level edge AI – smartphones, cars, IoT devices now run on‑device inference, 10 billion‑parameter models already running on‑device
Alagan’s insistence that “you don’t want to use LLM for everything” and the need to prune models for edge devices [213-214][230-236] is surprising given Sakshi’s claim that current smartphones already host 10 billion-parameter models on-device [262-264]. The contrast between a strong resource-constraint view and a claim of existing massive on-device models was not anticipated in the broader discussion.
POLICY CONTEXT (KNOWLEDGE BASE)
Feasibility of very large on-device models is questioned in technical analyses citing compute limits and power consumption challenges for edge hardware [S51][S54].
Human‑centric AI versus technology‑first edge strategy
Speakers: Ambassador Egriselda Lopez, Alagan Mahalingam
Human‑centric AI – AI should be close to people and services, improving speed, cost, and privacy, especially where connectivity is limited Edge for connectivity – focus on technical performance, speed, and data sovereignty, with less explicit reference to human‑centred outcomes
Ambassador Lopez emphasizes that AI must be “closer to where things happen… closer to people, services, communities” to improve speed, cost and privacy [313-318]. Alagan, while discussing edge deployments, concentrates on technical aspects such as connectivity, speed, and sovereignty without explicitly framing the solution around human-centred outcomes [202-215][230-236]. The shift from a human-rights framing to a purely technical framing was unexpected.
POLICY CONTEXT (KNOWLEDGE BASE)
Human-centred AI is advocated in health and ethics forums, contrasting with technology-first approaches, highlighting the need to align AI tools with user workflows and ethical principles [S44][S45].
Overall Assessment

The panel largely converged on the importance of edge AI for the Global South and on AI for Good as a multi‑pillar, year‑round movement. Disagreements emerged around the appropriate size of models for edge deployment, the need for formal standards versus rapid, task‑driven implementations, and the definition of evaluation metrics. Unexpected tensions appeared between claims of massive on‑device models and the resource‑constrained view, as well as between a human‑centred AI narrative and a technology‑first edge strategy.

Moderate – while there is strong consensus on the overall goal (bringing AI to underserved regions), the differing technical approaches and measurement frameworks indicate a need for coordinated policy and research to reconcile model size expectations, standardisation timelines, and metric definitions. These divergences could affect the speed and inclusivity of edge AI roll‑out if not addressed.

Partial Agreements
All speakers share the overarching goal of using edge AI to serve the Global South and improve critical services, but they propose different pathways: Brijesh stresses the technical convergence needed for safety‑critical haptics [91-95]; Alagan highlights connectivity‑driven deployments and hardware adaptations for rural farmers [202-215]; Ranjitha promotes federated learning as the privacy‑preserving mechanism [139-144]; Sakshi focuses on device‑level inference and broader benefits such as latency and cost reductions [274-279]. The consensus is on the importance of edge AI, yet the means—hardware design, connectivity, privacy‑preserving training, or device‑level inference—diverge.
Speakers: Brijesh Lal, Alagan Mahalingam, Ranjitha Prasad, Sakshi Gupta
Edge convergence – communication, compute, and control make edge essential for safety‑critical tasks like haptics Edge for connectivity – brings AI to under‑connected regions, improves speed and data sovereignty Privacy‑preserving intelligence – federated learning keeps raw user data at the edge while sharing only model updates Edge benefits – reduced latency, enhanced security, privacy, low cost, and energy efficiency for Global South applications
Takeaways
Key takeaways
AI for Good’s overarching goal is to unlock AI’s potential for humanity, organized as a year‑long movement built around three pillars – solutions, skills, and standards. Edge AI is essential because the convergence of communication, compute and control enables safety‑critical and context‑specific applications, especially in the Global South where connectivity is limited. Practical edge AI use cases were highlighted: XR‑enabled medical emergency response (public 5G), on‑premise industrial XR tours (private 5G), device‑level inference on smartphones, cars and IoT, federated‑learning‑driven traffic prediction and V2X, and low‑cost agricultural advisory systems. Task‑driven, lightweight models are preferred over large foundation models for edge deployments; model distillation, quantisation and pruning are key techniques. Federated learning is positioned as a privacy‑preserving enabler that keeps raw data at the edge while sharing model updates, reducing bandwidth and latency. Standards work is underway at ITU on AI‑native networks, future 5G/6G architectures, and quality‑of‑experience metrics for multimodal (including haptic) services. Human‑centric AI – placing intelligence close to people, services and communities – improves speed, cost, privacy and data sovereignty. The upcoming UN Global AI Governance Dialogue (July, Geneva) will focus on inclusive, practical outcomes, capacity‑building and avoiding fragmented approaches.
Resolutions and action items
Continue AI for Good activities throughout the year (online events, challenges, skill‑building sandboxes, standards work). Develop and publish AI‑native network standards, including 5G/6G and edge‑centric architectures, under ITU’s coordination. Promote open‑source AI models via the AI for Good sandbox to enable community testing and rapid deployment. Leverage Qualcomm’s Tech‑for‑Good programme to mentor and fund startups building edge AI solutions, especially in the Global South. Organise the first UN Global AI Governance Dialogue in July 2024, ensuring multi‑stakeholder participation and alignment with existing UN processes. Encourage participants to adopt federated‑learning approaches for privacy‑sensitive use cases such as traffic prediction and V2X. Panelists and researchers (e.g., Brijesh Lal, Alagan Mahalingam, Ranjitha Prasad) to contribute findings to ITU standardisation work and share best‑practice reports.
Unresolved issues
Specific metrics and benchmarks for evaluating edge AI deployments (e.g., hardware availability, latency thresholds, privacy safeguards) remain to be finalised. How to ensure interoperability of heterogeneous edge devices and haptic interfaces across different manufacturers is still an open challenge. Funding mechanisms and sustainable business models for scaling edge AI solutions in under‑connected regions were discussed but not concretised. The balance between on‑device training versus cloud‑based training for federated learning, especially for large‑scale models, needs further clarification. Mechanisms for coordinated data sharing and knowledge transfer between UN agencies, national governments and private sector partners were mentioned but not detailed.
Suggested compromises
Adopt a task‑centric approach: design lightweight, purpose‑built models for edge rather than deploying full‑scale LLMs, thereby reducing resource demands. Combine cloud and edge intelligence – keep heavy training in the cloud while moving inference and privacy‑preserving updates to the edge. Use open‑source model repositories to allow multiple stakeholders to test, fine‑tune and validate solutions before large‑scale rollout. Encourage regional pilots (e.g., in India, Sri Lanka, Portugal) that can be adapted to other contexts, avoiding a one‑size‑fits‑all solution. Align standards development with existing UN frameworks to prevent duplication and promote interoperability across regions.
Thought Provoking Comments
What if the last thing that humans ever invent is invention itself? … if AI is the last thing we ever invent, we must ensure it is for good.
Frames AI as the ultimate invention, raising the stakes of AI safety and aligning the entire summit around the responsibility of shaping AI’s purpose.
Set a philosophical tone that guided the rest of the session, prompting speakers to justify why AI must be directed toward societal benefit and leading to the emphasis on standards, governance, and ‘AI for Good’ as a movement.
Speaker: Fred Werner
We are entering a zero‑click world where agents act on our behalf without waiting for prompts.
Introduces the concept of autonomous AI agents, moving the discussion from AI as a tool to AI as an independent actor.
Shifted the conversation toward the need for edge intelligence and real‑time decision‑making, paving the way for the panel’s focus on edge AI, haptics, and low‑latency use cases.
Speaker: Fred Werner
The convergence of communication, compute and control makes edge capability essential, especially for haptics where latency and accuracy are life‑critical.
Highlights a concrete technical challenge (haptics) that illustrates why edge AI is not just a convenience but a safety requirement.
Prompted deeper technical discussion on split‑control architectures and intent‑based signal processing, influencing later speakers to address latency, privacy, and reliability.
Speaker: Brijesh Lal
Federated learning brings the code to the data, preserving privacy while enabling sub‑10 ms latency for mission‑critical tasks.
Connects a cutting‑edge ML paradigm directly to the core themes of edge AI, privacy, and bandwidth constraints, offering a practical solution to the problems raised earlier.
Expanded the conversation from hardware constraints to algorithmic strategies, leading panelists to discuss how models can be trained locally and only metadata shared.
Speaker: Ranjitha Prasad
In XR‑assisted medical emergencies, public 5G delivers real‑time vitals to responders, while private 5G enables on‑premise HCI for industry 5.0 applications.
Provides a vivid, real‑world example of edge AI improving health outcomes, illustrating both public and private network roles.
Illustrated the societal impact of edge AI, reinforcing the ‘AI for Good’ narrative and prompting other panelists to consider use‑case diversity across sectors.
Speaker: Mala Kumar
When we built a farmer‑advisory system for Portugal and tried to deploy it in Sri Lanka, we realized we needed edge compute because connectivity was unreliable; we even imagined putting a data centre on a tuk‑tuk.
Combines a personal story with a creative solution, emphasizing that edge AI is essential not only in remote areas but also in well‑connected regions when connectivity drops.
Served as a turning point that highlighted the universality of edge challenges, encouraging the audience to think beyond geographic stereotypes and to prioritize task‑first model design.
Speaker: Alagan Mahalingam
Training can stay in the cloud, but inference is moving to the edge—smartphones now run 10‑billion‑parameter models on‑device, enabling AI without any network connection.
Clarifies the distinction between training and inference and showcases the rapid hardware advances that make edge AI feasible at consumer scale.
Broadened the scope of the discussion from specialized industrial deployments to everyday devices, reinforcing the argument that edge AI is a global, not just a niche, phenomenon.
Speaker: Sakshi Gupta
HAI means using AI closer to where things happen—closer to people, services, and communities—so it can work better where connectivity is limited, reducing cost and preserving privacy.
Synthesises technical points into a policy‑oriented definition, linking edge AI directly to development goals and human‑centered design.
Unified the technical and governance strands of the session, setting the stage for the upcoming Global AI Dialogue and emphasizing the need for inclusive standards.
Speaker: Ambassador Egriselda Lopez
The Global AI Dialogue will focus on practical outcomes, inclusivity, and capacity‑building, avoiding duplication and fragmentation across national and regional approaches.
Articulates a concrete roadmap for international cooperation, translating the technical insights of the panel into actionable policy direction.
Provided a concluding turning point that moved the conversation from discussion to commitment, encouraging participants to see their technical work as feeding into a larger governance framework.
Speaker: Ambassador Reintam Saar
Overall Assessment

The discussion was driven forward by a series of pivotal remarks that moved the conversation from abstract philosophical concerns about AI’s ultimate role, through concrete technical challenges of edge computing, to real‑world applications and finally to policy and governance. Fred Werner’s opening question set a high‑level purpose, which was then grounded by Brijesh’s edge‑haptics argument, Ranjitha’s federated learning solution, and Alagan’s farmer‑centric deployment story. These insights reframed edge AI as a universal necessity rather than a niche technology. Subsequent contributions from Mala, Sakshi, and the ambassadors linked the technical possibilities to societal impact and global coordination, culminating in a clear call for inclusive standards and actionable outcomes. Collectively, these comments shaped a narrative that progressed from vision to implementation to governance, ensuring the panel remained focused, interdisciplinary, and outcome‑oriented.

Follow-up Questions
What metrics should be used to evaluate edge AI deployments (e.g., hardware availability, connectivity, privacy, data issues)?
Understanding appropriate evaluation metrics is crucial for assessing the effectiveness and scalability of edge AI solutions across different regions and use cases.
Speaker: Sakshi Gupta
How can XR applications be made open‑source and shared via platforms like ITU AI for Good to enable broader testing and fine‑tuning?
Open‑source XR models would allow the international community to validate, improve, and deploy solutions more rapidly, fostering collaboration and standardisation.
Speaker: Mala Kumar
Is it feasible to deploy edge compute resources on mobile units such as tuk‑tuks to serve remote villages, and what are the technical and economic implications?
Exploring mobile edge data centres could provide connectivity and AI processing in underserved areas, addressing a key challenge for the Global South.
Speaker: Alagan Mahalingam
What standards are needed for AI‑native networks, future (5G/6G) edge architectures, and quality‑of‑experience for multimodal applications?
Developing robust standards will ensure interoperability, security, and performance as AI moves from cloud to edge across diverse deployments.
Speaker: Fred Werner
How can latency and synchronization issues in haptic data transmission be mitigated to ensure reliable real‑time experiences?
Haptic feedback is latency‑sensitive; research is required to define QoE metrics and edge processing techniques that meet stringent timing constraints.
Speaker: Brijesh Lal
What are the privacy, latency, bandwidth, and architectural impacts of federated learning in telecom networks, and how can they be optimised for edge AI?
Federated learning promises privacy‑preserving AI at the edge, but its practical effects on network resources and model performance need systematic investigation.
Speaker: Ranjitha Prasad
How can member‑state expertise be collected and synthesised into a concrete roadmap for the UN Global Dialogue on AI Governance?
A structured methodology for gathering and translating stakeholder wisdom into actionable policy is essential for effective, inclusive AI governance.
Speaker: Ambassador Reintam Saar
What evidence is needed to demonstrate that edge AI improves development outcomes (speed, cost, privacy) in low‑connectivity settings?
Empirical studies showing tangible benefits of edge AI in the Global South will support policy decisions and investment in such technologies.
Speaker: Ambassador Egriselda Lopez

Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.