Digital Democracy Leveraging the Bhashini Stack in the Parliamen
20 Feb 2026 12:00h - 13:00h
Digital Democracy Leveraging the Bhashini Stack in the Parliamen
Session at a glance
Summary
This discussion centered on the launch of a policy report and developer toolkit for building open and responsible voice technology ecosystems in India, developed through an Indo-German partnership. The event featured presentations from various stakeholders including government officials, researchers, and industry practitioners who explored the challenges and opportunities in developing inclusive voice AI systems for Indian languages.
Amitabh Nag from Bhashini emphasized the importance of treating voice technology development as a continuous process rather than a static solution, highlighting the need for scalable systems that embrace diversity and inclusion by design. He stressed that unlike traditional digital systems, AI solutions require constant upgrades and must accommodate the vast linguistic and cultural diversity across different regions and communities.
Dr. Ariane Ahildur-Brandt from the German Federal Ministry presented the collaboration as an example of cooperative rather than competitive AI development, emphasizing how voice technology can serve as a gateway to digital inclusion for millions with limited literacy or device access. The partnership has created open voice technologies for nine Indian languages that can be used by NGOs, government agencies, and companies for applications in healthcare, agriculture, and other sectors.
Harleen Kaur outlined the policy framework built around four pillars: treating foundational datasets as public goods, institutionalizing sustainable open-source infrastructure, building open and representative models, and strengthening responsible deployment. The developer toolkit focuses on ensuring representation, data quality, and embedding responsible AI practices throughout the development lifecycle.
The panel discussion revealed several critical challenges in the voice technology ecosystem. Dr. Prasanta Ghosh highlighted the complexity of evaluation systems, noting that even humans from the same region often disagree on transcriptions, suggesting the need for multi-layered evaluation approaches rather than simple word-error-rate comparisons. Industry representative Kritika K.R. emphasized the importance of domain-specific adaptations and feedback loops for real-world applications, while also noting the need for scalable infrastructure and edge deployment capabilities.
Legal expert Thomas Vallianeth addressed the intersection of voice technology with copyright and privacy laws, advocating for careful documentation and safeguards from the beginning of data collection through deployment. He emphasized that while legal principles exist, the evidentiary standards for AI-related disputes are still evolving, making trust and proper documentation crucial for the ecosystem.
The discussion concluded with calls for more collaborative frameworks, including national-level evaluation standards and competitive yet cooperative approaches to advancing voice technology in Indian languages, ultimately aiming to create more inclusive and accessible AI systems for diverse populations.
Keypoints
Major Discussion Points:
– Scaling AI solutions with diversity and inclusion as core design principles: The discussion emphasized that unlike traditional digital systems that work on standards and exclude outliers, AI systems must be built with diversity and inclusion at their foundation, accommodating different languages, cultures, and contexts with continuous upgrades rather than static solutions.
– Treating foundational speech datasets as digital public goods: A key policy recommendation was to treat speech datasets as public infrastructure, similar to digital public goods (DPGs), requiring government stewardship, sustainable funding, and collaborative governance frameworks to support languages that may not be commercially viable.
– Challenges in evaluation and benchmarking for Indian language AI systems: The panel highlighted that current evaluation methods are inadequate for Indian contexts, noting that even humans from the same region often disagree on transcriptions, requiring new multi-layered evaluation frameworks that account for linguistic diversity and subjective interpretation.
– Legal and ethical considerations in voice AI development: The discussion covered the intersection of copyright law, data governance, and privacy concerns, emphasizing the need for proper documentation, consent protocols, and safeguards throughout the AI development lifecycle to build trust in the ecosystem.
– Creating sustainable feedback loops for continuous model improvement: Speakers discussed the importance of moving beyond static data collection to dynamic systems where user feedback, enterprise applications, and real-world deployments continuously contribute to improving AI models and datasets.
Overall Purpose:
The discussion served as a launch event for a policy report and developer toolkit focused on building an open and responsible voice technology ecosystem in India. The goal was to present findings from a German-Indian partnership and facilitate dialogue among stakeholders about best practices, challenges, and frameworks for developing inclusive speech AI systems that serve India’s diverse linguistic landscape.
Overall Tone:
The discussion maintained a collaborative and constructive tone throughout, characterized by academic rigor combined with practical industry insights. The atmosphere was optimistic about the potential for voice AI to enable digital inclusion, while being realistic about the complex challenges involved. The tone remained consistently professional and solution-oriented, with speakers building upon each other’s points rather than presenting conflicting viewpoints. There was a strong emphasis on partnership and cooperation, reflecting the international collaborative nature of the initiative being launched.
Speakers
Speakers from the provided list:
– Amitabh Nag – CEO of DIBD (Digital India Bhashini Division)
– Moderator – Event moderator (specific name not clearly identified in transcript)
– Ariane Ahildur (full name: Ariane Ahildur-Brandt) – Director General of the Department for Global Health, Equality of Opportunity, Digital Technologies and Food Security of the German Federal Ministry for Economic Cooperation and Development
– Harleen Kaur – Research Manager, Digital Futures Lab
– Kritika K.R. – Head Artificial Intelligence and Product Researcher, SanLogic
– Nihar Desai – Head of JNI, Panel Discussion Moderator
– Prasanta Ghosh (Dr. Prasanta Ghosh) – Associate Professor at the Indian Institute of Science
– Thomas J. Vallianeth (Thomas Valunith/Thomas Salenat in transcript) – Counsel, Trilegal
Additional speakers:
– Shailendra Pal Singh – Senior General Manager, Bhashani
Full session report
This discussion centered on the launch of a policy report and developer toolkit for building open and responsible voice technology ecosystems in India, developed through an Indo-German partnership. The event brought together government officials, academic researchers, industry practitioners, and legal experts to explore the challenges and opportunities in developing inclusive voice AI systems for India’s diverse linguistic landscape.
The Paradigm Shift: Embracing Diversity in AI Design
Amitabh Nag from Bhashini highlighted a fundamental shift in how AI systems must be approached compared to traditional digital technologies. “AI, and the reason for this is diversity. You know, each person is different. Each language is different. Each culture is different… unlike the earlier digital systems which used to work on only standards… Here, inclusion is the name of the, inclusion is part of the design, diversity is part of the design.”
This represents a departure from conventional engineering practices that typically standardize away outliers. Unlike traditional digital systems that could operate for years with minimal updates, AI solutions for diverse populations require constant upgrades with shelf lives as short as three to six months. This reality stems from the inherent diversity across languages, cultures, and contexts that must be accommodated rather than eliminated.
Indo-German Partnership: Cooperation Over Competition
Dr. Ariane Ahildur-Brandt, Director General of the Department for Global Health, Equality of Opportunity, Digital Technologies and Food Security of the German Federal Ministry for Economic Cooperation and Development, positioned the collaboration as an alternative to competitive AI narratives. “At a time when artificial intelligence is often framed as a global competition, this report offers a different narrative, and this is a narrative of cooperation.”
The partnership has created open voice technologies for nine Indian languages, with practical applications including voice assistance for health workers to improve women’s healthcare and agricultural advisory systems for farmers. These deployments demonstrate how voice technology can serve as a gateway to digital inclusion, particularly for populations with limited literacy or device access.
The collaboration aligns with international frameworks, including the Hamburg Declaration on Responsible AI for Sustainable Development Goals, which has support from more than 50 stakeholders.
Policy Framework and Developer Toolkit Launch
Harleen Kaur presented the comprehensive policy report and developer toolkit, emphasizing that “policy intent alone does not ensure inclusive AI systems.” The toolkit focuses on three broad areas designed to translate principles into practical implementation guidance for developers working in diverse linguistic contexts.
The policy framework addresses ecosystem-level challenges through multiple interconnected approaches, including treating foundational datasets as public goods, institutionalizing sustainable open-source infrastructure, building representative models through locally relevant benchmarks, and strengthening responsible deployment practices.
The developer toolkit serves as a bridge between policy intentions and practical implementation, providing concrete guidance for developers to embed inclusive practices throughout the development lifecycle rather than treating them as an afterthought.
The Fundamental Evaluation Challenge
One of the most significant insights emerged around the challenges of evaluating voice AI systems for diverse populations. Dr. Prasanta Ghosh from the Indian Institute of Science presented compelling evidence that traditional evaluation methods are inadequate: “if you give a piece of audio to two individuals, they never exactly agree on what they hear… two people from the same district… just three kilometers away in terms of their location, but still they did not agree how that should be written from the audio they hear.”
This observation fundamentally challenges the assumption that objective ground truth exists in speech recognition evaluation. If human annotators from the same geographic area cannot agree on transcriptions, then traditional word error rate comparisons become questionable as primary evaluation metrics.
Amitabh Nag reinforced this perspective by emphasizing audience acceptability over perfect accuracy: “Rather than looking at it only from a perspective of application or academics, we would have to look at it from a perspective of audience.” This user-centric approach acknowledges that language use varies significantly based on context, application, and user expectations.
Panel Discussion: Real-World Perspectives
The panel discussion, moderated by Nihar Desai, brought together diverse stakeholders to explore practical implementation challenges. Kritika K.R. shared industry perspectives on deploying voice AI across healthcare, manufacturing, and automotive sectors, emphasizing the importance of domain-specific adaptation and continuous feedback loops between deployed systems and model improvement.
Thomas J. Vallianeth provided legal insights, highlighting a critical misconception: “there is a common myth in India that anything that is public is freely available.” This observation underscored the complex intersection of voice technology with copyright law, data governance, and privacy regulations, emphasizing the need for careful consideration of licensing and provenance throughout development.
Continuous Improvement Through Systematic Feedback
A recurring theme was the need for systematic approaches to continuous improvement. Amitabh Nag outlined various approaches to ongoing dataset creation, including traditional field collection, feedback from deployed applications, and harvesting digital data from various sources.
The concept of creating “improvement corpus” from deployed systems represents a significant opportunity for enhancement. When users interact with AI systems and provide corrections, this feedback can be channeled through vetting pipelines to create training data for model improvements, though this requires conscious program design rather than ad hoc collection.
Towards National Coordination
Despite the emphasis on diversity and local adaptation, there was consensus on the need for national-level coordination. Dr. Ghosh proposed creating a national evaluation framework under Bhashini with annual assessments and leaderboards, similar to those that have driven progress in English language AI systems.
This approach reflects recognition that diversity and standardization can coexist when properly structured. A national framework could provide consistent evaluation methodologies while accommodating India’s linguistic and cultural diversity, creating competitive yet collaborative environments that drive systematic progress across the ecosystem.
Ongoing Challenges and Future Directions
The discussion revealed several unresolved challenges requiring continued attention. The fundamental question of evaluation when human perception varies significantly remains open, requiring experimentation with multi-layered evaluation approaches and audience-centric metrics.
Legal and regulatory frameworks continue evolving, particularly around evidentiary standards for AI-related disputes and appropriate safeguards for different applications. Sustainability models for maintaining voice technology infrastructure as public goods also need further development.
The complexity of these challenges requires ongoing coordination between academic researchers, industry practitioners, government agencies, and legal experts. The collaborative approach demonstrated by the Indo-German partnership provides a foundation for continued progress toward more inclusive and effective voice technology systems, offering insights relevant to multilingual societies and underserved populations globally.
The event ultimately presented a vision for responsible voice AI development that prioritizes inclusion, cooperation, and continuous improvement, while acknowledging the significant technical, legal, and practical challenges that remain to be addressed.
Session transcript
including, you know, Southeast Asia as well as Africa and other places. So from that perspective, it is very important that we scale these solutions. We have policies, standards, toolkits which are developed which can be actually replicated. And frankly speaking, in this area, in this situation, nothing is static. You have a shelf life which is sometimes three months or six months or even less. Yes. So we have to continuously upgrade the things as we go by. You know, we can’t be saying that this is what we have done, unlike a machine which we have built up and it works for six years or five years. There is no guarantee, no warranty in these kind of systems which we are building in AI.
AI, and the reason for this is diversity. You know, each person is different. Each language is different. Each culture is different. So there is… There is huge amount of diversity and we have to live with the diversity unlike the earlier digital systems which used to work on only standards. You know, they had standards and they would perhaps keep the outliers away. Here, inclusion is the name of the, inclusion is part of the design, diversity is part of the design. And we would perhaps have to go step by step to define those diversities so that they start becoming standards. Right. You know, it’s a very different kind of a setup which is there and happy to be part of this journey, happy to, happy and acknowledged to the help which is being provided.
And hopefully we are going to get across to the next level and higher steps in the journey as we go by in future. Thank you very much.
Thank you, Mr. Nag for your insightful words and also for your incredible support throughout the last year over the course of the program. Right. Thank you. I will now invite Dr. Ariane Ahildur -Brandt, Director General of the Department for Global Health, Equality of Opportunity, Digital Technologies and Food Security of the German Federal Ministry for Economic Cooperation and Development to deliver the keynote address. Thank you. Thank you.
Dear Mr. Naack, dear partners, distinguished guests, it is a great pleasure to welcome you to this launch today. We present to you the Policy Report and Developers Toolkit Building on Open and Responsible Voice Technology Ecosystem in India. The report and the toolkit are the impressive result of a very productive partnership between Germany and India. And it is the result of a joint effort involving a group of distinguished partners and experts. This is why I would like to start by thanking you, Mr. Nack, and your colleagues from Ascini, for the excellent cooperation. And I would like to thank the Digital Futures Lab, Art Park, TriLegal, and NASSCOM for their invaluable support. Dear guests, you will find that the report and toolkit that we are presenting today is full of best practices and lessons learned.
It will provide guidance and hands -on advice to policymakers and to the tech community alike. But for me, this report is more than useful and more than practical content. It also conveys a shared conviction, shared values, and a shared vision for digital inclusion. In fact, when it comes to inclusion, voice technology has a key role to play. For millions of people. Voice is the most natural and powerful interface to the digital world, especially for those with limited literacy or access to digital devices. When voice AI works in local languages and dialects, it will become a gateway to public services, healthcare, education, and economic participation. When it does not, AI risks reinforcing existing devices and may even become an instrument for exclusion.
This is why responsible, inclusive voice AI is not just a technical issue. As I said, it is part of a shared vision, a shared vision between India and Germany. At a time when artificial intelligence is often framed as a global competition, this report offers a different narrative, and this is a narrative of cooperation. The Indo -German Partnership on AI, and particularly on language, and voice technologies shows what is possible when we join forces. Together with Bashini and the Indian Institute of Science, our initiative Fair Forward has created open voice technologies for nine Indian languages. These language models can now be used by NGOs, state agencies and companies. For example, they can be integrated into voice assistance for health workers, which in turn can improve health care for women.
Or they can be used to advise farmers on crop management. This collaboration, based on the principles of openness, fairness and responsibility, is the foundation for AI that truly serves the common good. And it contradicts those who claim that only fierce competition can generate prosperity and innovation. Ladies and gentlemen, this approach, closely aligns with the principles articulated by the International Cooperation on Climate Change. in the Hamburg Declaration on Responsible AI for Sustainable Development Goals. This declaration, presented by BMZ, our ministry, and UNDP last year, has been endorsed by more than 50 stakeholders already, including governments, international organizations, NGOs, and companies. The declaration reminds us that AI should serve the people and the planet, strengthen inclusion, and support sustainable development.
And our report here is a very practical and relevant contribution to that agenda, translating shared principles into concrete guidance. So let us thus deepen cooperation, strengthen trust, and build voice technologies that truly speak to everyone. Thank you for your attention.
Thank you so much, Dr. Hillbrand. We shall now move on to the formal launch of the report and toolkit. I’ll invite all the representatives of the consortium from GIZ, Tri -Legal, Art Park, NASSCOM, Digital Futures Lab to please come on stage. And Mr. Nag to present the data. Thank you. Thank you. Thank you. Thank you. Now that we’re done with the formal launch of the report and policy toolkit, just to give you a brief overview, I invite Ms. Harleen Kaur, Research Manager, Digital Futures Lab, to present the report.
Good morning, everyone, and thank you for being present. on a Friday morning for the launch of this report, as well as the developer toolkit. So I’ve linked the outputs in case you’d want to see them. If you can take a quick photo, and I’ll move towards discussing the high points of the findings that we had both for our policy report as well as developer’s toolkit. So when we began this work last year, we found that the challenges that are there in the voice tech arena, they are not limited to data collection alone. So the challenges are multi -layered that start right at the data collection stage and curation stage, but then move on to model development, where we see linguistic diversity gaps, lack of standards, uneven documentation, unclear data ownership and structures being a problem.
But then when we move on to the, hosting and licensing aspect, long -term infrastructure costs, costs, governance of open source assets, as well as sustainability of shared resources is something that we felt was a very important problem that needed to be solved in a certain manner. And the last is downstream deployment and impact, where bias, exclusion and lack of accountability for misuse become more visible. All of these are essentially starting at the data collection stage, but they move on to the life cycle of the voice technology ecosystem in India, specifically when you feel like supporting an open voice ecosystem in India. To lay down our approach for this project, we thought about how can we move on from the traditional government systems where government has primarily acted as a regulator, it enforces rules, it corrects market failures, to a newer active role, and that we have seen with Bhashani.
We encourage governments across the world to adopt this framework where the government acts as a steward of public good. ecosystem convener, as well as a standard setter, not just through licenses, but actually through practice as well. This is the overview of our policy framework. Based on this approach, we have structured our policy framework around the four pillars that you see on the screen. The first is treating foundational data sets as public goods. Second is institutionalizing sustainable open source infrastructure. Third is building open and representative models. And finally, strengthening responsible deployment. And what do we mean when we say this? When we say treat foundational data sets as public good, we are saying that government should be encouraging both funding and convening for public good functions.
For example, supporting languages that are not commercially viable as such. Institutionalizing governance. Governance framework. Thank you. to strengthen RAI practices, for example, through procurement, etc. On open representative models, we believe that local and contextually relevant benchmarks that are curated by government bodies not just at the center, but at the relevant diversity ecosystem, whether it is state, district, etc., is important. Shared national compute infrastructure, preferential treatment to open source ecosystem is something that we propose. On open source infrastructure itself, standardization of documents and promoting collaborative data steward models is something that has already been written in the report. Strengthening responsible deployment, public value sharing is another aspect of the report. We believe that public value sharing comes not just from financial arrangements, but also a buy -in of communities into what kind of… uses of voice technology are there.
And of course, supporting public literacy to protect against misuse and preventing harms is the policy side of our suggestion. Moving on to developer’s toolkit. You know, policy intent alone does not ensure inclusive AI systems. So alongside the policy framework, we’ve developed a developer toolkit that translates some of these principles into practice for developers. So it focuses on three broad areas, representation being the foremost through diversity planning, et cetera. Second being data quality and evaluation. And the third one being embedding RAI practices throughout the lifecycle of development of open voice I’ll just give you a brief overview of what we mean when we say this. So for developers, we have a toolkit that includes best practices that we’ve seen in industry.
And we have a toolkit that we’ve seen in India and outside on what does it mean to ensure adequate representation. on what does it mean to ensure adequate representation. So we have a toolkit that we’ve seen in India and outside So we have a toolkit that we’ve seen in India and outside on what does it mean to ensure adequate representation. Things like having a diversity wish list, making sure that you’re not collecting data from one source, applying linguistic expertise, using synthetic data, training model for linguistic and environmental nuances, and also layered data strategy. Which again means that don’t just use one source of data. Don’t do active or passive collection alone. Use a hybrid layered structure to make your models more diverse.
Once the developer move on from data collection to curation, we suggest many, many ways. This is just a very bird’s eye view overview in which data quality can be enhanced in the constraints that we operate in, in countries like India. And there are suggestions to make the applications inclusive and useful in practice, including robust transcription standards, contextual benchmarks. using data cards, model cards that are standardized, as well as continuous post -deployment monitoring. You can find more details in the report itself. And the last aspect of the developer’s toolkit is actually embedding RAI practices. We’ve taken another lifecycle framework within this where we believe that RAI practices are not the domain of policy alone. At enterprise startup developer level, ensuring a framework that serves to support them by providing them clarity on what does it mean when we say your output should be responsible.
So things like be mindful of engagement with the communities from whom you are taking data, annotation is happening, consent protocols, privacy enhancing techniques. So this report essentially is compliance plus. It actually shares practices that we believe are useful to promote open, responsible AI voice technology ecosystem. Please feel free to engage with the reports We’ll be very happy to take your comments, suggestions Thank you so much
Thank you, Harleen We shall now move on to a short panel discussion On voice technologies in India Unpacking the present and future Of the voice AI application ecosystem For India and beyond Joining us today, I will invite to the stage Mr. Amitabh Nag, CEO of DIBD Dr. Prasanta Ghosh, Associate Professor At the Indian Institute of Science Ms. Kritika K.R., Head Artificial Intelligence And Product Researcher, SanLogic Mr. Thomas Valunith, Counsel Trilegal And this discussion will be moderated By the Board of Directors of the Indian Institute of Science And Product Researcher, SanLogic Mr. Nihar Desai, Head of JNI Thank you.
Hello. Hello. Am I audible? Okay. Thanks everybody for joining. So, I just delving right deep into it. My first question to you would be Mr. Nag. As we saw in the toolkit, we were arguing that data set like foundational data sets, speech data sets, must be treated as DPIs and DPGs and hence be available in general. From your experience in driving this ecosystem for about two years since I’ve been a part at least, what does it take to continue creation, ongoing facilitation of such innovations being put up as a digital public good while ensuring trust safety, right? And is there a way for us to have a flywheel of data sorts, data goods of sorts?
Yeah, that’s a very important aspect of what we should be doing. That means continue the creation of data sets because it will then improve the models as we go by. Now, continuation of creation of data sets are, I would say that these are going to be in two or three ways, you know. One is the way which we have been… doing, which is the brute data collection, which is going to the various fields and then picking up the data from there and then creating the diversity which is required to actually build the model. So that is one way of doing it and that will continue. We will have to keep the focus with respect to saying that now I am doing for this particular area, this particular dialect, this particular language, while as it will be for other language in some other way.
The second is to actually look at using the products which have been developed using these models and creating such open domain activities to create the digital data. So you are creating the digital data which you are speaking, automatically creating the parallel corpus and then finding a way to actually vet this out and annotate and label and saying that, okay, this is the improvement corpus. That is the second thing. So one, you are creating a primary corpus. Second is… you are creating an improvement corpus which can be again fed back to the model and say that this is what is to be used and that is a big area of work as we look at. Allied to that is a lot of also the digital data is getting created any which way in the open domain which we can actually use to build the corpus again.
So you know YouTube videos today the world is more digital than it was yesterday. But the conscious way of looking at it as a program is what is required. How do I look at it as a program that I will be creating a data corpus at various places and this need not be necessarily an open domain. Open domain is kind of an easy way to work upon it. It can be a closed domain as well that there is an application which is working in an enterprise or a government and the people there are given an option to give suggestions to the translations or the answers or the things which you have gone in and that can get into a wetting pipeline and you are able to create that.
So those applications which are related to this when we are looking at AI portfolio not only languages but otherwise AI portfolio is very important for us to be on a continuous improvement journey. The most important aspect hence would be that if a person for example is working on a enterprise system of mails for example and it is actually deriving some summary of a document in perhaps a known language also or not a known language. The summary differs from what he thinks as a manual activity. He should be able to put that down somewhere and that goes as a feedback to the model. Currently that is something which is a concept which which may or may not exist, some enterprise would have done it, other enterprise would not have done it.
So looking at these kind of interventions which can be run as a program in a conscious way that everybody is able to contribute into the system his or her own things and then take it back from the, you know, improve the model or improve the AI systems, because they still require a lot of interventions from each and every person. The knowledge still is deficient. Thank
So what I’m taking away is that data sets need to be more of lived in nature. It’s not static. It has to be built upon by users and by others. And also just the fact that the feedback itself could lead to better data quality and which is something that enterprises might be doing, but it could definitely be done more. Thank you for that input. But to his point on the first question on data set inclusivity, Prashanta, like in going back to your research activity. mostly on inclusive data sets. The toolkit also argues that inclusivity must be designed at the foundational data layer at the time of designing data sets. But still we do find data sets which do lack this aspect.
What’s your take on what are the gaps over here at the research and academia level in terms of designing better inclusive data sets that could hence lead to better applications down the road?
That’s a very deep and good question. So to cover the diversity and become more inclusive, one approach would be to cover in the data, right? But if we think about the diversity that is there in Indian languages, right, that is a function of the culture, caste, local knowledge and everything, right? And while we see the diversity, they are not independent elements. There are certain commonalities and certain uniqueness in each of these languages and dialects and accents that we talk about. So one important direction in modeling would be to think about this intrinsic basis components that finally leads to this diversity. Instead of a brute force way of covering data from all parts of the country.
So if you can discover, for example, just an example, I’m not an expert of linguistics, but if you look at the Indian languages, there are two broad, right? One is Indo -Aryan and the other is Dabirian. Now, while there are multiple languages within each of the streams, we may say, well, can we go and then to cater certain technologies? Two speakers of these languages. should we go ahead and collect a good amount of data in everything, each of those. That may not be the only way to think about. How do we balance and make a trade -off between the amount of data we collect? We know that’s challenging and costly as well, to a novel modeling where we start from those intrinsic basis components and then manifest into those individual diversities.
I think that may help us to jointly think about modeling and collection for catering to this diverse population.
If you could help the audience with one example of when you say balance both aspects. Let’s say if we could pick up one of your initiatives, Syspin, Respin or Wani or any other data set. How did you manage or balance inclusivity versus model building activities versus maybe other factors that might be coming into factor while designing specifications?
Yeah, so the aspect of modeling that I brought out is something I would say not very well established at this moment. But from my experience in the project ResPin, I can give a concrete example. For example, if you take the Telugu as a language, right, there are, we worked with four major dialectal variations. One is in the region of Krishna Guntur, another is Vishakapatnam Vizag, another is Anandpur Chittoor, another is Nalgonda. Now, when you look at their intrinsic variations, we see that there are some commonalities. And then there are some unique aspects in each of those dialects. So now think about a brute force approach that I collect thousand hours in each of them. Versus think of collecting certain kind of stimuli to cover the actuality.
Acoustics case of the speakers, maybe from one region that will automatically cater to the other region. And then collect something that will complement. in each of the other regions, right? So that way, our overall timeline, budget, cost will all go down. And there has to be a novelty in terms of having a model that will start from the intrinsic one and then naturally diversify itself to cater to those populations. So that has to become a region -anchored approach that we started later on in one.
I see. Okay. Thanks for that input. Just to summarize, what I’m taking away is that instead of having brute force approach, what we’re essentially saying is balancing across various parameters on the basis of which you would train a model, such as linguistic diversity, acoustic diversity, and then using some sort of a smart approach to dissect the current audience, ways of collecting data, to maximize the output while maximizing bang for the buck. Thanks for that input. But this… This is also slightly… you are coming from the perspective of academia I would like to switch to Dr. Krithika from the perspective of as an applied AI researcher you are also one of the people in this panel who has really deployed speech AI solutions what is your take on challenges that you faced with inclusivity either at the data set layer or the application layer
Kritika K.R.:
More towards on core of the enterprise applications, knowledge repo integrations are coming up, beta healthcare, or even the manufacturing automobiles. So voice being the go -to interface for different applications and enabling the workforce across the industries is coming up. So in that case, again, as I said, on the consistency with the various user scenario and more specific to the domain adoption. Specialized domain adoption is required. That feedback loop is more important while the system is in the practice or while the system is in progress. I would say that point. And more critical aspect is on giving the scalable and sustainable infrastructure that comes with more optimized models and also like bringing the edge deployments also.
So that the real adoption can be scaled across multiple… industries and the normal usage for… various sectors across the industry. So I’m talking more on the end user perspective and using, getting the data. Data is one source of it, but making it reliable across the infrastructure and also giving the required scalable model at the device intelligence level is also important when it comes to the real adoption of these AI models.
Thanks for the input. So I guess after all, industry is also using feedback as a tool. It’s a nice validation over here. Yeah, maybe coming to Thomas, switching tracks to slightly legal sites. We’ve seen that, at least in the toolkit also, we’ve argued that speech models and speech data sets are at the intersection of copyright law, you know, data governance and security, etc. And how do you propose, how do you propose balancing sort of innovation? versus caution on these sites, especially with all the researchers and practitioners in the room?
Thanks, Nihal. That’s, again, a very helpful question. I think Harleen had articulated it quite well in the beginning when we have to consider the entire ecosystem as a whole. There is a common myth in India that anything that is public is freely available. I think what we have to think about is also that, you know, all data sets operate at the intersection of privacy law and copyright law. Under privacy law, most publicly available data sets are essentially freely available to be used under, you know, even the new legislation. But under copyright law, even if it is publicly available, somebody else may own the copyright on that. So there has to be careful thought put in place right from the beginning itself in terms of what data sets you’re collecting, what is the copyright provenance of it, are you able to defer to, you know, freely licensed and open source kind of material to compile it, compile that data set, and if not, are you able to obtain the licenses to do so?
So the thought process from the beginning in terms of how you’re structuring the way to get this and also how to reduce the surface area of the impact of some of these laws. So for instance, in relation to privacy laws, if you’re collecting somewhat more private data sets, if you can use privacy enhancing technologies or you’re able to extract data such that no personal data is ultimately captured or stored at the point of data collection, all of these are various ways in which you can put in place mechanisms right from the start of when the ecosystem begins to ensure that downstream use cases are also protected in that sense. The second big aspect is, of course, the documentation, right?
Now, the data collector, the data creator is essentially the person who is the gateway to the entire ecosystem in some senses. The documentation has to be robust right from the beginning to enable everybody in the downstream chain to be able to use this data and to ensure that there’s a good and safe and trusted ecosystem created. with respect to that specific data set. So yes, there are flexibilities that are available under the law in terms of how you are able to use voice data sets, but at the same time, there’s some caution that you have to put in place right from the beginning and throughout the life cycle of this in terms of figuring out how to be able to use these data sets effectively.
Of course, the last kind of related aspect to this is to think about the various layers in which these legalities operate. So of course, you can think of the speech data set itself as being copyrighted, but equally, if they are reading out of a book passage or if they’re reading specific performance and so on, there may be separate rights that are allocated in relation to some of these other tangential elements as well. All of these are to be accounted for from the very beginning of the ecosystem itself such that downstream usage is not… in that sense impacted. So I would say, you know, the report’s argument in that sense is that think about it as a whole.
Don’t think of each action in isolation. Think about the entire impact downstream as well. And then account for both either enabling maneuvers under law in terms of documentation, privacy enhancing techniques and so on, or implement the appropriate cautionary mechanisms to ensure that downstream usage is also protected.
Yeah, at least in some of the hats that I wear, I am also collecting data sets and those are important points that we keep in mind. And hopefully we’ll be able to take the learnings out of toolkit to actually implement in our processes. Switching tracks slightly to Dr. Prashanta here, we’ve without measurement, right, we don’t really get anywhere in terms of implementing the right frameworks, implementing the right legal processes, etc., in terms of implementing, measuring quality. what you’ve also spoken about evaluations being broken as far as Indian context are concerned can you elaborate a little bit on what challenges we face on a day to day basis where do they come across and how do you foresee this sort of challenges either getting resolved or getting amplified again I think this is an important area that all of us together should explore and contribute to
so when we build something like an automatic speech recognition system that is being used in many many applications think of this to be yet another human who is listening to the audio and trying to spit out what is spoken in text now if you go out in the real world as we have realized that multiple number of times and experienced through multiple projects in ResPin as well as Vani and many other projects that I have done and that we are able to do and that we are able to do and that we are able to do and that we are able to do is that if you give a piece of audio to two individuals, they never exactly agree on what they hear.
And I’m telling from my experience, not from two different parts of the country, I’m talking in terms of, you know, two people from the same district. In fact, there was an incident where we realized that these two people were just three kilometers away in terms of their location, but still they did not agree how that should be written from the audio they hear. So what it tells us is there is an inherent variation or variability in the way as an individual, as an Indian, I perceive or I like to see the text as, right? Now, if we accept that fact that exists today, we need to think of building our systems and system evaluation to cater to that variation.
So we need to think of that variability and to be… robust to that variability. So if, as I said in the beginning, if we treat the system also as a human, it will also not agree with another human. So if we just go by word -by -word comparison of how the system performs compared to some of the humans, certainly it will not be 100 % accurate. Or in other words, we calculate using what we call word error rate, which is objective way of evaluating. So a word -based comparison is not probably the right way to go at this point. Maybe the ASR system is doing pretty well, but just because it made a mistake slightly in one of the words, we are penalizing and telling that it’s not doing well.
So now we have to think about how do we solve this problem. It could be that we have a multiple evaluation system where we just don’t use word error rate. That’s one aspect. Another way to think about this will be to build ASR so that it itself can give not just one output, rather multiple outputs. which could be potentially right and then evaluate that not just objectively but also subjectively through human because human can absorb that error and say yes still it’s okay third will be to take that to the downstream application where depending on what you are using could be an LLM or any other QNA system that can absorb that robustness so I think we need to break down the entire evaluation system into multi -layered evaluations and then they are not really independent we need to take feedback all the way down to the final application back to ASR and so on so forth so I guess here individuals from the application areas, individuals from the linguistic background, engineers everyone has to come together and
so what I am hearing is that to solve this is more of an ecosystem level challenge right and then And maybe before our ecosystem champion over here, Mr. Nag, before you come in on this, I would just like one industry perspective of Dr. Krithika, how do you solve this from an application standpoint? Prashanta explained this challenge from more of an academic or foundational research standpoint. But how does evaluation play a role in your daily application layer?
Kritika K.R.:
Yeah, so as I said, right, so the applications are varied. So now the adoption is at the conversational level, right from bringing the analytics out of the data. Then now it is more on the voice interface and the multilingual conversation. Now with the speech -to -speech translation, those things are more prevalent with the conversation right now. Now coming to the industry application, industry aspect of it, yeah, adopting these models to the custom data set is one way. And also right pick of sourcing the data. From the available open source so that this model will be more specialized to those particular tasks. and the work they are supposed to do it. So now coming with the LLMs, these models are more adaptable to the industry jargons or even the core of the industry workflow.
Now making AI with the ASR models also enabling with the LLM, you have various methods from the data creation perspective, leveraging the open source data, and also like custom tuning the data to the various industry use cases. Definitely with the required compliance and these open source models are also enabling the on -prem deployment of these models, which enables the security aspect when it comes to creating the model for different core industry applications so that the models can be much more fine -tuned or trained across the domain, keeping the compliance aspect and the security aspects intact.
so having heard both of these perspectives Mr. Nag, how do you just from your experience standpoint, how do we approach resolving this conflict where all of us sort of concur that evaluations need we need a better framework to evaluation but it’s also in some ways nobody’s problem at the moment so is there a way to break this
so let’s step back and let’s evaluate our conversation itself you know, is there a framework by which we can say that who was saying, who has spoken better language right, it was as good as other people understand it you know, if the audience is able to understand what I am speaking and what I am intending to speak that is what is going to be the final evaluation by any aspect. What we have to actually look at it is that we have to reach a level by which it is acceptable to the people who are sitting in front of me. I don’t think we will be able to ever reach a situation where we will be able to say that this is the best, second best, third best.
It is a situation, ultimately the audience decide whether they are in a position to do that. We are looking at few of the use cases where we have actually deployed these technologies and we incidentally, you know, go to various evaluations. One of them is grievance incidentally and when we were giving it to the last, to the person who is actually the owner of the system, the acceptance was supposed to be taken up by various ministries. So one ministry would say that this model is better. The other ministry would perhaps display. It’s a question of perception and ultimately the audience would decide. And some would like the tone of speaking, some would like the modality, some would like the pronunciation.
So it’s all based on what the person’s perception is. Now, is there a common way in which we can say that this is the acceptable thing? Right? But then also we will have differences. You know, many of the public figures, for example, when they speak, you know, Hindi or English or whatever language, there are gaps in the language, but still, you know, they are understood. They are able to connect to the people. So we have a difficult challenge. Rather than looking at it only from a perspective of application or academics, we would have to look at it from a perspective of audience. But then we also have some issues. You know, we have situations where we have a lot of people who are not aware of the situation.
And we have to look at the situation from a perspective of the audience. which require accurate and perfect transcriptions. Like, for example, if I’m arguing a case in a court, you know, I can’t have variations in terms of languages. If I am, for example, trying to be in a meeting where I am saying something, again, I cannot have variations. But for that also, we will perhaps have to step two steps back and look at purity of language with respect to the acceptance. Because most of our language has become impure because of the fact that we are, you know, using mixed code most of the time, especially in the cosmopolitan area. And in the other areas, even if we are having native language, dialects are taking over.
So it’s a very complex problem. It’s not an easy problem to solve. At this point in time, when we are looking at how do we actually take it forward, I would tend to say that we should look at what is acceptable to the audience and then start working back to define an acceptable way by which in which the models can go out in the market.
Yeah, that’s an important point that so far we’ve been looking at mostly, at least I have been looking at mostly from the lens of application versus academia, but maybe we need to go what works point of view and not really from just the traditional ranking point of view. But Thomas, in a world where, and this is, we’ve not talked about this and this might be a curveball, but in a world where evaluation is slightly subjective and no longer objective, how does law see this? How do you make decisions for procurement? How do you resolve arguments, differences between two opinions, and especially in cases where both might be right and it’s a gray area? Like, do you foresee these sort of scenarios coming in, especially with Gen AI, which is like… Like, do you foresee these sort of scenarios coming in, especially with Gen AI, to be fair I think the legal principles at least on this are somewhat more clear at least in terms of some of the more privacy facing or copyright facing principles they occur much before outputs for instance are produced or any of these methodologies are implemented and we have a body of law that existed for many years in India it’s just a question of how do you lead evidence in relation to some of these matters so if it ever comes to the question of is a specific output right or is a specific output implying this or a specific output implying that I think where we haven’t caught up as a country is in terms of how to evaluate the evidentiary standard in relation to that the principles of course are fairly laid out saying that this is how you would decide it but what you would show the court to say this is the evidence for that that’s something I think that’s still evolving but I think it also brings me to I think a larger point and I think we’re making on the in the report as well is that you know there is a measure of trust that needs to be put in place in the ecosystem as a whole, right?
Irrespective of what the outcome of evaluation may be, there are measures that you can put in place right from the get -go. And one example I can give you is in relation to harmful content, right? Now, if there is a debate in relation to whether content is harmful or not, and it is a subjective determination, you can avoid that question to some degree by putting in place the necessary rails and safeguards right from the beginning itself so that trust is engineered into the process already as opposed to having to face that choice kind of downstream. But yes, to your point, and if we’re coming to a place where we need to face that question, I think the principles exist, but how you lead evidence, how you show the court that one is the interpretation over the other, still developing and very, very subjective.
I think some of the cases that, you know, the prominent AI players have in the country will go a long way to develop. Some of those standards, but at least as of now, the court system is still trying to catch up. to some of these principles. Documentation goes a long way to show intent. Methodologies that you have implemented that go to the extent of showing that you assumed reasonably high enough safeguards, reasonably high enough principles. All of these go a large extent to show intent. And so the subjectivity, I think, in that sense is far reduced if you put in place some of these measures that bring trust in the entire ecosystem. So I think at that one flashpoint of failure perhaps is tough to look at for the courts as well.
But if you look at it from an ecosystem perspective, I think there’s a lot of that that may reduce those flashpoints of failure or those flashpoints of evaluation at least from a legal perspective.
I see. Thanks for that summarization. That the law as such is at a stage where it can accommodate some amount of subjectivity but there needs to be dialogue and more policy decisions to make it crisper and of course follow on into the application of the law. Thanks for that input. last question is leaving the floor open in terms of any inputs we do have the topic at hand is challenges and best practices for speech models and data sets at the ecosystem level or from your experiences any open points any arguments that you would like to make or any sort of a call out that you would like to make to the ecosystem right here it means the call out is means like you know many of the things which were indeterministic or unknown a few days back have started coming into a situation where we are able to crystallize it so I think we need to get into more workshops more discussions to think about it as how to do it take more use cases study more use cases in detail to figure out a framework by which acceptability and evaluations are properly benchmarked.
That’s a good point. Go ahead, Thomas. I have a point to add here, which is, you know, I think there is a certain sense of affinity in this ecosystem towards open source data sets or open models. I would be more thoughtful in terms of how and when these are suitable. Are there particular safeguards you need to put in place for open source data sets is something you need to think about. Are there end -use considerations that need to be tailored? And a good example is, you know, we have, I’ve seen an example where somebody is training a model to detect hate speech, right? Now the safeguards you would put in place to detect hate speech in a model is different from a data set and a model that you would develop to detect regular speech -to -speech translation.
So the decision as to what licensing frameworks, what documentation frameworks are fairly, need to be informed by what end -use case you’re doing, what unique… Thank you. attributes arise as a result of the specific data sets and applications that you are considering and finally on the basis of what downstream users you are expecting the choice needs to be made I think in a little bit more of a conscious fashion Nishant you wanted to say sure this your question actually stimulates me to think about you know English I mean you know sort of models that were built on American English so there have been always a standardization on evaluation in fact NIST evaluation if you look at there have been various protocols and there have been call out every year who beats the best baseline so far achieved I believe we have to do in our country in India at least for Indian languages and it’s very diverse as we just discussed so first of all thinking about how to evaluate and then creating a national level framework for evaluation.
And every year, let’s assess ourselves, all these stakeholders, right?
It could be general evaluation, could be application specific in each language or dialect. And then we really have a leaderboard, which, of course, you know, there are many individual leaderboards across the country, but let’s have only one under Varshini, let’s say, right? And that should be elaborate enough to cater to all languages and dialects. And maybe that’s not the right way, but you think through and make sure every year we make progress in each of those three. I think that has to be brought in in the system to bring competitiveness in a collaborative way, of course. And overall, that can help improve the voice technology in Indian languages. And the reason I’m saying it
mostly from my understanding and experience with the English that has happened, in the past. Yeah. interesting points, Prasanta, in terms of I hear you sort of speak passionately about evaluation and now you’re taking it one step further in terms of how do we really create a unified framework for evaluation within competitive but yet collaborative manner for the ecosystem housed under a central, unpartial entity like Bhashani. This is a great point. I hope the audience found some of these points helpful and enriching. Thank you so much for making time in what is sure to be a very busy event and hope you have a rest of a good day. Thank you. Invite Mr. Shailendra Pal Singh, Senior General Manager, Bhashani to felicitate the speakers.
Thank you. Mr. Amitabh Nag Dr. Prasanta Ghosh Dr. Krithika K.I. Mr. Thomas Salenat I’m Ms. Harleen Kaur Thank you to all our speakers for walking us through this rich tapestry of voice technologies and their life cycle in the Indian context and we hope you read our report and the toolkit and find it useful. Thank you so much. Thank you so much to the audience for staying with us patiently throughout this entire hour. Thank you. Thank you.
Ariane Ahildur
Speech speed
126 words per minute
Speech length
562 words
Speech time
266 seconds
Voice AI as gateway to public services and inclusion
Explanation
Ariane explains that when voice AI supports local languages and dialects it can unlock access to essential public services such as health, education and economic participation, especially for people with limited literacy or device access. This positions voice as a natural, powerful interface for inclusive digital transformation.
Evidence
“When voice AI works in local languages and dialects, it will become a gateway to public services, healthcare, education, and economic participation.” [1]. “Voice is the most natural and powerful interface to the digital world, especially for those with limited literacy or access to digital devices.” [6].
Major discussion point
Inclusive Voice AI Vision & Policy
Topics
Closing all digital divides | Social and economic development | Artificial intelligence
Harleen Kaur
Speech speed
143 words per minute
Speech length
1036 words
Speech time
432 seconds
Four‑pillar policy framework and developer toolkit
Explanation
Harleen outlines that the policy report is organized around four pillars and that a companion developer toolkit translates these principles into actionable guidance for developers, fostering an open and responsible voice‑AI ecosystem.
Evidence
“Based on this approach, we have structured our policy framework around the four pillars that you see on the screen.” [19]. “So alongside the policy framework, we’ve developed a developer toolkit that translates some of these principles into practice for developers.” [17].
Major discussion point
Inclusive Voice AI Vision & Policy
Topics
The enabling environment for digital development | Artificial intelligence | Data governance
Representation planning, layered data strategies and synthetic data for inclusive datasets
Explanation
Harleen proposes concrete measures such as diversity wish‑lists, linguistic expertise, synthetic data generation and a layered data strategy to ensure that voice datasets capture the full spectrum of languages and dialects, making models more inclusive.
Evidence
“Things like having a diversity wish list, making sure that you’re not collecting data from one source, applying linguistic expertise, using synthetic data, training model for linguistic and environmental nuances, and also layered data strategy.” [71]. “So it focuses on three broad areas, representation being the foremost through diversity planning, et cetera.” [82].
Major discussion point
Designing Inclusive Datasets for Linguistic Diversity
Topics
Data governance | Closing all digital divides | Artificial intelligence
Moderator
Speech speed
68 words per minute
Speech length
267 words
Speech time
232 seconds
Formal launch emphasizing Indo‑German partnership and report rollout
Explanation
The moderator signals the transition to the official launch of the policy report and developer toolkit, highlighting the collaborative Indo‑German effort that underpins the initiative.
Evidence
“We shall now move on to the formal launch of the report and toolkit.” [31]. “I will now invite Dr. Ariane Ahildur -Brandt, Director General … to deliver the keynote address.” [32].
Major discussion point
Inclusive Voice AI Vision & Policy
Topics
The enabling environment for digital development | Information and communication technologies for development
Amitabh Nag
Speech speed
162 words per minute
Speech length
1513 words
Speech time
558 seconds
Continuous creation of primary and improvement corpora
Explanation
Amitabh describes a two‑step data pipeline: first building a primary corpus, then generating an improvement corpus that is fed back into models for ongoing upgrades, ensuring the system learns continuously from product feedback.
Evidence
“So one, you are creating a primary corpus.” [37]. “Second is… you are creating an improvement corpus which can be again fed back to the model and say that this is what is to be used and that is a big area of work as we look at.” [38]. “So we have to continuously upgrade the things as we go by.” [41].
Major discussion point
Sustainable Data Ecosystem & Digital Public Goods
Topics
Data governance | Artificial intelligence
Enterprise feedback loops to refine models
Explanation
Amitabh stresses that feedback loops from real‑world deployments are crucial; users and enterprises must continuously build on model outputs to improve reliability and relevance.
Evidence
“That feedback loop is more important while the system is in the practice or while the system is in progress.” [44]. “It has to be built upon by users and by others.” [45].
Major discussion point
Deployment, Scalability, and Industry Adoption
Topics
Deployment, Scalability, and Industry Adoption | Artificial intelligence
Nihar Desai
Speech speed
131 words per minute
Speech length
1767 words
Speech time
804 seconds
Treat foundational speech datasets as public goods and create a data‑flywheel
Explanation
Nihar argues that foundational speech datasets should be regarded as public goods (DPIs/DPGs) and that a self‑reinforcing data‑flywheel can sustain ongoing innovation and accessibility.
Evidence
“The first is treating foundational data sets as public good.” [50]. “As we saw in the toolkit, we were arguing that data set like foundational data sets, speech data sets, must be treated as DPIs and DPGs and hence be available in general.” [51]. “And is there a way for us to have a flywheel of data sorts, data goods of sorts?” [53].
Major discussion point
Sustainable Data Ecosystem & Digital Public Goods
Topics
Data governance | Information and communication technologies for development
Trust‑by‑design, documentation and evidence standards for subjective evaluation
Explanation
Nihar highlights that while legal principles exist, evidentiary standards for subjective decisions are still evolving; embedding safeguards, robust documentation and trust mechanisms from the outset can mitigate subjectivity in evaluation and downstream use.
Evidence
“But yes, to your point, and if we’re coming to a place where we need to face that question, I think the principles exist, but how you lead evidence, how you show the court that one is the interpretation over the other, still developing and very, very subjective.” [107]. “Now, if there is a debate in relation to whether content is harmful or not, and it is a subjective determination, you can avoid that question to some degree by putting in place the necessary rails and safeguards right from the beginning itself so that trust is engineered into the process already as opposed to having to face that choice kind of downstream.” [109]. “And so the subjectivity, I think, in that sense is far reduced if you put in place some of these measures that bring trust in the entire ecosystem.” [111].
Major discussion point
Evaluation, Trust, and Legal/Ethical Frameworks
Topics
Human rights and the ethical dimensions of the information society | Data governance | Artificial intelligence
Prasanta Ghosh
Speech speed
160 words per minute
Speech length
1184 words
Speech time
443 seconds
Use intrinsic language‑family structures to balance data collection across dialects
Explanation
Prasanta stresses the need to balance data collection by considering language‑family structures and ensuring coverage across all languages and dialects, thereby avoiding over‑representation of any single source.
Evidence
“How do we balance and make a trade -off between the amount of data we collect?” [69]. “And that should be elaborate enough to cater to all languages and dialects.” [72]. “So to cover the diversity and become more inclusive, one approach would be …” [77].
Major discussion point
Designing Inclusive Datasets for Linguistic Diversity
Topics
Closing all digital divides | Data governance
Multi‑layered evaluation acknowledging human transcription variability
Explanation
Prasanta points out that human transcription variability requires evaluation frameworks that combine objective metrics with subjective, human‑centric assessments, using multi‑layered feedback throughout the pipeline.
Evidence
“which could be potentially right and then evaluate that not just objectively but also subjectively through human because human can absorb that error and say yes still it’s okay … we need to break down the entire evaluation system into multi -layered evaluations …” [101]. “So we need to think of that variability and to be… robust to that variability.” [102]. “Now, if we accept that fact that exists today, we need to think of building our systems and system evaluation to cater to that variation.” [104].
Major discussion point
Evaluation, Trust, and Legal/Ethical Frameworks
Topics
Artificial intelligence | Evaluation, Trust, and Legal/Ethical Frameworks
Kritika K.R.
Speech speed
138 words per minute
Speech length
432 words
Speech time
186 seconds
Scalable, edge‑ready infrastructure and domain‑specific model tuning for enterprise
Explanation
Kritika emphasizes the need for scalable, sustainable infrastructure that supports optimized models and edge deployments, enabling on‑premise, compliant and secure implementations tailored to specific industry domains.
Evidence
“And more critical aspect is on giving the scalable and sustainable infrastructure that comes with more optimized models and also like bringing the edge deployments also.” [86]. “Definitely with the required compliance and these open source models are also enabling the on -prem deployment of these models, which enables the security aspect when it comes to creating the model for different core industry applications so that the models can be much more fine -tuned or trained across the domain, keeping the compliance aspect and the security aspects intact.” [88].
Major discussion point
Deployment, Scalability, and Industry Adoption
Topics
Deployment, Scalability, and Industry Adoption | Artificial intelligence | Building confidence and security in the use of ICTs
Thomas J. Vallianeth
Speech speed
171 words per minute
Speech length
1138 words
Speech time
397 seconds
Copyright, privacy and robust documentation safeguards for open‑source voice data
Explanation
Thomas stresses that open‑source voice datasets must be accompanied by robust documentation, careful copyright provenance checks, and privacy‑enhancing techniques from the outset to create a safe, trusted ecosystem.
Evidence
“The documentation has to be robust right from the beginning to enable everybody in the downstream chain to be able to use this data and to ensure that there’s a good and safe and trusted ecosystem created.” [65]. “So there has to be careful thought put in place right from the beginning itself in terms of what data sets you’re collecting, what is the copyright provenance of it, are you able to defer to, you know, freely licensed and open source kind of material to compile it, compile that data set, and if not, are you able to obtain the licenses to do so?” [67]. “All data sets operate at the intersection of privacy law and copyright law.” [68].
Major discussion point
Evaluation, Trust, and Legal/Ethical Frameworks
Topics
Human rights and the ethical dimensions of the information society | Data governance | Artificial intelligence
Agreements
Agreement points
Continuous data creation and improvement through feedback loops
Speakers
– Amitabh Nag
– Kritika K.R.
– Nihar Desai
Arguments
Foundational datasets should be treated as digital public goods with ongoing creation through multiple approaches including field collection and product feedback
Industry applications require domain-specific adaptation and feedback loops for real-world deployment success
Creating a flywheel of data goods requires systematic approaches to continuous dataset creation while ensuring trust and safety
Summary
All speakers agree that data creation must be an ongoing process with continuous feedback loops from users and applications to improve models over time, rather than static one-time collection efforts.
Topics
Data governance | Artificial intelligence | Information and communication technologies for development
Evaluation frameworks need fundamental reform beyond traditional metrics
Speakers
– Amitabh Nag
– Prasanta Ghosh
– Nihar Desai
Arguments
Evaluation should focus on audience acceptability rather than perfect accuracy, as language use varies by context and application
Multi-layered evaluation systems beyond word error rates are needed, including subjective human assessment
Evaluation frameworks need to balance multiple parameters including linguistic and acoustic diversity to maximize effectiveness
Summary
There is strong consensus that current evaluation methods are inadequate and need to move beyond traditional objective metrics to include subjective assessment and audience acceptability.
Topics
Monitoring and measurement | Artificial intelligence
Inclusivity and diversity must be designed from the foundation
Speakers
– Amitabh Nag
– Harleen Kaur
– Nihar Desai
Arguments
AI systems must embrace diversity and inclusion as core design principles, unlike traditional digital systems that relied on standards
Diversity planning and layered data strategies are essential for adequate representation in voice models
Inclusivity must be designed at the foundational data layer rather than addressed later in the development process
Summary
All speakers emphasize that inclusivity cannot be an afterthought but must be built into the very foundation of system design and data collection from the beginning.
Topics
Closing all digital divides | Artificial intelligence | Human rights and the ethical dimensions of the information society
Government should play an active stewardship role beyond regulation
Speakers
– Harleen Kaur
– Ariane Ahildur
Arguments
Government should act as ecosystem steward and convener rather than just regulator, supporting commercially non-viable languages
Indo-German partnership demonstrates cooperation over competition in AI development, creating open voice technologies for nine Indian languages
Summary
Both speakers advocate for government taking an active role in stewarding the voice technology ecosystem, supporting underserved languages and fostering international cooperation.
Topics
The enabling environment for digital development | Information and communication technologies for development | Closing all digital divides
Trust and safety must be engineered into systems from the beginning
Speakers
– Thomas J. Vallianeth
– Harleen Kaur
– Nihar Desai
Arguments
Trust should be engineered into processes from the beginning through proper safeguards and documentation
Responsible AI practices must be embedded throughout the development lifecycle, not just at the policy level
Creating a flywheel of data goods requires systematic approaches to continuous dataset creation while ensuring trust and safety
Summary
All speakers agree that trust and safety considerations cannot be addressed as an afterthought but must be built into the development process from the very beginning.
Topics
Human rights and the ethical dimensions of the information society | Building confidence and security in the use of ICTs | Artificial intelligence
Similar viewpoints
Both speakers advocate for intelligent, adaptive approaches to model development that account for linguistic diversity without requiring exhaustive data collection from every possible variation.
Speakers
– Prasanta Ghosh
– Amitabh Nag
Arguments
Smart modeling approaches should identify intrinsic basis components rather than using brute force data collection across all dialects
Solutions must be continuously upgraded with short shelf lives of 3-6 months due to diversity requirements
Topics
Artificial intelligence | Closing all digital divides | Capacity development
Both speakers emphasize that challenges in voice technology development are systemic and require comprehensive approaches that consider the entire lifecycle from data collection to deployment.
Speakers
– Harleen Kaur
– Thomas J. Vallianeth
Arguments
Data collection challenges span the entire lifecycle from curation to deployment, requiring multi-layered solutions
Legal frameworks must balance innovation with caution, considering copyright, privacy, and documentation requirements from the beginning
Topics
Data governance | Human rights and the ethical dimensions of the information society | The enabling environment for digital development
Both speakers from industry/implementation backgrounds emphasize the importance of practical deployment considerations and continuous improvement through real-world usage feedback.
Speakers
– Kritika K.R.
– Amitabh Nag
Arguments
Industry applications require domain-specific adaptation and feedback loops for real-world deployment success
Foundational datasets should be treated as digital public goods with ongoing creation through multiple approaches including field collection and product feedback
Topics
Social and economic development | Artificial intelligence | The digital economy
Unexpected consensus
Subjectivity in evaluation is acceptable and necessary
Speakers
– Amitabh Nag
– Prasanta Ghosh
– Thomas J. Vallianeth
Arguments
Evaluation should focus on audience acceptability rather than perfect accuracy, as language use varies by context and application
Current evaluation methods are inadequate as human annotators from the same region often disagree on transcriptions
Legal frameworks must address the intersection of speech models with copyright, data governance, and security while balancing innovation with caution
Explanation
It’s unexpected that speakers from technical, academic, and legal backgrounds all agree that subjective evaluation approaches are not only acceptable but necessary, moving away from traditional objective metrics that have dominated the field.
Topics
Monitoring and measurement | Artificial intelligence | Human rights and the ethical dimensions of the information society
National-level coordination and standardization is needed despite diversity emphasis
Speakers
– Prasanta Ghosh
– Harleen Kaur
– Amitabh Nag
Arguments
A national-level evaluation framework under Bhashini could create competitive but collaborative progress tracking
Government should act as ecosystem steward and convener rather than just regulator, supporting commercially non-viable languages
Solutions must be continuously upgraded with short shelf lives of 3-6 months due to diversity requirements
Explanation
Despite strong emphasis on diversity and inclusion throughout the discussion, there’s unexpected consensus on the need for centralized coordination and standardization through national frameworks, showing recognition that diversity and standardization can coexist.
Topics
The enabling environment for digital development | Monitoring and measurement | Artificial intelligence
Overall assessment
Summary
The speakers demonstrate remarkable consensus across multiple critical areas: the need for continuous data improvement, fundamental reform of evaluation frameworks, foundational design for inclusivity, active government stewardship, and built-in trust and safety measures. There’s also agreement on balancing diversity with practical implementation needs.
Consensus level
High level of consensus with significant implications for voice technology development in India. The agreement spans technical, policy, legal, and implementation perspectives, suggesting a mature understanding of the challenges and a shared vision for solutions. This consensus provides a strong foundation for coordinated action across the ecosystem, though implementation will require sustained collaboration among all stakeholders.
Differences
Different viewpoints
Approach to evaluation frameworks – objective vs subjective measures
Speakers
– Prasanta Ghosh
– Amitabh Nag
Arguments
Multi-layered evaluation systems beyond word error rates are needed, including subjective human assessment
Evaluation should focus on audience acceptability rather than perfect accuracy, as language use varies by context and application
Summary
Ghosh advocates for systematic multi-layered evaluation frameworks with both objective and subjective measures, while Nag emphasizes that evaluation should primarily focus on audience acceptability and context-dependent needs rather than standardized metrics
Topics
Monitoring and measurement | Artificial intelligence
Data collection strategy – smart modeling vs comprehensive collection
Speakers
– Prasanta Ghosh
– Amitabh Nag
Arguments
Smart modeling approaches should identify intrinsic basis components rather than using brute force data collection across all dialects
Foundational datasets should be treated as digital public goods with ongoing creation through multiple approaches including field collection and product feedback
Summary
Ghosh advocates for intelligent modeling that identifies common underlying components to reduce data collection needs, while Nag emphasizes comprehensive ongoing data collection through multiple channels including traditional field collection
Topics
Data governance | Artificial intelligence | Capacity development
Open source adoption approach – cautious vs embracing
Speakers
– Thomas J. Vallianeth
– Harleen Kaur
Arguments
Open source approaches need thoughtful consideration of safeguards and end-use cases rather than blanket adoption
Government should act as ecosystem steward and convener rather than just regulator, supporting commercially non-viable languages
Summary
Vallianeth advocates for careful, case-by-case consideration of open source approaches with appropriate safeguards, while Kaur promotes broader government support for open ecosystems and public goods
Topics
The enabling environment for digital development | Human rights and the ethical dimensions of the information society | Data governance
Unexpected differences
Role of government in ecosystem development
Speakers
– Harleen Kaur
– Thomas J. Vallianeth
Arguments
Government should act as ecosystem steward and convener rather than just regulator, supporting commercially non-viable languages
Legal frameworks must balance innovation with caution, considering copyright, privacy, and documentation requirements from the beginning
Explanation
While both speakers support responsible development, Kaur advocates for active government stewardship and support for public goods, while Vallianeth emphasizes the need for careful legal frameworks and caution. This represents an unexpected tension between proactive government intervention versus regulatory prudence
Topics
The enabling environment for digital development | Data governance | Human rights and the ethical dimensions of the information society
Standardization vs diversity accommodation
Speakers
– Prasanta Ghosh
– Amitabh Nag
Arguments
A national-level evaluation framework under Bhashini could create competitive but collaborative progress tracking
AI systems must embrace diversity and inclusion as core design principles, unlike traditional digital systems that relied on standards
Explanation
Ghosh advocates for standardized national evaluation frameworks similar to English language models, while Nag emphasizes that AI systems must move away from standardization to embrace diversity. This creates tension between the need for comparable metrics and the principle of diversity-first design
Topics
Monitoring and measurement | Artificial intelligence | Closing all digital divides
Overall assessment
Summary
The main areas of disagreement center around evaluation methodologies, data collection strategies, the role of government, and the balance between standardization and diversity accommodation
Disagreement level
Moderate disagreement level with significant implications – while speakers share common goals of inclusive voice technology development, their different approaches to evaluation, data collection, and governance could lead to fragmented implementation strategies. The disagreements suggest need for more coordinated discussion on unified frameworks that can accommodate different perspectives while maintaining coherent ecosystem development
Partial agreements
Partial agreements
All speakers agree that better evaluation frameworks are needed for voice technology, but disagree on implementation – Ghosh wants standardized national frameworks, Nag prefers audience-focused acceptability measures, and Kaur emphasizes multi-layered approaches throughout the development lifecycle
Speakers
– Prasanta Ghosh
– Amitabh Nag
– Harleen Kaur
Arguments
A national-level evaluation framework under Bhashini could create competitive but collaborative progress tracking
Solutions must be continuously upgraded with short shelf lives of 3-6 months due to diversity requirements
Diversity planning and layered data strategies are essential for adequate representation in voice models
Topics
Monitoring and measurement | Artificial intelligence | The enabling environment for digital development
All speakers agree on the importance of responsible AI development, but differ on approach – Vallianeth emphasizes legal safeguards and documentation, Kaur focuses on lifecycle integration of RAI practices, and Nag emphasizes design-level inclusion principles
Speakers
– Thomas J. Vallianeth
– Harleen Kaur
– Amitabh Nag
Arguments
Trust should be engineered into processes from the beginning through proper safeguards and documentation
Responsible AI practices must be embedded throughout the development lifecycle, not just at the policy level
AI systems must embrace diversity and inclusion as core design principles, unlike traditional digital systems that relied on standards
Topics
Human rights and the ethical dimensions of the information society | Artificial intelligence | The enabling environment for digital development
Similar viewpoints
Both speakers advocate for intelligent, adaptive approaches to model development that account for linguistic diversity without requiring exhaustive data collection from every possible variation.
Speakers
– Prasanta Ghosh
– Amitabh Nag
Arguments
Smart modeling approaches should identify intrinsic basis components rather than using brute force data collection across all dialects
Solutions must be continuously upgraded with short shelf lives of 3-6 months due to diversity requirements
Topics
Artificial intelligence | Closing all digital divides | Capacity development
Both speakers emphasize that challenges in voice technology development are systemic and require comprehensive approaches that consider the entire lifecycle from data collection to deployment.
Speakers
– Harleen Kaur
– Thomas J. Vallianeth
Arguments
Data collection challenges span the entire lifecycle from curation to deployment, requiring multi-layered solutions
Legal frameworks must balance innovation with caution, considering copyright, privacy, and documentation requirements from the beginning
Topics
Data governance | Human rights and the ethical dimensions of the information society | The enabling environment for digital development
Both speakers from industry/implementation backgrounds emphasize the importance of practical deployment considerations and continuous improvement through real-world usage feedback.
Speakers
– Kritika K.R.
– Amitabh Nag
Arguments
Industry applications require domain-specific adaptation and feedback loops for real-world deployment success
Foundational datasets should be treated as digital public goods with ongoing creation through multiple approaches including field collection and product feedback
Topics
Social and economic development | Artificial intelligence | The digital economy
Takeaways
Key takeaways
Voice technology ecosystems require continuous evolution with short 3-6 month upgrade cycles due to linguistic and cultural diversity requirements
Foundational speech datasets should be treated as digital public goods (DPGs) with government acting as ecosystem steward rather than just regulator
Current evaluation frameworks for Indian language voice systems are inadequate – human annotators from the same region often disagree on transcriptions, requiring multi-layered evaluation beyond word error rates
Inclusive AI design must embed diversity and representation at the foundational data layer, using smart modeling approaches that identify intrinsic basis components rather than brute force data collection
Legal frameworks must balance innovation with caution by considering copyright, privacy, and documentation requirements from the project inception
Industry adoption requires domain-specific adaptation, feedback loops, and scalable edge deployment infrastructure for real-world success
The Indo-German partnership demonstrates that international cooperation can be more effective than competition in developing open voice technologies for underserved languages
Resolutions and action items
Create a national-level evaluation framework under Bhashini with annual competitive but collaborative progress tracking across Indian languages and dialects
Develop continuous data creation programs through multiple approaches: field collection, product feedback loops, and digital data harvesting
Implement multi-layered evaluation systems that include subjective human assessment alongside objective metrics
Establish workshops and discussions to crystallize frameworks for acceptability and evaluation benchmarks
Focus evaluation on audience acceptability rather than perfect accuracy, recognizing contextual language variations
Engineer trust into AI processes from the beginning through proper safeguards, documentation, and responsible AI practices embedded throughout the development lifecycle
Unresolved issues
How to standardize evaluation when human perception of speech varies significantly even within the same geographic region
Balancing the trade-off between comprehensive diversity coverage and practical resource constraints in data collection
Determining optimal licensing frameworks and safeguards for different types of open source voice datasets based on end-use cases
Establishing evidentiary standards for AI-related legal disputes in Indian courts
Creating sustainable funding models for maintaining and updating voice technology infrastructure as public goods
Defining acceptable accuracy thresholds for different application contexts (legal transcription vs. general conversation)
Scaling personalized domain adaptation across diverse industries while maintaining security and compliance requirements
Suggested compromises
Use hybrid approaches combining brute force data collection with smart modeling based on intrinsic linguistic components to balance coverage and efficiency
Implement layered data strategies using multiple sources (active collection, passive collection, synthetic data) rather than relying on single approaches
Focus on audience acceptability as the primary evaluation criterion while maintaining objective metrics for specific high-accuracy use cases like legal or medical applications
Adopt ‘compliance plus’ frameworks that exceed minimum legal requirements while remaining practically implementable
Create collaborative competitive environments where organizations share evaluation benchmarks while competing on implementation quality
Balance open source principles with thoughtful safeguards tailored to specific use cases and risk profiles
Thought provoking comments
AI, and the reason for this is diversity. You know, each person is different. Each language is different. Each culture is different. So there is… There is huge amount of diversity and we have to live with the diversity unlike the earlier digital systems which used to work on only standards. You know, they had standards and they would perhaps keep the outliers away. Here, inclusion is the name of the, inclusion is part of the design, diversity is part of the design.
Speaker
Amitabh Nag
Reason
This comment fundamentally reframes how we think about AI system design – moving from standardization that excludes outliers to inclusion-by-design that embraces diversity as a core principle. It challenges the traditional engineering approach of creating uniform standards.
Impact
This set the philosophical foundation for the entire discussion, establishing that voice AI for India requires a paradigm shift from conventional approaches. It influenced subsequent speakers to focus on practical implementations of inclusive design rather than debating whether inclusion is necessary.
Instead of a brute force approach that I collect thousand hours in each of them. Versus think of collecting certain kind of stimuli to cover the actuality. Acoustics case of the speakers, maybe from one region that will automatically cater to the other region. And then collect something that will complement… So that way, our overall timeline, budget, cost will all go down.
Speaker
Prasanta Ghosh
Reason
This introduces a sophisticated approach to data collection that balances inclusivity with practical constraints. Rather than simply collecting more data, it proposes understanding the underlying linguistic structures to optimize coverage efficiently.
Impact
This shifted the conversation from theoretical discussions about inclusivity to practical methodologies. It influenced other panelists to think about smart approaches rather than resource-intensive solutions, and led to discussions about feedback loops and continuous improvement.
If you give a piece of audio to two individuals, they never exactly agree on what they hear. And I’m telling from my experience, not from two different parts of the country, I’m talking in terms of, you know, two people from the same district. In fact, there was an incident where these two people were just three kilometers away in terms of their location, but still they did not agree how that should be written from the audio they hear.
Speaker
Prasanta Ghosh
Reason
This observation fundamentally challenges the assumption that there can be objective ‘ground truth’ in speech recognition evaluation. It reveals the inherent subjectivity in human perception of speech, which has profound implications for how AI systems should be evaluated.
Impact
This comment created a pivotal moment in the discussion, forcing all participants to reconsider evaluation frameworks. It led to discussions about multi-layered evaluation systems, subjective vs. objective metrics, and ultimately influenced the conversation toward audience-centric approaches rather than purely technical benchmarks.
Rather than looking at it only from a perspective of application or academics, we would have to look at it from a perspective of audience. But then also we have some issues… It’s a question of perception and ultimately the audience would decide whether they are in a position to do that.
Speaker
Amitabh Nag
Reason
This shifts the entire evaluation paradigm from technical metrics to user acceptance, acknowledging that the ultimate measure of success is whether the technology serves its intended users effectively, regardless of technical perfection.
Impact
This comment synthesized the evaluation challenge discussion and redirected it toward practical deployment considerations. It influenced the legal perspective discussion that followed, where Thomas discussed trust-based approaches rather than purely objective legal standards.
There is a common myth in India that anything that is public is freely available. I think what we have to think about is also that, you know, all data sets operate at the intersection of privacy law and copyright law.
Speaker
Thomas J. Vallianeth
Reason
This comment addresses a critical misconception that could undermine the entire open voice technology ecosystem. It highlights the complex legal landscape that developers must navigate, challenging assumptions about data availability.
Impact
This introduced a sobering reality check to the discussion, shifting focus from technical and design challenges to legal and compliance considerations. It led to discussions about documentation, provenance tracking, and the need for legal safeguards to be built into the system from the beginning.
At a time when artificial intelligence is often framed as a global competition, this report offers a different narrative, and this is a narrative of cooperation. The Indo-German Partnership on AI, and particularly on language, and voice technologies shows what is possible when we join forces.
Speaker
Ariane Ahildur
Reason
This reframes AI development from a zero-sum competitive game to a collaborative endeavor focused on public good. It challenges the dominant narrative of AI as a race between nations and instead positions it as a tool for inclusive development.
Impact
This comment established the cooperative framework that influenced the entire discussion’s tone. It set the stage for speakers to focus on shared challenges and collaborative solutions rather than competitive advantages, influencing the emphasis on open-source approaches and shared standards throughout the panel.
Overall assessment
These key comments fundamentally shaped the discussion by establishing three critical paradigm shifts: (1) from standardization to inclusion-by-design in AI development, (2) from objective technical metrics to user-centric evaluation frameworks, and (3) from competitive to collaborative approaches in AI development. The conversation evolved from theoretical principles to practical implementation challenges, with each insightful comment building upon previous ones to create a comprehensive framework for responsible voice AI development. The discussion successfully bridged academic research, industry application, policy considerations, and legal frameworks, largely due to these thought-provoking interventions that challenged conventional assumptions and introduced new perspectives. The overall impact was a nuanced understanding that voice AI for diverse populations like India requires fundamentally different approaches than traditional AI development paradigms.
Follow-up questions
How can we create a continuous feedback loop system where enterprise applications automatically contribute improvements back to foundational AI models?
Speaker
Amitabh Nag
Explanation
This addresses the need for systematic improvement of AI systems through user feedback, which is currently inconsistent across enterprises and could significantly enhance model performance if implemented as a conscious program.
How can we develop novel modeling approaches that start from intrinsic basis components of languages rather than using brute force data collection methods?
Speaker
Prasanta Ghosh
Explanation
This could revolutionize how we approach linguistic diversity by identifying commonalities between language families (like Indo-Aryan and Dravidian) to create more efficient and cost-effective models while maintaining inclusivity.
How do we establish a national-level evaluation framework for Indian languages with annual assessments and leaderboards?
Speaker
Prasanta Ghosh
Explanation
Current evaluation methods are inadequate for Indian language diversity, and a unified framework under Bhashini could drive competitive yet collaborative improvement across the ecosystem.
How can we develop multi-layered evaluation systems that move beyond word error rates to accommodate inherent variability in human perception of speech?
Speaker
Prasanta Ghosh
Explanation
Traditional objective evaluation methods fail to capture the reality that even humans from the same region disagree on speech transcription, requiring new evaluation paradigms.
What specific safeguards and licensing frameworks should be implemented for different types of open source datasets based on their intended use cases?
Speaker
Thomas J. Vallianeth
Explanation
Different applications (like hate speech detection vs. general speech translation) require different safeguards, and the choice of open source frameworks should be more consciously tailored to specific end-use cases.
How can we establish evidentiary standards for AI-related legal cases, particularly for subjective determinations in court proceedings?
Speaker
Thomas J. Vallianeth
Explanation
While legal principles exist, the court system lacks established methods for evaluating AI-generated evidence and outputs, which will become increasingly important as AI adoption grows.
How do we define and measure ‘acceptability’ from an audience perspective rather than purely technical metrics?
Speaker
Amitabh Nag
Explanation
Current evaluation focuses on technical accuracy, but real-world success depends on user acceptance, which varies based on context, audience, and application requirements.
What methodologies can be developed to balance data collection costs with linguistic and acoustic diversity coverage?
Speaker
Prasanta Ghosh
Explanation
There’s a need for smart approaches that maximize model performance across diverse populations while minimizing the time, budget, and resources required for comprehensive data collection.
How can we create standardized documentation and data stewardship models that ensure proper copyright and privacy compliance throughout the AI development lifecycle?
Speaker
Thomas J. Vallianeth
Explanation
Robust documentation from data collection through deployment is crucial for legal compliance and building trusted ecosystems, but current practices are inconsistent.
What frameworks are needed to enable scalable edge deployment of optimized voice AI models across different industries?
Speaker
Kritika K.R.
Explanation
Real adoption requires models that can work reliably at the device level across various sectors, but current infrastructure and optimization approaches may not be sufficient for widespread deployment.
Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.
Related event

