Founders Adda Raw Conversations with India’s Top AI Pioneers

20 Feb 2026 13:00h - 14:00h

Founders Adda Raw Conversations with India’s Top AI Pioneers

Session at a glanceSummary, keypoints, and speakers overview

Summary

The session, organized by Archana Jahargirdar, was framed as a product-only showcase where founders present technical details without discussing business or funding, and speakers were asked to balance jargon with accessibility for non-AI audiences [2-8].


Ravindra Kumar introduced Technodate AI, describing its ambition to “automate automation” by using agentic AI to make industrial robotics and automation DIY-friendly, offering three modules for concept design, deployment, and troubleshooting [28-32]. He explained that while a foundational model would be ideal, limited funding in India forced the team to first engage customers, run pilot deployments, and only later recognize the need for such a model, yet they have already delivered solutions to Fortune-500 firms and the Indian Air Force [33-41]. Technodate’s credibility is reinforced by collaborations with experts such as Dr. Sumit Chopra and a team drawn from IITs, and a demo was promised to illustrate the end-to-end workflow [45-48][49-53].


Vaibhavath Shukla presented Quonsys AI, positioning it as a voice-infrastructure platform that removes humans from call-center loops, enabling end-to-end automation of customer support across Indian languages and leveraging a partnership with OpenAI for data generation [67-76][70-73]. He illustrated a real-estate lead scenario where the AI agent can answer calls, record interest, and schedule visits, and noted that the service is billed per minute of usage rather than a subscription [95-101].


Pradyum Gupta described Papri Labs’ visual-data mapping solution that continuously updates maps using dash-cam and CCTV feeds, processing petabytes of video to provide real-time information for applications such as billboard pricing, autonomous-vehicle safety, and public-transport optimization [134-142][148-154]. In response to data-privacy concerns, he stated that raw video is never released publicly, faces and number plates are blurred, and the system runs on bare-metal servers in Europe to comply with India’s DPDP regulations [207-214][225-233]. Pricing is offered on a per-tile basis, with a 25 km² tile costing 1.5 lakh rupees for a single-day license, scaling with volume [267-271].


Meenal Gupta introduced Imagix AI, an AI-driven precision imaging platform for cancer treatment planning that is HIPAA-compliant, ISO-13485 certified, holds four patents, and has achieved 92-99 % accuracy after training on a 5-million-image dataset that includes 30 % Indian data [288-295][332-340][336-342]. She emphasized that the system assists rather than replaces radiologists, keeping a human-in-the-loop for final approval to build trust in clinical settings [348-352].


Vivek Gupta then outlined Indus Labs AI’s voice operating system, a DIY, no-code platform that provides speech-to-text, text-to-speech, and LLM services optimized for Indian dialects with sub-400 ms latency and up to 70 % cost reduction compared with global providers [360-371][380-384]. The platform includes emotion detection, integrates with CRM workflows, offers per-second billing, and is hosted on Indian sovereign infrastructure, with partnerships for telecom connectivity and international white-labeling [389-403][417-424].


The session concluded with Archana thanking the founders and encouraging further one-on-one discussions, underscoring the emphasis on practical product deployment and compliance across diverse AI applications [468-469].


Keypoints

Major discussion points


Product-only presentation format: The moderator stresses that the summit is strictly for sharing product details, not business pitches or funding talks, and asks presenters to balance technical jargon with accessibility for non-AI audiences. [2-8][9-10]


AI-driven solutions targeting distinct industry problems:


Industrial automation: Technodate AI aims to “automate automation” with an agentic AI that helps users conceptualize, deploy, and troubleshoot robotics solutions, and notes the need for a foundational model after early customer experiments. [20-32][33-41]


Voice-first call-center automation: Quonsys AI builds a “voice infrastructure” that can run end-to-end call-center operations, handling inbound leads, booking appointments, and charging per-minute usage. [68-78][99]


Real-time map updating: Papri Labs uses city-wide dash-cam and CCTV feeds to create instantly refreshed visual maps and offers use-case-specific pricing (tiles of 25 km²). [118-138][267-270]


AI-assisted cancer treatment planning: EasyOPI’s Imagix AI provides HIPAA-compliant, ISO-certified imaging analysis that reduces manual contouring time from up to 90 minutes to 5-15 minutes, with reported 92-99 % accuracy across multiple Indian states. [288-336][337-345]


Voice-platform as an operating system: Indus Labs AI builds a DIY, low-latency voice stack (STT, TTS, LLM, emotion detection) for Indian languages, promising up to 70 % cost reduction and integration with CRM and telephony systems. [355-363][380-384][389-393]


Technical and regulatory challenges around foundational models, data, and compliance: Several founders discuss the difficulty of building or accessing large foundational models, the importance of proprietary data engines, and strategies for scaling while remaining compliant with data-privacy laws (DPDP) and medical regulations. [33-41][109-113][207-216][225-236][290-352]


Business models, pricing, and deployment considerations: Presenters outline their revenue approaches-per-minute usage for voice agents, tile-based licensing for mapping data, and subscription models for automation-while fielding audience questions about integration, incentives for data contributors, and cost structures. [99][267-270][244-247][99-101]


Overall purpose or goal of the discussion


The session is a founder-focused showcase at an AI summit where each startup presents only its product (no fundraising or market-size pitches) to enable peer learning, surface practical implementation issues, and foster collaboration among AI innovators. [2-8][9-10]


Overall tone


The conversation begins with a courteous, instructional tone as the moderator sets expectations. It then shifts to an enthusiastic, technical tone as founders detail their innovations, followed by a more interactive and inquisitive tone during the Q&A, where practical concerns (pricing, compliance, scaling) are raised. Throughout, the atmosphere remains supportive and collaborative, with occasional defensive nuances when addressing challenges (e.g., data-privacy compliance). [2-8][55-60][207-216][225-236][442-445]


Speakers

Archana Jahargirdar – Moderator/host from Rukam Capital; facilitates founder presentations and Q&A sessions. [S4]


Meenal Gupta – Founder of EasyOPI Solutions; expertise in AI-driven precision imaging and treatment planning for cancer (HIPAA-compliant, medical-device software). [S3]


Vaibhavath Shukla – Founder and CEO of Quonsys AI; focuses on voice infrastructure and AI-powered call-center automation. [S6]


Pradyum Gupta – Founder/representative of Papri Labs; builds real-time mapping and visual-analytics platform using dashcam/CCTV data. [S7]


Ravindra Kumar – Representative of Technodate AI; works on agentic AI for automation, robotics conceptualization, deployment and troubleshooting.


Vivek Gupta – Founder and CEO of Indus Labs AI; develops a voice operating system (speech-to-text, text-to-speech, LLM) for Indian languages with low-latency voice agents. [S11]


Audience – General participants asking questions; no specific role or title provided.


Additional speakers:


Weber – Mentioned by Archana as the next presenter; no further details available.


Karan – From Rukam Capital (mentioned alongside Archana); likely a partner or investor at Rukam Capital.


Dr. Sumit Chopra – Ph.D. collaborator referenced in the discussion; expertise in AI research.


Full session reportComprehensive analysis and detailed insights

The session began with moderator Archana Jahargirdar establishing a strict “product-only” format, asking founders to discuss only the technical aspects of their solutions and to avoid any mention of business models, funding or revenue. She also invited presenters to use jargon where appropriate but to simplify where possible for audience members who are not AI specialists [2-8][9-10].


Ravindra Kumar (Technodate AI) introduced the company’s ambition to “automate automation” by deploying agentic AI that makes industrial robotics and automation accessible as a DIY task [11-13]. He described three core modules: (i) conceptualising engineering solutions, (ii) deploying and commissioning them-including robot programming, and (iii) troubleshooting any failures [30-32]. Although an ideal foundational model would accelerate development, limited funding in India forced the team to first engage customers, run pilot deployments and only later recognise the need for such a model [33-41]. Kumar highlighted collaborations with Dr Sumit Chopra and a team drawn from IITs [45-48] and announced a live demo of the end-to-end workflow [49-52]. He also disclosed upcoming deployments with the Indian Air Force [53-55], partnerships with several Fortune 500 companies [45-48], and noted that Technodate will exhibit at Hall 14 for further discussions [58-60].


During the brief Q&A, Kumar argued that even if a super-intelligent model (ASI) were available, the real value lies in building the application layer that solves specific customer problems, rather than relying solely on the model itself [55-60][61-64].


Vaibhavath Shukla (Quonsys AI) positioned his venture as a “voice infrastructure” that can fully automate call-centre operations, removing humans from the loop and handling tasks such as inbound lead qualification, appointment booking and follow-up [68-78]. He cited partnerships with OpenAI, Paytm, CRED and PropBotX [84-86][95-101] and described a pricing scheme based on per-minute usage rather than a fixed subscription [95-101]. To overcome data scarcity, Quonsys built a proprietary data engine that generates synthetic training data and also powers Indic-language voice models in collaboration with OpenAI [78-80][109-115]. Shukla reported cost-saving figures of 70-90 % for large BPO or enterprise customers [119-122] and outlined plans to increase concurrency from the current 50 requests to thousands as the model scales [124-131]. In the subsequent audience interaction, Shukla reiterated that the proprietary data engine is central to scaling and that the system can dynamically adapt to varied use-cases (e.g., real-estate lead handling) by integrating with web-socket handshakes and CRM tools [95-101][124-131].


Pradyum Gupta (Papri Labs) described a visual-data mapping platform that continuously refreshes maps using dash-cam and CCTV feeds deployed across metro cities [134-142]. The platform processes petabytes of video to support use-cases such as dynamic billboard pricing, autonomous-vehicle safety checks, optimisation of Delhi Transport Corporation’s bus fleet, and automated news generation [148-152][150-156]. Pricing is offered on a per-tile basis (25 km² tiles at ₹1.5 lakh per day, with volume discounts) [267-271]. When questioned about data-privacy under India’s DPDP regime, Gupta clarified that raw video never leaves the company, that faces and number plates are blurred, and that all processing runs on bare-metal servers in Europe rather than on hyperscalers [207-214][225-233]. He also stated that contributors (e.g., dash-cam owners) are not paid incentives; instead, the platform charges them for the service [246-247].


Meenal Gupta introduced EasyOPI Solutions’ “Imagix AI”, an AI-driven precision imaging platform for cancer treatment planning. She highlighted the acute shortage of oncology experts in India and explained how the system assists radiologists by automatically contouring organs at risk, reducing manual processing time from up to 960 minutes to 5-15 minutes [332-340][336-345]. The product is HIPAA-compliant, ISO 13485 certified and holds four patents, with an accuracy range of 92-99 % after training on a 5-million-image dataset that includes 30 % Indian data [288-295][337-342]. Trust is reinforced by keeping a human-in-the-loop for final approval [348-352]. Gupta also mentioned an invitation by Bill Gates at Microsoft to showcase the technology [332-340].


Vivek Gupta (Indus Labs AI) presented a DIY, no-code voice operating system that provides speech-to-text, text-to-speech, large-language-model and speech-to-speech capabilities optimised for Indian dialects. The platform delivers sub-400 ms latency, supports emotion detection and integrates end-to-end with CRM workflows, promising up to 70 % cost reduction compared with global providers such as 11 Labs [362-368][370-384][389-393]. Data residency is ensured by hosting all components on Indian sovereign infrastructure [390-393]. The company has already partnered with telecom operators (Airtel, Geo) and international white-label partners in Dubai and Germany [417-424], and the system also supports Arabic, German, French and Mandarin languages [424-426]. The system is billed per second of usage, with a recharge-based model for customers [423-425]. In the Q&A, Gupta demonstrated how a user can define a lead-handling journey by linking nodes to Google Calendar, enabling the AI agent to book meetings automatically [435-440]. He also described the company’s origin story: after encountering pronunciation issues with third-party TTS, the team built its own stack, first using public data and then creating a proprietary data engine to achieve scalability up to 1 000 concurrent requests within ten minutes [451-466].


The presenters largely agreed that domain-specific application layers supported by proprietary data pipelines are more critical than investing in large, generic foundational models, and that such pipelines help meet data-privacy and regulatory requirements while delivering significant cost savings [55-60][109-115][162-165][362-368][207-214][225-233][290-294][390-393]. Divergences emerged around three points: (1) Kumar argued that a foundational model would eventually be required for industrial automation [33-40], whereas Shukla maintained that a custom data engine suffices [109-115]; (2) Papri Labs stores raw video on European bare-metal servers to satisfy DPDP, while Indus Labs insists on keeping all voice-AI data within India for sovereign control [207-214][225-233][390-393]; and (3) pricing strategies differ, with Papri Labs using a per-tile, per-day licence [267-271], Quonsys charging per minute of AI usage [95-101], and Indus Labs adopting a per-second, recharge-based model [423-425].


Most presenters largely respected the product-only guideline, though a few references to pricing, partnerships, or commercial arrangements slightly breached the rule [43][267-271][74-76].


The session concluded with Archana thanking the founders, encouraging attendees to continue one-on-one conversations, and inviting everyone for a group photograph [468-469], reinforcing the summit’s goal of fostering collaborative, product-centric dialogue among AI innovators while highlighting shared challenges of data ownership, regulatory compliance and cost-effective deployment across diverse Indian sectors.


Session transcriptComplete transcript of the session
Archana Jahargirdar

Thank you. Thank you. you you Thank you. Thank you. Thank you. so how do founders learn about these changes the only way you can learn or maybe the best way to learn at a conference like this at a summit like this is by listening to each other so it’s not a pitch there’s not going to be any talk about business there’s going to be no funding conversation it’s only about product so I’m going to request all the founders who are presenting to come up and then we’ll go sequentially and the other request on the presentations to all the founders who are presenting is please come I mean feel free to come up is that please use jargon because the intent is that the audience will understand it however also be mindful that if you want more people to learn who may not be AI natives may not be AI people may not be technologists may not be AI people but it’s still important for them to learn and understand.

So if you can simplify it, it’s fine. If you don’t want to simplify it, it’s also okay. So the format we’ll follow is that each one of you takes a little bit of time to talk about your product. But like I said again, only product. No business, no pitching, no money, nothing. So I’m going to request to start with, if I could request Ravindra Kumar to talk about what is it that you’re building. So quick introduction and then the product that you’ve built.

Ravindra Kumar

Hi everyone, this is Ravindra from Technodate AI. And we are aiming to automate automation itself. Everybody says AI won’t take away any jobs. We’re like, let us do something about it.

Archana Jahargirdar

Do you want to stand there at the podium or you want me to start? Whatever, whatever. No, no. Yeah, you can start your presentation. Shall we do that? Yeah, yeah, we should. Because I want people to really get into the product.

Ravindra Kumar

Can I use the clicker? What generally happens is that there are already very sophisticated automation equipment available out there in the market. Before starting Technodate, I have been working with this company, which happens to be world’s largest manufacturer of industrial robots. Way back in the year 2010 or so, they achieved 100 % automation, which means no human on the shop floor. Still, the manufacturing is happening at 100 % capacity. On the other side, if you see globally, including India, manufacturing is not even successful. People are not able to. use automation to the fullest extent. That is something which Technodate is aiming to solve. We want to make automation as easy as DIY using agentic AI. So what it does is basically help you in three ways.

First, to conceptualize an engineering robotics and automation engineering solution on your own. Then to deploy and commission that including robot programming etc. etc. And then eventually it also helps you to troubleshoot when something doesn’t work. For this, we started as okay, we have to do something like this because the idea of this discussion is how do we go from experimentation to real world deployment. So when I came up with this idea, the first thought was that you need to build a foundational model. But we are in India. It’s not that easy to raise money to build a foundational model. But then how do you approach this? The idea is okay, let us go talk to customer.

Let us experiment with what all options are available out there and then figure out in the process, do we need a foundational model? So we started working, started talking to customers, started doing some initial deployments. Today we stand again back where we started from that we need a foundational model for this. But in the process, we have already started deploying application, including with people like Fortune 500 companies. This is how the team comes from. We are working with some people that I’m sorry, Archana, but being a founder, some pitching comes in by default. But then, yeah, this is how the team looks like. We are collaborating with people like Dr. Sumit Chopra, a Ph .D. under the godfather of AI and Lincoln.

He worked at a fair earlier. We are exploring or we are rather going to deploy a use case very soon with Indian Air Force itself. Of course, the team comes from IITs. I have a small demo to show to everyone. There’s some music to this. But it is. It’s just music. There’s no audio in any case. So what it does is, as I said, three modules. it helps you to conceptualize robotics and automation solution it helps you to build that the agentic AI what it does is it acts mimics automation expert it really finds what it takes to deploy that solution in a real world scenario it gives you the complete architectures it gives you the programs it gives you the step by step procedures how do you put these systems together and then eventually in the final I mean of course you can also ask it to make changes what happens is in industrial scenario you change one equipment it has to talk to all other equipment so everything changes so it does all that on its own autonomously by using agents in the background you can also see how that solution will look like on your factory which has been conceptualized by the agents and then when it comes to robot programming many people ask me that Chad Jibadi can’t do this why do you want to build a foundational model for this when it comes to robotics necessarily we are interacting with the real world you have to understand what is the object what needs to be done how the robot needs to move all that data needs to be injected into the systems only then robotic programming can be done anyways uh then there is something called cnc programming cncs are the mother machine so every aerospace component every automotive engine be it two -wheeler four -wheeler they’re all machined on cnc machines for that matter to build other machines you need a cnc machine so all those programs can also be generated by using uh agentic or generative ai in this case in defense use case it’s like for example this is a case of aero engine where you just say the error code the 3d model explodes and the generative ai tells you where and what steps to take to solve that particular problem for example you will be able to see you said the error code it shows you where in the whole machine that error belongs and these are the steps you need to take to solve these problems so yes this is it from me we are exhibiting at hall 14 see you all there who want to discuss more

Archana Jahargirdar

so anybody has questions on the product you including founders sitting on this panel can ask questions on the product any question yes please

Ravindra Kumar

we’ll use their model see I have no I am not fond of building foundational model my aim is to solve the problem of my customer absolutely so one thing that this is this will never be in a human history these kind of tasks will never be simple chat response kind of scenario you need common play workflows, right? So even if, let us say, OpenAI wants to do it, he will have to build a custom application for this, right? So this is an application layer. Model can become ASI, the super intelligence level. You still will have to build the application. So that is our first approach. Second is for industrial domain. Even if OpenAI wants to do today, he will have to build a foundational model for this, separately.

Because it is related to industrial world. The 3D actual world, the data is proprietary, customer doesn’t share it with you. So your application has to run on his premises or on the virtual clouds.

Archana Jahargirdar

Okay, thank you. Weber, you are next.

Vaibhavath Shukla

Thank you so much. Thank you. Thank you. Thank you. first of all I would like to thank Karan and Archana from Rukam Capital for giving me this opportunity India doesn’t need more wrappers we need infrastructure and that’s what we are building at Quonsys AI my name is Vaibhavath Shukla I’m the founder and CEO of Quonsys AI we are building the voice infrastructure for India and so India is the customer support capital of the world it is 55 billion dollar industry for us it’s roughly 2 % of India’s GDP and the problem is that this entire model is outdated in the agentic era so that’s what we are solving we ask this question to ourselves if we can automate the call centers itself and the call centers can automate and completely run by themselves so for that we started solving this problem and we started building from scratch for what exactly is required to automate the entire call center pieces and that’s what we initiated with Quonsys AI.

So Quonsys is the default layer wherein you don’t need humans in the loop which can automate the entire call center and BPO infrastructure and we can completely run end -to -end for the processes. So these systems can listen, understand, act, respond and solve the entire purpose for any particular use case. So it’s not a concept anymore. We have been working with some of the top enterprises like Paytm, CRED, PropBotX. We are also partnered with OpenAI for the infrastructure. We are working with them on voice and Indic languages infrastructure which we have developed by our own digital data engine and we can generate data at scale. So we are different in a way because we have solved the entire layer for be it your application layer, for orchestration layer, then organization, on the model layer and the data layer itself.

So for example, anything and everything that is required we are basically making the entire suite of the… automation layer for call centers. So it’s completely, you can say call centers are completely running on itself. We have built companies before. We have a really good research team which is helping us in developing the entire foundational layer of it. And we have deployed some of the use cases when we have already worked with some of the large enterprises already. Yeah, and I’m happy to answer

Archana Jahargirdar

Any questions on the product?

Audience

Yes, yeah. So the call lands on the somebody’s phone.

Vaibhavath Shukla

Correct.

Audience

So it’s like again a kind of thing.

Vaibhavath Shukla

Correct. those kind of scenarios yes it can be can you be more specific on the use case

Audience

yeah for example uh i got generated a lead on google ads or say a training uh on digital marketing right

Vaibhavath Shukla

yeah

Audience

so that customer is calling to a particular number

Vaibhavath Shukla

correct

Audience

this lands on say in this phone

Vaibhavath Shukla

yeah

Audience

so can i put this agent into this phone which can attend that call and answer according to my requirements

Vaibhavath Shukla

yeah it can definitely do so what it will do in the back end and then you can you know have a handshake handshake of web sockets when your number and the other number that we have uh we can basically merge together and the conversation can float from there and it can answer the questions because these are dynamic questions it’s not a fixed kind of question

Audience

right

Vaibhavath Shukla

right it can so all the knowledge that you can you’re going to give it so for example i’ll give you a use case of real estate right so uh if somebody’s making an inquiry about the real estate project you basically fill the form you get the number there the AI agent will make the call it is already trained on the entire data set of your real estate project where is it what is the per square feet size the cost of it what are the amenities and all those things locality all those things which is already trained on that it will talk to you on the basis of all that information it will record the interest level from you whether you want to visit the site or not and then it will automatically book the site visit as well and you can trigger SMS WhatsApp email whatever you require so everything that was previously done by a call center agent is completely automated using AI agents and it’s end to end process so basically the purpose that you have given it it can completely solve for that

Audience

and can like institutes can take this or companies can take this on stand alone basis or you have put in a subscription mode kind of thing

Vaibhavath Shukla

so it’s more like with charging per minute kind of subscription at this point so you set it up one time then or whatever the number of minutes that you consume with us you pay for that

Archana Jahargirdar

Okay. Any other question?

Audience

Yeah. I mean, you talked about building foundational models before the ending language, right?

Vaibhavath Shukla

Yeah.

Audience

So if you could tell me how you’re scaling on the same because foundational models are very good for demos, but when we scale, we have even seen Servam breaking.

Vaibhavath Shukla

Right.

Audience

So how are you…

Vaibhavath Shukla

That is right. We basically gave a demo with Servam and… Guys.

Audience

Well, that was too loud, but then, yeah, how are you thinking of combating that scenario?

Vaibhavath Shukla

The main thing is basically the data engine, right? So, I mean, data that you have basically trained it on, that’s the most important piece. Initially, when we tried it with Bhashini and Google data sets, all the public libraries that are available, we basically tried to fine -tune and generate, basically train the model on that data set. But unfortunately, like you mentioned, there are so many problems with that. So that’s why we built our own data engine. As you can see, we… We won an award from Prime Minister Modi as well. It was right here in the Bhatman room last year. so we basically generated data generate data from our own data data engine and that is what we are basically putting it in the model so for use case by use case for example Paytm that is working with us at scale we are making tens of thousands of all with Paytm for those kind of use cases we basically take it what exactly is the use case on right for example merchant is a very complex use case yeah right right right so we are working I mean concurrency currently is around 50 now we are going to increase that as I mean as the model grows we will increase the concurrency like right right right right right right right right right correct so there are two kinds of uh problems right so there are smaller companies which are employing five guys ten guys so if you talk about that that’s not something where we are currently focusing on and it’s not the industry can’t focus on that i mean the pricing will come down drastically in the next couple of years but there are companies like sbi insurance they are employing tens of thousands of people in particular building right so from real estate from managing the security the parking spaces uh the hr management team managements all those things subscription headsets machinery all that those things if you take it down to the last minute so that costs roughly 25 to 30 minute rupees per minute for this thing this particularly maybe cost three rupees per minute so that’s more like 90 percent of the cost saving for those kind of companies so that’s where the current market is and that’s what we are basically focusing on

Archana Jahargirdar

thank you thank you very much I request guys a round of applause for all the founders I request Pradyum Gupta to now come and present be generous with the applause at the questions both please yeah I mean founders are taking time out to talk about their product

Pradyum Gupta

thank you ma ‘am for providing me this opportunity okay hi Hi everyone, my name is Pradyum. I am representing Papri Labs here. So, just giving a simple example what Papri Labs actually do is that, for example, you are today all coming to Bharat Mandapam. Now you must be using, maybe if you are here from Delhi, you might not be using a map, but I am from outside Delhi, so I was using a map from IIT Delhi to Bharat Mandapam. Now what happened was that it said to me that these gates are open, but these gates were all of them were closed. I was just looking around all over the parking areas. And normally this is a common problem in the map system today.

What map system have done is that they have brought a great navigation system. So, you want to go at a particular place that could be anywhere all over the city, you can go down there. But that navigation will never be so much aware that the kind of awareness that you require. So, for example, that there could be a ploy. There could be a place that the gates are closed. There could be something which is happening. Maybe a very heavy fog is there. Now that is not updated. What our company does is, it is not updated. is that we update the map, any kind of existing mapping system in a very instant way. How we do it?

We work on a visual system. So for example, any kind of a vehicle which are having on the ground, they have our cameras placed in. So these are simple dashcams, CCTVs, anything which is visual, we basically place all over the cities and we work only in the metro cities as on date or only on the major highways. We take out all the data and we plot it over the map and then we not only place the videos or the images, we categorize them. So like for example, right now in Delhi, we work with a local transport here, DTC is a Delhi Transport Company. So we plotted about 8 ,000 units and then we were getting like 100 petabytes of data from all this thing.

And then we will categorize them so fast that you basically see the entire Delhi life. And the best part was that you can search from them, what’s going on there. So now, what are the use cases that we… we brought probably three use cases in the market so for example there’s a company called JC de Cox they own about 4 ,000 billboards all over New Delhi the problem in all these billboards is these billboards come at a standard price so for example you own like 10 ,000 billboards but you don’t you usually sell it only on a like a basis of if it’s a posh area then I will charge maybe more if it’s a less posh I will charge them less what we brought as a new kind of pricing mechanism that you charge on the base of impression count what digital arts brought for them and that’s how we were able to increase the revenue for about 40 -45 percent because now they were charging more on the revenue basis we work with the company called autonomous cars of this mg motors they had a hectare hectare vehicle which when they were entering in India back then they were bringing that thing in the autonomous set that one they were bringing internet edge inside one of the problems that they had was that they had luxury passengers but they wanted to know that okay what is happening on the road Like instantly, even if it has a fog, they need to know that, okay, divider is broken or not.

Am I safe in there or not? So we started to update. There’s a company called MapMyIndia. We started to update their systems very fast. Third, we worked with a company like BCG. BCG is a consulting firm which basically consults government to take decisions on the ground. What we told them that this is where the demand is there. This is where the capacity is high. That’s how we brought a root rationalization algorithm. That’s helped DTC on the ground to basically manage all their 8 ,000 buses to where they need to actually deploy more buses so that they can increase the revenue. But the second perspective was that more passengers can actually board the bus. So we update the map on the cases that they want.

Now, we have been penetrating in news. So normal daily newspaper that you usually read on a regular basis, there’s an image tree that is attached to it. In that image, there’s an about. 8 ,000 to 10 ,000 people usually just on the ground. do a basic job is to collect these images. What we do is because we have a huge volume of videos that we have, we are just updating them and they are creating a news out of it. So like you want to search anything, any news you want to create, you can create from there. How we do it? So because this is a more of a product business, so one of the problems that we faced in India when we were trying to scale this product is that even everyone is talking about AI, but today if I am just going to be asking any single passenger just to put a phone in their car and provide us data, none of you will do it.

And this was a very basic problem. We realized that people want it. In India there is this perception is that they want to absorb, they absorb the technology really fast. But to give that information is very hard. So we created a mechanism in which the customer started to supply the data by itself. So for example, when we started to deal with passenger bus service, we started to deal with passenger bus service. So we started to give them passenger bus service. So we started to give them passenger bus service. So we started to give them passenger bus service. So we started to give them passenger bus service. So we started to give them passenger bus service.

So we started to give them passenger bus service. So we started to give them passenger bus service. So we started to give them passenger bus service. counting because the problem was that they didn’t know how many people come inside the bus on a particular bus stand where we need to run more buses. When we started to deal with packages like logistic companies, logistic companies use digital locks. So today any truck which goes all over the country, the problem is that they put a digital lock and then they expect that the truck is safe. But that digital lock even gets opened, any kind of thing is not evidence in a court. What we added was a small camera in the particular container.

That thing counted how many goods were getting inside and how many goods were not, like what was the exact tally value. That’s how we deployed in our highway sector. So here if there are three sides of the data part, one is the passenger, that means we are getting city data. If we are deploying in the logistic sector, we are getting city data. We are getting highways data. And if you are deploying in the normal commercial cars, we are getting the lane information. And the perspective was. to just to get the front imagery. The back imagery they use it for themselves and we are certainly not interested in that part. That’s how we formed this entire information.

One of the problem that we faced to one of our customers was so we reached to one of so Delhi police and we started to sell this entire platform. They started to Google that like basically they want to search everything that where people are not wearing helmets because they want to cut the chalance very instantly. Now these things was that we created layers but we didn’t have a system that we can create very dynamic layers for that particular person. So what we did was we added that’s where the LLM thing came in that we started to describe every image and then internally we were searching everything for them. So like you have anything any idea in your mind you want to for example a person comes to me he says to me that find me all the CCTVs.

In New Delhi find me street lights which are not working just prompt it up. internally we are a video analytics company like we are so we keep we are running on a bare metal like hundred petabytes and then we’re just processing them really fast and you can then the best part that we brought was let’s start to compare like what changed now and what was previously before like six days back year back what was the development going on and this is how the basically the end customer gets so for example if it’s a local bus passenger company they wants to know that how many passengers actually board so we we provide a system to them but internally we use a front camera system so they use a fleet management system then we brought popular over the top so this is an example that we brought with DTC this this was funded by JICA JICA is an investment corporation which funds the government of New Delhi and that’s how we scaled in entire New Delhi second thing was that if you want to know any count like where the people whether how many cars been gone through or how many buses pass through that particular portion or where the two wheelers are or where the ambulances actually cross, we started to take out every information all over New Delhi and this is all real time.

So if you want to do it today, you want to compare it for like last six months and then you will track and target them, you can do all of it. So we brought these pay systems because JCD Cox was the organization for. So we are

Audience

Yes. Hello. So thank you so much for presenting. I am curious about knowing that you mentioned a certain petabyte of data that you are using and data you know is a very debatable topic right now after DPDP. So how are you handling that? How are you DPDP compliant? Because you are going to give this to certain other businesses also. So because you are getting a lot of personal data too, like getting images of people, getting images of the car numbers and all of these things. So how are you dbdp compliant and ensure that?

Pradyum Gupta

So there are two things. One is that inside videos we never take out for the public information. Even though the clients are ready to pay even 10 times over that value. Second thing is like this is the rule of property labs internally. Second thing is only front data is used. Front camera data faces are blurred. Number plates are also blurred. Second, third thing is that right now we don’t run on any AWS. So we don’t use hyperscalers right now. We only use bare metal servers. So bare metal is stacked in which we keep everything in Europe right now. So Europe, Hetzner, we have taken a portion of their data centers. And the second thing, so in India there’s a big problem.

Audience

how are you handling that? How are you DPDP compliant? Because you are going to give this to certain other businesses also. So because you are getting a lot of personal data too, like getting images of people, getting images of the car numbers and all of these things. So how are you DPDP compliant and ensure that?

Pradyum Gupta

So there are two things. One is that inside videos, we never take out for the public information, even though the clients are ready to pay even 10 times over that value. Second thing is, like this is the rule of popular. It loves internally. Second thing is only front data is used. Front camera data faces are blurred. Number plates are also blurred. Second, third thing is that right now, we don’t run on any AWS. So we don’t use hyperscalers right now. We only use bare metal servers. So bare metal means stacks in which we keep everything in Europe right now. So Europe, Hetzner, we have taken a portion of their data centers. And the second thing, so in India, there’s a big problem.

problem. One of the things that people say that, okay, GPUs are a lot, but the reality is that GPUs are, the companies which are actually selling these GPUs never purchase from them, rather purchase from CDAC. So CDAC is an organization called Aravat. Aravat is providing us supercomputers and like a bare, dirty price. So if you’re going to be searching on Aravat, just purchase their GPUs and then keep data on bare metal on your security premises, then it’s very safe. And it’s very cheap.

Archana Jahargirdar

Okay, one more question. And in the end, we’ll take more questions because once everyone’s done their presentations, please go for it.

Audience

I want to ask, because I’m a performative on the product, and I just want to ask, like, what are the incentives you are giving to the dashcam holders? Like, I heard you are giving incentives to the local DTDC buses or like…

Pradyum Gupta

So we don’t pay incentives, they pay us.

Audience

So like, what is the leverage you are holding for them to…

Pradyum Gupta

For example… So this company… DTC, Dairy Transport Corporation, they burn about 80 crores every year on not providing the timely bus service. And they had a very low revenue. Like they had a revenue loss of about 800 crores as I had a talk with A .S. Sachin Shinde back then he was there. Now A .S. Jitendraji came in. Now when we came into the system, we actually reduced on the revenue loss for them. So for example, if you see this number, 27 is the demand which is there and 25 is the capacity. So in India what happened was that when Amadvi Party came in, they provided female passengers as a free bus service. Now every party started to criticize all these parties that you are providing free for the bus service to the females.

We were the first company which actually gave them a mandate that 1 % is the actual female passengers are operating. So that’s how they were able to save their lives. So that’s when, you know, when we came in, we saw there are a lot of… operational issues.

Audience

We can access it through our apps or like…

Pradyum Gupta

Nothing is possible. We are a pure B2B company. We never intend to be B2C.

Archana Jahargirdar

Okay. We’ll do questions at the end. Let’s finish the presentations and… Okay, quickly. But short answer on your part. Yeah, but short answer, please.

Pradyum Gupta

So, we sell on tile basis. So, for example, this particular area, this comes at 25 by 25 square kilometers. This starts at 1 .5 lakh rupees. Per tile. This is for only valid for one day. And this usually multiplies at a volume at the company comes in. Thank you so much.

Archana Jahargirdar

So, now I’ll request Meenal to come and present, please.

Meenal Gupta

Hello everyone, I am Meenal Gupta from EasyOPI Solutions and so nice to see you over here. Who all are founders over here? Oh wow, so many. So we… Founders are here. So founders should be here. Who all are founders over here? So I love to be with founders. I know they share the journey and the struggle, they know it very well. So we all three women, mostly known as GreenDeviya because this name was given to us by Mr. Narendra. Okay. So I am Narendra Modi. So Meenal Gupta, I am the founder. I am the founder. Noor for… Noor for… and Sheetal Tarkas. We all started this journey. Our platform, we have named it as Imagix AI.

It’s an AI driven precision imaging to treatment planning for cancer. So we are HIPAA compliant. We have four patents in hand. We are ISO 13485 certified company. We also have SEDESCO license. Talking about SEDESCO, people who are from medical field, they might be knowing that there is a license which is required when we want to take our solution to hospitals. So there is a ICMR agency that certifies your product that is software as a medical device. Once it is certified, then you can take it to any hospital. You can actually commercialize post that. So we are a company. We are SEDESCO certified. So talking about the problem, we have a lot of people who are from medical field.

We know there are around 20 million new cancer cases every year. And there is not like that doctors don’t have the intent of solving the problem or treating cancer. But the main problem is the shortage of clinical experts. We can increase the devices like diagnosis devices can be increased, imaging can be increased. But the problem that is facing is the shortage of expertise for oncology. So and finally the treatment planning. So talking about the problem, once a cancer is being detected, the patient is sent for CT scan. Once the CT scan MRI. Once it is being done, then. Tumor board decides whether the patient has to go for radiation therapy or they have to go for surgery.

or the combination of both. I can understand everyone can relate it because almost every family in India or world are having someone very near and dear who are facing through cancer and they have gone through just such challenges. So what happens is because of this shortage, it cause life. It cause life or the stage of the cancer changes. Either it changes from first level to second stage. It changes because of this unavailability. So this is where our solution comes in. So this was our own personal experience where all three founders have personally experienced cancer to our near ones and we have gone through this radiation therapy where we had to wait in queue because of unavailability of specialist, because of unavailability of treatment planning.

So this was a very big bottleneck. You can see over here. So this was a very big bottleneck. You can see over here. So this was a very big bottleneck. So this was a very big bottleneck. So this was a very big bottleneck. So this was a very big bottleneck. So this was a very big bottleneck. So this was a very big bottleneck. Once a patient is recommended to go for radiation therapy, there is a planning to do for that radiation therapy. For this planning, there is a manual process where wherever there is a tumor, all the surrounding organs of tumor, they are to be segmented. And this is a manual process. I can proudly say that in India, there is no one who is solving this problem and we are the only one who have this solution.

Here we contour all the, contouring is the masking. We mask all the organs which are on risk and are surrounding tumor. So here the main purpose is that all the organs that are surrounding tumor, they should be saved because they are healthy organs. And radiation therapy should be as less as possible on those healthy organs. the manual process it used to take somewhere around 960 to 90 minutes we have reduced it to at the max of 15 minutes the reason is uh for complex uh radiation therapy when it is head and neck cancer it takes lots of time so maximum 15 minutes and minimum 5 minutes so we have reduced it here uh what we do is once the patient is diagnosed with cancer he goes for city scan tumor board this city scan is being done it is uploaded on our cloud tumor board have the access of this city scan through our own uh dicom viewer uh dicom is the format uh through which this images can be seen they decide whether they have to go for city scan radiation therapy planning or surgery and we have various suits we do ai analysis where do we do first level of analysis where we mention the load of the tumor and second level of analysis is we have various suits we have XraySuite, NeuroSuite and OncoSuite which works on this scans and finally we give the final report.

So this is our product. We have trained our AI on about 5 million of data set in which around 30 % is Indian data set and this 30 % we gathered it from northeast region. So northeast region is very tough terian and we got support from Niti Aayog where we went and collected data. It is very tough terian. Means taking off AI over there is very challenging because 4G has not reached there yet. So we had to implement it on -premise solution so that we can get the data and we can help them solve that. So 30 % of data we have almost deployed it in 14 states in India. We got our data 30 % from that. Accuracy is around 92%. So it is around 92 % to 99 % depending upon the data.

complexity you can see this data we are working in Gujarat in seven district where we are helping to scan to do CXR chest and lung analysis where we have helped we have helped we have made somewhere around 1 million of scans we have detected around 4 ,000 TB positive cases till yet in which there were around six lung cancer cases where early intervention is still possible we have done around thousand of radiation radiotherapy plans till yet and talking about in last three months we had done around five fifty thousand of chest x -rays where twenty seven hundred of TB were flagged so we could save TB these are live photos where handle x -rays and all are being done so I this was our solution was recognized first by Mr.

Naren Indra Modi and day before yesterday we were invited in Microsoft by Bill Gates to show our solution to him.

Audience

In health tech, I’ve observed that trust is a very big factor in terms of AI adoption and you seem to be implementing it across India. So how do you make sure that the technology and the science behind it is trusted by the people who are being benefited by it?

Meenal Gupta

Yeah, I understand. So here, our solution, we are not replacing doctors. We are just assisting doctors. We have made their manual process easy process. But final approval has to be done by radiologists. So it is human in the loop. We are not claiming that directly our AI will solve it.

Archana Jahargirdar

Thank you. I’ll request Vivek now to come

Vivek Gupta

Hi everyone. So first of all, thank you so much team Rukam Capital for organizing such a vibrant event and the energy is full of high in this room, I can see. It can be higher though. Yeah. I think Yes. So my name is Vivek Gupta. I’m the founder and CEO of company called Indus Labs AI. So we are building the voice architecture of India. So we are basically building the whole layer of . operating system of voice, where all the layers like speech to text, text to speech, the LLM, speech to speech, all of this infrastructure we are building, right? So it is a common platform where anyone, I said anyone can come on this platform and build their own voice agent.

As sir was asking the question about whether you are running a campaign on Google, you have put a number, you can build your own agent by yourself, right? So it’s a DIY platform where we are training, we are primarily focusing on Indian languages because the problem, linguistic problem in our country is if you see after each 20 kilometer dialect changes, right? So we are working with couple of banks, NBFCs, right? And whenever they run a campaign, they run a basically cold calling in let’s say Mojaffarnagar region, in UP, right? And whenever they call in Gorakhpur region, the Hindi is totally different. But global players like 11 Labs or maybe, you know, the Indian language, they are all different.

You know, some different global players, players like Azure and Google, they are providing a generic Hindi, but we need. a company in India who can build the infrastructure of voice in our country based on our directs. And again, so while we are building the infrastructure on our own GPUs and servers, hyperscalers we have inbuilt in our system, so that’s how we are able to reduce the latency. So we have some sub 400, 500 millisecond around latency into the system, which is like more human conversation, you can feel it. And you know, the complete analysis of the call is there, right? As soon as the call gets disconnected, you will find the sentiment analysis of the call, the outcome of the call, and it will log into the system.

So what is the expected outcome of the call? It will go into your CRM. So the journey starts from your CRM and ends with CRM. You trigger the calls from CRM and it ends with the CRM. So again, as I said, like native dialect mastery we have, and we are ultra low latency. And again, you know, call. somebody was also asking question regarding how effective it is in terms of cost. So if I talk about existing system cost we are reducing the cost up to 70 % right. So up to 70 % cost can be reduced and operationally you are enabled like 24 7 availability of the system is there. System is multilingual right. You don’t need to have multiple people for different languages.

Single system can handle 24 7 and that’s how you are able to reduce cost and operationally efficient. And you know the important part is emotional handling right. So like one year back I started this company 2 .5 years back right. I used to be director of an engineering company in software company in Bangalore and 2 .5 years back I quit and started Indus Labs and my background is from IT Delhi right. So the core problem was emotions right when I started this company. I always thought like if somebody is laughing over the call how would AI system would recognize the person is happy or angry. That’s how the agent would say sorry or congratulate you. Right because you need to understand the emotions right.

So we were working on this. is speech -to -text model since last 1 .5 years and on the 16th of this month in the department only we launched this model called it’s basically no emotion of your STD so we launched this model here in part of the monthly and we are basically you know distributing it to our customers existing customers now so that they can start using it’s a PUC phase right now and the good part is since we are an Indian company the whole the data is going to reside here on our sovereign feel is there right so we are pure Indian origin company so as I said like if I if I compare with now global pairs like Google and 11 labs so we are cost so like let’s say I hope many people knows what 11 labs is right so their cost is somewhere around eight rupees per minute right but we are seventy percent lower we sell at two rupees per minute right and we are superior in terms of Indian dialect accuracy and we are superior and streaming latency is somewhere around three hundred to four hundred millisecond and emotional expressiveness is already there in our system as we recently launched it and Indian data residency clause is obviously there because we are an Indian company.

So, I mean, so we are a huge case agnostic platform. We don’t say like we are having mastery in this huge case. You can come on our platform. So, like as of today, we are working with multiple use cases, right? We are working with banks. We are working with enterprises in FMCG. We are working with, you know, customer support people. They are building their own voice agents, but they use our STT and TTS through API, right? So, not anyone can build STT and TTS. So, instead of using 11 labs, they use us because we are cost effective and obviously good in terms of Indian dialect mastery. So, it’s DIY platform. Anyone can come and build their own agent.

We have different flows. Workflow is already there. You can create nodes. Each node can be connected with webhooks or API and that can be used to build your own voice agent. So, we are working with customers. We are working with customers. We are working with customers. We are working with customers. We are working with customers. We are working with customers. We are working with customers. So it’s completely guided journey and you can create and you can also integrate your voice agent with telephony. So we have already partnership with Airtel and Geo. So that’s how you can give the SIP channels through there and connect with your voice agent. So it’s complete end to end journey.

And again, so the core market is B2B enterprises. Right. And we are also platform for developers. So developers can use our APIs into their existing systems wherever they want to use. Right. So you just based API and it is again per second based costing is there. So as how many seconds you will use, you will be the credits would be detected. So it’s a recharge based system. You recharge and you can use it. And also we are into we are building channel partners as well. So like we have a couple of partners. One is in Dubai. One is in Germany. So we are white white labeling our platform for them. Right. So they can onboard their clients on their platform internally.

They would become our client shared. and we can share the revenue so we have right now we have developed four to five partners globally so we are building from Bharat for the globe so we have foreign languages as well we have Arabic we have German we have French and Mandarin Mandarin is in building stage right now so we have core and English all the accents we have or still in accent British accent American accent all the male female voices and again you can clone your voice as

Archana Jahargirdar

thank you any questions quick questions any questions

Vivek Gupta

yep absolutely it’s a no -code platform it’s a journey based platform you need to trigger what you want to do what you want to basically build out of it so let’s say you are you want to build an agent for inbound agent for your leads right so anybody who is calling from Google Ads would land on this voice agent. So you will define the journey and how you want to integrate. Let’s say you want to, some meeting has been fixed for your product, right? So it is connected with Google Calendar, your Google Calendar. So as soon as AI agent books a meeting, you will get an email on your Google Calendar. This meeting has been fixed and your Google Calendar will be blocked.

Audience

So my question is like, I have to make this like, there are some nodes, I have to connect these and make a flow or you have all things are made up, we just have to click and the agent will start walking.

Vivek Gupta

Yeah, it’s like DIY platform and we have tutorials as well, right? If you are stuck, you can see the tutorials. So and still you are feeling you are not able to build it, you can support the customer center. So our team will help you in that case.

Archana Jahargirdar

Okay, quick question.

Audience

Yeah. How did you start? When you start, left your job?

Archana Jahargirdar

We don’t have so much time. You can talk to them offline, but the data is a question.

Vivek Gupta

I’ll make it short. So, journey started 2 .5 years back. So, I initially started building voice agent and started using TTS of somebody else. Then we figured out that this TTS is having this issue. How can I solve this? Because my customer will ask the issues complaint to me. So, we were able to solve this problem of pronunciation of some words because these issues were there already in the third -party system. So, we thought of initially building our own infrastructure. And we pivoted in a model that people will use our APIs. That’s what we want to build it, right? So, then, firstly, we used public data, publicly available data. Then we started creating our own data, right?

So, we create the data and we have multiple hyperscalers available. And scalability -wise, our system is so much scalable that you can put a thousand requests at a time, it will handle. So, it will scale 0 to 1000 within 10 minutes. So, that’s how we built it. Thank you very much.

Archana Jahargirdar

thank you so much for all the engaging questions that everybody did make the effort to ask we are time constrained over here so I want to thank all the founders for sharing your product and any questions you have for the founders please do connect with them and do continue the conversation it’s just that we need to leave the room and I request all of us to do a quick picture together thank you Thank you. Thank you. Thank you.

Related ResourcesKnowledge base sources related to the discussion topics (16)
Factual NotesClaims verified against the Diplo knowledge base (2)
Confirmedhigh

“Moderator Archana Jahargirdar emphasized a product‑only format, asking founders to discuss only technical aspects and avoid business model, funding, or revenue details.”

The knowledge base states that the session was moderated by Archana Jahargirdar, who emphasized that presentations should focus purely on product details, confirming the product‑only directive.

Additional Contextmedium

“Ravindra Kumar described Technodate AI’s ambition to “automate automation” by making industrial robotics and automation accessible as a DIY task.”

The knowledge base notes that before starting Technodate, Kumar worked with the world’s largest manufacturer of industrial robots, providing background on his experience in industrial automation that underpins the “automate automation” claim.

External Sources (86)
S1
Scaling Innovation Building a Robust AI Startup Ecosystem — -Shri Ashok Gupta: Title – Director STPI Gurugram; Role – Dignitary presenting mementos Hi, I’m Meenal Gupta, founder o…
S2
https://dig.watch/event/india-ai-impact-summit-2026/founders-adda-raw-conversations-with-indias-top-ai-pioneers — Accuracy is around 92%. So it is around 92 % to 99 % depending upon the data. complexity you can see this data we are wo…
S3
Founders Adda Raw Conversations with India’s Top AI Pioneers — Hello everyone, I am Meenal Gupta from EasyOPI Solutions and so nice to see you over here. Who all are founders over her…
S4
Founders Adda Raw Conversations with India’s Top AI Pioneers — -Archana Jahargirdar- Conference moderator/host from Rukam Capital, facilitating the founder presentations and Q&A sessi…
S5
https://dig.watch/event/india-ai-impact-summit-2026/leveraging-ai4all_-pathways-to-inclusion — That’s great. I think moving into the curriculum always helps that you’re planting the seeds early on for the training. …
S6
S7
Founders Adda Raw Conversations with India’s Top AI Pioneers — Pradyum Gupta from Papri Labs showcased a real-time mapping system that updates existing maps using visual data from das…
S8
Comprehensive Report: Preventing Jobless Growth in the Age of AI — He’s a commissioner at the European Commission. His focus is on the economy and productivity. Ravi Kumar is the CEO of C…
S10
https://dig.watch/event/india-ai-impact-summit-2026/founders-adda-raw-conversations-with-indias-top-ai-pioneers — And again, so the core market is B2B enterprises. Right. And we are also platform for developers. So developers can use …
S11
Founders Adda Raw Conversations with India’s Top AI Pioneers — Hi everyone. So first of all, thank you so much team Rukam Capital for organizing such a vibrant event and the energy is…
S12
Driving Indias AI Future Growth Innovation and Impact — And for this, I’m delighted to welcome two very eminent leaders who are instrumental in shaping the journey, both from p…
S13
WS #280 the DNS Trust Horizon Safeguarding Digital Identity — – **Audience** – Individual from Senegal named Yuv (role/title not specified)
S14
Building the Workforce_ AI for Viksit Bharat 2047 — -Audience- Role/Title: Professor Charu from Indian Institute of Public Administration (one identified audience member), …
S15
Nri Collaborative Session Navigating Global Cyber Threats Via Local Practices — – **Audience** – Dr. Nazar (specific role/title not clearly mentioned)
S16
Internet standards and human rights | IGF 2023 WS #460 — Engaging with technical standard bodies demands technical expertise, but acknowledging the need for inclusivity without …
S17
AI Transformation in Practice_ Insights from India’s Consulting Leaders — The tone was pragmatically optimistic and refreshingly candid. Both speakers were honest about challenges and uncertaint…
S18
Agentic AI and the new industrial diplomacy — Chileoffers a very different entry point into the same technology.Codelco, the state-owned copper giant and one of the w…
S19
AI for Good Technology That Empowers People — The professor emphasised that whilst foundation models attempting to solve universal problems receive significant attent…
S20
Conversation: 02 — Companies like us and others who are starting to make, we have been doing that for a few years, where they’ve been makin…
S21
WS #462 Bridging the Compute Divide a Global Alliance for AI — Ivy Lau-Schindewolf highlighted OpenAI’s Stargate infrastructure project as an example of private sector leadership mobi…
S22
Box 4.1: Adapting ITU’s price data collection to ICT developments — 1. The prices of the operator with the largest market share (measured by the number of fixedtelephone subscriptions) are…
S23
Day 0 Event #171 Legalization of data governance — He Bo: Thank you. Good afternoon, everyone. I’m He Bo from China Academy. Academy of Information and Communication T…
S24
Panel #3: « Gouverner les données : entre souveraineté, éthique et sécurité à l’ère de l’interconnexion » — Patricia Egger Merci. Oui, donc, c’est particulier d’être ici parce qu’en effet, j’ai la casquette, je représente Proton…
S25
Israel’s Policy on Artificial Intelligence Regulation and Ethics — Privacy– The development and use of AI systems necessitates the use of large quantities of data, some of which could inc…
S26
Closing remarks — Minimal to no disagreement present. This transcript represents a closing ceremony where speakers (Doreen Bogdan Martin, …
S27
Building Trusted AI at Scale Cities Startups & Digital Sovereignty – Keynote Ebba Busch Deputy Prime Minister Sweden — This transcript contains a single keynote speech by Deputy Prime Minister Ebba Bush with only brief introductory comment…
S28
Laying the foundations for AI governance — Low to moderate disagreement level. The speakers largely agreed on problem identification but differed on solutions and …
S29
Setting the Rules_ Global AI Standards for Growth and Governance — Yes. I’ll take it back to what Chris was talking about in terms of collective action problems. So some of the mitigation…
S30
Table of Contents — Advanced manufacturing addresses the transformation of the manufacturing and automation industry to a new level of intel…
S31
INCREASING ACCESS TO DATA ACROSS THE ECONOMY — Industrial policy objectives attempt to improve the business environment for specific sectors or technologies t…
S32
Global Data Partnership Against Forced Labour: A Comprehensive Discussion Summary — Integration with existing regulatory frameworks and compliance systems across different jurisdictions presents complex t…
S33
E-Commerce Legal and Regulatory Framework for Data Governance in Developing Countries ( Nigeria Customs Service) — In conclusion, startups face challenges when it comes to sharing data with regulatory agencies, particularly in terms of…
S34
Open Forum #64 Local AI Policy Pathways for Sustainable Digital Economies — Abhishek Singh: One part is that, of course, the way the technology is evolving, there is IP-driven solutions and there …
S35
OVERVIEW — – -The propensity to act fast including when ‘testing in the wild’ and deploying innovations at scale in ways that can u…
S36
Networking Session #37 Mapping the DPI stakeholders? — ## Audience Contributions A significant challenge he identified was the lack of visibility into deployment impact. Most…
S37
Main Session | Policy Network on Meaningful Access — Oscar G Leon Suarez: Hello. This is Oscar León, Executive Secretary of the Inter-American Telecommunication Commissio…
S38
Founders Adda Raw Conversations with India’s Top AI Pioneers — -Real-time Map Intelligence and Urban Analytics: Papri Labs demonstrated their visual data processing system that update…
S39
WS #257 Data for Impact Equitable Sustainable DPI Data Governance — Malik Payal: Thank you, Priya. And yes, as you mentioned about that T20 policy brief, which we did, it was really great …
S40
Turbocharging Digital Transformation in Emerging Markets: Unleashing the Power of AI in Agritech (ITC) — The business model for AI in farming can be particularly challenging, especially for smallholder farmers in emerging eco…
S41
GEO-politics/economics/emotions in the AI era — Paradoxically, as technology developed, it became increasingly tied to geography. Once connected, users’ physical locati…
S42
Keynote_ 2030 – The Rise of an AI Storytelling Civilization _ India AI Impact Summit — The speaker calls for a fundamental redesign of monetisation, moving away from advertising‑only and subscription models …
S43
NRIs MAIN SESSION: DATA GOVERNANCE — Additionally, there is an advocacy for appropriate data protection legislation and policies. Data is subject to the laws…
S44
The Challenges of Data Governance in a Multilateral World — An advocate in the discussion strongly supports data governance models that prioritize cooperation, privacy, and the com…
S45
Dare to Share: Rebuilding Trust Through Data Stewardship | IGF 2023 Town Hall #91 — The speakers also emphasized the importance of extending beyond first-generation rights when it comes to data governance…
S46
Driving Social Good with AI_ Evaluation and Open Source at Scale — The conversation then shifted to the growing problem of AI-generated code submissions to open source projects. Sanket Ve…
S47
Tessl secures $125M for AI-powered code platform — London-based startup Tesslhas raised$125 million in funding, achieving a valuation exceeding $500 million. Led by founde…
S48
From principles to practice: Governing advanced AI in action — Ya Qin Zhang: I thought the National AI Safety Institute and a lot of the NGOs have played a very constructive and posit…
S49
AI That Empowers Safety Growth and Social Inclusion in Action — Well, I mean, I think in general we have sort of corporations are incentivized to put products on market that are safe a…
S50
Summit Opening Session — The summit’s emphasis on practical guidance, from streamlining permitting processes to strengthening repair readiness, d…
S51
(Interactive Dialogue 1) Summit of the Future – General Assembly, 79th session — Tunisia: Mr. Chairman, an objective review of the current shape of our organization stresses the need of a deep reform…
S52
Background — – review and assess progress at the international and regional levels in the implementation of action lines, recommendat…
S53
Ad Hoc Consultation: Friday 9th February, Morning session — Additional Observations: – The focused nature of the statement, omitting counterarguments or challenges from other membe…
S54
Survival Tech Harnessing AI to Manage Global Climate Extremes — Professor Amit Sheth opened the discussion by explaining the origins of IRO, which emerged from a December 2023 meeting …
S55
The Foundation of AI Democratizing Compute Data Infrastructure — Given the volume of funds available, I would focus a lot more on capability development of people to be able, their abil…
S56
Report of the Special Rapporteur on the promotion and protection of the right to freedom of opinion and expression — ; Association for Progressive 28. The scale and complexity of addressing hateful expression presents long-term …
S57
PREAMBLE — – -The Signatories of this Code recognise the importance of diluting the visibility of Disinformation by i…
S58
OVERVIEW — 1. Technology company business models, and the commercial underpinnings of 21st century technological advances…
S59
Founders Adda Raw Conversations with India’s Top AI Pioneers — This was a founder showcase event organized by Rukam Capital where AI startup founders presented their products to an au…
S60
Closing remarks — Minimal to no disagreement present. This transcript represents a closing ceremony where speakers (Doreen Bogdan Martin, …
S61
Day 0 Event #178 Ethical Procurement in the Digital Age — As this is a single-speaker presentation, there is no consensus to assess among multiple speakers. However, the speaker …
S62
Strategic Action Plan for Artificial Intelligence — Many large Dutch companies are already working on deepening their knowledge of AI and using it to improve their services…
S63
Strategy — ‘Foster the use of AI in vital developmental sectors using partnerships with local beneficiaries and local or foreign te…
S64
Multistakeholder Partnerships for Thriving AI Ecosystems — LLMs solve only part of the problem; industry-specific, company-specific, and context-specific solutions still require s…
S65
MASTERPLAN FLAGSHIP PROGRAMMES — To create this plan, the government will convene an interagency AI task force comprised of National Government agencies,…
S66
MASTERPLAN FLAGSHIP PROGRAMMES — To create this plan, the government will convene an interagency AI task force comprised of National Government agencies,…
S67
DC-Blockchain Implementation of the DAO Model Law:Challenges & Way Forward | IGF 2023 — The frustration faced in movement and change across various legal systems is acknowledged. Overall, the analysis provide…
S68
Exploring Emerging PE³Ts for Data Governance with Trust | IGF 2023 Open Forum #161 — Automation is widely regarded as a crucial component in privacy management. It allows for scaling efforts and addressing…
S69
E-Commerce Legal and Regulatory Framework for Data Governance in Developing Countries ( Nigeria Customs Service) — In conclusion, startups face challenges when it comes to sharing data with regulatory agencies, particularly in terms of…
S70
Open Forum #64 Local AI Policy Pathways for Sustainable Digital Economies — Abhishek Singh: One part is that, of course, the way the technology is evolving, there is IP-driven solutions and there …
S71
WS #225 Bridging the Connectivity Gap for Excluded Communities — Christopher Locke presented community networks as viable alternatives to traditional telecommunications models, emphasiz…
S72
OVERVIEW — – -The propensity to act fast including when ‘testing in the wild’ and deploying innovations at scale in ways that can u…
S73
About the Authors — Modularity also has several important implications from a supply-side perspective. First, the same task can be accomplis…
S74
AI Infrastructure and Future Development: A Panel Discussion — And of course, Sora, because now we have multimodal. So the product platform is multidimensional. And then finally, the …
S75
Invest India Fireside Chat — -Moderator: Event moderator introducing the session participants
S76
AI for social good: the new face of technosolutionism — Birhane concluded her presentation by acknowledging that being allowed to “take centre stage here and to speak about thi…
S77
Al and Global Challenges: Ethical Development and Responsible Deployment — Dr. Shukla further discussed the importance of transparency in AI applications, which would enable better understanding …
S78
Democratizing AI: Open foundations and shared resources for global impact — Repeatedly invited audience participation, encouraged reaching out to the presenters, and emphasized the openness of the…
S79
Comprehensive Summary: The Future of Robotics and Physical AI — And so there are plenty of challenges, of technical challenges. And yet, if we look at what the machines can do today, w…
S80
https://dig.watch/event/india-ai-impact-summit-2026/keynote-by-vivek-mahajan-cto-fujitsu-india-ai-impact-summit — But then this technology, the compute networks, as well as the AI platform stack, comes together in edge devices. Robots…
S81
AUDA-NEPAD White Paper: Regulation and Responsible Adoption of AI in Africa Towards Achievement of AU Agenda 2063 — Strengthening the digital component of education entails a good foundation for scientific education at the tertiary leve…
S82
From Innovation to Impact_ Bringing AI to the Public — If we don’t make for it, our all compounded historical knowledge will be lacking in the next generation. So instead of a…
S83
How the Global South Is Accelerating AI Adoption_ Finance Sector Insights — Data residency requirements and lack of cutting-edge model infrastructure in India create deployment barriers
S84
https://dig.watch/event/india-ai-impact-summit-2026/driving-indias-ai-future-growth-innovation-and-impact — Thank you, Mridu, and thank you, everyone, for joining us for the unveiling of this important blueprint. As we have hear…
S85
The Innovation Beneath AI: The US-India Partnership powering the AI Era — Thank you. Thank you, Joel. Thank you, everybody, for being here this morning. Let me first start by putting the AI. Tha…
S86
Sovereign AI for India – Building Indigenous Capabilities for National and Global Impact — Brandon Mello from GenSpark identified adoption challenges, noting that 95% of AI pilots fail to reach production due to…
Speakers Analysis
Detailed breakdown of each speaker’s arguments and positions
A
Archana Jahargirdar
2 arguments69 words per minute569 words488 seconds
Argument 1
Emphasis on product‑only pitches, no business or funding talk
EXPLANATION
Archana instructed the founders to keep their presentations strictly about the product, avoiding any discussion of business models, funding, or revenue. This rule was set to ensure the summit focuses on technical product insights rather than commercial pitches.
EVIDENCE
She outlined the format, stating that each founder should talk only about their product and that there should be no business, pitching, or money discussion during the presentations [5-8].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The session moderation notes state that Archana instructed founders to keep presentations strictly about the product and avoid any business or funding discussion [S3].
MAJOR DISCUSSION POINT
Product‑only presentation rule
DISAGREED WITH
Ravindra Kumar, Pradyum Gupta, Vaibhavath Shukla
Argument 2
Guidance to presenters to balance technical jargon with accessibility
EXPLANATION
Archana asked presenters to use appropriate technical language but also to simplify explanations for audience members who may not be AI experts. She emphasized that both jargon‑heavy and simplified talks are acceptable as long as the audience can follow.
EVIDENCE
She requested presenters to use jargon if the audience can understand it, while also being mindful of non-AI natives and offered the option to simplify the language [2]; she explicitly said simplifying is fine [3] and not simplifying is also okay [4].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Archana asked presenters to use jargon only when the audience can follow, while also allowing simplification for non-AI natives, as highlighted in the discussion summary [S17] and the session guidelines [S3].
MAJOR DISCUSSION POINT
Balancing jargon and accessibility
R
Ravindra Kumar
5 arguments161 words per minute1033 words382 seconds
Argument 1
Goal to “automate automation” using agentic AI; three modules: conceptualize, deploy, troubleshoot
EXPLANATION
Ravindra presented Technodate AI’s vision of making automation as easy as DIY by leveraging agentic AI. The solution is structured into three modules that help users design, implement, and maintain automation systems.
EVIDENCE
He explained that Technodate aims to make automation DIY using agentic AI and described three modules: conceptualizing robotics solutions, deploying and commissioning them, and troubleshooting when issues arise [28-32].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Ravindra described Technodate’s agentic AI platform with three modules (conceptualize, deploy, troubleshoot) in the founders’ conversation [S3], and the broader context of agentic AI in industry is discussed in an external analysis of industrial diplomacy [S18].
MAJOR DISCUSSION POINT
Agentic AI for automation
Argument 2
Need for a foundational model despite funding challenges; iterative customer‑driven approach
EXPLANATION
Ravindra noted that building a foundational AI model is essential for their product, but raising funds for such a model in India is difficult. Consequently, they adopted an iterative approach, engaging customers early and experimenting before deciding on the need for a foundational model.
EVIDENCE
He described the initial thought of building a foundational model, the difficulty of raising money in India, and the decision to talk to customers and experiment, eventually realizing a foundational model was still required [33-40] and highlighted funding challenges [34-36].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
He noted difficulty raising funds for a foundational model in India and adopted an iterative, customer-driven approach, as recorded in the raw conversation transcript [S3].
MAJOR DISCUSSION POINT
Foundational model funding dilemma
DISAGREED WITH
Vaibhavath Shukla
Argument 3
Foundational models are optional; focus should be on solving specific customer problems at the application layer
EXPLANATION
Ravindra argued that building a foundational model is not always necessary; the priority should be delivering solutions that address concrete customer needs through application‑level development.
EVIDENCE
He stated that he is not fond of building foundational models and that the aim is to solve the customer’s problem, emphasizing the importance of the application layer over the model itself [55-60].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
He argued that building foundational models is not always necessary and emphasized application-layer solutions, aligning with perspectives on context-specific AI versus universal foundation models [S19].
MAJOR DISCUSSION POINT
Application‑layer focus over foundational models
Argument 4
Strategic partnerships with the Indian Air Force and Fortune 500 companies validate the platform’s impact
EXPLANATION
Ravindra highlighted collaborations with high‑profile customers, including a forthcoming deployment with the Indian Air Force and existing applications for Fortune 500 firms, demonstrating market traction and potential societal impact.
EVIDENCE
He mentioned that they are exploring a use case with the Indian Air Force and have already deployed applications with Fortune 500 companies, indicating strong validation of their technology [47] and [41].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
He mentioned an upcoming deployment with the Indian Air Force and existing Fortune 500 customers, confirming market validation in the discussion notes [S3].
MAJOR DISCUSSION POINT
Strategic partnerships and market validation
Argument 5
Agentic AI can automatically generate CNC programming for manufacturing and defense use cases
EXPLANATION
Ravindra explained that their platform extends automation beyond robotics by using generative AI to produce CNC programs required for aerospace, automotive, and defense components, thereby streamlining complex manufacturing workflows.
EVIDENCE
He described CNC programming as essential for aerospace and automotive parts and stated that such programs can be generated using agentic or generative AI, providing an example of an aero-engine error diagnosis powered by AI [53-55].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The platform’s capability to generate CNC programs for aerospace and defense using generative AI is described in the founders’ conversation [S3].
MAJOR DISCUSSION POINT
AI‑driven CNC programming for industrial automation
V
Vaibhavath Shukla
7 arguments163 words per minute1130 words414 seconds
Argument 1
Building a complete voice infrastructure that removes humans from the loop; partnerships with OpenAI and large enterprises
EXPLANATION
Vaibhavath described Quonsys AI’s end‑to‑end voice platform that automates call‑center operations without human intervention. The company collaborates with OpenAI and serves major enterprises such as Paytm, CRED, and PropBotX.
EVIDENCE
He introduced Quonsys AI as a voice infrastructure that eliminates the need for humans in the loop, mentioning partnerships with OpenAI and work with top enterprises like Paytm, CRED, and PropBotX [68-76].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Vaibhavath presented Quonsys AI as an end-to-end voice infrastructure with OpenAI partnership and enterprise customers such as Paytm and CRED, as detailed in the raw conversation summary [S3].
MAJOR DISCUSSION POINT
Voice AI for call‑center automation
Argument 2
Pricing model based on per‑minute usage; scalability plan with custom data engine and concurrency growth
EXPLANATION
The pricing strategy charges customers per minute of AI usage, with a subscription‑like model. Vaibhavath also highlighted plans to increase concurrency and scale the system as demand grows.
EVIDENCE
He explained a per-minute charging model and described scaling concurrency from 50 upwards, noting cost reductions of up to 70 % for large customers [99-107] and detailed pricing per minute [119-122].
MAJOR DISCUSSION POINT
Per‑minute pricing and scalability
DISAGREED WITH
Pradyum Gupta, Vivek Gupta
Argument 3
Creation of a proprietary data engine to generate training data; avoiding public datasets that cause reliability issues
EXPLANATION
After encountering problems with public datasets, Vaibhavath’s team built their own data engine to generate high‑quality training data for their voice models, ensuring better performance and reliability.
EVIDENCE
He described building a proprietary data engine after public datasets proved problematic, generating data internally and even receiving an award from the Prime Minister for this effort [109-115].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
He emphasized building a proprietary data engine to avoid unreliable public datasets, reflecting the need for domain-specific data highlighted in analyses of AI strategy and model development [S19].
MAJOR DISCUSSION POINT
Proprietary data engine development
Argument 4
Demonstrated cost savings of 70‑90 % for large BPO/enterprise customers
EXPLANATION
Vaibhavath claimed that Quonsys AI delivers substantial cost reductions for large enterprises, citing savings ranging from 70 % to 90 % compared with traditional call‑center operations.
EVIDENCE
He mentioned that the solution can save 70-90 % for big customers, providing an example of reduced per-minute cost and overall cost efficiency for enterprises like Paytm [119-122].
MAJOR DISCUSSION POINT
Significant cost savings
Argument 5
Building a proprietary data engine is essential for domain‑specific performance; foundational model development is secondary
EXPLANATION
Vaibhavath reiterated that a custom data engine is crucial for achieving high performance in their specific domain, reducing reliance on generic foundational models.
EVIDENCE
He emphasized that the proprietary data engine is central to their approach because public datasets caused reliability issues, and that this engine underpins their domain-specific performance [109-115].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The priority given to a custom data engine over generic foundational models mirrors arguments about context-specific AI solutions versus universal models [S19].
MAJOR DISCUSSION POINT
Domain‑specific data engine priority
DISAGREED WITH
Ravindra Kumar
Argument 6
Inquiry about routing calls to AI agents and dynamic use‑case handling
EXPLANATION
An audience member asked whether an AI agent could be placed on a phone to answer incoming calls according to specific requirements. Vaibhavath confirmed this capability and explained the technical handshake involved.
EVIDENCE
The audience asked if an agent could be embedded in a phone to answer calls, and Vaibhavath responded that it can be done via a backend handshake using web sockets, allowing dynamic question handling [94-95].
MAJOR DISCUSSION POINT
Dynamic call routing to AI agents
Argument 7
Question on scaling challenges of foundational models and concurrency limits
EXPLANATION
The audience raised concerns about scaling foundational models and handling high concurrency. Vaibhavath addressed these concerns by describing their data engine, concurrency handling, and plans to increase capacity.
EVIDENCE
The audience queried scaling and concurrency issues, and Vaibhavath answered by discussing the proprietary data engine, current concurrency of 50, and plans to scale up to handle thousands of requests, emphasizing cost-effective scaling [103-105] and his detailed response on scaling strategy [109-115].
MAJOR DISCUSSION POINT
Scaling and concurrency strategy
P
Pradyum Gupta
5 arguments190 words per minute2346 words739 seconds
Argument 1
Collecting massive visual data via dash‑cams/CCTVs to update maps instantly; use cases in billboard pricing, autonomous vehicle safety, bus fleet optimization
EXPLANATION
Pradyum explained that Papri Labs gathers visual data from dash‑cams and CCTVs across metro cities, processes petabytes of footage, and updates maps in real time. This data supports applications such as dynamic billboard pricing, safety for autonomous vehicles, and optimizing bus fleet deployment.
EVIDENCE
He described deploying cameras on vehicles to collect visual data, handling around 100 petabytes, and using it for use cases like billboard pricing based on impressions, autonomous vehicle safety, and bus fleet capacity optimization [121-144].
MAJOR DISCUSSION POINT
Real‑time visual mapping platform
Argument 2
Business model based on selling “tiles” of mapped area on a per‑day basis
EXPLANATION
Papri Labs monetizes its mapping service by selling geographic tiles (25 km × 25 km) to customers, charging a fixed fee per tile per day.
EVIDENCE
He stated that the company sells mapped areas in tiles of 25 km by 25 km, priced at 1.5 lakh rupees per tile per day [267-271].
MAJOR DISCUSSION POINT
Tile‑based pricing model
DISAGREED WITH
Vaibhavath Shukla, Vivek Gupta
Argument 3
Blurring faces and number plates; keeping raw video internal; using bare‑metal European servers, no hyperscalers
EXPLANATION
Pradyum outlined privacy safeguards: raw video footage is never released, personal identifiers such as faces and license plates are blurred, and all data is stored on bare‑metal servers in Europe, avoiding public cloud providers.
EVIDENCE
He explained that they never expose raw video, blur faces and number plates, and store data on bare-metal European servers (Hetzner), without using hyperscalers like AWS [208-214].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The privacy measures-blurring personal identifiers and storing raw video on bare-metal European servers-are consistent with data-governance practices discussed in recent policy literature on data protection and sovereignty [S23][S25].
MAJOR DISCUSSION POINT
Data privacy and compliance measures
DISAGREED WITH
Vivek Gupta
Argument 4
Concern about DPDP compliance and personal data handling in visual mapping
EXPLANATION
An audience member asked how Papri Labs complies with India’s DPDP regulations given the personal data they collect. Pradyum responded by detailing their anonymisation practices and secure storage architecture.
EVIDENCE
The audience raised DPDP compliance concerns regarding personal images and number plates, and Pradyum answered that they blur faces and plates, keep raw video internal, and host data on bare-metal European servers, ensuring compliance [208-214] (reiterated in [225-231]).
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Their compliance approach aligns with privacy regulation considerations outlined in AI privacy studies and data-governance frameworks, which stress anonymisation and secure storage for visual data [S25][S23].
MAJOR DISCUSSION POINT
DPDP compliance strategy
Argument 5
Difficulty in obtaining crowdsourced visual data because individuals are reluctant to install dash‑cams or share phone data
EXPLANATION
Pradyum noted that despite the need for large volumes of visual data, convincing individual passengers or vehicle owners to place dash‑cams or share phone data is a major obstacle, limiting the scalability of their data collection approach.
EVIDENCE
He remarked that asking a single passenger to put a phone in their car to provide data meets resistance, with none willing to comply, illustrating the challenge of crowdsourced data acquisition [162-164].
MAJOR DISCUSSION POINT
Challenges in crowdsourced data acquisition
M
Meenal Gupta
3 arguments154 words per minute1230 words478 seconds
Argument 1
AI‑assisted contouring reduces manual radiotherapy planning from 90‑960 min to 5‑15 min; HIPAA, ISO, and SEDESCO certifications ensure regulatory compliance
EXPLANATION
Meenal described how EasyOPI’s Imagix AI automates the contouring step in radiotherapy planning, cutting processing time dramatically. The solution is certified for medical use, meeting HIPAA, ISO 13485, and SEDESCO standards.
EVIDENCE
She explained that manual contouring takes 90-960 minutes, while their AI reduces it to 5-15 minutes, and highlighted certifications such as HIPAA, ISO 13485, and SEDESCO [332-335] and [290-294].
MAJOR DISCUSSION POINT
AI acceleration of radiotherapy planning
Argument 2
Trust built through human‑in‑the‑loop validation; AI provides assistance, not autonomous decisions
EXPLANATION
Meenal emphasized that their AI system assists clinicians but final decisions remain with radiologists, ensuring a human‑in‑the‑loop approach that builds trust in the technology.
EVIDENCE
She stated that the AI assists doctors, but final approval must be given by radiologists, maintaining a human-in-the-loop process [348-352].
MAJOR DISCUSSION POINT
Human‑in‑the‑loop trust model
Argument 3
Large‑scale deployment has yielded over one million scans and identified thousands of TB and lung‑cancer cases, demonstrating tangible health impact
EXPLANATION
Meenal highlighted the extensive reach of Imagix AI across multiple Indian states, processing more than a million chest X‑ray scans and flagging thousands of TB‑positive and lung‑cancer cases, thereby showing concrete public‑health benefits.
EVIDENCE
She reported that the platform has processed around one million scans, detected approximately 4,000 TB-positive cases and 2,700 TB-flagged cases, and performed hundreds of thousands of chest X-rays, underscoring its real-world impact [334-340].
MAJOR DISCUSSION POINT
Real‑world health impact of AI‑driven imaging
V
Vivek Gupta
4 arguments193 words per minute1679 words519 seconds
Argument 1
DIY, no‑code voice platform covering STT, TTS, LLM, speech‑to‑speech with sub‑400 ms latency; native dialect mastery
EXPLANATION
Vivek presented Indus Labs AI’s platform as a no‑code, DIY solution that provides end‑to‑end voice capabilities—including speech‑to‑text, text‑to‑speech, and large‑language‑model integration—with low latency and support for diverse Indian dialects.
EVIDENCE
He described the platform as a DIY, no-code solution that includes STT, TTS, LLM, speech-to-speech, achieves sub-400 ms latency, and handles native Indian dialects across regions [362-368] (also noted in [363]).
MAJOR DISCUSSION POINT
DIY multilingual voice platform
DISAGREED WITH
Pradyum Gupta, Vaibhavath Shukla
Argument 2
End‑to‑end integration with CRM, cost reduction up to 70 %, data residency on Indian servers
EXPLANATION
Vivek explained that the platform integrates directly with CRM systems, provides sentiment analysis, reduces operational costs by up to 70 %, and ensures that all data remains within India for sovereignty.
EVIDENCE
He detailed CRM integration, sentiment analysis, cost reductions of up to 70 %, and Indian data residency, noting latency and multilingual support [374-382].
MAJOR DISCUSSION POINT
CRM integration and cost efficiency
Argument 3
Request for details on how to construct voice‑agent flows on the platform
EXPLANATION
An audience member asked whether the platform provides pre‑built flows or requires users to manually connect nodes. Vivek clarified that the platform is DIY, offering tutorials and support for building custom flows.
EVIDENCE
The audience queried the flow-building process, and Vivek responded that the platform is DIY with tutorials, and that support is available if users encounter difficulties [441-445].
MAJOR DISCUSSION POINT
DIY flow construction guidance
Argument 4
Commitment to Indian data sovereignty by keeping all platform data on Indian servers
EXPLANATION
Vivek emphasized that the entire data pipeline resides within India, ensuring compliance with sovereign data requirements and enhancing trust for Indian enterprises and regulators.
EVIDENCE
He stated that the data will reside in India, describing the platform as a pure Indian company with data residency on Indian soil, reinforcing the focus on data sovereignty [390-393].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
Vivek’s emphasis on Indian-only data residency reflects broader policy discussions on data sovereignty and national regulatory compliance in AI deployments [S25][S23].
MAJOR DISCUSSION POINT
Data sovereignty and compliance
DISAGREED WITH
Pradyum Gupta
A
Audience
5 arguments161 words per minute492 words183 seconds
Argument 1
Concern about DPDP compliance for large‑scale visual data collection
EXPLANATION
The audience raised questions about how Papri Labs complies with India’s Data Protection and Data Privacy (DPDP) regulations given the massive amount of personal visual data they collect, such as images of people and vehicle number plates.
EVIDENCE
Audience members asked how the company handles DPDP compliance and personal data after hearing that they process petabytes of video and capture images of individuals and car numbers, prompting a clarification on anonymisation and storage practices [199-206].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The audience’s DPDP concerns echo the same data-protection guidelines and best-practice recommendations on handling personal visual data in large-scale projects [S25][S23].
MAJOR DISCUSSION POINT
Data privacy and regulatory compliance
Argument 2
Question about incentive mechanisms for dash‑cam data contributors
EXPLANATION
The audience inquired what incentives are offered to owners of dash‑cams or other devices that supply visual data, highlighting the need for a sustainable model to encourage data contribution.
EVIDENCE
An audience member asked about incentives for dash-cam holders, and Pradyum responded that they do not pay incentives; instead, the data providers pay the company for the service [244-247] and [246].
MAJOR DISCUSSION POINT
Incentive structures for data contributors
Argument 3
Skepticism about scaling foundational AI models after demo‑level success
EXPLANATION
Audience members expressed doubts that foundational models, which work well in demos, can be scaled reliably, citing the failure of the Servam model at larger scale and asking how the startup plans to address such scalability issues.
EVIDENCE
The audience referenced seeing Servam break when scaling and asked how the company would combat similar scenarios, leading to a discussion about their proprietary data engine and concurrency scaling plans [103-108].
EXTERNAL EVIDENCE (KNOWLEDGE BASE)
The skepticism mirrors expert commentary on the challenges of scaling foundation models from prototype to production, as discussed in analyses of AI model scalability and context-specific deployments [S19].
MAJOR DISCUSSION POINT
Scalability challenges of AI models
Argument 4
Inquiry about deploying AI agents directly on end‑user phones
EXPLANATION
The audience asked whether an AI agent could be installed on a phone to answer incoming calls according to specific business requirements, probing the feasibility of on‑device AI solutions.
EVIDENCE
Audience members asked if an agent could be placed into a phone to attend calls and answer according to requirements, and Vaibhavath confirmed it is possible via a backend handshake using web sockets [94-95].
MAJOR DISCUSSION POINT
On‑device AI agent deployment
Argument 5
Emphasis on trust as a critical factor for AI adoption in health technology
EXPLANATION
The audience highlighted that trust in the underlying technology and scientific methodology is essential for the acceptance of AI‑driven health solutions across India.
EVIDENCE
An audience participant asked how the company ensures that the technology and science behind its health AI product are trusted by beneficiaries [346-347].
MAJOR DISCUSSION POINT
Building trust in health AI solutions
Agreements
Agreement Points
All presenters adhered to the moderator’s rule to keep pitches strictly product‑focused, avoiding business, funding or revenue discussion.
Speakers: Archana Jahargirdar, Ravindra Kumar, Vaibhavath Shukla, Pradyum Gupta, Meenal Gupta, Vivek Gupta
Emphasis on product-only pitches, no business or funding talk (Archana) [5-8] Presentation of Technodate AI focused on product modules (Ravindra) [28-32] Presentation of Quonsys AI focused on voice infrastructure product (Vaibhavath) [68-76] Presentation of Papri Labs focused on visual-mapping product (Pradyum) [121-144] Presentation of EasyOPI Solutions focused on Imagix AI product (Meenal) [288-335] Presentation of Indus Labs AI focused on voice platform product (Vivek) [362-368]
Archana set a clear guideline that founders should talk only about their product and not about business or funding, and every founder respected this instruction, keeping their remarks centred on technical capabilities and use‑cases rather than commercial details.
Consensus that delivering domain‑specific applications and building proprietary data assets is more critical than investing in large foundational models.
Speakers: Ravindra Kumar, Vaibhavath Shukla, Pradyum Gupta, Vivek Gupta
Foundational models are optional; focus should be on solving specific customer problems at the application layer (Ravindra) [55-60] Creation of a proprietary data engine to generate training data; avoiding public datasets (Vaibhavath) [109-115] Collecting massive visual data via own dash-cams and processing it internally rather than relying on generic models (Pradyum) [162-165] Building a DIY platform with own infrastructure rather than depending on external foundational models (Vivek) [362-368]
All four speakers argued that the priority is to build specialised, application‑level solutions supported by in‑house data pipelines, and that the expense and difficulty of creating a generic foundational model can be bypassed.
POLICY CONTEXT (KNOWLEDGE BASE)
This view aligns with expert commentary urging focus on small, domain-specific niche models rather than large foundation models, emphasizing democratized AI development [S55].
Strong emphasis on data privacy, regulatory compliance and sovereignty across different domains.
Speakers: Pradyum Gupta, Meenal Gupta, Vivek Gupta, Vaibhavath Shukla
Blurring faces and number plates; storing raw video on bare-metal European servers; DPDP compliance (Pradyum) [208-214] HIPAA, ISO 13485 and SEDESCO certifications ensure medical data compliance (Meenal) [290-294] Indian data residency; all data kept on Indian servers for sovereignty (Vivek) [390-393] Proprietary data engine to control data quality and avoid reliance on public datasets (Vaibhavath) [109-115]
Each speaker highlighted concrete measures to protect personal data and meet national or sectoral regulations, signalling a shared commitment to privacy and data‑sovereignty.
POLICY CONTEXT (KNOWLEDGE BASE)
The emphasis mirrors multilateral data-governance frameworks that stress data sovereignty, compliance with national laws, and robust protection policies [S43][S44][S45].
Cost reduction is presented as a primary value proposition for AI‑enabled automation.
Speakers: Vaibhavath Shukla, Vivek Gupta, Ravindra Kumar
Demonstrated cost savings of 70-90 % for large BPO/enterprise customers (Vaibhavath) [119-122] End-to-end integration can reduce operational costs up to 70 % (Vivek) [380-382] Automation of industrial processes reduces labour and improves efficiency (Ravindra) [28-32]
All three speakers framed their solutions as ways to achieve substantial cost efficiencies for enterprises, whether in call‑center operations, voice platforms or industrial automation.
Promotion of DIY / no‑code platforms that enable non‑technical users to build AI‑driven solutions.
Speakers: Ravindra Kumar, Vaibhavath Shukla, Vivek Gupta
Make automation as easy as DIY using agentic AI (Ravindra) [28-32] Voice infrastructure is a DIY platform for building voice agents (Vaibhavath) [68-76] DIY, no-code voice platform with low latency and dialect mastery (Vivek) [362-368]
Each of these founders positioned their product as a self‑service, low‑code environment that lowers the barrier for organisations to create AI solutions without deep technical expertise.
Similar Viewpoints
Both argue that the core competitive advantage lies in owning the data pipeline and tailoring models to the specific problem domain, reducing dependence on large, generic foundation models.
Speakers: Ravindra Kumar, Vaibhavath Shukla
Need for proprietary data / domain-specific models rather than generic foundational models (Ravindra) [55-60] Proprietary data engine to generate high-quality training data (Vaibhavath) [109-115]
Both stress the importance of controlling data collection and processing to ensure quality, security and regulatory compliance.
Speakers: Pradyum Gupta, Vaibhavath Shukla
Privacy safeguards (blurring, internal storage) for large visual datasets (Pradyum) [208-214] Proprietary data engine to avoid unreliable public datasets (Vaibhavath) [109-115]
Both see trust‑building measures—whether through human oversight or data sovereignty—as essential for adoption of AI solutions in sensitive sectors.
Speakers: Meenal Gupta, Vivek Gupta
Human-in-the-loop validation to build trust in health AI (Meenal) [348-352] Data residency and sovereign hosting to foster trust in voice AI (Vivek) [390-393]
Unexpected Consensus
Both a health‑imaging startup (EasyOPI) and a voice‑AI startup (Indus Labs) highlighted the need for human‑in‑the‑loop or sovereign data handling as a trust mechanism, despite operating in very different domains.
Speakers: Meenal Gupta, Vivek Gupta
Human-in-the-loop validation ensures trust in radiotherapy planning (Meenal) [348-352] Indian data residency guarantees sovereign control and trust (Vivek) [390-393]
While one focuses on clinical validation and the other on national data residency, both converge on the principle that trust is achieved by keeping a human or jurisdictional safeguard over AI decisions.
POLICY CONTEXT (KNOWLEDGE BASE)
Their trust-by-design approach reflects governance recommendations for human oversight and sovereign data stewardship to build user confidence [S43][S44].
Agreement between a visual‑mapping company (Papri Labs) and a voice‑AI company (Quonsys AI) on the necessity of building a proprietary data engine to overcome limitations of public datasets, even though their products serve unrelated markets.
Speakers: Pradyum Gupta, Vaibhavath Shukla
Difficulty of crowdsourced visual data and reliance on own data collection (Pradyum) [162-165] Creation of a proprietary data engine after public datasets proved unreliable (Vaibhavath) [109-115]
Both founders independently arrived at the conclusion that owning the data generation pipeline is essential for performance, showing a cross‑domain convergence on data strategy.
POLICY CONTEXT (KNOWLEDGE BASE)
Papri Labs demonstrated a proprietary visual data engine to address public-dataset gaps, illustrating the broader industry trend toward private data assets for domain solutions [S38][S55].
Overall Assessment

The discussion revealed a clear convergence around product‑centric, application‑layer AI solutions that prioritize proprietary data, privacy compliance, cost efficiency and user‑friendly DIY interfaces. Speakers from diverse sectors (industrial automation, voice call‑center automation, visual mapping, health imaging) repeatedly stressed the same strategic pillars: avoid heavyweight foundational models, protect data, demonstrate tangible cost savings, and empower users through low‑code platforms.

High consensus on strategic approach (application focus, data ownership, privacy, cost reduction, DIY enablement). This alignment suggests that future AI deployments in the Indian context are likely to follow a model of domain‑specific, privacy‑by‑design products that are accessible to non‑technical users and deliver clear economic benefits.

Differences
Different Viewpoints
Whether a foundational AI model is required for the product
Speakers: Ravindra Kumar, Vaibhavath Shukla
Need for a foundational model despite funding challenges; iterative customer‑driven approach Building a proprietary data engine is essential for domain‑specific performance; foundational model development is secondary
Ravindra states that after experimenting with customers they realized they still need a foundational model for their automation platform [33-40], while Vaibhavath argues that a custom data engine is sufficient and a foundational model is not essential for delivering performance [109-115].
POLICY CONTEXT (KNOWLEDGE BASE)
The debate echoes ongoing discussions in AI policy circles about preferring niche, domain-specific models over large, generic foundation models [S55].
Approach to data residency and storage for large‑scale visual data
Speakers: Pradyum Gupta, Vivek Gupta
Blurring faces and number plates; keeping raw video internal; using bare‑metal European servers, no hyperscalers Commitment to Indian data sovereignty by keeping all platform data on Indian servers
Pradyum explains that all raw video is stored on bare-metal servers in Europe and never exposed publicly, emphasizing privacy and avoiding hyperscalers [208-214]. Vivek, in contrast, stresses that all data for his voice platform resides within India to satisfy data-sovereignty requirements [390-393].
POLICY CONTEXT (KNOWLEDGE BASE)
Data-residency considerations are guided by principles that data remains subject to the laws of its country of origin and must respect sovereignty, as highlighted in recent data-governance policy statements [S43][S41].
Pricing and monetisation models for AI‑driven services
Speakers: Pradyum Gupta, Vaibhavath Shukla, Vivek Gupta
Business model based on selling “tiles” of mapped area on a per‑day basis Pricing model based on per‑minute usage; scalability plan with custom data engine and concurrency growth DIY, no‑code voice platform covering STT, TTS, LLM, speech‑to‑speech with sub‑400 ms latency; native dialect mastery
Pradyum proposes a tile-based fee of 1.5 lakh rupees per 25 km × 25 km tile per day [267-271]. Vaibhavath charges customers per minute of AI usage and plans to increase concurrency as demand grows [99-107]. Vivek also uses a per-minute pricing model but at a lower rate (≈2 rupees per minute) and highlights cost reductions up to 70 % [380-382]. The three founders therefore disagree on the optimal monetisation strategy.
POLICY CONTEXT (KNOWLEDGE BASE)
Challenges of AI service monetisation for low-resource users and calls for new, sustainable models have been documented in agritech AI deployments and AI-storytelling summit recommendations [S40][S42].
Adherence to the summit’s “product‑only” presentation rule
Speakers: Archana Jahargirdar, Ravindra Kumar, Pradyum Gupta, Vaibhavath Shukla
Emphasis on product‑only pitches, no business or funding talk Ravindra acknowledges that “being a founder, some pitching comes in by default” Business model based on selling tiles; discussion of revenue and partnerships Building a complete voice infrastructure … includes partnership with OpenAI and pricing details
Archana explicitly instructs presenters to avoid any business, funding or revenue discussion and focus solely on the product [5-8]. Ravindra, however, admits that pitching elements slipped into his talk [43]. Pradyum and Vaibhavath both describe commercial aspects such as pricing, partnerships and revenue models during their presentations [267-271] and [74-76], respectively, creating tension with the moderator’s rule.
Unexpected Differences
Founders’ inclusion of commercial details despite a moderator‑enforced product‑only format
Speakers: Archana Jahargirdar, Ravindra Kumar, Pradyum Gupta, Vaibhavath Shukla
Emphasis on product‑only pitches, no business or funding talk Ravindra acknowledges that “being a founder, some pitching comes in by default” Business model based on selling tiles; discussion of revenue and partnerships Building a complete voice infrastructure … includes partnership with OpenAI and pricing details
The moderator’s clear instruction to keep presentations strictly technical [5-8] was unexpectedly breached by multiple founders who introduced business-related content (pricing, partnerships, revenue models). This tension was not anticipated given the session’s stated purpose.
Overall Assessment

The discussion revealed several substantive disagreements: (1) the necessity of a foundational AI model versus reliance on bespoke data engines; (2) contrasting data‑sovereignty strategies (European bare‑metal vs Indian‑hosted servers); (3) divergent monetisation approaches (tile‑based, per‑minute, or low‑cost per‑minute pricing); and (4) tension between the moderator’s product‑only rule and founders’ inclination to discuss commercial aspects. While all participants agree on the broader aim of AI‑driven automation, they differ markedly on technical architecture, data governance, and business models.

Moderate to high – the disagreements span technical design choices, regulatory compliance strategies, and presentation norms, indicating that consensus on implementation pathways is limited. These divergences could affect collaboration, standard‑setting, and policy formulation within the AI‑driven automation ecosystem.

Partial Agreements
All speakers share the overarching goal of leveraging AI to automate complex, domain‑specific processes (industrial automation, call‑center operations, real‑time visual mapping, voice infrastructure, radiotherapy planning). However, they diverge on the technical route to achieve this—Ravindra emphasises a foundational model, Vaibhavath relies on a custom data engine, Pradyum uses crowdsourced visual data, Vivek builds a DIY multilingual voice stack, and Meenal focuses on AI‑assisted medical imaging with strict regulatory compliance.
Speakers: Ravindra Kumar, Vaibhavath Shukla, Pradyum Gupta, Vivek Gupta, Meenal Gupta
Goal to automate automation / voice / mapping / radiotherapy planning using AI Building a proprietary data engine is essential for domain‑specific performance; foundational model development is secondary Collecting massive visual data via dash‑cams/CCTVs to update maps instantly; use cases in billboard pricing, autonomous vehicle safety, bus fleet optimisation DIY, no‑code voice platform covering STT, TTS, LLM, speech‑to‑speech with sub‑400 ms latency; native dialect mastery AI‑assisted contouring reduces manual radiotherapy planning from 90‑960 min to 5‑15 min; HIPAA, ISO, SEDESCO certifications ensure regulatory compliance
Takeaways
Key takeaways
The summit adopted a strict product‑only presentation format: founders were asked to discuss only their technology, avoiding business, funding, or sales pitches. Presenters were encouraged to balance technical jargon with accessibility so non‑AI audiences could understand. Technodate AI (Ravindra Kumar) aims to ‘automate automation’ using agentic AI with three modules – conceptualization, deployment, and troubleshooting – and highlighted the need for a domain‑specific foundational model despite funding constraints. Quonsys AI (Vaibhavath Shukla) is building an end‑to‑end voice AI platform to fully automate call‑center operations, leveraging a proprietary data engine, per‑minute pricing, and partnerships with OpenAI and large enterprises. Papri Labs (Pradyum Gupta) offers a real‑time visual mapping and video‑analytics platform that aggregates dash‑cam/CCTV footage to keep maps current and provides B2B services such as dynamic billboard pricing and fleet optimization, sold on a per‑tile, per‑day basis. Papri Labs addressed data‑privacy concerns by blurring personally identifiable information, keeping raw video internal, and hosting on bare‑metal European servers rather than hyperscalers. EasyOPI Solutions (Meenal Gupta) delivers AI‑assisted cancer imaging and radiotherapy treatment planning, reducing contouring time from up to 960 minutes to 5‑15 minutes, and emphasizes regulatory compliance (HIPAA, ISO 13485, SEDESCO) and a human‑in‑the‑loop model for trust. Indus Labs AI (Vivek Gupta) provides a DIY, no‑code voice‑architecture platform covering STT, TTS, LLM and speech‑to‑speech with sub‑400 ms latency, native Indian dialect support, end‑to‑end CRM integration, and a cost model up to 70 % cheaper than global alternatives. Founders across the board stressed that solving concrete customer problems at the application layer is more critical than building generic foundational models.
Resolutions and action items
Founders were invited to continue one‑on‑one conversations after the session (implicit action to follow up with interested parties). Papri Labs clarified its pricing model (tile‑based, per‑day) in response to audience queries. Quonsys AI explained its per‑minute usage pricing and plans to increase concurrency as the model scales.
Unresolved issues
How Quonsys AI will reliably scale its foundational model to handle higher concurrency without the failures observed in other systems (e.g., Servam) – no concrete scaling plan was detailed. Detailed DPDP compliance mechanisms for Papri Labs beyond blurring faces and number plates and using European bare‑metal servers remain unclear. Incentive mechanisms for dash‑cam owners or bus operators supplying visual data were not fully explained; the claim that they pay the platform leaves the motivation question open. Specific steps for non‑technical users to construct voice‑agent flows on Indus Labs’ platform were only broadly described; a concrete UI/UX walkthrough was not provided. Long‑term sustainability of Technodate AI’s foundational model development given funding challenges was not resolved.
Suggested compromises
Archana asked presenters to use jargon if needed but also to simplify for non‑AI audiences – a compromise between technical depth and accessibility. Ravindra Kumar offered presenters the choice to simplify or retain technical language, respecting both preferences. Papri Labs chose to keep raw video data internal and only sell processed, anonymized outputs, balancing data utility with privacy compliance. Quonsys AI adopted a per‑minute pricing model rather than a fixed subscription, allowing customers to pay only for actual usage.
Thought Provoking Comments
The format we’ll follow is that each one of you takes a little bit of time to talk about your product. But like I said again, only product. No business, no pitching, no money, nothing.
Sets a clear, disciplined scope for the session, emphasizing knowledge sharing over fundraising, which frames the entire discussion and encourages technical depth.
Established the tone of the meeting, prompting founders to focus on product details. This led to more technical explanations (e.g., foundational models, data engines) rather than sales pitches, shaping the subsequent flow of the conversation.
Speaker: Archana Jahargirdar
Model can become ASI, the super intelligence level. You still will have to build the application.
Highlights the distinction between raw AI capability and practical productisation, reminding the audience that even the most advanced models need domain‑specific application layers to deliver value.
Shifted the discussion from abstract AI hype to concrete engineering challenges. It prompted follow‑up questions about foundational models, data ownership, and on‑premise deployment, deepening the technical debate.
Speaker: Ravindra Kumar
India doesn’t need more wrappers we need infrastructure and that’s what we are building at Quonsys AI… we can automate the entire call‑center and run it end‑to‑end without humans in the loop.
Frames the problem as a missing foundational layer rather than incremental features, positioning voice AI as essential national infrastructure.
Redirected the audience’s focus to large‑scale, systemic challenges (data generation, latency, cost). Sparked a series of questions about deployment, scaling, and pricing models, moving the conversation toward practical implementation.
Speaker: Vaibhavath Shukla
We built our own data engine because public datasets were insufficient; we generate data at scale and fine‑tune on it.
Identifies a core bottleneck—data scarcity—and presents a self‑sufficient solution, illustrating a strategic approach to AI development in a resource‑constrained environment.
Introduced the theme of data sovereignty and scalability, leading to deeper inquiries about model performance, concurrency limits, and cost structures.
Speaker: Vaibhavath Shukla
We never take out front‑camera video for public use; faces and number plates are blurred. We run on bare‑metal servers in Europe, not on hyperscalers.
Addresses privacy and regulatory compliance (DPDP) head‑on, showing a concrete governance framework for handling massive visual data.
Turned the discussion toward legal and ethical considerations, prompting the audience to probe further on incentives for data contributors and the business model, thereby expanding the scope beyond pure technology.
Speaker: Pradyum Gupta
We are not replacing doctors. We are just assisting them; final approval has to be done by radiologists. It’s a human‑in‑the‑loop system.
Acknowledges trust issues in health‑tech AI and offers a pragmatic mitigation strategy, reinforcing credibility and ethical responsibility.
Reassured the audience about safety and trust, leading to a concise Q&A about adoption barriers and reinforcing the product’s positioning as a supportive tool rather than a black‑box replacement.
Speaker: Meenal Gupta
We are building the voice operating system of India – low latency (sub‑500 ms), Indian dialect mastery, emotional handling, and sovereign data residency.
Combines technical performance metrics with cultural relevance and data sovereignty, presenting a comprehensive value proposition that differentiates from global players.
Created a pivot point where the conversation moved to comparative analysis with global solutions, cost advantages, and the importance of localized AI, prompting questions about no‑code flow and integration.
Speaker: Vivek Gupta
Overall Assessment

The discussion was shaped by a handful of strategic comments that repeatedly redirected the conversation from generic product pitches to deeper, systemic issues—such as the necessity of application layers over raw AI models, data sovereignty, regulatory compliance, and trust in high‑stakes domains like health. Archana’s opening rule set the disciplined, product‑centric tone, while each founder’s standout remark introduced a new dimension (foundational models, infrastructure gaps, data engine creation, privacy safeguards, human‑in‑the‑loop design, and localized voice OS). These insights triggered focused Q&A rounds, broadened the scope to include legal, ethical, and scalability concerns, and ultimately elevated the dialogue from superficial descriptions to a nuanced exploration of how AI products can be responsibly and effectively deployed in India.

Follow-up Questions
Can the AI agent be deployed directly onto a phone number to answer inbound calls according to specific requirements?
Clarifies technical feasibility of integrating the AI call‑center solution with existing telephony infrastructure, crucial for practical adoption.
Speaker: Audience (unidentified participant)
What is the pricing and subscription model for the AI call‑center solution (per‑minute vs subscription)?
Understanding the cost structure is essential for scaling the product and for potential customers to evaluate ROI.
Speaker: Audience (unidentified participant)
How will the voice AI platform scale reliably, especially given observed failures like Servam when scaling foundational models?
Addresses concerns about robustness and performance of large‑scale AI models, a key factor for enterprise deployment.
Speaker: Audience (unidentified participant)
How does Papri Labs ensure compliance with DPDP (data privacy) when handling personal data such as faces and vehicle number plates in its mapping solution?
Legal compliance and privacy protection are critical for operating in regulated markets and maintaining user trust.
Speaker: Audience (unidentified participant)
What incentives are offered to dash‑cam or vehicle owners to contribute data for the mapping platform?
Sustainable data collection depends on effective incentive mechanisms; understanding this helps assess scalability of data acquisition.
Speaker: Audience (unidentified participant)
How does EasyOPI ensure trust and validation of its AI‑driven cancer treatment planning among clinicians and patients?
Trust is a major barrier in health‑tech adoption; mechanisms for validation and clinician oversight are vital for acceptance.
Speaker: Audience (unidentified participant)
Is the voice‑AI platform a fully no‑code solution where users can simply click to start an agent, or must they manually connect nodes and build flows?
Usability determines adoption speed for non‑technical users; clarity on the level of required configuration is needed.
Speaker: Audience (unidentified participant)
What motivated the founders to leave their previous jobs and start their AI ventures, and what challenges did they face early on?
Founder stories provide insight into entrepreneurial pathways and potential hurdles for future founders.
Speaker: Audience (unidentified participant)
Is building a proprietary foundational model necessary for industrial automation, or can existing models suffice?
Determines the strategic direction and resource allocation for developing AI solutions in manufacturing.
Speaker: Ravindra Kumar
Can agentic AI be effectively used for CNC programming and automated error diagnosis in aerospace/defense equipment?
Explores a high‑impact application area where AI could streamline complex engineering processes.
Speaker: Ravindra Kumar
What methods does Quonsys AI use to generate large‑scale synthetic training data via its data engine, and how does this affect model performance?
Understanding data generation pipelines is key for replicating success and improving model robustness.
Speaker: Vaibhavath Shukla
What challenges arise when deploying AI‑driven medical imaging solutions in low‑connectivity regions, and how can on‑premise deployments address them?
Highlights infrastructure constraints in remote areas, informing strategies for broader healthcare AI rollout.
Speaker: Meenal Gupta
What are the advantages and trade‑offs of using bare‑metal servers versus hyperscalers for security, cost, and compliance in AI deployments?
Infrastructure choices impact data sovereignty, latency, and operational expenses, influencing deployment decisions.
Speaker: Pradyum Gupta
How does Indus Labs achieve low latency and accurate multi‑dialect support for Indian languages in its voice AI platform?
Technical solutions for dialect diversity and latency are critical for user experience in a linguistically varied market.
Speaker: Vivek Gupta
How does emotional detection (affect recognition) in voice AI improve customer interactions, and what metrics are used to evaluate its effectiveness?
Affective computing can enhance satisfaction; measuring its impact guides product refinement.
Speaker: Vivek Gupta
What is the pricing strategy for Papri Labs’ geospatial data (tile‑based pricing), and how does it scale with larger geographic coverage?
Understanding pricing models for spatial data informs business sustainability and market penetration.
Speaker: Pradyum Gupta

Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.