Waves of infrastructure Open Systems Open Source Open Cloud

20 Feb 2026 18:00h - 19:00h

Waves of infrastructure Open Systems Open Source Open Cloud

Session at a glance

Summary

This discussion centered on Proximal Cloud’s launch in India and the future of AI infrastructure, featuring presentations from Renu Raman and several industry partners. Raman, drawing from his experience at Sun Microsystems, outlined how computing transitions occur every 15-30 years, arguing that we are currently experiencing a shift from CPU-only systems to heterogeneous AI computing similar to the distributed systems revolution of the 1990s-2000s. He emphasized that AI will impact 95% of work compared to the surface-level productivity improvements of the SaaS era, creating massive demand for computing infrastructure.


Proximal Cloud’s strategy focuses on bringing compute closer to data through partnerships with AMD for CPU-GPU hybrid systems and collaboration with UC San Diego for research in health sciences and education. The company aims to address India’s need for extremely low-cost, population-scale computing infrastructure. Several partners presented their integration with Proximal’s platform, including PharmEx’s agricultural AI solutions using sensors and autonomous tractors, Divium’s model optimization platform that reduces AI costs by 30-60%, and Instant System’s venture building capabilities for AI startups.


The panel discussion explored whether India could produce major technology corporations like NVIDIA or SAP through this AI transition. Participants noted that India’s planned 10-gigawatt AI infrastructure buildout could generate $250 billion in hardware demand, potentially supporting an entire ecosystem of systems companies. However, they acknowledged the funding gap between Indian startups (receiving crores) versus global AI companies (investing billions per engineer). The consensus was that India’s opportunity lies in combining domain expertise with advanced technology to achieve higher gross margins than traditional service models, requiring sustained long-term investment similar to ISRO’s development approach.


Keypoints

Major Discussion Points:

Technology Infrastructure Evolution and AI Computing Demands: The discussion centers on the massive shift from traditional CPU-based computing to AI-driven heterogeneous systems, with speakers emphasizing the need for new distributed computing architectures to handle the exponential growth in AI workloads. They highlight the transition from training-focused to inference-focused computing systems.


India’s AI and Semiconductor Opportunity: A significant focus on India’s potential to become a major player in the AI infrastructure space, with discussions about the country’s plan to scale from less than 1 gigawatt to 10 gigawatts of AI computing capacity, representing a $250 billion hardware opportunity and the potential for “population-scale computing.”


Proximal Cloud’s Business Model and Partnerships: Renu Raman presents Proximal Cloud’s approach to bringing compute closer to data through sovereign, on-premises AI infrastructure solutions, showcasing partnerships with UC San Diego, AMD, and various Indian companies across agriculture, education, and enterprise sectors.


Practical AI Implementation Challenges: Multiple speakers address real-world barriers to AI adoption, including the “90% of Gen-AI pilots never make it to production” problem, issues with model selection, cost optimization, data security, hallucinations, and the need for reliable inference systems that can operate at sub-second response times.


Investment and Scaling Challenges in India: The panel discusses the significant funding gap between Indian startups (receiving crores) versus global tech companies (investing hundreds of millions per engineer), emphasizing the need for sustained, long-term investment similar to India’s ISRO model to build competitive AI and semiconductor companies.


Overall Purpose:

The discussion serves as a launch event for Proximal Cloud’s AI infrastructure offerings in India, combining company introduction with broader industry analysis. The goal is to position the company within the context of major technology shifts while demonstrating practical applications through partner showcases and addressing the strategic opportunity for India to develop sovereign AI computing capabilities.


Overall Tone:

The tone is predominantly optimistic and forward-looking, with speakers expressing excitement about India’s potential in the AI space. The discussion maintains a professional, technical atmosphere throughout, with Renu Raman setting an educational tone through historical technology parallels. While acknowledging significant challenges (funding gaps, technical hurdles, market realities), the overall sentiment remains bullish on India’s prospects. The tone becomes more interactive and collaborative during the Q&A segments, with industry participants sharing practical insights and reinforcing themes of opportunity and innovation potential.


Speakers

Speakers from the provided list:


Renu Raman – Main presenter, hardware/software expert with background at Sun Microsystems, founder/leader at Proximal Cloud, focuses on distributed systems and AI infrastructure


Jensen Huang – CEO of NVIDIA (quoted in video/audio clip about data processing and accelerated computing)


Michael Dell – CEO of Dell Technologies (quoted in video/audio clip about AI factories and enterprise customers)


Lalit Bhatt – Director heading India office for PharmEx, works in agricultural AI technology with sensors, imaging, and autonomous systems


Sandeep Kumar – From Instant System, Silicon Valley-based venture builder, works on AI conversation software and financial domain solutions


Abhishek Singh – Founder of ZetaVault, specializes in LLM acceleration and custom silicon for AI inferencing


Audience – Multiple audience members asking questions, including one identified as Arya Bhattacharjee from Infosys (Senior VP driving semiconductor and AI vision)


Additional speakers:


Bharat Jain – Director at Divium (mentioned in introduction but actual presentation was given by someone else discussing Divium’s inference optimization platform)


Full session report

This comprehensive discussion centered on Proximal Cloud’s strategic launch in India and the broader transformation of AI infrastructure, featuring detailed presentations from industry leaders and extensive analysis of India’s potential role in the global AI economy. The event served as both a company introduction and a forward-looking examination of how distributed computing architectures must evolve to meet the unprecedented demands of artificial intelligence at population scale.


Technology Evolution and Infrastructure Context

Renu Raman opened the discussion by positioning current AI developments within historical technology cycles, drawing from his extensive experience at Sun Microsystems during the semiconductor boom. He argued that major technology shifts occur predictably every 15-30 years, with semiconductors driving innovation in the 1980s-90s through Moore’s Law, followed by the cloud phenomenon of the last two decades. The current AI revolution represents a convergence of bottom-up innovation from companies like NVIDIA at the silicon level and top-down innovation from language models and higher-order AI functions.


Raman emphasized the philosophical foundation underlying Proximal’s approach, referencing the principle that “people who are serious about software should make their own hardware” and its corollary: “people who are serious about AI should make their own cloud.” This philosophy drives the company’s focus on sovereign, distributed computing solutions that bring AI capabilities closer to where data resides.


The discussion highlighted a critical transition occurring in AI workloads, moving from training-focused systems requiring massive scale-up architectures to inference-focused computing that can be more distributed. Raman noted that while the SaaS era delivered productivity improvements, AI will impact 95% of work, creating economic effects far beyond previous technology waves. This shift has driven infrastructure spending from $50 billion to $300 billion, with projections reaching even higher levels as AI deployment scales globally.


Jensen Huang’s insight that “data processing is a CPU job” was specifically highlighted by Raman as crucial to understanding why hybrid architectures combining CPUs and GPUs will be essential for practical AI deployment, particularly when dealing with enterprise data that requires complex processing before AI analysis.


Proximal Cloud’s Integrated Platform Approach

Proximal Cloud’s strategy emerged as a comprehensive response to identified market needs, focusing on sovereign, on-premises AI infrastructure solutions. The company’s partnership with AMD provides access to both x86 CPU capabilities and GPU roadmaps that enable hybrid architectures supporting traditional data processing alongside AI workloads. Raman noted that AMD’s higher memory capacity GPUs can support substantial workloads for most customers, enabling more distributed systems approaches.


The collaboration with UC San Diego adds research credibility and practical application development across education, research, and industry contexts. This partnership enables work spanning hardware-level optimization, compute kernel development, and real-world application validation. Renu demonstrated an education use case during the presentation, though technical issues prevented a complete showing.


The platform addresses the critical challenge that Michael Dell identified in his video message: with more than 90% of enterprise data remaining on-premises and continuing to be generated locally, the traditional approach of moving data to cloud-based AI becomes impractical. This reality necessitates bringing AI capabilities to where data resides.


Partner Ecosystem and Real-World Applications

The partner presentations demonstrated practical applications across diverse sectors. Lalit Bhatt from PharmEx presented their agricultural solutions, focusing on sensor networks for precision irrigation and farming applications. Their work addresses cost pressures in agriculture where efficiency improvements must maintain farmer affordability while delivering measurable value.


Bharat from Divium addressed a critical deployment challenge: 90% of generative AI pilots never reach production, not due to poor demonstrations or weak models, but because of undefined quality standards, unpredictable costs, and constantly changing model selection criteria. Divium’s platform demonstrates 30-60% cost reductions through intelligent routing and quality-based model selection, directly addressing these production deployment challenges.


Sandeep Kumar from Instant System clarified their role as a venture builder rather than an incubator, focusing on creating companies that can reach mid-market and enterprise customers. He described their work on financial AI systems requiring 99% reliability without hallucinations, with advanced architectures managing data privacy and access control. These requirements demonstrate that production AI systems must meet enterprise-grade reliability standards far exceeding typical pilot project expectations.


India’s Strategic AI Infrastructure Opportunity

The panel discussion positioned India’s AI infrastructure development as both a unique challenge and unprecedented opportunity. Arya Bhattacharjee, driving semiconductor and AI vision for Infosys, argued that India’s success lies not in competing directly with premium players like Palantir or accepting traditional service margins, but in achieving higher value through domain expertise combined with advanced technology.


India’s requirement for extremely low-cost computing at population scale—serving 1.4 billion people—presents engineering challenges that could drive breakthrough innovations with global applicability. Abhishek Singh from ZetaVault, working on LLM acceleration, posed the specific challenge of serving 1.5 billion people at ₹200 per month, highlighting the extreme cost constraints that could drive innovation.


Arya provided crucial context from semiconductor manufacturing, noting that modern fabs represent $10 billion facilities where every day saved represents $10 million in value. With fabs generating massive amounts of data requiring real-time AI processing to optimize yields, the intersection of AI and semiconductor manufacturing presents specific use cases where India’s software expertise can deliver immediate, measurable value.


The discussion emphasized India’s opportunity to follow a “make in India for global” model rather than China’s more closed approach, potentially enabling Indian companies to compete globally on innovation rather than just cost.


Technical Performance and Economic Challenges

Renu proposed ambitious performance targets, suggesting a 120-millisecond response time standard for any query, drawing parallels to Google’s historical 20-millisecond standard that drove massive infrastructure investment and innovation. Achieving this target at population scale would require breakthrough advances in both computing resources and algorithmic efficiency.


The economic scale of India’s AI ambitions became clear through specific projections. India’s planned expansion to 10 gigawatts of AI computing capacity represents a massive hardware opportunity that could support multiple system companies and sustain an entire semiconductor ecosystem, potentially enabling India to move beyond traditional outsourcing models to capture higher-value segments of the AI value chain.


However, significant structural challenges exist in building world-class AI and semiconductor companies from India. The contrast between available startup funding and the hundreds of millions to billions required for competitive AI infrastructure companies highlights the scale mismatch between available capital and global competition requirements.


Investment and Ecosystem Development

The funding discussion revealed both challenges and opportunities. While traditional venture capital may be insufficient for the scale required, India’s public markets value technology companies favorably for certain types of infrastructure businesses. The discussion emphasized that going public should be viewed not as an exit strategy but as a mechanism for raising capital to scale businesses over the long investment horizons required for infrastructure companies.


The ISRO model emerged as a template for sustained technology development, demonstrating how continuous support over decades can build world-class capabilities despite initial failures. This suggests that similar sustained investment in AI and semiconductor technologies could yield comparable results.


Future Implications and Strategic Questions

Several critical questions emerged that will shape India’s AI infrastructure development trajectory. The technical challenge of achieving sub-second response times at population scale while maintaining extreme cost efficiency remains unsolved, requiring breakthrough innovations in computing architecture and algorithmic efficiency.


The economic development question of how India can capture projected hardware opportunities locally, rather than simply importing solutions, requires coordinated policy and investment approaches. Success will likely require new funding models that combine government support, private investment, and public market access in ways that sustain long-term technology development cycles.


The discussion concluded with recognition that India’s AI infrastructure opportunity extends beyond domestic markets. Success in solving population-scale computing challenges at extreme cost points could establish technological leadership with global applications, potentially enabling Indian companies to compete not just on cost but on innovation and capability. This transformation from service provider to technology leader represents both the opportunity and the challenge facing India’s AI ecosystem development.


Session transcript

Renu Raman

Announcements and a lot of activities going on here this week. Excited about it. We are excited about introducing what we do and what we do more in the context of India. We just launched our offering and we’ll be talking more of what we do with our partners in the coming weeks and months. But today, I’d like to introduce ourselves. But before we introduce, we want to set the context of where we fit in, both in the industry trends and the ecosystem and what category we go after from an enterprise private cloud infrastructure. And then we’ll get into sharing some of our partners that we work with and then a Q &A at the end of it with a presentation from Bharat Jain and from Zeta Bolt.

We’ll have an interactive Q &A on some key top three questions or end questions that we think need to be answered. With that, let me start with the first. I want to… thank our sponsors and our collaborators and partnerships at UC San Diego, where they have an initiative for public -private partnership at UC San Diego for AI for education, AI for research, and AI for industry. And we are one of the early industry partners. There’s a newly constituted data science and data center institute called School of Computing, Information Sciences, and Digital Sciences. And we’ll talk a little bit more about it downstream. But this collaboration enables us to not only work on technologies, but also look at key use cases, particularly on health sciences, because San Diego has got one of the largest health science, both hospital system as well as clinical research and variety of health and biotech research.

With the thesis that fundamentally computing is going to be driven by biology and health, it’s a very key partnerships that we hope to work with. going forward. With that, let me step back. This is my standard slide I use in any presentations in terms of long -term reminders, what happens in technology. So where we fit in, we’ll just walk through for the next 20, 30 minutes about what we are doing from a systems innovation, but the systems innovation is going to be punctuated or represented in the context of where the technology shifts that have occurred and will occur as we go forward. So simple reminders are we, as humanity, underestimate. We overestimate what can be done in two years, but we underestimate what can be done in 10 years.

You can go back in history, look at self -driving cars, look at neural link. I remember a slide I had put at UC Berkeley, a conference about programming languages and productivity languages and kind of a very tongue -in -cheek thought, and you just have to think and write and get confused. And I thought, well, I’m going to record out. that was in 2014. I’ll put the slides out later. I thought it would be science fiction, never happened for hundreds of years. But guess what? You can think, you can put a neural link, and probably have cursors generate code for you today. That I never thought about in 2014. So never underestimate what will happen. The big technology shifts that occur every 30 years, 15 years, 7 years.

But the key thing is semiconductors drove the technology innovation in the 80s and 90s, thanks to Moore’s Law. And the cloud phenomenon happened in the last two decades. I do see the pattern now as you are innovating, as you can see, where NVIDIA is innovating tremendously from the silicon side up. And of course, there are innovations going from the top -down, from the use case, from the language models, and higher order functions in AI. And both are coming at the same time, together. A third bullet I would say is, people who are serious about software should make their own hardware. The corollary is, people who are serious about hardware should also make their own software.

So I’m a hardware guy who’s done software, and this venture, I should be doing software first, going to the hardware later, kind of reverse model. this is the last one day one thing I’ll say about myself my professional life has been shaped by luckily I didn’t realize where between 1980s and 2000 there was a peak of Moore’s law there was an exponential part but happened to be lucky to have been part of the semiconductor innovation cycle having developed and delivered a number of world class microprocessors so today we talk about models there are only 4 or 5 guys who could do microprocessors there’s a difficult very small teams 150 person if you look at model foundation models today it’s the same characteristic there are hard problems of course it’s a lot more money you need a billion dollars and lots of GPUs but you still need the same 150 people to do the models it’s not like everybody can do the models so there is a similarity between what happened in the 90s about microprocessors and what I see today in model building it’s the same level of complexity where you need the best and brightest roughly about it’s not me it’s some altman coding that it’s 120 people and I think that’s the difference and you need to have them with the right resources computing.

We also need to have a lot of computing resources to go build the models. So with that let me start I think the next wave and we hope to drive the innovation and disrupt in terms of systems building going forward but the context is why it’s economically interesting and valuable is I think everybody knows if you look at GDP we’ve gone in the last 20 years from 33 trillion dollars to almost 100 trillion dollars by all accounts the GDP could improve by 2x or 4x in the next 20 years but the SAS era was really a productivity improvement so it really scratched the surface about productivity whereas AI is going to impact 95 % of work so the time is much bigger the impact is much bigger, the blast radius is much bigger than the last 20 years.

That’s why the computing also is needed much more. We have 300 billion dollars of infrastructure we’re in from 50 billion dollars I believe in 2000 So we’re going to be in from 50 billion dollars in the next 20 years. So we’re going to be in from 50 billion dollars in the next 20 years. So we’re going to be in from 50 billion dollars in the next 20 years. to now about $250 to $300 million of capital expenditure spent for infrastructure. So power in, capital spent. So every dollar of power you spend ends up being $3 to $5 of capital for compute, memory, network storage. And from there, you do the upper layers of software and then applications.

So that $50 billion was $300 billion, but if you look at all the spending, we’re already at $400, $500 billion, and all accounts in the next 5 to 10 years will be almost $2 trillion of spend. That creates, obviously, there’s a big demand -supply gap. The great thing about programming is every time there is a layer of abstraction, the programming gets simpler, which means it brings more people to the party to be able to compute. I think what LLMs and natural and transform models have done is bring everybody to be able to program. We all are logical. We can algorithmically think. We can program makers, but not everybody could program. finally we have a tool to be able to program in the natural language your mother, your grandmother can also talk to the computer and tell what steps to take and it will do the steps for you or it will tell you what steps to do so that’s the fundamental shift which means at population scale you’re going to have computing for everybody, that creates a huge gap, it’s not even 1000x as Jensen would say it’s a billion x absolutely true, but it creates a big technology gap, supply gap and increasingly because of model and languages and data the sovereignty gap also that’s appeared, that’s the theme of the conference that continues to drive tremendous amount of demand now we have seen a little bit of this before, I have been through the first two cycles of innovation in semiconductors in my first job as at Sun Microsystems and then the dot com era and then now and there was always a demand supply gap in one of these transitions and But we solved it in one way.

It doesn’t mean you can solve it the same way, but we are at the crux of solving it also in a similar way, but with a different set of boundary conditions, if you will. So what we solved between 1990 and 2000, if you look at, we went from clock rate, single CPU, to fundamentally shifting to multicore threaded and distributed systems, and that was the cloud phenomena. I have a slide later to show what the transition was. I’ll probably skip this slide. I think everybody knows we need lots of power, and one interesting point is I think India is going from almost nothing less than a gigawatt to about 10 gigawatt buildup, while the U.S. is going from 25 to 125 gigawatt in other regions, and China.

EMEA is going to be on a comparative basis, on a relative basis, a lot less. But the need for AI -ready geolocal data centers we already see. Everybody is building out. And what is the infrastructure? What is the architecture to support that? there’s certainly reference architectures inside the hyperscalers Google has got a TPU based and AWS has got Tranium based infrastructure Tranium plus general purpose computing and of course Microsoft has got Maya and of course NVIDIA so those are probably and of course AMD but increasingly over time you want to have an open multi -vendor strategy that’s probably where we’ll check we’ll talk more about so why do I believe these transitions and distributed systems are drivers of new innovation up and down the stack this is not new this has happened in history starting from VAX 11 780 was disrupted by of course at that time PC but more so in the enterprise side was Sun and the workstation if you think of the first distributed system in the modern era it was Sun Microsystems where Ethernet was used to build a distributed system network file system and that was version one and over time it’s like evolution you gain more mass, more momentum more weight in your capabilities and you end up building big monstrous machines in E10K that drove the internet and the dot com era but that was also was an Achilles heel because that was not going to enable the scale that people had to go build at much bigger so Google was probably the epitome of the next big shift I’ll talk about that and similar thing we see today is we’ve gone from CPU only dual socket x86 memory clusters to heterogeneous compute but also gone to a fairly large scale up now the interesting transition today was then is training and inferencing as you’ve seen the news lately with the Grok acquisition by Nvidia and others there’s clearly a separation between the training kind of workloads and inference type of workloads and what kind of systems you want to support because inference is going to drive a lot more of the compute so the one way I think about is inference and biology or workloads related to biology and healthcare are going to be the drivers of computing like it was for graphics in the 1990s.

So this is back again to reinforce the point that between 1994 and 2005, we saw the shift going from the version 1 .0 of distributed system to version 2 .0, which is open source. So the first one was open systems in the first 20 years. And open source came and enabled a new way to build distributed system because from an economics, it removed the cost of middle age of software. Everybody got access. In this case, Linux. This is Solaris. But that also enabled to build truly hundreds of thousands of machines in a single cluster. And out of it came Borg, Kubernetes, a whole bunch of other distributed file systems, all kinds of innovations that happened.

So the proposition here is I think we are at the cusp of similar things on the infant scale computers. And I think we are at the cusp of similar things on the infant scale computers. And I think we are at the cusp of similar things on the infant scale computers. And I think we are at the cusp of similar things on the infant scale computers. So just a reminder. And the punctuation that happens every time turns out to be, if you look back in history, it’s Ethernet. Yes, the network is the computer, but more important is 10 megabit was the onset of replacing big mainframes or miniframes like VAX to workstations and network of workstations.

Then right at the point of 10 gigabit Ethernet coming around 2000 to 2002 timeframe, along with it was multi -core, was enabled the new distributed building block. We are at the same point. We have got 800 gigabit Ethernet going to 800 gigabit Ethernet and probably a terabit Ethernet networks. And that’s hopefully, and that will be the enabler, and that’s a bet we are making. So the other element of the system is its network and then the memory. And do you build a full scale -up system at data center -wide? Certainly you need for training for backpropagation and forward. but inference can be much more distributed, shardable and it’s time to rethink what kind of systems you want for inference only dominating infrastructure.

The other dimension to think about is we’ve gone from a single memory type to multiple memory types so do we need four light types of memory to deal with a variety of layers or just two or one? That’s a lot of debate in the technical community but that’s a critical decision that will happen. So a way to think about this, we think of the entire system not from flops and GPU and compute. GPU compute, CPU compute are needed but really what does the memory hierarchy or memory system look like because there is a physical view because that dominates the cost function and the power function but equally at the same time you have to represent that especially from a performance standpoint you are caching lots of different data.

for computing. So think of the KV caches for the LLM side, the in -memory representations of many of the data from a performance standpoint. So that’s a layer that is continuous, is rich in innovation and technical innovation that we hope to have an influence as well as probably make a mark. And then the large part is the logical view of memory, especially deep context. You want to go from session to session, location to location, and you want to have your memory state. You want to be able to switch models and have some state of the memory state. All of these consume various layers of the logical and the physical layers of memory. So that’s what we think about.

So net, putting all this together, we think of taking a bet with interrnet, taking a bet with memory, and build an infant -scale compute for population scale, like in this case India, but also in certain key verticals like health sciences and others. So… So there’s another important element we want to highlight. I can’t take a quote from Jensen.

Jensen Huang

One of the applications that my favorite is just good old -fashioned data processing, structured data and unstructured data, just good old -fashioned data processing. And very soon we’re going to announce a very big initiative of accelerated data processing. Data processing represents the vast majority of the world’s CPUs today. It still completely runs on CPUs. If you go to Databricks, it’s mostly CPUs. If you go to Snowflakes, mostly CPUs. SQL processing at Oracle, mostly CPUs. Everybody’s using CPUs to do SQL, structured data.

Renu Raman

So taking a cue from what he’s saying, historically, databases, SQL, all run on CPUs. And that will remain the case for a variety of reasons. so that’s an important metric in terms of why we believe the new systems that we compose going forward needs to have a happy blend especially the ways to design systems for the hyperscalers but also the whole category of use case and customers in the private side where they don’t need to have 100 ,000 machines but smaller scale machines but it needs to have a happy blend of CPUs and GPUs that’s the main point in terms of so in that context we have taken a position to start working in partnership with AMD because they’ve got the x86 CPU assets and a compelling GPU roadmap as well as an architecture that supports both from the network side as well as the memory side they have higher memory capacity for LLM so it started with 256GB of HPM which supports 128 billion parameter models at least now it’s going to go to 288 and 512 and no time which means we can fit fairly sizable models so that enables one to do more kind of classical distributed systems principles of single node that captures most of the workload for most customers and be able to optimize it on that.

So coming back, before we get into what we do in Proximal, I want to emphasize the partnership with UC San Diego. They have a data center, as well as I told, it’s a supercomputing data center for research for NSF and DARPA, where we are doing some of the work in terms of the hardware level at the middle layer, in terms of the compute kernels, as well as in the inference engines, as well as the use cases, as I said, because there’s a data science institute, AI for education, to transform the undergrad and graduate level programs using the same tools to have advanced research capability, as well as for health sciences. So with that, I think that is a part to set the motivation for the future of the field.

Thank you. what we’re doing in Proximal Cloud. The next phase we want to go into specifically what we are launching in the four layers, the key components of what we are building and delivering to many cloud partners in India, starting with. There’s also a why India question. I think I’ll say one aspect is India demands an extremely low -cost infant -scale compute at population scale, and that’s a challenge. So we really are excited to work on that problem to start with. So the first thesis, why do we need compute other than the cloud? I think the best way to quote is Michael Dell telling you what he sees. To the beginning here.

Michael Dell

Yeah so we in the last year you know delivered a little over 3 ,000 of these still AI factories and you know those are increasingly to enterprise and commercial customers that want to bring the AI to their data not the data to the AI and you know there’s just a ton of data that is still on -prem and being generated on -prem and it turns out to the beginning here

Renu Raman

If you have a particular question in the domain that you understand, we can try it out after this. So we enable with this interactive learning for the students, contextualized intelligence, and, of course, instructor empowerment in that. And the way it will look and feel will be like a Jupyter notebook on the extension side will be the research content, the archive papers for them to use. It’s an add -on thing. It doesn’t have to be integrated. It’s a commercial AI chat, if you will. Then the next example would be MRI images. Unfortunately, I’m not able to log in remotely onto that right now. The other one I had a local copy. I’m not able to show the MRI images right now.

So at this point, I want to summarize saying that what is Proximal? The word Proximal brings compute closer to your data. The word Proximal means it’s sovereign to the nation or the region or the business that cares about it. And the word Proximal also means we bring compute closer to memory. We bring compute closer to where the business is. so that was the thesis we are not doing this alone we are doing it with some technology partners as well as we have some key customers and partners so with that let me give an example for a given education use case. Let’s go to I’ll bring Lalit Bhatt here to talk about who is director here heading the India office for PharmEx the key partner

Lalit Bhatt

Thanks Renu So what I’ll do is and thanks to Proximal Cloud for giving the stage out here what we’ll do is that basically first I’ll little bit talk about what PharmEx stands for and then why in this space and different space why local compute and all these things are becoming important so PharmEx is basically a comprehensive AI stack so if you see on the left hand side we have lot of infield sensors and So we have a complete comprehensive platform in terms of not only soil moisture sensor but dendro meters and multiple sensors. We also have imaging capabilities where we can take images using satellite, using drones. And we also have an autonomy stack and we just now have acquired an autonomous electric tractor.

Basically these are pretty big machines. They might look like transformers but they are like almost 70, 80 horsepower machines. And we are putting our autonomy stack here so that they will go completely as autonomous ones. So what I’ll do is that I’ll just run a small clip. And I’ll just run a small clip. Thank you. Thank you. Thank you. again I think this is probably very standard everyone understands this you need to do AI you need data these two things we need that one then what becomes important is how efficient you are in terms of running those inferences using those data and we are also dealing with huge amount of data and that’s where we are looking into this technologies where we can reduce our cost everyone understands that in agriculture it is very difficult to ask lot of money from the farmer so where we can really make our operation more efficient if we start like making sure that we are very efficient very effective in terms of dealing with a large amount of data and able to run inferences on top of that one but essentially that’s what happens we get a lot of data both from the imaging side both from the sensor side and then we have our all engines running which basically leads to diagnostic and recommendations and this is just an example of the kind of thing that we do with our customers.

You would see here like complete or autonomous irrigation scheduling. A lot of data points would go into those models basically to create those schedules, anomaly identifications, crop stresses, yield predictions, frost predictions, and even we have worked on this soil percolations model as well. It depends on what all sensors you take. So in India I can tell you like we sell one, there is two feet four sensing probe, which is like four sensing, it goes two feet one, and with the whole controller unit it says 45 ,000 per unit basically. Usually in India we recommend one unit in one hectare, but again it will change based on the variation of the soil and these things like that, but this has been a good ballpark basically.

So yeah, I guess that’s it. And I think the whole theme is that we also are looking for really reduce our inference cost and that’s where Proxima Cloud comes into picture. Thank you.

Renu Raman

Thank you, Bharat. Okay, next we’ll have Bharat, Director at Divium, who is a key partner, and as I mentioned earlier, about model selection and runtime optimization that is integrated or will be integrated into a stack. So, Bharat.

Lalit Bhatt

Hey, good afternoon, everyone. So, let’s address the… hard truth out there. 90 % of Gen -AI pilots never make it to production. Not because the demo was bad or the models were weak or bad. It’s primarily because of three reasons. Number one, quality is undefined. What’s good for one use case is not necessarily good for another one. There’s no standardized way for evaluation or regressions. Number two, the costs are unpredictable. Be the cheapest model or the best model, you can see the price of these models ranging different from like 10 to 50x. The moment your application goes into production and hits real traffic, the costs spike up. There are AI engineers who are running experimentations and trying to tune this.

But model selection is always a moving target. There are always new models coming which are trying to fix something and are breaking something else. So without addressing all three, it’s very difficult for an enterprise to take their pilots to production. And that’s why we built Divium. So Divium is the only inference layer built on quality. Thank you. Divium defines measurable evals aligned against each use cases And it optimizes every incoming query to select the model Which is giving you the best quality per dollar The other part is that Divium automates the entire model selection process By continuously evaluating new models Deprecating the previous old ones And migrating you to new ones If we find something better, we auto -upgrade without breaking production Evals first, routing second And that’s what makes Divium different from every other routing platform out there Divium is the only inference layer with customer -specific intelligence Your apps can be AI agents, rack pipelines, or multi -agent workflows And the LLMs can be from the standard OpenAI, Anthropic Be your own fine -tuned models or deployed open -source models We sit right in between We provide you a single API.

We are continuously evaluating each and every incoming request, routing it to the model which is giving you the most optimal performance and also giving you detailed visibility on what models are working, how is your agent performing and what’s the overall quality. Remember, DVM is trained on your data, your agents and your quality. There’s nothing generic out there. And this is just a theory. We’ve already proven it across multiple deployments. For the India’s largest travel aggregator, which runs a conversational shopping assistant in their application homepage, we were able to cut down the cost by more than 60%. For one of the leading e -pharmacies of India, the customer support chatbot had a little bit lower latency.

So we ended up reducing the cost by 30%, reducing the latency, latency by 30%. leading to a case resolution improvement of 95%. As you can see, different use cases, different industries, but the result is the same. Lower cost and better outcomes. And we understand enterprise realities. You can keep your data secure. We have flexible deployment options, be it SaaS, privately hosted, or on -prem clusters. You stay in control. If you’re trying to take your AI pilots to production, feel free to reach out to us. Thank you.

Renu Raman

Thank you, Bhatt. So we talked first about application use case, one in education and agriculture. Second one, how we are bringing optimization to the system stack. Some of it we do and some of it with our partners. Third, we want to bring in how do we get customers, many of them mid -market, small, as well as large ones, enabled on our platform. I’m happy to introduce Sandeep Kumar. Coming is part of… venture builder instances. It’s a company that we partnered with in Delhi here to take this to a variety of customers, small, medium, large, with a higher velocity. Let him describe what they can do and how we partner.

Sandeep Kumar

Hello, everyone. I’m Sandeep Kumar from Instant System. We are a Silicon Valley -based venture builder. We do not just build startups. We grow them. We are partners in every domain of a startup, be it engineering, be it product, be it marketing. We just give them full blueprint to be a successful startup. We co -invest in the startup so that we are there in every journey of them for them to be a successful startup. We are a venture builder. Sometimes it’s often confused with the incubator, but we are a venture builder. We are a venture builder where we actually help in every step of your startup to be a successful startup. Part of the engine system I am mostly responsible for a company called VanEye though we usually do not disclose the name of the company that we are partners with to protect the IP and the confidentiality but just to give you a use case that you know what our capabilities are and what we have been able to build so far so this is a use case that I am taking this company has got nearly around 200 million dollars funding from the top investors including South Bank we are building an AI conversation software here and we are dealing with real use case real challenges of you know for a mostly like financial domain or financial based industries but all of these solutions are also generic for the analytical based industries as well so I am going to talk about you know some of the challenges that are actually common to every problem or every AI -based solution.

But we’ve been able to identify these challenges and we’ve been able to solve these challenges for this particular use case. So one of the most challenged that every AI -based software face is hallucinations. So LLMs always try to answer to your question irrespective of how much of the context it does have. We’ve been able to solve this problem up to a very good extent and our system is almost 99 % reliable. They do not hallucinate. That was the biggest problem that we’ve been able to solve. Next challenge that we face is disambiguation. So in spite of providing the context, sometimes the system is not able to understand how to disambiguate between some specific terms which may exist in different domains.

So that’s also the problem that we’ve been able to solve. As the theme of the system, and it’s very closely related to the theme of the system, because data security and the data privacy is one of the major industry concerns that we’ve been able to solve. So we’ve been able to address and challenge this problem so that the data privacy and the data access control is being managed at the raw level or in a very technical term I would say at the object level. We’ve been able to tackle that problem and solve that problem efficiently and that’s already running there and working fine. Evaluation and the quality management, that’s also one of the key areas.

That we need to solve as part of the venture that we are building. That’s also that we’ve been able to solve very efficiently. Another thing is the reliability because since we are talking about the financial system, the system has to be reliable. It has to be reliable every time. You cannot send a million dollars to someone’s account by mistake. That doesn’t work in the financial world. Or you cannot report data where you could show losses. instead of revenue or vice versa because you cannot survive in that world with hallucinated or the data which is not correct or factual. So being able to, with our advanced architecture system, we’ve been able to solve that problem as well.

There’s a long list of the pointers that we’ve been able to solve, but I’ll just cut down short. The system that we’ve been building, our performance, the reliable, we’ve been able to keep a check on the cost and efficiency of the system. That’s how we’ve been able to serve to the different audience, different customers from the different niche. So that’s what our theme is. We are a venture builder. Please feel free to reach out to us, and we’d love to talk to you about your startup. We don’t pick a selected startup to work with, but you all are free. You’re welcome to reach out to us, and we can discuss all the stuff that we are doing.

working on. Thank you so much.

Renu Raman

Thank you, Sandeep. I think that ends what we’re doing in Proximal and what our partners and our customers are working with in the early phases. We have partners in the U.S. like UCSD and Life Sciences, Health Sciences, Education, and here in Agriculture and soon to other, particularly we’re going to focus more on, it turns out to be that the Government of India initiative of Education, Health and Agriculture coincidentally aligned. It was not planned. It turned out to be that way. With that, I can go back to any questions. We’ve had a small panel session we can go to. I don’t know if Piyush has come here or not, but I think there’s a question here.

I’m here from ARIA, from Infosys, Senior VP at Infosys. please

Audience

Hello excuse me my name is Arya Bhattacharjee I am from entrepreneur from Silicon Valley so right now like Renu said I am driving this semiconductor and AI vision for Infosys from the United States and India also so the reason I am here is because like Renu said very correctly a very important question that what’s in future for India how can India capitalize or make a mark in this journey so not a small answer but I can tell you what we are trying to do at Infosys because if India is going to win this Semicon 3 .0 or 4 .0, 2 .0, I don’t know, it has to be in software, it has to be in AI.

The chip building -wise is going to take some time. So Renu said that 80 % of the data is on -premise. And what we are working on together is to see on the semiconductor price, this is true, absolutely true, more than 90 % of the data is on -premise. Yes. So the whole journey of how to take the data and how to create solutions through agentic AI approach, through distributed computing, and actually by owning the architecture to lower inferencing cost is a main challenge. So to answer the question which Renu asked me, what’s the future of India? I think that India… what we’re going to do is we’re going to look at a domain. So at Infosys at least we have selected domain and semiconductor, I was talking to him also, that is a large domain and we have taken the leadership with some major clients right now, I can’t talk about details, we’re using an agentech AI on premise and delivering productivity solutions, AI solutions and by cutting our productivity for chip making at least by 25 % and every day in a semiconductor fab you save $10 million, benchmark for a 7 nanometer type of technology, not even 1 .9.

So with that, good luck to Renu and I look forward to collaborating. Thank you.

Renu Raman

Thank you, Arya. Now we welcome Abhishek Venjan but before, just come on board. To summarize what we do, graph, that is underneath what I think is the most important AI factors. Organizing the data layer turns out to be probably the most complicated thing, which spans the enterprise such that it can meet the intelligence. And so that’s the stuff that I think we’ll probably do a lot of. We still don’t really have deep research in a corporate context. We do. That’s what Copilot is about. But most people day -to -day do not have this. So are they just underusing AI that exists? Yes. In fact, it’s interesting you brought that up because to me that is the killer feature.

So the biggest thing we did was we took this graph that is underneath what I think is the most important database in any company, which is underneath your email, your documents, your Teams calls, what have you. It’s the relationships that, by the way, own AI factors. Organizing. Organizing the data layer. so that’s a best summary obviously Satya wants to do it in the cloud and that will happen but also you need to have it in your on -prem, near -prem isolated from other sovereign as well and have the same capabilities, in a sense that’s what we bring to the enterprise if you will any other questions before we go to a panel session

Abhishek Singh

Thanks for having me here this is Abhishek, founder of ZetaVault we did a lot of work on the LLM acceleration, what it means is that we offload the large language models to the specific chips and custom silicon and thereby get the inferencing states and all we have Renu here who has wealth of experience on the distributed computing side and we were supposed to have a panel discussion but I thought I would pick his brain on what the challenges and what kind of changes he has seen in the industry. So, Renu, like, you have been part of, like, Sun Microsystems and early sort of, like, pioneers of distributed computing. So from Sun, which was maybe the distributed systems 1 .0 to Linux, which pretty much democratized the entire competition space and brought the Linux and x86 and now almost every embedded device, every competition pretty much happens on Linux.

So that was the distributed systems 2 .0. And now coming to the distributed computing space with the open models, right? Open source has played a lot of sort of role in the proliferation of the distributed computing. What do you foresee or what do you envision the open models are going to do for the distributed computing? Are we going to see a distributed computing 3 .0?

Renu Raman

Hello. Yeah. So that’s a fundamental thesis in that. I mean, we are, in a way, in part of that continuum to some extent. If you look at… not to take anything away from how NVIDIA designs, but there is a clear bifurcation going on right now as we speak on, as we said, training versus inferencing. And then there is open source models, and a variety of customers’ use cases would use and need the open source models. There’s always been the history of open and close in every transition. I mean, if you go back in the 80s and you go back, I mean, if you look at what enabled the cloud was hypervisors. There was KVM and VMware.

The same thing will apply. There will be open models and closed models. But the way I like to think about it is models is a new abstraction layer that separates between the underlying computing needs and everything above. Hypervisors separated the physical machine to a virtual machine, and then operating systems unix at that time also did that. The same thing is models are the abstraction layer that provides a higher degree of innovation both from closed and open models. The closed one will probably be innovated within OpenAI and Google, but the rest of the world will take the open source model, like what happened with Linux, and innovate. It’s not just going to be an NVIDIA GPU or AMD GPU, or there could be a plethora of GPUs, country -specific, region -specific, domain -specific.

Anything can happen over time.

Abhishek Singh

That’s a very wonderful take. One of the things that we have been wondering is about the latency you talked about in the various scientific and other applications you’re working. When we build the solutions for our customers, we build a lot of, actually, natural language to query processing kind of solutions. Like, we have been able to do maybe a sub -minute kind of a solution, which is acceptable to the customer because from weeks or days, he is able to answer or get the answers to their queries in less than a minute, right? But even a minute is not sufficient, right? When you talk about, like, really interactive queries and all, you want, like, sub -millisecond. Or maybe, like, sub -second kind of response.

what are your thoughts on that? Is it even possible that to a population or to a large customer base that we have in India about 1 .5 billion people at a very low cost, maybe like 200 rupees per month, you can provide query processing at a scale which is like within sub second?

Renu Raman

I think that’s a very good question. Sometimes scaling the problem is more important than the answer. This is an interesting way to frame the problem is, if you go back and look in history, why did Google succeed? A fundamental decision that was made by them on the toolbar is every query response has to be in 20 milliseconds. Now, nobody thought about it prior to that. It’s obvious today. But that key proposition or definition or the question that was asked, maybe by Larry Page or Sergey Brin, whoever it was, led to what we see as Google today in the back end, which is a huge amount of infrastructure to satisfy the 20 millisecond response to any query.

so to me the same thing applies today, maybe 20 is too hard I’m just going to arbitrarily pick I have a simple demo or animation thing I was trying to show every 120 milliseconds you want to have the answer today if you go ask a question it will take seconds sometimes longer than that we are all impatient, we want the answer in quick order, when I ask you a question you don’t say let me think and come back, you want to give the answer if you don’t know how to think and come back, ok, you can but that’s a very deep question you can think and come back but we can throw more computing resources to get that so what it tells you is you can throw a lot more computing to get to the answer, it’s not just hardware, it’s going to be algorithmic improvements other improvements, but to me that’s the benchmark, get 120 milliseconds to any query for anybody so there’s a global context and India context, India provides an ample opportunity for 1 .4 billion people if you can deliver at a cost point and you can deliver at a cost point like 200 rupees a month at 120 seconds and any query to be handled which is a long road to go but if you can meet the objective in 10 to 20 years it serves a lot of people but it also will drive tremendous amount of innovation that’s why when somebody says population scale unique India has a unique thing about the population scale problem and the cost problem so hopefully there are enough people within as Arya said here semiconductor 3 .0 and other innovations that can drive to build India’s own sovereign lowest cost, shortest answer to any language to the question that you asked

Abhishek Singh

interesting take on that one of the things that keeps coming is about the scale of the global corporations or the size that they have been able to reach with AI gaining mainstream in India. And there is a parallel actually theme going on, which is on the semiconductor side, right? Like we are putting, the government is putting a lot of focus. Private players are putting a lot of focus. We have an audience like esteemed, like our Infosys guest. And the question that everybody keeps wondering about is with the AI and AI speeding up things, he’s talking about productivity gains and all. Like, can we, like what kind of corporations can come out of India? Can we see like NVIDIA is coming out or, I don’t know, like Palantir or Supermicro or even a new version of some microsystems coming out just because there is so much emphasis on AI or the Semicon side.

I’ll let Renu talk and then maybe you can also have your take on this particular question, right? What kind of corporations can come out, right? Your take.

Renu Raman

transition, can there be an SAP coming out of this transition in the AI? Why not? To give you some raw numbers, every gigawatt of power will require $25 billion worth of compute memory network storage. So if India is going to do 10 gigawatts, that’s $250 billion of hardware. That brings multiple super micros, or that sustains a semiconductor ecosystem at that scale. So certainly the investments going in for power, which is a long lead item, is important, but the next layer provides the economic value to host the hardware systems company, the HP, the Dells, and Supermicro can emerge. You can go each layer of the stack. The next layer is the application in your tier. Proximal is that.

Maybe we could become the SAP of tomorrow. That could be a Palantir, which is the application tier, not just Palantir, any other company. So both the scale and if you can solve the technology, the scale, and the cost economic it’s not just restricted India it can be global unlike the China model where it ended up being a very close garden wall I think India has the opportunity to be make in India and make for global it’s much better but you just have to think bigger take more important is take bolder bets and go for the long haul not just work for it for 5 years 7 years these are 10 20 years cycles to change very interesting take and thanks a lot for actually and I would like to have your opinion on this particular question do we see NVIDIA SAP’s and Oracle’s coming out thanks sir so I think the semiconductor data for example recently working with more than large companies I want to give some specific examples they are just ingesting data right now just I’m talking about a fab not design they have got 7 petabytes of data ingested and they don’t know what to do with it And like I said, a typical fab manufacturing facility is worth at least $10 billion.

And it’s got thousands of steps. It takes about 120 days to make a chip. So $10 billion for 120 days producing a wafer, and there are defects, design issues, things. So if you take the data just from basic information, run -time, real -time data, defects, soft defects and hard defects, because, you know, just because a chip fails doesn’t mean it’s slow. Slow means no money. That’s a failure. So collecting all that data, classificating the data, understanding and using agents in an edge computing way. You cannot solve this in a server. And then feeding it back to design infrastructure. So the design time also has shrunk a lot. And the yields are going up. 30 years ago when I was in Intel, we were talking about die sizes of like maybe one centimeter by one centimeter.

Today, in a 300 millimeter wafer, NVIDIA’s latest wafer level ship is about 20 centimeter by 20 centimeter. That level of yield and reliability is unimaginable without the use of AI. So I can go on and on, but I think if India has to win, I don’t think India needs to become a Palantir. And India does not want to become a slave shop. So the way I explain that in a one level page, the Palantir’s gross margin is 95%. Indian company’s gross margin is 30%. Can we build a business at 50 % gross margin where the amount of domain expertise India provides with the amount of data is available to take these technologies we talk about to implement them in real practice?

That’s what India can win. it is the execution with the best technology thank you thank you

Abhishek Singh

thanks everyone I have one question for the venture side so like all these like technologies require a lot of investment before they can actually become fruitful right I heard somewhere that the government of Karnataka and I don’t want to demean them by the way I put like 20 crores of fund for actually funding the startups and all that right the single engineer which actually Meta is hiring right now they are throwing how much 100 million dollars at that engineer 20 crore for funding like hundreds of startups versus 100 billion dollar like being given to one engineer right there is a huge mismatch now the question is do we like for Indian companies do the venture capitalists or this or the private equity do they have such deep pockets to fund them for like continuously fund them for hundreds of millions of billions of dollars so that an Nvidia or AMD or I don’t know like the Sun Microsystems can

Renu Raman

I want you to take. Answer this question. Actually, I would like to take your. I don’t want to answer. I want you to answer the question. Answer your own question.

Abhishek Singh

I want to answer my own question. Yes, it will require that kind of investment, right? I think this was one topic I touched upon a long time back. ISRO has been funded like continuously, right? Initially, the ISRO rockets would all land up in the ocean, right? Or the sea or whatever. But over a period of time, they gained competence. They are among the top four in the world right now, right? I think that kind of like continuous and continued support is needed for whatever industry we are picking, whether it is AI or whether it is Semicon. Like we need the private players. We need the government to support it like till the end, right? And that’s when maybe the key players and the winners will emerge.

Renu Raman

I think your question has got two parts. I think the first part is that the government is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going to be the one that is going Sorry, there was a public announcement.

There’s an interruption here. I think there are two parts. One is there’s a mismatch between what a demand supply gap and skills in the model companies, if you

Abhishek Singh

Hopefully, yeah.

Renu Raman

So why did you do it and what do you think? That’s why I asked the question back to you.

Abhishek Singh

It’s good. It’s fun to build for India, by the way, and build from India, right? Build for India, build from India. That’s why we are here. And that’s why all this conference is here. All the discussions are happening.

Renu Raman

But thanks a lot, Renu, for all the wonderful insights. Last call. Anybody from the audience wants to ask any question to Renu?

Audience

Yes, sir. Thanks, sir. So my question is that you shared that if 10 gigawatt business comes to India, that means $250 billion worth of equipment will be purchased or something will happen in India. so how can we ensure that 10 gigawatts just leave 10 let’s start with 1 so how will that business come to India how will that I am just sharing that if 1 gigawatt business you said will come so how will that business come how will that business come

Renu Raman

today we already see most of the hardware is either broadened by the hyperscalers who got some capacity and then Dell HP are the largest OEMs as I understand super micro is behind I guess most of the hardware level systems are manufactured in Taiwan and other places and brought here and there are emergent players like VVDN Sanmina has got a manufacturing plant in Chennai who is going to come and do make in India I don’t want to steal the thunder so there are emergent ones seeing the economic value of that scale to start designing but we have already seen I don’t know the details of all the phone manufacturing that’s happened So the ecosystem of building chassis systems, board design, design capability was there, but manufacturing and operations support and all that was not there.

So I do expect that to start happening. That’s why we started working with CDAC and VVDN to some extent. We do see the opportunity that there’s at least a $300 to $500 million opportunity. If you look at the interesting aspect is the Indian public market is also valuing these things fairly high. Look at NetWeb and others. So you can’t go and raise money in the public market for these kinds of businesses in NASDAQ. You can certainly do that in India. So it’s an interesting point in time where there’s a demand, there’s a need, there needs to be enough people willing to invest. And there’s also probably a way to scale the business. I don’t view going public as an exit.

But really? I’m viewing going public as a way to raise money to scale the business. So there’s enough financial muscle that’s getting built at all stages. But the question is, are there enough people funding at the early phases to fund some of these, right? That I think they have to come together. I’m on the entrepreneur side, not the venture side. I’ve played both, but that has to come. That’s my point of answer to Abhishek’s question is, at a 10 gigawatt, it’s going to be multi -hundreds of billion dollars worth all the layers of the stack, and there should be enough investments going in. And if you look at what has happened in China, there’s a different way to drive that capitalistic structure, right?

They have taken a centralized model, but enabled a lot of districts and regional people to go invest. Look at the cars. How many car companies are there? I’m not saying you should follow the same model, but there should be enough early stage at various layers of the stack to be invested. So, the opportunity, the exit

Abhishek Singh

Thank you. Thank you, Renu, for all that, and everybody who has participated. Thanks for coming, guys. We’ll close the session here, so thanks a lot. Thank you. Thank you.

R

Renu Raman

Speech speed

174 words per minute

Speech length

6044 words

Speech time

2073 seconds

AI impact on work and projected spend

Explanation

Renu highlights that AI will affect the vast majority of jobs, creating unprecedented demand for compute resources. She forecasts that global AI-related spending could approach $2 trillion in the next decade.


Evidence

“AI is going to impact 95 % of work” [1]. “…in the next 5 to 10 years will be almost $2 trillion of spend” [2].


Major discussion point

AI‑driven compute demand and infrastructure scale


Topics

Artificial intelligence | The digital economy


AMD partnership for balanced CPU/GPU blend

Explanation

Renu explains that partnering with AMD provides both x86 CPUs and a strong GPU roadmap, enabling a “happy blend” of compute resources for diverse workloads.


Evidence

“we have taken a position to start working in partnership with AMD because they’ve got the x86 CPU assets and a compelling GPU roadmap … it needs to have a happy blend of CPUs and GPUs” [24].


Major discussion point

Hardware‑software co‑design and strategic partnerships


Topics

Artificial intelligence | The enabling environment for digital development


Proximal Cloud sovereign, low‑cost compute

Explanation

Renu describes Proximal Cloud as a solution that brings compute close to data while being sovereign to a nation or region, targeting infant‑scale, affordable compute for mass adoption.


Evidence

“what we’re doing in Proximal Cloud” [76]. “The word Proximal brings compute closer to your data” [77]. “The word Proximal means it’s sovereign to the nation or the region or the business that cares about it” [81].


Major discussion point

Sovereign, low‑cost compute for India and vertical use cases


Topics

Artificial intelligence | The enabling environment for digital development


Memory hierarchy critical for affordable inference

Explanation

Renu stresses that the design of memory hierarchies and systems is a dominant factor in both cost and power for AI inference, and must be optimized alongside compute.


Evidence

“GPU compute, CPU compute are needed but really what does the memory hierarchy or memory system look like because there is a physical view because that dominates the cost function and the power function…” [18].


Major discussion point

Model selection, inference optimization, and cost efficiency


Topics

Artificial intelligence


10 GW power translates to $250 B hardware spend

Explanation

Renu points out that scaling AI infrastructure to 10 GW of power would require roughly $250 billion in hardware, creating a massive market for Indian OEMs and system integrators.


Evidence

“So if India is going to do 10 gigawatts, that’s $250 billion of hardware” [122].


Major discussion point

Ecosystem building, investment, and future Indian AI/semiconductor companies


Topics

Financial mechanisms | Artificial intelligence | The enabling environment for digital development


Open models as an abstraction layer

Explanation

Renu argues that open (and closed) AI models act as a new abstraction layer separating compute needs from applications, enabling broader innovation across the stack.


Evidence

“models is a new abstraction layer that separates between the underlying computing needs and everything above” [62]. “The same thing is models are the abstraction layer that provides a higher degree of innovation both from closed and open models” [63].


Major discussion point

Open vs. closed models and the next phase of distributed computing


Topics

Artificial intelligence


J

Jensen Huang

Speech speed

149 words per minute

Speech length

86 words

Speech time

34 seconds

CPUs dominate current data processing

Explanation

Jensen notes that the majority of data‑processing workloads still run on CPUs, underscoring the scale of existing compute infrastructure.


Evidence

“Data processing represents the vast majority of the world’s CPUs today” [12].


Major discussion point

AI‑driven compute demand and infrastructure scale


Topics

Artificial intelligence | The digital economy


Accelerated data processing with GPUs

Explanation

Jensen announces an upcoming initiative to accelerate data processing, signalling a shift toward GPU‑based workloads.


Evidence

“And very soon we’re going to announce a very big initiative of accelerated data processing” [14].


Major discussion point

Hardware‑software co‑design and strategic partnerships


Topics

Artificial intelligence


M

Michael Dell

Speech speed

126 words per minute

Speech length

73 words

Speech time

34 seconds

AI factories bring AI to the data

Explanation

Michael describes how AI factories deliver on‑prem AI, moving compute to where the data resides and reducing latency and cloud dependence.


Evidence

“those are increasingly to enterprise and commercial customers that want to bring the AI to their data not the data to the AI” [27].


Major discussion point

AI‑driven compute demand and infrastructure scale


Topics

Artificial intelligence | The enabling environment for digital development


L

Lalit Bhatt

Speech speed

123 words per minute

Speech length

1070 words

Speech time

519 seconds

Local compute cuts inference cost in agriculture

Explanation

Lalit explains that bringing compute close to the field reduces inference cost and latency, making AI‑driven irrigation and yield prediction affordable for farmers.


Evidence

“…it’s very difficult to ask lot of money from the farmer so where we can really make our operation more efficient…” [39]. “So we ended up reducing the cost by 30%, reducing the latency, latency by 30%” [28].


Major discussion point

Sovereign, low‑cost compute for India and vertical use cases


Topics

Social and economic development | Artificial intelligence


Dynamic model routing and selection (Divium)

Explanation

Lalit details a system that continuously evaluates incoming queries, routes them to the optimal model, and auto‑updates to maintain best quality‑per‑dollar performance.


Evidence

“Divium defines measurable evals aligned against each use cases And it optimizes every incoming query to select the model Which is giving you the best quality per dollar” [110]. “Divium is the only inference layer with customer‑specific intelligence…” [111].


Major discussion point

Model selection, inference optimization, and cost efficiency


Topics

Artificial intelligence | Financial mechanisms


S

Sandeep Kumar

Speech speed

158 words per minute

Speech length

769 words

Speech time

290 seconds

Venture‑builder end‑to‑end support

Explanation

Sandeep outlines the venture‑builder model that provides startups with engineering, product, and marketing assistance throughout their journey.


Evidence

“We are a venture builder where we actually help in every step of your startup to be a successful startup” [90]. “We are partners in every domain of a startup, be it engineering, be it product, be it marketing” [92].


Major discussion point

Sovereign, low‑cost compute for India and vertical use cases


Topics

The enabling environment for digital development | Financial mechanisms


Addressing hallucinations, disambiguation, and data privacy

Explanation

Sandeep points out key reliability challenges—hallucinations, disambiguation, and data security—that must be solved to ensure trustworthy AI inference.


Evidence

“one of the most challenged that every AI‑based software face is hallucinations” [116]. “Next challenge that we face is disambiguation” [117]. “You can keep your data secure” [119].


Major discussion point

Model selection, inference optimization, and cost efficiency


Topics

Artificial intelligence | Human rights and the ethical dimensions of the information society


A

Audience

Speech speed

136 words per minute

Speech length

428 words

Speech time

188 seconds

India’s strength lies in software and AI

Explanation

An audience member stresses that India’s competitive advantage will come from software and AI capabilities rather than hardware manufacturing.


Evidence

“it has to be in software, it has to be in AI” [101].


Major discussion point

Sovereign, low‑cost compute for India and vertical use cases


Topics

Artificial intelligence | The digital economy


Massive hardware investment needed for 10 GW AI business

Explanation

The audience highlights that a 10 GW AI deployment in India would require hundreds of billions of dollars in equipment, underscoring the scale of required investment.


Evidence

“if 10 gigawatt business comes to India, that means $250 billion worth of equipment will be purchased” [123]. “it’s going to be multi‑hundreds of billion dollars” [124].


Major discussion point

Ecosystem building, investment, and future Indian AI/semiconductor companies


Topics

Financial mechanisms | Artificial intelligence


A

Abhishek Singh

Speech speed

157 words per minute

Speech length

955 words

Speech time

364 seconds

Sub‑second query processing for 1.5 billion users

Explanation

Abhishek questions whether India can deliver AI query responses within sub‑second latency to its 1.5 billion population at an affordable cost.


Evidence

“Is it even possible that … 1 .5 billion people … you can provide query processing at a scale which is like within sub second?” [40]. “When you talk about, like, really interactive queries … sub‑millisecond” [41]. “… you want to have the answer in 120 milliseconds…” [42].


Major discussion point

AI‑driven compute demand and infrastructure scale


Topics

Artificial intelligence | The digital economy


Open models driving Distributed Computing 3.0

Explanation

Abhishek envisions that open‑source AI models will enable a new era of distributed computing, which he terms “Distributed Computing 3.0”.


Evidence

“And now coming to the distributed computing space with the open models” [67]. “What do you foresee … open models are going to do for the distributed computing?” [68]. “Are we going to see a distributed computing 3 .0?” [150]. “we saw the shift … version 2.0, which is open source” [151].


Major discussion point

Open vs. closed models and the next phase of distributed computing


Topics

Artificial intelligence | Internet governance


Agreements

Agreement points

Most enterprise data remains on-premises requiring AI solutions to come to the data

Speakers

– Renu Raman
– Michael Dell

Arguments

Most enterprise data (80-90%) remains on-premises, requiring AI solutions that bring compute to data rather than data to cloud


Most enterprise data (80-90%) remains on-premises, requiring AI solutions that bring compute to data rather than data to cloud


Summary

Both speakers strongly agree that the vast majority of enterprise data still resides on-premises and continues to be generated locally, necessitating a fundamental shift in AI deployment strategy to bring computing capabilities to where data exists rather than moving data to centralized cloud systems


Topics

Data governance | Artificial intelligence | The digital economy


Hybrid CPU-GPU architectures are necessary for enterprise AI systems

Speakers

– Renu Raman
– Jensen Huang

Arguments

Current transition from training-focused to inference-focused workloads requires rethinking system architecture for distributed, shardable inference computing


Traditional data processing (SQL, databases) still runs primarily on CPUs and will continue to do so, requiring hybrid CPU-GPU systems


Summary

Both speakers recognize that enterprise AI systems cannot rely solely on GPU computing but must incorporate hybrid architectures that effectively combine CPU and GPU capabilities, as traditional data processing workloads continue to run on CPUs while AI workloads benefit from GPU acceleration


Topics

Artificial intelligence | Information and communication technologies for development


India has unique opportunities for population-scale computing innovation

Speakers

– Renu Raman
– Abhishek Singh

Arguments

India demands extremely low-cost infant-scale compute at population scale, presenting unique engineering challenges and opportunities


Sub-second query processing at population scale (1.4 billion people) for ₹200/month requires breakthrough innovations in cost and latency


Summary

Both speakers acknowledge that India’s massive population creates unprecedented opportunities for computing innovation, requiring breakthrough solutions that can serve 1.4 billion people at extremely low cost points while maintaining high performance standards


Topics

Closing all digital divides | Information and communication technologies for development | Artificial intelligence


Long-term sustained investment is crucial for technology development success

Speakers

– Renu Raman
– Abhishek Singh
– Audience

Arguments

10 gigawatt power infrastructure in India could drive $250 billion in hardware systems, creating opportunities for multiple system companies and semiconductor ecosystem


India needs sustained long-term investment similar to ISRO’s model, with continuous government and private support over decades


India can win in semiconductor 3.0/4.0 through software and AI rather than chip manufacturing, focusing on domain expertise with 50% gross margins versus traditional 30%


Summary

All speakers agree that achieving success in AI and semiconductor industries requires sustained, long-term investment approaches spanning decades, with both government and private sector commitment, similar to successful models like ISRO


Topics

Financial mechanisms | The enabling environment for digital development | Social and economic development


Similar viewpoints

Both speakers emphasize the critical importance of cost optimization in AI systems while maintaining high performance, recognizing that economic constraints in sectors like agriculture require innovative approaches to make AI viable

Speakers

– Renu Raman
– Lalit Bhatt

Arguments

Agriculture sector needs efficient AI inference for sensor data, imaging, and autonomous systems while maintaining low costs for farmers


Target of 120 millisecond response time for any query (similar to Google’s 20ms standard) requires massive computing resources and algorithmic improvements


Topics

Social and economic development | Artificial intelligence | Closing all digital divides


Both speakers recognize that managing and processing massive amounts of enterprise data is fundamental to successful AI implementation, whether in general enterprise contexts or specialized manufacturing environments

Speakers

– Renu Raman
– Audience

Arguments

Data organization and creating intelligent enterprise graphs from emails, documents, and communications is the most critical AI infrastructure challenge


Semiconductor manufacturing generates 7 petabytes of data requiring real-time edge AI processing to improve yields and reduce defects in $10 billion facilities


Topics

Data governance | Artificial intelligence | Social and economic development


Both speakers emphasize that production-ready AI systems require extremely high reliability standards and sophisticated quality management approaches, far beyond what is typically achieved in pilot projects

Speakers

– Lalit Bhatt
– Sandeep Kumar

Arguments

90% of Gen-AI pilots never make it to production due to undefined quality standards, unpredictable costs, and constantly changing model selection


Financial systems require 99% reliability without hallucinations, with advanced architecture solving data privacy and access control at object level


Topics

Artificial intelligence | Building confidence and security in the use of ICTs | The digital economy


Unexpected consensus

Open source models as drivers of distributed computing innovation

Speakers

– Renu Raman
– Abhishek Singh

Arguments

Open source models will drive distributed computing 3.0, similar to how Linux democratized computing, enabling country-specific and domain-specific innovations


India needs sustained long-term investment similar to ISRO’s model, with continuous government and private support over decades


Explanation

The unexpected consensus emerges around the idea that open source AI models will have a democratizing effect similar to Linux, enabling smaller countries and specialized domains to innovate independently of major tech companies. This represents a shift from centralized AI development to distributed innovation ecosystems


Topics

Artificial intelligence | The enabling environment for digital development | Information and communication technologies for development


Memory hierarchy as the dominant factor in AI system design

Speakers

– Renu Raman
– Jensen Huang

Arguments

Memory hierarchy design with multiple memory types and caching strategies (KV caches, in-memory representations) dominates cost and performance functions


Traditional data processing (SQL, databases) still runs primarily on CPUs and will continue to do so, requiring hybrid CPU-GPU systems


Explanation

There’s unexpected alignment on the idea that memory architecture, rather than pure compute power, has become the critical design factor in AI systems. This challenges the common focus on GPU compute capabilities and highlights the importance of data movement and storage optimization


Topics

Artificial intelligence | Information and communication technologies for development


Overall assessment

Summary

The speakers demonstrate strong consensus on several key areas: the need for on-premises AI solutions due to data locality, the requirement for hybrid computing architectures, India’s unique position for population-scale innovation, and the necessity of long-term investment strategies. There’s also alignment on the critical importance of cost optimization, data management challenges, and quality/reliability standards for production AI systems.


Consensus level

High level of consensus with significant implications for AI infrastructure development, particularly regarding the shift from cloud-centric to edge/on-premises AI deployment strategies, the recognition of India as a unique testing ground for population-scale computing solutions, and the understanding that successful AI implementation requires sustained, multi-decade investment approaches rather than short-term initiatives.


Differences

Different viewpoints

Approach to AI deployment – centralized cloud vs distributed edge computing

Speakers

– Renu Raman
– Michael Dell

Arguments

Most enterprise data (80-90%) remains on-premises, requiring AI solutions that bring compute to data rather than data to cloud


Most enterprise data (80-90%) remains on-premises, requiring AI solutions that bring compute to data rather than data to cloud


Summary

While both speakers agree on the data location reality, there’s an implicit disagreement on deployment strategy. Renu advocates for distributed, sovereign computing solutions while Michael Dell focuses on enterprise AI factories, suggesting different architectural approaches to the same problem.


Topics

Data governance | Artificial intelligence | The digital economy


India’s competitive positioning strategy in global markets

Speakers

– Renu Raman
– Audience

Arguments

India demands extremely low-cost infant-scale compute at population scale, presenting unique engineering challenges and opportunities


India can win in semiconductor 3.0/4.0 through software and AI rather than chip manufacturing, focusing on domain expertise with 50% gross margins versus traditional 30%


Summary

Renu focuses on India’s unique population-scale computing challenges requiring breakthrough cost innovations, while the Infosys representative argues for a middle-ground approach targeting 50% gross margins through domain expertise rather than competing on pure cost or trying to match premium players.


Topics

Social and economic development | The enabling environment for digital development | Artificial intelligence


Unexpected differences

Investment and funding approach for technology development

Speakers

– Renu Raman
– Abhishek Singh

Arguments

Open source models will drive distributed computing 3.0, similar to how Linux democratized computing, enabling country-specific and domain-specific innovations


India needs sustained long-term investment similar to ISRO’s model, with continuous government and private support over decades


Explanation

While both support India’s technology development, there’s an unexpected disagreement on approach – Renu emphasizes open source democratization and market-driven innovation, while Abhishek advocates for sustained government-led investment models. This represents different philosophies on how technological breakthroughs should be achieved.


Topics

Financial mechanisms | The enabling environment for digital development | Artificial intelligence


Overall assessment

Summary

The discussion shows relatively low levels of direct disagreement, with most speakers sharing common goals around AI infrastructure development, India’s technological advancement, and the need for cost-effective solutions. The main disagreements center on strategic approaches rather than fundamental objectives.


Disagreement level

Low to moderate disagreement level. The speakers generally align on the opportunities and challenges but differ on implementation strategies, economic models, and technical approaches. This suggests a healthy diversity of perspectives within a shared vision, which could lead to complementary rather than competing solutions.


Partial agreements

Partial agreements

Both agree that hybrid computing architectures are necessary, but they emphasize different aspects – Renu focuses on memory hierarchy optimization while Jensen emphasizes the continued importance of CPU-based data processing alongside GPU acceleration.

Speakers

– Renu Raman
– Jensen Huang

Arguments

Memory hierarchy design with multiple memory types and caching strategies (KV caches, in-memory representations) dominates cost and performance functions


Traditional data processing (SQL, databases) still runs primarily on CPUs and will continue to do so, requiring hybrid CPU-GPU systems


Topics

Artificial intelligence | Information and communication technologies for development


Both agree on the need for extremely fast AI response times, but differ on the specific targets and economic constraints – Renu sets 120ms as the benchmark while Abhishek focuses on sub-second responses at extremely low cost points for population scale.

Speakers

– Renu Raman
– Abhishek Singh

Arguments

Target of 120 millisecond response time for any query (similar to Google’s 20ms standard) requires massive computing resources and algorithmic improvements


Sub-second query processing at population scale (1.4 billion people) for ₹200/month requires breakthrough innovations in cost and latency


Topics

Artificial intelligence | Closing all digital divides | Financial mechanisms


Both see massive opportunities in India’s semiconductor and AI infrastructure development, but approach it from different angles – Renu focuses on the hardware systems opportunities from power infrastructure while the audience member emphasizes the data processing and manufacturing optimization aspects.

Speakers

– Renu Raman
– Audience

Arguments

10 gigawatt power infrastructure in India could drive $250 billion in hardware systems, creating opportunities for multiple system companies and semiconductor ecosystem


Semiconductor manufacturing generates 7 petabytes of data requiring real-time edge AI processing to improve yields and reduce defects in $10 billion facilities


Topics

The enabling environment for digital development | Social and economic development | Artificial intelligence


Similar viewpoints

Both speakers emphasize the critical importance of cost optimization in AI systems while maintaining high performance, recognizing that economic constraints in sectors like agriculture require innovative approaches to make AI viable

Speakers

– Renu Raman
– Lalit Bhatt

Arguments

Agriculture sector needs efficient AI inference for sensor data, imaging, and autonomous systems while maintaining low costs for farmers


Target of 120 millisecond response time for any query (similar to Google’s 20ms standard) requires massive computing resources and algorithmic improvements


Topics

Social and economic development | Artificial intelligence | Closing all digital divides


Both speakers recognize that managing and processing massive amounts of enterprise data is fundamental to successful AI implementation, whether in general enterprise contexts or specialized manufacturing environments

Speakers

– Renu Raman
– Audience

Arguments

Data organization and creating intelligent enterprise graphs from emails, documents, and communications is the most critical AI infrastructure challenge


Semiconductor manufacturing generates 7 petabytes of data requiring real-time edge AI processing to improve yields and reduce defects in $10 billion facilities


Topics

Data governance | Artificial intelligence | Social and economic development


Both speakers emphasize that production-ready AI systems require extremely high reliability standards and sophisticated quality management approaches, far beyond what is typically achieved in pilot projects

Speakers

– Lalit Bhatt
– Sandeep Kumar

Arguments

90% of Gen-AI pilots never make it to production due to undefined quality standards, unpredictable costs, and constantly changing model selection


Financial systems require 99% reliability without hallucinations, with advanced architecture solving data privacy and access control at object level


Topics

Artificial intelligence | Building confidence and security in the use of ICTs | The digital economy


Takeaways

Key takeaways

AI represents a fundamental technology shift occurring every 15-30 years, with current demand increases of billion-fold scale requiring massive infrastructure investment ($2 trillion projected in 5-10 years)


Enterprise AI faces critical adoption barriers with 90% of pilots failing to reach production due to undefined quality standards, unpredictable costs, and model selection challenges


India has a unique opportunity to lead in AI/semiconductor 3.0 through software and domain expertise rather than chip manufacturing, leveraging population-scale computing demands


The shift from training-focused to inference-focused workloads requires new distributed computing architectures, with open source models driving innovation similar to Linux’s impact


Most enterprise data (80-90%) remains on-premises, necessitating solutions that bring compute to data rather than moving data to cloud


Memory hierarchy design and network infrastructure (800GB to terabit Ethernet) are critical enablers for next-generation distributed computing systems


Target performance requirements include 120-millisecond query response times at population scale for ₹200/month, requiring breakthrough cost and latency innovations


Resolutions and action items

Proximal Cloud launched their offering in India focusing on enterprise private cloud infrastructure with partnerships in education, agriculture, and health sciences


Partnership established with UC San Diego for AI research in education, health sciences, and industry applications


Collaboration agreements in place with technology partners including PharmEx (agriculture), Divium (model optimization), and Instant System (venture building)


Focus on Government of India initiatives in Education, Health, and Agriculture as primary market segments


Integration of AMD-based systems for CPU-GPU hybrid architecture to support both traditional data processing and AI workloads


Development of domain-specific solutions including agricultural sensor systems, educational AI tools, and financial sector applications


Unresolved issues

Funding gap between Indian startup investment (₹20 crores for hundreds of startups) versus global AI investment scales (hundreds of millions per engineer)


How to achieve 10 gigawatt power infrastructure buildout in India and ensure associated $250 billion hardware business benefits domestic companies


Technical challenge of achieving sub-second query processing at population scale (1.4 billion people) while maintaining ₹200/month cost point


Scaling from pilot projects to production deployment across various industry verticals while maintaining quality and cost effectiveness


Long-term sustainability of venture funding for hardware and infrastructure companies requiring 10-20 year development cycles


Integration challenges between multiple memory types and caching strategies for optimal performance and cost balance


Suggested compromises

Hybrid CPU-GPU systems approach rather than GPU-only infrastructure to balance traditional data processing needs with AI workloads


Gradual scaling approach starting with 1 gigawatt infrastructure before targeting 10 gigawatt buildout


50% gross margin business model for Indian companies (between traditional 30% and Palantir’s 95%) to balance competitiveness with profitability


Public market funding strategy in India rather than NASDAQ for raising capital to scale infrastructure businesses


Partnership model combining international technology expertise with local domain knowledge and cost optimization


Flexible deployment options (SaaS, privately hosted, on-premises) to address varying enterprise security and sovereignty requirements


Thought provoking comments

We overestimate what can be done in two years, but we underestimate what can be done in 10 years… I thought it would be science fiction, never happened for hundreds of years. But guess what? You can think, you can put a neural link, and probably have cursors generate code for you today. That I never thought about in 2014.

Speaker

Renu Raman


Reason

This comment provides a profound framework for understanding technological progress and sets the philosophical tone for the entire discussion. It challenges linear thinking about innovation and introduces the concept that breakthrough technologies often emerge faster than expected once foundational elements align.


Impact

This opening insight established the discussion’s forward-looking perspective and justified the ambitious scope of their AI infrastructure vision. It primed the audience to think beyond current limitations and consider transformative possibilities, setting up the entire presentation’s credibility for discussing seemingly ambitious goals.


People who are serious about software should make their own hardware. The corollary is, people who are serious about hardware should also make their own software.

Speaker

Renu Raman


Reason

This challenges the traditional separation between hardware and software development, advocating for vertical integration. It’s particularly insightful given the current AI landscape where companies like NVIDIA are succeeding precisely because they control both layers.


Impact

This comment justified Proximal’s approach of building integrated solutions rather than focusing on just one layer. It influenced the subsequent discussion about their partnerships with AMD and their full-stack approach, making their comprehensive strategy seem necessary rather than overly ambitious.


AI is going to impact 95% of work… whereas the SaaS era was really a productivity improvement so it really scratched the surface about productivity

Speaker

Renu Raman


Reason

This comment reframes AI not as an incremental improvement but as a fundamental transformation of work itself. It provides economic justification for massive infrastructure investments by positioning AI as qualitatively different from previous technology waves.


Impact

This insight shifted the discussion from technical capabilities to economic transformation, providing the business case for the infrastructure investments being discussed. It elevated the conversation from ‘how to build AI systems’ to ‘how to prepare for economic transformation,’ influencing subsequent discussions about scale and investment needs.


90% of Gen-AI pilots never make it to production. Not because the demo was bad or the models were weak or bad. It’s primarily because of three reasons: quality is undefined, costs are unpredictable, and model selection is always a moving target.

Speaker

Lalit Bhatt (Divium)


Reason

This comment cuts through the AI hype to identify the real practical barriers to AI adoption. It’s particularly insightful because it focuses on operational rather than technical challenges, revealing why AI success requires more than just good models.


Impact

This observation shifted the discussion from theoretical capabilities to practical implementation challenges. It validated the need for the optimization and management layers that the partners were building, and introduced a more realistic perspective on AI deployment that influenced subsequent discussions about enterprise adoption.


Every query response has to be in 20 milliseconds… so to me the same thing applies today, maybe 20 is too hard I’m just going to arbitrarily pick… 120 milliseconds you want to have the answer

Speaker

Renu Raman


Reason

This comment draws a powerful parallel between Google’s success and AI infrastructure requirements, suggesting that user experience constraints (response time) should drive infrastructure design rather than technical capabilities driving user experience.


Impact

This insight reframed the entire infrastructure discussion around user experience requirements rather than technical specifications. It provided a concrete performance target that influenced how the audience thought about the scale and sophistication of infrastructure needed, making the ambitious infrastructure investments seem not just justified but necessary.


India does not want to become a slave shop. So the way I explain that… the Palantir’s gross margin is 95%. Indian company’s gross margin is 30%. Can we build a business at 50% gross margin where the amount of domain expertise India provides with the amount of data is available?

Speaker

Arya Bhattacharjee (Infosys)


Reason

This comment addresses a critical strategic question about India’s positioning in the global AI economy. It challenges the traditional outsourcing model and proposes a middle path that leverages India’s strengths while capturing more value.


Impact

This observation elevated the discussion from technical implementation to national economic strategy. It influenced the conversation about what kinds of companies could emerge from India and how they should be positioned, adding a geopolitical and economic development dimension to the technical discussion.


If India is going to do 10 gigawatts, that’s $250 billion of hardware. That brings multiple super micros, or that sustains a semiconductor ecosystem at that scale.

Speaker

Renu Raman


Reason

This comment provides concrete economic scale that transforms abstract infrastructure discussions into tangible business opportunities. It demonstrates how infrastructure investments can create entire ecosystems of companies and economic value.


Impact

This quantification shifted the discussion from whether India could compete in AI infrastructure to how it could build an entire ecosystem around that infrastructure. It influenced subsequent questions about funding, manufacturing, and the potential for creating major technology companies, making the ambitious vision seem economically viable.


Overall assessment

These key comments shaped the discussion by progressively building a comprehensive vision that moved from philosophical foundations to practical implementation to economic transformation. Renu Raman’s opening insights about technological progress and hardware-software integration established credibility and ambition. The practical challenges identified by partners like Divium grounded the discussion in real-world implementation issues. The economic scale discussions and strategic positioning comments elevated the conversation to national competitiveness and ecosystem building. Together, these comments created a narrative arc that justified ambitious infrastructure investments not just as technical necessities, but as economic and strategic imperatives for India’s position in the global AI economy. The discussion evolved from a product presentation into a broader conversation about technological sovereignty and economic development strategy.


Follow-up questions

How to achieve 120 millisecond response time for any query at population scale (1.4 billion people) at a cost point of 200 rupees per month?

Speaker

Abhishek Singh and Renu Raman


Explanation

This represents a fundamental technical and economic challenge that would require significant algorithmic improvements and computing resources to solve, potentially driving major innovation in India’s AI infrastructure


What is the optimal memory hierarchy design for AI systems – do we need four different types of memory or just two or one?

Speaker

Renu Raman


Explanation

This is described as having ‘a lot of debate in the technical community’ and represents a critical decision that will impact cost and performance of AI systems


How can India ensure that the projected 10 gigawatt AI infrastructure business actually comes to India and benefits local companies?

Speaker

Audience member


Explanation

This addresses the practical implementation of India’s AI infrastructure ambitions and how to capture the associated $250 billion hardware opportunity locally


Can Indian companies achieve 50% gross margins (between Palantir’s 95% and typical Indian company’s 30%) by combining domain expertise with advanced AI technologies?

Speaker

Arya Bhattacharjee (Infosys)


Explanation

This explores a potential business model for Indian AI companies to compete globally while leveraging India’s strengths in domain knowledge and execution


How to solve the 90% failure rate of Gen-AI pilots making it to production, particularly around quality definition, cost predictability, and model selection?

Speaker

Bharat (Divium)


Explanation

This addresses a critical industry problem that prevents AI adoption at scale and represents a significant market opportunity for solutions


What kind of venture capital funding structure and depth is needed in India to support AI and semiconductor companies that require hundreds of millions to billions in investment?

Speaker

Abhishek Singh


Explanation

This highlights the funding gap between what’s available in India versus what’s needed to build world-class AI and semiconductor companies, using the example of Karnataka’s 20 crore fund versus Meta’s $100 million engineer hiring


How to effectively utilize the 7 petabytes of manufacturing data being ingested by semiconductor fabs to improve yields and reduce defects?

Speaker

Arya Bhattacharjee (Infosys)


Explanation

This represents a specific use case where AI can provide significant value in semiconductor manufacturing, with potential savings of $10 million per day in a typical fab


Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.