From KW to GW Scaling the Infrastructure of the Global AI Economy
20 Feb 2026 15:00h - 16:00h
From KW to GW Scaling the Infrastructure of the Global AI Economy
Session at a glance
Summary
This discussion focused on India’s rapidly expanding AI infrastructure and the development of sovereign AI capabilities at gigawatt scale. The conversation featured industry leaders from companies like NVIDIA, Vertiv, Google, and government organizations discussing the transformation from traditional cloud computing to AI factories that generate new data and outputs in real-time rather than simply retrieving stored information.
A central theme was India’s potential to become a global AI hub, with projections suggesting the country will scale from 1.5 gigawatts to 10-12 gigawatts of AI infrastructure capacity within three years, far exceeding current industry estimates of 5-6 gigawatts. Speakers emphasized that India generates 20% of the world’s data but currently has only 3% of global data center capacity, highlighting a significant infrastructure gap that needs addressing for data sovereignty.
The discussion revealed how AI workloads require fundamentally different infrastructure approaches compared to traditional data centers. While conventional racks operated at 10-15 kilowatts, AI applications now demand 130-240 kilowatts per rack, with future generations potentially reaching one megawatt per rack. This dramatic increase necessitates liquid cooling systems and purpose-built “AI factories” rather than retrofitted traditional data centers.
Participants stressed the importance of starting infrastructure design from the GPU level rather than working from the power grid down, emphasizing speed of deployment as critical since delayed monetization of expensive GPU investments significantly impacts return on investment. The conversation also highlighted successful AI implementations in India, including government applications for fraud detection in food subsidies and financial transactions, as well as multilingual translation services processing 100 million requests per hour. The discussion concluded with recognition that building this infrastructure at scale requires significant skill development programs and international collaboration to avoid repeating mistakes made in other regions.
Keypoints
Major Discussion Points:
– India’s AI Sovereignty and Infrastructure Development: Discussion of India’s potential to become a global AI hub, with emphasis on building sovereign AI capabilities including indigenous models like Bharat GPT, data residency requirements, and the need for local processing of India’s data (20% of world’s data with only 3% of global data center capacity).
– Scaling AI Infrastructure at Unprecedented Speed: Focus on the massive infrastructure scaling required – from current 1.5 gigawatts to projected 10-12 gigawatts in 3 years, representing a doubling of capacity annually. Emphasis on “speed at scale” deployment and the shift from traditional 18-month project cycles to 4-6 month cycles.
– Fundamental Shift in Data Center Design Philosophy: Moving from traditional “grid-to-chip” thinking to “chip-to-grid” approach, starting with GPU requirements (240+ kilowatts per rack, potentially reaching 1 megawatt per rack) and designing infrastructure around compute needs rather than the reverse.
– Liquid Cooling and High-Density Computing: Technical discussions about the transition from air cooling to liquid cooling systems, managing extreme power densities, and the infrastructure challenges of supporting AI workloads that are fundamentally different from traditional cloud computing.
– Real-World AI Applications and Economic Impact: Examples of AI implementation in India including IRCTC’s traffic management, government fraud detection systems saving millions daily, multilingual translation services (Bhashini), and UPI payment fraud prevention, demonstrating tangible economic benefits.
Overall Purpose:
The discussion aimed to educate stakeholders about the technical, infrastructural, and strategic requirements for building large-scale AI data centers in India, with particular focus on achieving AI sovereignty while leveraging global expertise and partnerships between companies like NVIDIA and Vertiv.
Overall Tone:
The tone was consistently optimistic and collaborative throughout, with industry experts sharing technical knowledge and best practices. Speakers demonstrated enthusiasm about India’s AI potential while maintaining a practical, solution-oriented approach to addressing infrastructure challenges. The discussion maintained a professional, educational atmosphere with participants actively sharing lessons learned from global deployments to help India avoid common pitfalls and accelerate development.
Speakers
Speakers from the provided list:
– Ankush Sabharwal – CEO/Founder working on Bharat GPT, focuses on AI with purpose and trust, works with enterprise partners on AI solutions
– Akanksha Swarup – Moderator/Host conducting interviews and panel discussions
– Nitin Gupta – Works at Google, manages solution and engineering for India in AI/cloud infrastructure, focuses on sovereignty and innovation
– Sudeesh VC Nambiar – Works with IRCTC (Indian Railway Catering and Tourism Corporation), deals with railway ticketing systems and AI implementation
– Srirang Deshpande – Part of strategy for India, managing Vertiv strategy and market development
– Peter Panfil – Senior Vice President for Technical Business Development at Vertiv, based in US, involved in large scale data center and gigawatt designs
– Jigar Halani – Works at NVIDIA, manages solution and engineering for India, focuses on AI ecosystem, infrastructure applications, and use cases
– Sanjay Kumar Sainani – Senior Vice President, Technical Business Development at Vertiv, 35+ years experience in leadership roles across multiple regions
– Srikanth Cherukuri – Works on large-scale AI factory implementations and data center deployments
– Moderator – Various moderators conducting different sessions and managing discussions
– Audience – Multiple audience members asking questions during Q&A sessions, including Ani, Shlom, Dal Bhanushali
Additional speakers:
None – all speakers mentioned in the transcript were included in the provided speakers names list.
Full session report
This comprehensive discussion brought together industry leaders from NVIDIA, Vertiv, Google, and Indian government organisations to examine India’s rapidly evolving AI infrastructure landscape and the country’s ambitious journey towards AI sovereignty at gigawatt scale. The conversation featured two main panel discussions that revealed a fundamental transformation occurring in how nations approach AI development, moving from traditional cloud computing models to purpose-built “AI factories” that generate new data and outputs in real-time rather than simply retrieving stored information.
India’s Strategic Position in Global AI Development
The opening panel discussion featured insights from Ankush Sabharwal (CoRover), Nitin Gupta (Google), Sudeesh VC Nambiar (IRCTC), and Akanksha Sharma (MeitY), who collectively painted a picture of India’s unique positioning in the global AI landscape. Ankush Sabharwal made the bold assertion that India would become “the hub of AI development for the world” within months rather than years, citing the country’s aspirational nature and readiness to adopt new technology.
This optimistic projection was supported by compelling data points, particularly the observation that India generates 20% of the world’s data whilst possessing only 3% of global data centre capacity. This stark disparity illustrates both the challenge and the enormous opportunity facing India’s AI infrastructure development.
The speakers emphasised that India’s position as a consumer nation—being the largest ChatGPT user base globally and ranking highly for other AI platforms—creates a unique foundation for AI sovereignty. Rather than viewing sovereignty as protectionist, the discussion reframed it as an efficiency optimisation, with the argument that the most efficient and effective place to process data is at the source of the data. This perspective suggests that local processing isn’t merely about regulatory compliance but represents optimal resource utilisation and performance enhancement.
Google’s Nitin Gupta reinforced this collaborative approach to sovereignty, emphasising that “sovereignty and innovation must run together” rather than being competing choices. Google’s strategy includes building data centres in India whilst providing indigenous solutions for critical data, including indigenous data solutions that remain within customer premises whilst being fully powered by AI capabilities.
Real-World Applications Demonstrating Economic Impact
The discussion included compelling examples of AI applications already delivering significant economic benefits in India. Government implementations showcase the practical value of AI infrastructure investments, with fraud detection systems for subsidised food distribution saving “couple of millions of dollars per day” through automated quality and quantity verification calls to citizens. Additionally, UPI transaction fraud prevention systems demonstrate the immediate economic value of AI infrastructure investments.
The IRCTC example, presented by Sudeesh VC Nambiar, illustrated the complexity of managing AI workloads during peak demand periods, such as tatkal booking times when traffic dramatically increases. He described this as a “constant cat and mouse game” requiring advanced AI solutions to combat automated booking tools whilst maintaining service availability for legitimate users. The railway booking system faces unique challenges where legitimate users compete with sophisticated bots, requiring continuous AI-powered countermeasures.
The Bhashini multilingual translation service represents another significant application, processing 100 million requests per hour across India’s 22 official languages and numerous dialects. This service alone demonstrates the massive compute requirements for real-world AI applications at national scale, requiring substantial data centre infrastructure to support its operations.
Ankush Sabharwal’s approach with Bharat GPT emphasises practical, domain-specific solutions rather than attempting to create universal AI applications. His philosophy of “AI with purpose and trust” focuses on working with enterprise partners who understand their specific domains, allowing AI companies to provide technical infrastructure whilst leveraging existing domain expertise. The discussion also highlighted India’s development of 10 foundation models as part of the broader AI sovereignty initiative.
Unprecedented Infrastructure Scaling Requirements
The technical discussions revealed the massive scale of transformation required for AI infrastructure. Projections suggest that India will cross 10-12 gigawatts of AI infrastructure capacity within three years, significantly exceeding some industry reports. This projection is driven primarily by inference workloads, reflecting India’s role as a consumer nation that must process its own data locally due to emerging data protection regulations under the DPDP law and sovereignty requirements.
The infrastructure demands represent a fundamental shift from traditional data centre design. The speakers noted that whilst conventional cloud racks operated at much lower power densities, current AI applications demand dramatically higher power per rack, with future generations potentially requiring even more substantial power infrastructure. As one speaker noted, “A few years ago, the whole data centre was one megawatt… Now this has flipped.”
This dramatic increase in power density necessitates liquid cooling systems and purpose-built facilities rather than retrofitted traditional data centres. The speakers emphasised that attempting to retrofit existing infrastructure for AI workloads presents significant challenges, particularly at enterprise level, where the complexity of implementing high-power and liquid cooling systems can be “pretty scary” for enterprise CTOs.
Revolutionary Design Philosophy: Chip-to-Grid Thinking
A central theme throughout the technical discussion was the need to fundamentally rethink data centre design philosophy. The speakers advocated for abandoning traditional “grid-to-chip” thinking in favour of a “chip-to-grid” approach, starting with GPU requirements and designing infrastructure around compute needs rather than the reverse. This represents a complete inversion of conventional data centre planning methodologies.
The speakers stressed the importance of reference designs and standardised approaches to achieve this transformation efficiently. NVIDIA and Vertiv have collaborated on reference designs that can support multiple generations of GPUs, allowing for seamless transitions between different compute platforms without requiring complete infrastructure overhauls. These designs focus on pod-level thinking rather than individual rack optimisation, creating “bounding boxes” that can accommodate future technological evolution.
The importance of future-proofing was emphasised, noting that infrastructure investments must accommodate the rapid pace of GPU evolution whilst maintaining economic viability. The goal is to create designs where transitions between GPU generations require minimal infrastructure changes, with “all you have to do is reconfigure the cabinets, nothing else, everything else stays the same.”
Speed at Scale: Compressed Deployment Timelines
The discussion revealed a dramatic compression in project timelines, with deployment schedules moving from 18-month cycles in the cloud world to 4-6 months in the GPU world. This acceleration is driven by economic imperatives, as expensive GPU investments—often worth billions of dollars—require rapid monetisation to achieve acceptable return on capital employed.
The concept of “speed to token” emerged as a critical success factor, emphasising that infrastructure must be deployed quickly to begin generating revenue from AI workloads. This urgency has led to innovations in prefabricated systems and off-site integration, moving testing and complex assembly work away from data centre sites to controlled factory environments.
The speakers highlighted that India can benefit from lessons learned in other regions over the past 18 months, avoiding costly mistakes and accelerating deployment. This knowledge transfer represents a significant advantage for India’s AI infrastructure development, allowing the country to implement proven solutions rather than experimenting with unproven approaches.
Technology Integration and Sovereignty Framework
The discussion provided a framework for understanding AI sovereignty through a five-layer stack: energy, infrastructure, compute, models, and applications. India has made significant progress in four of these layers, with indigenous chip manufacturing representing the remaining challenge for complete sovereignty through initiatives like the Semicon mission.
NVIDIA’s contribution to India’s AI ecosystem includes sharing reference designs for AI factories, open-sourcing control plane technology for local inferencing, and providing the technical foundation for numerous AI applications. The partnership approach demonstrates that achieving sovereignty doesn’t require complete isolation from global technology ecosystems, but rather involves building local capabilities whilst maintaining beneficial partnerships.
The speakers emphasised that sovereignty and innovation must work together, with local processing capabilities being developed through collaborative approaches that leverage global expertise whilst building indigenous capabilities.
Energy Efficiency and Sustainability Considerations
The speakers addressed energy efficiency as a critical component of sustainable AI infrastructure development. India’s advantage in renewable energy generation—with significant portions of current energy production from green sources—provides a strong foundation for sustainable AI infrastructure scaling.
However, the discussion revealed complexities in measuring and optimising energy efficiency in AI environments. Traditional PUE (Power Usage Effectiveness) metrics were criticised as potentially misleading, with simple temperature adjustments to improve PUE calculations potentially increasing total power consumption by forcing server components to work harder.
The speakers advocated for more sophisticated approaches to energy optimisation, including thermal and load cycle management, and the integration of telemetry from chip level to data centre level. Purpose-built AI factories could enable automated optimisation without human intervention, moving beyond traditional risk-averse manual control room operations.
The emphasis was placed on maximising power delivery efficiency from source to GPU to optimise compute performance per watt consumed, with every watt saved at the source eliminating the need for generation, distribution, and heat rejection.
Workforce Development and Implementation Challenges
The discussion acknowledged significant challenges in scaling the workforce required for AI infrastructure operations. Specialised skills for operating advanced cooling systems and high-density environments cannot be easily taught in traditional educational settings, creating a potential bottleneck for infrastructure scaling.
Comprehensive skill development programmes have been initiated, including extended training courses for data centre operations and maintenance. These programmes cover both off-site and on-site training components and are being made available through online platforms.
The speakers suggested that prefabricated systems and standardised designs could help address skill gaps by moving complex integration work to controlled factory environments where specialised expertise can be concentrated. This approach reduces the need for highly skilled technicians at every deployment site whilst maintaining quality and reliability standards.
Economic Implications and Investment Requirements
The discussion revealed the massive financial implications of AI infrastructure development. Individual data centres may cost hundreds of millions of dollars, but the GPU investments within them often represent even larger investments in compute hardware. This inversion of traditional data centre economics—where the IT equipment significantly exceeds the infrastructure cost—fundamentally changes investment and deployment strategies.
These investments must be monetised quickly to achieve acceptable returns, driving the need for compressed deployment timelines and reliable operations. The cost of downtime in AI environments is dramatically higher than traditional computing, with GPU node failures potentially costing substantial amounts due to the need to restart training workloads from previous checkpoints.
Future Outlook and Remaining Challenges
Whilst the discussion painted an optimistic picture of India’s AI infrastructure potential, several challenges remain. The timeline for indigenous semiconductor manufacturing capabilities requires further development, and the regulatory framework for data sovereignty under the DPDP law requires continued clarification as it comes into enforcement.
The integration of existing traditional data centres with new AI factory requirements presents ongoing challenges, particularly for organisations seeking to retrofit existing facilities rather than building purpose-built AI infrastructure.
The speakers acknowledged that the rapid pace of technological change creates ongoing challenges for infrastructure planning, with equipment potentially requiring updates within short timeframes. This reality necessitates careful consideration of future-proofing strategies and flexible design approaches that can accommodate multiple generations of compute technology.
Conclusion
The discussion revealed India’s unique position at the intersection of massive data generation, growing technical capabilities, and supportive policy frameworks for AI development. The convergence of these factors, combined with lessons learned from global AI infrastructure deployments, positions India to potentially achieve its ambitious goal of becoming a global AI hub.
The transformation from traditional data centres to AI factories represents more than a technical evolution—it embodies a fundamental shift in how nations approach digital sovereignty and economic development in the AI era. The speakers’ consensus on the importance of speed, scale, and sustainability suggests that India’s AI infrastructure development will serve as a model for other nations seeking to balance sovereignty with innovation.
The success of this transformation will depend on continued collaboration between global technology providers and local capabilities, sustained investment in both infrastructure and human capital development, and the ability to maintain the rapid pace of deployment required to capitalise on the current window of opportunity in global AI development. The real-world applications already demonstrating economic impact provide concrete evidence that this infrastructure investment translates into tangible benefits for citizens and the economy.
Session transcript
having the complete sovereignty in terms of AI and not just the platform. I think India being so aspirational and ready to adopt new technology for the welfare of themselves and the welfare of the businesses, I think we would be the hub of AI development for the world. You will start seeing that happening in a few months, not years.
It’s actually heartwarming to hear that from someone who’s actually fronting India’s AI story at the moment. Nitin, as someone who is at Google, how do you see this for India? Do you think India has the right infrastructure, the right resources to build its own sovereign AI at the moment?
First of all, thank you, Corvo team, Ankush, for inviting me here. And, you know, I’ll be very happy to share my views. from Google perspective and from my personal perspective I feel yes sovereignty is very important but at the time with the sovereignty it is not a question between sovereignty or innovation it is sovereignty and innovation they have to run together they can’t be one choice versus the other one and with that Google while we have our entire data centers in India you have heard three months back we announced that we are going to be building big data centers in Vizag the announcement happened so we are ensuring that let’s say if any innovation and any data residency things are there they are being kept within the boundaries of India But then those data centers are definitely empowering the lot of AI, but they are for everyone, for all type of personas, whether they’re government, enterprises, startups, students, colleges, universities.
We understand that, you know, sometime there are going to be critical data which needs to stay even more secure. And for that, Google has created a completely indigenous data box which completely stays inside the customer premise and is fully powered by AI. So imagine that you have the full potential to run what you’re running in a Google data center, but inside your own premise. And that has full Google Gemini AI services. And that’s the definition we have for sovereignty, where you are. Also. Also controlling the hardware, not only what’s running on that hardware.
All right. So this IRCTC is one of the most heavily used websites in India. my data research says close to 50 million users visiting every month on an average correct me if I am wrong but how are you incorporating or leveraging the use of AI especially in peak periods when you look at say tatkal booking time when the traffic actually dramatically peaks up
yeah so we had tremendous mismatch of demand and supply as far as railway ticketing is concerned so we have the peak the morning 8 o ‘clock when the ticket is opened for the 60 days hence travel then 10 o ‘clock the AC tatkal and 11 o ‘clock the sleeper tatkal so there is a lot of huge demand and there is a demand supply mismatch as of today so people try to misuse and use the automated tools for accessing it So this is a constant, I would say, cat and mouse game we are sort of playing. And we are using AI also. We are using AI of very advanced AI solution. Maybe they are said to be the best in the world solution we are using
Any indigenous models are used?
Indigenous, of course, we have a layer of indigenous. There is a startup also who are doing the analysis, data analysis, and they constantly monitor the social media and see what is happening, what is the strategy. So it is basically a collaboration between the Indian startups and the global technology strength of a global company. So we are using AI and ML -based model. The model constantly learns and tries to… mitigate those automated…
Okay. Ankush, what differentiates Bharat GPT in terms of its vision when you compare it to say global models like ChatGPT or even Gemini and especially how is it curated for Indian citizens and enterprises? What is that differentiating factor?
Yeah, see, our tagline is AI with purpose and trust, right? So whatever we are doing, so I had read that book, Seven Habits of Highly Effective People very early on in my career. So begin with end in the mind. We always think what’s the use case? What’s the problem you’re going to solve? And then see what kind of model you need, tiny, small, medium, large. And then you see, okay, from where the data would come. See, the Bharat GPT family of models, right? It’s not the large language model. It’s not ready for consumers yet, right? So we work with our partners, get their data and train the model for their users because we believe we… is easy for us to solve the problem of enterprises because the enterprises say like IRCTC, they already know their domain.
We cannot learn, right? And if you say, hey, I can create travel AI solutions, very, very difficult, right? So they know travel, they know railways. So it would be, I think, much better to work with them, learn from them. They already know, they are already solving a lot of problems and they also know the problem, the real problem. They don’t have existential crisis, right? So they are not just in the game of valuation. So they are solving the real world problem.
That’s why we have him on stage with you today. He’ll share those precious tips. My last question, since we are running short of time, Nitin, I think this is also not to highlight the achievements, it’s also to perhaps highlight the concerns. And right now, one concern which the Indian Prime Minister has also highlighted is that of inclusivity. How is Google trying to bridge that divide as far as you can see? As far as digital divide is concerned, how do you make Google more accessible for the underprivileged, for those in rural areas? Nathan, before you answer, it should be shortened. I have my colleagues from other team, Vertex. I would like to apologize to them for this delay, but allow us just to wind this up.
Yeah, I’ll take a minute. Okay. So, great question. And, you know, Google has always been, you know, in the forefront of inclusivity, whether you call it Gmail, whether you call it search. You know, it is empowering billions of users every day. And just to summarize and give a recent example, we have very recently Sundar Pichai has announced that the JEE main exams, mock exams are available on Gemini free of cost for any student to try. That’s the inclusivity we want. We want to make sure that student at his home can keep on trying the mock test at free.
All right. Amazing. Amazing. Which is inclusive. Inclusive and democratic. Many thanks to you three gentlemen. It was a pleasure having you all over here. Thank you so much.
Good morning to all of you. As Rakesh has already introduced, two companies are planning for a lot of things together. As I said, I am part of strategy for India and managing Vortiv strategy and market development. The important thing, what we are bringing today for you is, as we see a lot of gigawatt infrastructures are getting announced, and that poses a lot of challenges for us. Till this time, data centers are getting built from outside in approach. and then now time is there or time has come where data centers are getting filled from inside out approach. So it’s first GPU gets decided or the workloads get decided and then the whole infrastructure gamut comes into the picture.
To discuss this, I have two friends, two industry veterans from Vertiv and NVIDIA to discuss the Fireside Chat. So we have a Jigar, I think by this time Jigar is already known to the industry because immense contribution Jigar has done for the AI ecosystem in India working with all the ecosystems, all the layers, infrastructure application, use cases and so on and so forth. He managed solution and engineering for India in NVIDIA and I have another my friend Peter Panfil . Peter is Encyclopedia in Vertiv. He is based in US. He is our senior vice president for technical business development. And he’s the one who’s involved into many designs, a large scale data centers and gigawatt designs.
I would request Jigar and Peter, please come on the stage. Let’s have a round of applause for Jigar and Peter. So, Jigar and Peter, it’s all yours now. Go to
Thank you. Thank you. Thank you. So, my friend, we got our introductions. Let’s see. Are we on? Are we on? You guys can all hear us? Yeah? Can you hear us? Good? We’re good? Okay. All right. So, my friend, great to see you. Great to see you. So, I got to start with how we would normally end. I believe that any discussion like this should start. with us telling you what we think you’re going to get out of this. So what key message or messages do you think this audience needs to hear before we get started? And then we can spin off of that and go into the kinds of details we really need to.
So where do you think, what do you think this audience is the most interested in?
Okay. Am I audible? Okay, great. So I think as the topic suggests as well, my view, what you will get to hear us next 30 -35 minutes is about why this AI is becoming so much of notion for every country. What is that is the building blocks of these AI factories and the sovereignty aspect of it? What is it two of us are trying to contribute in this journey for everyone for that matter? And how do we scale? and make it work for everyone, to make AI for all, how India wants to call it as, AI for all, is what I feel we should be discussing about here. Because that will be most relevant for the conference, for the audience, and what we can contribute back to the humanity as well.
What are your thoughts?
I agree with you completely. So the three things I feel are most relevant are speed at scale. Now, it’s not just the speed of the compute. It’s the speed of deployment. The faster we can get the GPU structures in place, the faster we can benefit from it. And scale, you and I talked about the scale. And you’re going to quote some numbers, I think, that the tops of their heads are going to blow off. But speed at scale. The second thing is we’ve got to stop. We’re not thinking the way we thought in the cloud world. In the cloud world, we were thinking a high -density rack was 10 kilowatts. And that we would start at the source, at the grid, and work our way to the chip.
What I’m here to advocate for you to do is start at the GPU. Start at the chip. Let’s start at the chip, define the most economical, most efficient, fastest from a compute perspective, and figure out how to deploy that as a pod, then replicate that pod, and achieve the speed. And the third is, don’t be scared. We got it covered. We got you covered. We know how to do this. This, we made a big, I got to just tell you, I told you this in the hallway. Vertiv made a big bet. with NVIDIA. We made a big bet. I actually reassigned myself. I was leading what we call a GSA, Global Strategic Account Pursuit Team, and I said, if we’re going to do this right, we’ve got to immerse ourselves in GPUs, understand how to deploy them, understand what drives our customers, and how we’re going to make them successful.
And I think that that has worked to both of our benefits.
Absolutely. And through the humanity as well, right? We are fundamentally changing everything that has been pursued so far, and you bring out the cloud part of it. I was just thinking while putting my hand on my beard that only a few hairs were white back then. It’s not that far that I’ve seen the retrieval clouds. We store the information and we’re just retrieving it to process the application to get us the information out, right? to the world of now generating every single time a new data and processing it right there to give you all the time a new input and a new output, right? Because prompts are new, the outputs are new, and thereby the world sees every time something different which is getting processed and being delivered to the customers, right?
So such an amazing and a fastest -paced change of how these clouds have emerged and what are your thoughts in terms of what this space is all about, how our customers are keeping up with this, and what are we contributing in that journey, if you can throw some light towards that.
Sure, that’s great. So first of all, it comes with understanding and having a transparent provider that says, here is what I’m producing today, here’s what I think I’m going to be producing a year from now, here’s what I think I’m going to be producing two years from now. Now, our goal is to make every deployment that you take on an AI factory. We all know what an AI factory is, right? An AI factory, think of it as a car factory, washing machine factory. Just, it’s a data factory, okay? And so our goal, I will just tell you, our goal along with your team is start as an AI factory. Yes, you might want to have mixed mode CPU and GPU workloads in your facility, but you’ve got to pilot the GPU configurations, at least pilot them.
When I say I reassign myself, I was working primarily with cloud providers, mostly hyperscalers, and they had a prescriptive formula. You know, they had their hacks, their number of racks. They would deploy them. We all knew which ones they were. Now, we can take a GPU pod, design it once, build it many, and apply it to the GPU that we need from that generation. It’s a complete change in the way we think about how to deploy the IT.
That’s so true. By the way, did you notice, every time we are talking about GPU, the screen is blinking. There you go. I think that’s a good message.
I think it’s because I owe somebody a nickel every time I use the letters GPU. It must be trademarked somewhere, all right? So I owe them a nickel. Okay, all right.
No, so I think the transition that we see because it’s generating something new every single time, the compute demand because of which is just exploding, right? And thereby, the possibility of what we could do more and new is every time becoming bigger and better, essentially. Right? And with that, I think the journey of data center is also evolving much more faster than what we have thought, right? You mentioned it, 10 kilowatt to 15 kilowatt, not that far. We were talking about this about four or five years back. To 40 kilowatt, what we transitioned it to it, to now to 120, 130 kilowatts. And as we announced it in January, we are now talking about 240, 230, 210 kilowatt per rack, which means this size hall could probably run a great portion of India with so many services that is probably never imagined before.
So I think it’s interesting that you comment about that, because one of the things that we’ve heard back from our customers who first do a lot of research, how do they take their critical infrastructure from CPU -based to GPU -based? And I think that’s something that we’re seeing a lot of growth in. First, there’s that transition to liquid. Don’t worry about it. We’ve been doing liquid cooling for 40 years. We know exactly how to manage it. Then there’s the density of the compute itself. I’m amazed at how quickly and easily our customers understood the move from a 10 -kilowatt rack to a 130 -kilowatt rack. I credit you all. So if you’ve already made that transition, I credit you.
You’re doing a spectacular job. Our job is to prepare you to have that go up by an order of magnitude. Not right away, but in future generations of compute. And so what we try to do is we try to prepare you for future -ready thinking. I know you don’t want to think three years down the road. You can do it. You can do it. You can do it. You can do it. You can do it. You can do it. let’s at least think three years down the road in three years based on the rate of what you’re seeing what we’re seeing both here in India and around the world
my perspective is I think all reports are talking about 5, 6 gigawatts kind of a number over the next three years my personal understanding from the lens I look at it both from NVIDIA as well as what industry and government is trying to do my anticipation is we will cross 10 to 12 gigawatts in the next three years and that’s not far and I’m not going by any of the announcements that has been made in the last three years. I know where the reality stands in terms of what inferencing and training workloads. I repeat, I started with inferencing. I did not start with training.
Yep. I noticed that.
The reason is we are a consumer country. Make a note of that. Yes. Right? He started with inferencing, not learning. Yes. All right? Because we are a consumer country. We have always been in the mode of first to consume, then to build. And thereby, we are the largest chat GPT consumer base for the globe. We are the largest for public city. We are the largest for even for Gemini as well. Right? I think we were number two about a month or so back. But my view is with this geo announcement, we should have crossed by now number one position. But the delta was pretty small. Right? What does that mean? That means. If that entire compute capacity.
that is currently not getting processed in the country should come back to India because of the DPDP law that has got enforced last month or so, then this number will be even higher. And we are very democratic that way. You know, we are not closing the doors for any businesses. We have never done that. I’m sure, knowing the country, we will never do that with this leadership that we have from Prime Minister Modi. That means we will still allow these processing to happen outside of India, but at the same time, we will do the regulatory reasons of some of the verticals, say, fintech, say, healthcare, defence and so on and so forth, or some of the government, you know, bodies.
Even if they start to do influencing locally, this number will easily touch 10 plus. And I’ve not included the industry at scale yet, which is what Anthropic and J &J, Gemini and others are trying to even capture it from that market perspective. So my understanding, it should cross 10. while all the reports are talking 5, but India will
So it’s amazing. We didn’t compare this note before we got on this stage. What was the number you gave me just 20 minutes ago? 10, right? So let’s think about that just for a second. We’re at 1 .5 now. We’re going to get to 10. So to get to 10 in that three – to five -year horizon, we’re going to have to scale pretty far, pretty fast. We’re going to have to draw on our shared expertise. And by drawing on our shared expertise, we’re going to help be a trusted advisor to you. Who’s your trusted advisor? I’ve got my trusted advisors. Somebody. Somebody I can always go to, and they’re always going to give me.
the right answer. It might not be the answer I like, but they give me the right answer. So what we want to do is make sure that you know, we understand how to scale. You understand how to scale. We understand how to scale. I think if we’re talking about doubling, getting to 10 in three years says we double every year starting this year. My one and a half goes to three, my three goes to six, my six goes to 12. We’re doubling every year. Now, if I was to take you outside of India now, North America market, when the North America market first started becoming aware of GPUs, there was a wide variety of acceptance.
There were the folks that said, yep, I want to be there. And I want to be there, and I want to do a pilot with you. I want to design a pilot that I can replicate into all of my either hyperscale or multi -tenant data center environments. The other thing they wanted to do was no data center left behind. They didn’t want to leave behind any capacity because they knew capacity was going to be the currency. They knew power and land and GPUs, that’s where they needed to be. The third thing was their project scales moved. We used to live in the cloud world at project scales of 18 months. We live now in the GPU world of project scales of between four and six months.
So a dramatic compression of schedules, a dramatic increase in capacity, what does that mean? We’ve got to build capacity at a faster rate. Than we ever have before. and I know we’re up to it. We’ve added the capacity we need to be able to support that kind of demand.
Peter, that actually brings to a very good question. When we talk about this at scale, and you said that in the U.S., you guys have already started to build this at scale because you see this as a great opportunity, essentially, and India is yet to build, right? In all fairness, I think some of the largest clusters are in 10Ks of GPU, essentially, right? While in the U.S., we’re talking about millions of GPU in a single data center, essentially. Would you like to throw some light on some of the learnings and, Kodi Kakar, a quick bite for the audience to know what are those quick things that India could do in terms of having these things done in, let’s say, three to eight months’ time frame, not just the project planning, not just the understanding of BOQs, not just the understanding of… who is going to deploy my project and how does the project look like and the 3D version of that.
How do I get the entire project done in, let’s say, six to eight months’ time frame? Including from land is what I have, and from there onwards, GPUs running and hugging and making the production environment happen.
shifted to 250. Now, along the way, we said, okay, let’s take these 10s and put them together and make a 50, and let’s take the 50s and put them together and make 100, and let’s put the 100s together and make a 220. Shoot me now. What we found is, let’s pick an optimum building block that supports the number of GPUs that is, I’ll call it, reasonable at scale, don’t take a design that has never been created before. Let’s take a design that we have a good basis on. For example, the pod. You just published some standards on pods. Reference designs.
Reference designs, okay.
We worked closely with your team on reference designs. We came up with the magic numbers that are reference designs that minimize the amount of, of underutilization, so maximize the utilization and make them the most efficient. I’ve been an advocate for efficiency within the data center space my entire career if you save a watt that’s a watt you don’t have to generate at the source you don’t have to distribute it you don’t have to reject it so the fewer watts you lose and the more watts you can put into the compute the more tokens I can generate and so our goal our goal in working with the GPU I’ll call it the AI factory mentality is how much power can we deliver from the source to the GPU as much power as we possibly can and how can we deploy that physically as quickly as we possibly can and it boils down to take the reference designs we’re not saying all the designs are going to be the same we know that’s not going to be the case so but I could show you a pod design it’s part of the reference design I could show you a pod design that supports three generations of GPUs so this year, next year, next year after that three generations of GPUs just by changing the way those pods are populated on the compute side and in fact we’ve got one customer who wants to be able to seamlessly mix GPU platforms within a pod he says I’m going to have one compute line up number one as one generation of GPUs, pod two is another generation of GPUs third generation of GPUs, so they want to be able to seamlessly move between GPU generations because at some point they’re going to optimize particular functions and particular outputs and services against a GPU platform.
You just brought up a perfect point, right? So a few things are why it’s important to follow the reference design. Just to bring everybody on the same page, CPU world was very different. Having a node down means few hundred dollars getting downtime. A GPU node down translates to few thousands of dollars going into a downtime, right? And the fortunate or unfortunate part is if your training workload is running, if a node fails, you start from the checkpoint that you have done it. Assume that your checkpoint was done eight hours before. For eight hours of, say, 4 ,000 GPUs of time multiplied by that much is what you have lost the compute time in the cloud.
Unfortunately, that translates to…
Real money.
Hundreds of thousands of dollars. Real money. Right? Real money. So while you as a cloud provider might be thinking, and I’m talking about both the sides, that, hey, let me do a little bit of cut corners, do something here, something there, and I’m still making the cluster up and running. But you know what? That’s going to cause a lot. And customer may not have SLAs with you in that direction because these are not the standard SLAs we’re talking about, right? What the world has seen in the typical cloud world. These are different type of SLAs that customer signs with you. And, you know, if it’s an inferencing workload and if it’s critical with the enterprises, we’re talking about down times, which is, again, by all law of cloud, is not acceptable.
But the key question could come in that, hey, why do I need these large -scale clusters only for training? Is that the only thing I do it? The answer is no. I don’t know, but I’m sure most of you might be following what Jensen talks about it. The three scaling laws that we have it. We’ll not go. We’ll go into detailing of it. I think Jensen has mentioned it like at least 100 times of his keynote. But in a simple term, if I have to tell you, let me take one or two good examples from the country itself, right, and what we announced in the last three days. So, taking a very simple example, as everybody knows, we are 1 .4 billion people, right?
Half of the audience is associated with farming in the country. Half of the audience or the, you know, citizen base is associated with the, you know, farming, and thereby one -third of the families of the country are completely aligned to the farming aspect of the story, right? They contribute just 15 % of our GDP, but half of the population does and associates with the farming, right? Now, government of India has two simple applications that has been launched, right? One is to check the subsidized… Food, you know, that government gives it to these half of the, you know, audience, half of the citizens in the country today. subsidized to the level which is a cent or two.
In Indian rupees, it is one rupee to five rupee, how government gives it. And a feedback call goes to all these citizens, asking how was the quality, did you get the right quantity, have they done any kind of fraud, and so on and so forth. A call per day, if government has been able to scale in the last one month to about 50 ,000 calls a day to citizens through a bot, which is talking in a local language, has been able to save a fraud work of around per day, and I’m talking about per day, in the range of a couple of millions of dollars. Okay, take that fraud. Talk about financial fraud. This would be another one, right?
Because we are the world’s largest online payment transaction country. We contribute 50 % of the digital, and that’s… by the NPCI data, globally acceptable, and it does at free of cost in this country. We call it as UPI, right? Most of the Indian people would know about it, right? And imagine the innovation that takes place in the fraud, you know, when the UPI transactions are taking place, right? And I do this transaction using mobile, from your mobile to your mobile, in a fraction of milliseconds. That data is in hundreds of millions, right? To prevent this fraud is where the AI is getting used. Now, if I’m putting a couple of hundreds of millions of dollars for five years as an initial investment, think of the economic benefits and the money back that I’m giving it to the citizens by not having these frauds.
And thereby for each of these applications, right? I have another good example, which is we have 22 official languages speaking in 500 dialects in the country, unofficial languages. So, officially, we have over 100 plus languages in the country. right government of India has an application called bashingi which does basically the translation you know and ASR and TTS in all different languages of India and government of India and state government has about 10 ,000 websites that government runs it we have only touched 1 ,000 and we are already hitting 100 million requests per hour right and this translates to in a simple term roughly about 2 million 2 megawatts of data center consumption per minute right in in 2 megawatt if I’m able to cater to 100 million requests a minute that’s massive
yeah look at the productivity improvement and I’m bringing it to the nation’s and this is just thousands of those websites of government of India so we yeah so we we talked we started with scale at speed Okay. That’s where we began. It’s not just the scale of the data center environment. It’s the scale of the applications and the benefit that they’re going to bring when they get fully populated.
Absolutely.
So I’m going to put you on the spot. How much do you think, on the journey right now, where are we? Are we at 3%, 5%, 10 %? I will tell you, I cannot wait for AI to take every mundane task I have to do in every day of my life and just do it. Okay? And then once those mundane tasks are out of the way, I can use every gray cell up here for productive work.
Absolutely. Absolutely.
So where do you think we are in terms of that scale?
But you touched upon a good point, how Meta calls it as personalized AI for everyone. So we are getting there, right? But in terms of data, and I think even Minister made that announcement yesterday when he was talking at the inaugural. He gave a nice statistics talking about we as India generate 20 % of world’s data, and what data center capacity the country has today is 3 % of the world’s data center capacity. So which means even if I don’t assume data generation speed over the next 3 to 5 years, even if I keep the data rate at 20 % only, and we are a young population, so we are bound to generate more data, and ours is the cheapest data rate in the world for 5G that we have, but assume that we don’t.
We don’t generate much of data, right, and we restrict it. We still have a long, long way to go in building the… large -scale data centers just to make sure our own data we process it by ourselves. Right? And that’s where the whole theme of this sovereignty, what government is talking about at least let’s protect our data. It’s more critical, not the general data. And that’s where the gigascale is more important.
But I don’t look at the sovereign data center approaches so much as a protection. I look at it as, where’s the most efficient place to process the data? It’s where the data is generated. The most efficient and effective place to process data is at the source of the data. Absolutely. And we are limited by energy. So we want to protect that layer as much as possible. And so I see a world where the data gets generated, it gets processed as closely and as quickly after the generation of the data. It’s used to further improve the performance and generation of subsequent data. So that data gets cleaned up as it goes. It gets more refined and more accurate.
We all know we make good decisions with good data. We know that. We make bad decisions with bad data. So the real issue here is we’ve got to take the data. I won’t say that our data is not clean now, but it’s not. Okay? I mean, you spend, give the audience an idea when a model is being put together. How much of the time is actually in cleaning the data and pre -processing of it, and how much of it actually goes into the language model itself?
So just to build, because India just announced 10 of their foundation models, cleaning of data typically is three to six months of journey on thousands of GPUs. for a language model that we are trying to build it in, right? If it’s a specific model for a particular task or a vertical that we are trying to build it in, and if the data is more notorious in terms of having more videos and images and stuff, it could be even longer.
Got it.
Right? And then comes the foundation model building itself and, you know, conversion of the model. That’s another 6 to 12 months of journey. It depends on the model size and type that you are trying to build.
So it could be a third of the time of realizing my language, my large language model, is cleaning of the data.
That’s correct.
Processing that data. Now, once it’s there, I’ve got a solid foundation of data to use for future models.
That’s correct.
Okay. So, again, I think if we’re talking about it in terms of percentage, are we 5 % there? Are we 10 % there?
So I would not put percentage. The reason is what type of model we are trying to build it in. depends on that. If it’s a language model, I think specifically to India per se, I will not comment about other countries because it depends on where their journey is in their data building. But India, in my view, has already nailed the data creation for a mid -sized model, how we call it, as a small to a mid -sized model in play. And they’re going to make it open source as well, how it has been announced. So, I will not claim it that we have a very large data set for a very large model, but a small to mid -sized, I think in the last one, one and a half years due to this India AI mission, we have been able to generate a pretty good amount of data and a pretty amazing clean data.
Perfect. Alright. So, I’m getting a hook from the guys in the front row. Okay. Yeah. I run long. I’ll always run long.
Sorry, I’m pausing you there. And I want to diverge a little bit right, asking as an India person. and I’m an Indian first, then I’m an Indian, what is that Vertiv is trying to contribute in this journey for, let’s say, India to begin with, and for the globe as well, you know, in the building blocks of these things that we are trying to do for these gigawatt -scale data centers? If you can throw some light.
Sure.
I know it’s a little bit silly question.
No, it’s not a silly question.
No, no, it’s not. We want to push manufacturing. We want to push, you know, India ecosystem, be more indigenized as much as possible, be more self -reliant. I want to know what Vertiv is trying to do.
So Vertiv is investing in people, in process, in production capacity.
Amazing.
Our goal, our goal is to build as much of the critical infrastructure here in India as we possibly can. And what that, it starts with. Working with our partners and our customers on first pilots and then production. So that production, you’re going to benefit. I will just tell you, India, you’re going to benefit from the mistakes that have been made in other regions for the last 12 to 18 months. So you’re going to benefit from that. You’re going to be able to jump right to it, all right? All right, so here’s the sum up. I asked you to give what you thought the panel should, this discussion, they should get out of it. What should they have heard from us that you want them to keep in their minds for the rest of the day?
For the rest of the day?
Rest of the day.
My view is, and I know it’s going to be a mix of audience here, you should be listening to it. How are these building blocks of AI factories that are getting to learn from the globe that India could adopt it fast, followed by? Followed by what’s happening in the modern world, because that’s the most and the fastest. building things that is happening in the world and the most fascinating because changing the world is so fast, followed by how are these models getting deployed, right? And what are the applications which are changing our world on a day -to -day basis? And fundamentally, the businesses are getting challenged on how they have been operating it for decades or centuries, right?
Versus how they could do this business today. If I would be you as audience, and that’s what I’m trying to do, being the audience as well, trying to constantly learn from this conference, what’s that the people who have done this at scale, what can I learn from there that I can deploy back in my country, in my profession, in my day -to -day life as a learning, is what I’m trying to do. And that’s what I would recommend everyone else to do as well.
Perfect. So let me add on top of that. It’s scale at speed. And it’s not just speed of build, it’s speed of compute, it’s speed of adoption. Yes. Second, stop thinking grid to chip and start thinking chip to grid and let the chip help us define what that critical infrastructure needs to look like and the third is we’re going to make it as sustainable as we possibly can because a lot that I don’t waste is one I don’t have to generate transmit or reject alright I think I think you’re up next Any questions? Can we have time to take questions from this? Okay, we have. Can we? Okay. Okay, we have one end up. She’s going to run a mic over to you.
Hi, my name is Ani. I have a question. As I can see…
Use your outside voice. That’s what my family always says.
As I can see, everywhere is AI. And in today’s era, it is totally about AI. So, as you also said that this is a… AI whereas everywhere industry and company and education in every sector using AI. So the day is not getting far once AI humans is totally dependent on the AI and once AI is in the subconsciousness as humans thinking as humans. There is any chance where humans and AI both are in the same niche?
So I think that early on AI got a bad rap. It was going to be the computers were going to take over and blow up the earth. That’s not what we’re finding. What we’re finding is that AI makes our life easier. life better every single day. I know that traffic systems in the city that I’m in now use AI to look at traffic congestion and traffic patterning. And they actually time the lights to improve the throughput on particular roads at particular times a day. Now, that’s where AI is going to really benefit society. It’s going to benefit it in transportation, in medicine, in research. I’m not so worried about the data being used for evil.
I’m really excited about the data being used for good because that’s where I think we’re going to get the most benefit.
True. But what if AI get their own subconsciousness? They don’t need humans to just act.
I wish you see that day. Somebody told me when I started my journey with the phone. That’s what is going to happen. You will lose touch with your family. You will always be busy with the phone. And so I don’t think so. We have even touched that level as a surface, even after having this phone with me for 20 years.
Here’s the example I like to give. Do you think about breathing and blinking? No. You do it automatically. So let’s let AI take those autonomous functions and do them for you automatically so that you don’t have to think about them. And then if I don’t have to think about breathing and blinking, then all of a sudden I can use my brain matter to do other things. So many things. So. I look at it as it’s going to free us. It’s going to free us from the mundane tasks that breathing and blinking. Come on, you’re laughing at me. But do you think about breathing? No. You only think about breathing when you’re trying to hold your breath.
Okay? So I think what’s going to happen is AI is going to become to us like breathing and blinking. It’s going to become an autonomous function that just runs in the background of our lives constantly and makes it better. It’s going to learn what we do and how we do it and how to improve that performance and give us more freedom to do what we really should be doing, and that is making the world better.
Thank you.
Thank you. That’s a good question. I’m glad you asked that question. We have one more. We have one more? Yeah. Hey, hi. We’re going to that side. Hello. Big one.
This is Shlom. I was watching the interview of Mr. Jensen Hong from NVIDIA, and he explained AI as a five -player stack, you know, energy -cheap infrastructure model and application. Which layer do you think, he also explained how US and China are working on different layer and how they are, you know, ahead of us many years in different layers. Which layer do you think India can excel with them or match with them in upcoming years?
So, I think we are already doing that, right? It’s a great question. When we talk about sovereignty, these are the layers we should be sovereign, essentially, right? We cannot be importing energy from anybody. We need to generate by ourselves. Otherwise, how will we run these lights and so many functions and how will we power these data centers, right? So, the good news and I think Prime Minister gave this answer so nicely yesterday in his keynote. He explained, no, the Minister said about it, sorry. He explained these five layer cake once again and I am proud to say he made that statement which we all know it. Half of our energy today is generated which is a green energy, right?
So, that layer is sorted. And I and you have a lesson to learn. We have to contribute more of the companies who have to contribute by having solar, hydro and air and other methods, right? Where NVIDIA is trying to contribute to the nation today is on the top three layers, right? We are helping the nation with AI factories. building it with all the learnings what Peter also mentioned. You don’t have to learn from all our mistakes of last 18 months that we have undergone in other regions because they were ahead. India was slightly delayed by at least 12 months or so. But we have put in all those learnings and the factories have come up way too faster than anywhere else in the world, right?
By all means. The second layer is one of the layer is the serving layer when you build these applications. How do you do inferencing? You’ll be surprised to hear Indian cloud providers never had a control plane, right? We were dependent on other nations to give us a control plane to run these cloud inferencing stack. NVIDIA has open sourced that work and shared that with government of India and that was the announcement that Sarvam did with the product named as how we call it Prava, if I’m not mistaken. I hope I’m pronouncing it correctly. And that layer is now completely owned by government of India and an Indian company to do the entire inferencing locally. right?
And the last piece, which is the application, right? I’m sure you would have visited the booth downstairs on the Hall 5. I don’t think that we have left any booth. Every booth is powered by NVIDIA open source stack that we have given it to build the agent -DIG AI platforms and formulation models. That’s the contribution we have done it for the nation. And India is right there. I think what’s missing, and I will fully agree with that, is we are missing with our own chips, right? And that’s the autonomy that every country is trying to drive across. I’m again proud to say that NVIDIA is fabulous. We don’t produce, right? We outsource that to Taiwan and a few other countries, essentially, right?
We have opened up partnerships in many countries, and we are very open to partner with India as well to give away our technology. Thank you. We will do the modifications and do the manufacturing by themselves. That’s the last piece which is left, and I’m very confident with this Semicon mission. this is going to happen very soon, even if we NVIDIA with somebody else.
Thank you so much. The past year time, we’ll have to get into our next session. 10 megawatt, 12 megawatt, and today we have…
Gigawatts. Gigawatts, baby. Gigawatts. Gigawatts.
I just wanted to leave important information. It took about 8 years time to build one 5 gigawatt. And another 10 gigawatts is going to happen next year. So look at the speed and scale. We both have to work together. And as Jigar rightly mentioned, all 5 layers will have a tremendous opportunity to work. Energy, infrastructure, compute, models, application, and so on and so forth. Huge amount of resources required. Huge amount of support required. And very exciting time ahead. Thank you so much.
And it’s going to be a system approach. System. Systems. Think systems. we as an industry have thought boxes for too long. We think, I got this compute box or that compute box. It’s now a system. It’s a platform. And that platform generates tokens. The new measure should be tokens for watt per dollar.
Absolutely. Absolutely. Very well said. Thank you so much.
He’s one of the guiding principles to implement a lot of large -scale data centers for Vertiv or all the entire ecosystems. Let me welcome Srikanth on the stage. A good round of applause for Srikanth. And another gentleman we have from Vertiv. He’s about 35 years of experience in leadership roles in Europe, Middle East, Africa, India, Southeast Asia, Asia, you name the region. He’s been there for many years. His name is Sanjay Sainani. He joined us as a senior vice president, technical business development. He’s the one strategizing all technical strategies for Vertiv and a business development area. Let me welcome Sanjay on the stage. A good round of applause for Sanjay. And I’ll be asking some questions on behalf of you.
I would also open the floor maybe sometime later. Welcome. Yeah, am I audible? Okay, so let me start, Srikanth, from you. Last question first. First, what is the one learning you want to give it to the audience from your experience of implementation when you build a large AI -scale factories? That was my last question, but I want to ask you first. One piece of advice or experience? Experience, out of your experience, because you already have good hands -on on implementation. So from a sustainability standpoint, implementation standpoint, what is one learning you want to give it to us in India when we’re building a scale of the factories and things like that?
Yeah, it’s an interesting question, right? Like when… One year back or one and a half year back, I came to India to review some data centers. And when I was asked to do that, one of the first things that crossed my mind is, wow, India is building data centers at scale? Because when we were growing up, power used to be a big issue. The reliability of the power used to be a big issue. The availability of power used to be an issue. And when I came here, I was amazed at how far, you know, I have been away from the ecosystem for a little bit, but I was amazed at how far things have come in terms of availability of power and the reliability of that power.
And the second thing I was amazed at is also just the knowledge here in the ecosystem as well as everything related to everything from safety to speed of light construction and the product ecosystem has come such a long way. I think the next step in terms of where India is going in this AI factory build -out is if you look at the U.S., it’s a little further ahead in terms of gigawatt scale and high -density racks, deploying high -density racks, high -density liquid -cooled racks. There’s a lot more experience over there. And I think our combined companies have created that experience. Like I’ve been working with WordUp for the last four to five years in the R &D work, engineering work, and then eventually the deployment work.
So now we have actually matured a lot in what we consider AI factories versus data centers. So there is a lot of advantage for India to draw from that experience. Our combined knowledge pool, again, it’s the same company. Whether you go to Europe or… US or India, it’s still Word of an NVIDIA. It has to be a strong cross -pollination between the ecosystem in the US and here, a strong knowledge sharing. And we are in year two or year three of this AI factory build -out worldwide. And as India is picking up pace in this journey, there’s a huge opportunity to not relearn all those hard lessons, or the hard way, but instead share that knowledge, share, you know, our combined teams share that knowledge and build it much faster here.
That’s first, as a thought leader, both sides we need to do that, we need to equip the market for those kind of things. And let me also tell you, on Vertiv’s side, whatever innovations we are doing in the US, we are real -time bringing to India, so that there’s no latency here, and absolutely whatever is going to happen in the US, we want to bring it to India. That takes me to our next question to Sanjay. Sanjay, we have heard about speed, and so far we have heard about speed of clock. Now, Peter, sometime back, and Jigar spoke about speed at scale. what is your thought process about speed at scale or ramp up of infrastructure happening at the speed level what’s your thought process
I mean most of us who are in the space of mission critical applications and then within IT and if you’re dealing with semiconductors we all know Moore’s law and that was pretty much a 10x almost a couple of years in terms of performance and while performance was 10x the energy required to reach that performance was probably 2 -2 .5x every generation so you were getting amazing efficiency in terms of performance because you were getting a 10x performance with 2 -2 .5x kind of additional energy usage and that’s what you saw for the past many many decades and we all thought that Moore’s law is kind of now reached a plateau there’s not much happening … and this is where companies like NVIDIA, working with other semiconductor ecosystem, came up with multi -tiered chip structures.
When you look at today, some of the chipsets, these are three -story, four -story, six -story buildings. If you had to look under a microscope, there are layers and layers of transistors, billions of transistors layered together. And the innovation that is happening now or that has kick -started now is again kind of retracing Moore’s law. So if you look at what NVIDIA is announcing in terms of the new generation of chipset, there’s a humongous amount of performance improvement every generation. While the performance generation is 10x, 20x, 50x, the energy consumption is also jumping up. It’s not 10x, but it’s 2x, it’s 2 .5x. So like Jigar and Peter mentioned a little while ago, you have the current generation.
The current generation at 130, 140 kilowatt per cabinet, while the next one is 250, 260. and the one down the road is 400, 500 kilowatt per rack. And while I don’t want to give away a bit too much, but one megawatt rack is not too far away. People are already testing it. So now think about it, one megawatt of rack. A few years ago, the whole data center was one megawatt. The white space would have 200 racks of five kilowatt each, and you had generators, chillers, transformers, facilities supporting that one megawatt. So the white space was 80 % of your footprint. The rest of this stuff was 20 -30 % of your footprint. Now this has flipped.
You have only one cabinet. But you still need all of that. You still need one megawatt worth of power, generators, chillers, transformers, everything. So in that context, if you see, we are innovating at tremendous speed. If you invest anything, today it’s outdated two years down the road. So that’s number one. That’s a challenge. The second challenge is that it costs a lot of money. Jigar mentioned the cost of a data center may be a billion dollars, or let’s make the numbers a bit more reasonable, $100 million. But the amount of GPUs sitting inside is probably $2 billion worth of GPUs. So now if I place an order today with $2 billion of GPUs, I want to monetize this project very, very quickly.
If you build a project, and in olden days we used to build a home in India, not just in India, in most other parts of the not -so -developed or developing countries, would have people carrying bricks on their heads and building a house. It takes two years to build a home. Now as a homeowner, you don’t see that as a problem. You’re trying to save $5 a year and $2 a year. You’d rather have a person taking a brick on the head rather than bringing a cement mixer because you thought you were saving money. In this world, you’re losing money. because the money you are spending is still going to be the same, probably 10 % cheaper, but your return will start after two years because you will monetize that investment after two years, after three years, because only when you turn on the switch, only when you turn on the tokens, that’s when you make money on your investment.
So it’s speed to token. Whether you spend $100 million or $1 billion, you need to spend it fast, get the factory up and running very fast, so that the token comes out very fast, so that you can get your return on your capital employed. So if you are anyone here who is from the finance industry, Rocky, return on capital employed is a seriously important KPI for money. So that’s speed. And the third is scale. The demand is so heavy. Jigar and Peter in their conversation talked about a few kind of areas where they have high applications. Think of agentic AI as what it can do for you and in how many areas of your daily life it can affect you.
The scales are crazy. And so not only we need to work on the degree of difficulty in terms of density, we need to deploy it tomorrow morning and we want to deploy it at massive amount of scale. And that’s the kind of problem statement or opportunity that we have.
So Sanjay, when you say speed at scale and that’s an idea which you have given because every month or a week save to deployment is going to be a go -to -market fast, right? And generally when you have to speed at scale, you also have to design for scale. And that’s where the blueprint discussion starts. Now why, Srikant, when it is a blueprint, it starts from a GPU architecture or GPU cluster architecture. What is your thought process? When you say… Scale for design, you not have to scale for it first which GPU you want to go with today and then you scale for that. What’s your thought process when you talk about why the GPU has to start with the… Why the blueprint of any data center has to start with a GPU cluster?
Could you repeat the last part again?
Okay, when I say when we have to speed at scale, we have to design for the scale. And that’s where the blueprint of GPU starts. GPU is the first thing we need to start with. And why is that?
Yeah, I think there’s a couple of things, right? When we first started designing, you know, the early phase of AI factories, we were relying on, you know, general purpose built data centers. And we were changing them rapidly into what… They weren’t even really AI factories, but we were trying to figure out how to make it work, right? It was not designed at scale. It wasn’t designed for… It was not purpose built designs. But I think the moment came on us so quickly. And again, NVIDIA and WordUp together foresaw that moment. We didn’t foresee the scale. We foresaw the moment. And we went… We went from very quickly from 10 megawatts to… Now we’re talking about gigawatts.
And so infrastructure doesn’t move at that speed. Infrastructure moves, you know, the design can move at that speed, but someone has to actually build out the AI factory. Someone has to build out the data centers. We have to make so many CDUs. So we were in a phase where we made it work, but we made it work in a very, we had to make it work way, right? If we had to do it all over again, that’s not how we would do it. So now we have a moment where we say, okay, if we were to do it the right way, now we know what the future looks like. We know, that’s why we’ve redefined the data center as an AI factory, which is a fully integrated, you know, where you go from a chip design to system design to the liquid cooling design or the power design.
It’s all, in fact, even the shell and the campus is all purpose -built as an AI factory. So we have to start thinking both in terms of design as well as manufacturing, as well as delivery, as well as operation. We have to think and start thinking about it at that scale. and I think we’ve already started doing that at the design. Like, you know, NVIDIA has a DSX reference design now, which is actually based out of word of, you know, smart run products and large -scale CDUs. So now we have to start deploying it that scale. That is one of the things that NBIS’s focus is, is how do we deploy it speed of light.
Everything from logistics to operations, everything is being redefined. So that’s why we have to, like, you know, you have to think of it as an end -to -end integrative product.
So you say about we have to design for the future. That means every design what we do has to have a future proof. What are two important ingredients you want to suggest to our audience or all of us when you talk about future proof from a design standpoint?
Yeah, I think the biggest one that I still have to repeat sometimes because it hasn’t caught on is we used to think of, I mean, Jigar and others, have spoken so much about rack density. we have to stop thinking about rack density we have to start thinking about row and data hole level density because how we almost are slowly retrofitting the entire footprint to match an AI factor design we will not be doing that generation to generation that’s just very expensive if we keep changing the technology it’s going to be very expensive for you’re not only spending a lot on building it you’re spending a lot on retrofitting it we don’t want that because that’s going to eat into the ROI so we have to start thinking about I’m at 30 today, I’m going to do this 40 tomorrow, I’m going to do this 100 tomorrow, I have to do something else 200 or 1 megawatt, I have to do something completely different we have to stop that mindset we have to start thinking about it in bounding boxes data hole level or row level bounding boxes and that’s what our latest reference designs do which is start looking at the entire pod as one big block don’t change the technology optimize it with a future proof mindset, right, will this work for that one megawatt rack and today with the digital twins you don’t need to actually build it to do it, you can actually simulate that so that’s number one I would say is take those bounding boxes and take that bounding box mentality now map that technology wise map that right up from the chip to the utility which is same redundancy this redundancy for compute this redundancy for network and have that cluster mindset where you say you map the cluster to the power and thermal perfectly so that every watt goes into maximizing tokens versus going into redundancy and your old school way of thinking so I think if you combine both of those elements you would get into a future proof data center again it’s the hyperscalers have have mastered over the last 10 -15 years.
Again, we pretend like AI is the first time we’re doing infrastructure build -out, but it’s not. The hyperscalers have been doing since the late 2000s, right? So they have mastered the concept of a reference design, a global reference design, where you once lock in that design, you generation -wise you stay in consistency. You build a template and you just feed it out.
I would like to ask the same question to you, Sanjay. From your perspective, what are two things you would like to offer from a design standpoint, infrastructure standpoint, when you want to give a future -proof design for at least for two or three generations, which Peter spoke about?
I think whether we like it or not, the speed of change in the semiconductor IT AI world is very different than the speed of change on the physical world in terms of power and cooling. And even the life cycles and depreciation cycles are very different. So, for example, compute storage or, you know, in the IT world is depreciated every three to five years because that’s the pace of evolution. generators, chillers, transformers, UPS batteries are deprecated 10 to 15 year cycle. So you got to figure out a way how do you run 2 to 3 cycles of IT within one cycle of infrastructure. This is a requirement. If you don’t do that to the point that was just made, Srikanth made that you would be keeping on investing and that’s not good business at all.
So now how do you do that? And in the cloud world again, we mastered that to the sense that in very simple English, how are we doing it today in the cloud space? We have a 30 megawatt data center. We have 2 to 3 to 4, 5 megawatt per data hall. Then we don’t worry about what’s inside the data hall. How does it matter? I have a 5 megawatt power, 5 megawatt cooling capacity. Bring whatever you want as long as it’s 5 megawatt, you’re good to run. The only thing that you are probably retrofitting, if at all you have a generational change, is the final mile of cable or connectors. Now, that becomes a slightly more complicated in the AI world because your densities are much higher while providing power is relatively easy.
Pumping a lot of air or now pumping a lot of liquid is not as simple. There’s much more piping happening. In fact, I joke with the people, the future is of electricians and plumbers, believe me. There’s so much plumbing now in a data center, you will need plumbers in the data center. So, the only way to do it is to again look at what was mentioned in the previous discussion also, is look at certain capacity pods, 2 .4 megawatt pod, 6 megawatt pod. So, now you have a pod. It fits certain number of GPUs of today’s generation. It has certain power capability and liquid capability and it’s done. All the upstream to that in terms of transomers, generators, utility connections is designed for 6 .2, 6 .4, whatever the case.
Now let’s say over the next three years, generations change. Well, all you have to do is reconfigure the cabinets, nothing else, everything else stays the same. Precisely what we are doing in the cloud world. It took us a couple of years to figure this out because this was all being done for the first time. But now this will definitely be the way to go going forward.
So, Sanjay, let me bring to a very different topic now on energy efficiency. When we are talking about gigawatt scale, energy conservation or energy saving is the most important piece. Now, we as a country are a tropical, right? We have a temperature right from 10 degree to 48 degree. So in such a span, what do you think the right approach to improve the PUE? Okay, maybe water usage or what are the important best practices you would like to suggest to the market when it comes to saving energy efficiency or improvising the PUE? Of course, because of liquid adoption, it is anyway has scaled down to an extent of what was there for the normal. But what would be the next stage of best practices you would like to suggest through your experience?
I think the word PUE is, I don’t know if this is the right word, but probably a very abused word in the industry. It’s used so commonly, thrown out there so easily that everyone believes, well, I have a lower PUE. Well, first of all, I can give you better PUE without doing anything. I can increase the air temperature. Suddenly your PUE is much better. You think your PUE is better, but right now your computer fans, your server fans will speed up. The temperature is more. They need to throw more. So the IT load increases. but you increase the temperature so your electrical load reduces, I mean your cooling load reduces, you suddenly have a better calculation.
But in reality, your total power increase, which you don’t realize, so the PUE is better. So PUE is a bit of a, you know, thrown out word, but here is how I look at this. I think the PUE in the data hall in the white space, irrespective where you build it, is the same. Because I need liquid at a certain temperature, I need air at a certain temperature, it needs to enter the rack. The rack is doing what it is doing, it doesn’t matter whether you build in Mumbai, Singapore, I live in Dubai, you build in Timbuktu, it’s exactly the same. The question is, how do you throw the heat out? Because now that depends on the environment outside.
So are you in Singapore? Rains all the time. Are you in Iceland? It’s never more than 20 degrees any time of the year. or are you in, you know, Dubai where it reaches 52 degrees in summer. At least that’s what we designed for 52 degrees. And that’s where the different technologies need to be adopted. Now, whether it is, you know, air -cooled chillers, whether in some markets you can have, you know, water -cooled chillers. One of the unique solution sets that we have started to see is that, especially in India, you have our cities and the way we are located in between the latitudes, our thermal variation or our temperature variation during the year is different.
We have very hot in the summer and we have reasonably good weather in the winter. So there are some entitlements that you can get in the winter. So, for example, we can use chiller technologies where during the winter months we are able to use a bit more free cooling. And in the summer months or during demand months, we add a bit more of, you know, chillers. Chiller technologies, I mean DX technologies, comparator elements that come in and help us to add that extra cooling factor when required. and so what we could do is optimize the way we cool across the thermal cycles of the year and bring down the annual PUEs of the year because at the highest point of temperature you will need that cooling whether you like it or not and so it’s this management of PUE through thermal cycles and some optimization also through load cycles because load also especially in the AI world may not be like a cloud business uniform throughout the year, throughout the day, through every month and so again certain optimizations in how you use your CDUs or fan wall units to bring that energy down will help us to improve the PUE.
One thing I would say about that is the design is there, right? Based on whether it’s the water temperatures, we’re all designing to the same targets. The design is there. Where it becomes extremely manual is again we’re still, in the traditional mode of operation in data centers where we have a large control room and we are optimizing for… for uptime and safety, and safety in the sense there’s no risk of downtime. We’re very risk -averse. But we haven’t, even if we have to do what Sanjay just suggested, which is optimize that, there is no automated way of doing that because the chip -level telemetry doesn’t talk to the data center -level telemetry. And that’s what NVIDIA’s reference design is looking to change today, is, again, if you were to retrofit a brownfield facility, this will be harder.
But if you were to build a purpose, of course, this is an opportunity for India, if you’re building an AI factory today, there is no reason why you can’t integrate telemetry from chip to chip. There’s no reason why you cannot simulate how to optimize that and simulate a traditional sample workload and see how you save energy. I’m sure that simulation will tell you that you’ll save a ton of energy without any human intervention.
You spoke about retrofit. So there have been normal cloud services or normal workloads have been working, let’s say about 5, 10, 15, 15 kilowatt of load. what do you see when it comes to AI augmentation or a GP augmentation in a same platform or a same aisle how the retrofit will be easy or difficult or what could be your one or two tips to do that like if you are talking about AI optimization for telemetry specifically there is already an existing workload which is working with a very small medium to small densities but in that row you want to put a GPU or a liquid could GPU or air GPU which means you are retrofitting some amount of passive infrastructure how difficult or easy would be that actually?
I think again if you go back to that journey even the design and the retrofit was extremely commercial even today I think at enterprise level it is extremely difficult if I was an enterprise CTO looking to deploy AI compute I might actually and I look at our experience in the last one year I might actually be a little you’re looking at a very cumbersome everywhere from design to following local regulations for the high power and liquid cooling having secondary loop built out that’s going to be that could be pretty scary at the end of the day but I think what Verte was doing for example with smart runs you know fully integrated mechanical electrical system that can be purpose built for any pod size that can track our you know our most scalable reference designs I think that would be the way to go right like that’s the importance why even Jigar mentioned that you know our following our reference design as closely as possible all these innovative designs and offerings will improve the adaptability part for the future change is what I can say
my last question to you the future of the the future of the the future of the there seems to have some NVIDIA -ready design offerings or certification offerings. Would you like to say, talk, or would you like to give some insight about that? Certification programs of NVIDIA -ready data centers. But NVIDIA -ready designs.
Yeah, I think whether it’s a colo or whether it’s at a cloud scale, an NCP scale, what we’ve been doing from the beginning is, you know, just like we’ve been enabling other partners, we’ve been enabling a lot of colo partners to build NVIDIA -ready data centers. Okay. And that optimizes for, you know, the water temperatures that we’re recommending, the port sizes that we’re recommending, the redundancy that we’re recommending, the integration between telemetry that we’re recommending. So for the partners that have followed that design, we have, you know, whether it’s DGX -ready or NVIDIA -ready. Now, the only thing I would encourage these partners and also those who are looking forward to this vertical, who is actually doing that?
at speed of light, in a sense. Like, you know, a lot of the data center industry is still in the mode of, you know, they’re thinking more like real estate developers, you know, waiting for, for example, you know, you have these tranches of data centers that you’re purpose -building for everyone. That is a traditional way of thinking and saying, I’m giving this space, this cage to you, and I’m going to build it out the way you want it. But you can’t wait. Like, the way the industry is operating, no one can wait for that, right? So the partners who are building purpose -built AI factories, they are part of, or want to be part of that future, building at large scale, and then whether they give those tranches or not, but they’re built on NVIDIA design, so when the customer comes, you already have built basically according to the specs.
That’s really insightful. Many of our Colo customers will take good insight from that. With this, I would come to an audience for any other question for them.
Hi, I’m Dal Bhanushali. Thanks for the talks, this one and the previous one. We have been talking about how we will scale India in the future. We also need to scale the talent. I wanted to get some viewpoints from you, from your experiences as we double capacities. You also need those people to run the data centers. We need DC ops as special. We can run the NVIDIA optimized containers in our laptops, but those word -to -you chillers, those skills are not common and cannot be easily taught in schools today. So what’s the plan? How do you think we should be going in the future? Especially double every year. is a huge challenge, right?
So I’ll just take this question for a while. So at Vati, we realized this challenge much ahead of time, and we started with a lot of skill development program. So the first thing first is about operation and management of the infrastructure, okay? That’s something which we have started with in collaboration with Indian Institute of Technology, Chennai, where Diploma and BTEC are graduate engineers. We train them for managing how to manage operation and maintenance of data centers. That’s about eight to 12 weeks program, extensive programs, off -site as well as on -site. So this is one part. And there are many other programs which are on the cards to develop design, engineering, and many other things, actually.
That’s what I can tell. And these programs are already available on the web. Anybody can have a look and enroll for that, okay? Any other thing which maybe anybody would like to say about skill development or any other thing which maybe anybody would like to say about skill development or any other development activity which NVIDIA would like to do with the need of ours when we are scaling so high?
I think that’s a question you also want to have. Could you repeat the last part of the question, if you don’t mind?
So he’s asking about how the scale is going up. There’s a lot of resources required, and the skill development is also a big challenge. So while NVIDIA is taking care of an operation and management piece, we are developing a lot of people through colleges and engineering institutions. What are the initiatives NVIDIA also…
managing how to manage operation and maintenance of data centers. That’s about 8 to 12 weeks program, extensive programs, off -site as well as on -site. So this is one part. And there are many other programs which are on the cards to develop design, engineering, and many other things actually. That’s what I can tell. And these programs are already available on the web. Anybody can have a look and enroll for that. Any other thing which maybe NVIDIA would like to say about skill development activity which NVIDIA would like to do with the need of ours when we are scaling so high? I think that’s a question you also want to have.
Could you repeat the last part of the question if you don’t mind?
So he’s talking about scale is going up. There’s a lot of resources required. And the skill development is also a big challenge. So while NVIDIA is taking care of an operation and management piece, we are developing a lot of people through colleges and engineering institutions. What are the initiatives NVIDIA is also taking to develop the skills within the ecosystem?
Yeah. I think a couple of things I would say. One is, you know, as you keep going up on the scale, the prefab systems that Vertebra is developing are going to be absolutely critical because, like, when I was talking about the enterprise -level difficulties right now, all that can be solved with. But it’s, you know, a lot of times you’re waiting for the data center, you’re waiting for the data hole to get ready before you can deploy the compute systems. Yeah. And each of them have dependencies on each other that all are centered around that space, right? When you’re doing off -site prefab integration, you’re doing prefab, you know, manufacturing, you can do that all in parallel.
You can do that all at scale in parallel and then bring it all into one place. And in the meantime, you could do the testing off on the factory. A lot of the testing is done today in the data hall. So you could avoid all that, move it all to the left by bringing it all outside of the data hall and then bring it all into the data hall once the data hall is ready, once the shell is built up, and you could really condense that build -out.
Srikant, as you say it rightly, as we are taking a lot of activity, it is supposed to happen on -site, taking to off -site, which by means of pre -engineering it, developing and building at the scale and getting deployed at the site. So that’s a way forward. Any more questions? Otherwise, we can hold it here.
Ankush Sabharwal
Speech speed
172 words per minute
Speech length
287 words
Speech time
99 seconds
AI Sovereignty and Hub Vision
Explanation
Ankush stresses that India must achieve complete AI sovereignty, not just platform control, and envisions the country becoming a global hub for AI development and deployment.
Evidence
“having the complete sovereignty in terms of AI and not just the platform” [1]. “I think India being so aspirational and ready to adopt new technology for the welfare of themselves and the welfare of the businesses, I think we would be the hub of AI development for the world” [2].
Major discussion point
AI Sovereignty and Indigenous Development in India
Topics
Artificial intelligence | The enabling environment for digital development
Bharat GPT Purpose‑Driven Model
Explanation
He introduces Bharat GPT as a family of models built with partner data, focused on purpose, trust, and solving concrete enterprise problems for Indian citizens and businesses.
Evidence
“See, the Bharat GPT family of models, right?” [40]. “Yeah, see, our tagline is AI with purpose and trust, right?” [43]. “So we work with our partners, get their data and train the model for their users because we believe we… is easy for us to solve the problem of enterprises because the enterprises say like IRCTC, they already know their domain” [42].
Major discussion point
AI Sovereignty and Indigenous Development in India
Topics
Artificial intelligence | Social and economic development
Akanksha Swarup
Speech speed
166 words per minute
Speech length
321 words
Speech time
115 seconds
Questioning Infrastructure for Sovereign AI
Explanation
Akanksha asks whether India currently possesses the necessary infrastructure and resources to build a sovereign AI ecosystem.
Evidence
“Do you think India has the right infrastructure, the right resources to build its own sovereign AI at the moment?” [5].
Major discussion point
AI Sovereignty and Indigenous Development in India
Topics
Artificial intelligence | The enabling environment for digital development
Digital Divide and Google Accessibility
Explanation
She probes how Google plans to make its AI services accessible to under‑privileged and rural populations, highlighting concerns about the digital divide.
Evidence
“As far as digital divide is concerned, how do you make Google more accessible for the underprivileged, for those in rural areas?” [105]. “How is Google trying to bridge that divide as far as you can see?” [106].
Major discussion point
AI Inclusivity and Accessibility
Topics
Closing all digital divides | Artificial intelligence
Nitin Gupta
Speech speed
140 words per minute
Speech length
366 words
Speech time
156 seconds
Google Data Centers and Indigenous Data Box
Explanation
Nitin explains that Google’s new Indian data centers and the on‑premise “Data Box” enable AI innovation while keeping data residency within India’s borders.
Evidence
“from Google perspective and from my personal perspective I feel yes sovereignty is very important but at the time with the sovereignty it is not a question between sovereignty or innovation it is sovereignty and innovation they have to run together they can’t be one choice versus the other one and with that Google while we have our entire data centers in India you have heard three months back we announced that we are going to be building big data centers in Vizag the announcement happened so we are ensuring that let’s say if any innovation and any data residency things are there they are being kept within the boundaries of India But then those data centers are definitely empowering the lot of AI, but they are for everyone, for all type of personas, whether they’re government, enterprises, startups, students, colleges, universities” [3]. “And for that, Google has created a completely indigenous data box which completely stays inside the customer premise and is fully powered by AI” [21]. “So imagine that you have the full potential to run what you’re running in a Google data center, but inside your own premise” [22].
Major discussion point
Infrastructure and Data Center Strategies for AI
Topics
Artificial intelligence | The enabling environment for digital development
Free Gemini JEE Mock Exams
Explanation
He highlights Google’s initiative to provide free Gemini‑powered JEE mock exams, expanding AI‑driven educational access for students across India.
Evidence
“And just to summarize and give a recent example, we have very recently Sundar Pichai has announced that the JEE main exams, mock exams are available on Gemini free of cost for any student to try” [103]. “We want to make sure that student at his home can keep on trying the mock test at free” [104].
Major discussion point
AI Inclusivity and Accessibility
Topics
Closing all digital divides | Artificial intelligence
Sudeesh VC Nambiar
Speech speed
136 words per minute
Speech length
205 words
Speech time
90 seconds
Collaboration with Indian Startups for Indigenous AI Layer
Explanation
Sudeesh describes a partnership model where Indian startups work with global technology firms to embed an indigenous AI layer for use‑cases such as railway ticketing security.
Evidence
“So it is basically a collaboration between the Indian startups and the global technology strength of a global company” [31]. “Indigenous, of course, we have a layer of indigenous” [37].
Major discussion point
Collaboration Between Global and Indian Entities
Topics
Artificial intelligence | The enabling environment for digital development
Use of Advanced AI Solutions
Explanation
He affirms that his organization is leveraging very advanced AI solutions for its operations.
Evidence
“We are using AI of very advanced AI solution” [7].
Major discussion point
Use Cases and Applications of AI in India
Topics
Artificial intelligence | Social and economic development
Srirang Deshpande
Speech speed
124 words per minute
Speech length
274 words
Speech time
131 seconds
Shift to Inside‑Out Data Center Design
Explanation
Srirang notes that data centers are moving from an outside‑in construction model to an inside‑out, AI‑centric approach.
Evidence
“and then now time is there or time has come where data centers are getting filled from inside out approach” [54]. “Till this time, data centers are getting built from outside in approach” [55].
Major discussion point
Infrastructure and Data Center Strategies for AI
Topics
Artificial intelligence | The enabling environment for digital development
Gigawatt‑Scale Data Center Experience
Explanation
He points out his involvement in designing large‑scale, gigawatt‑level data centers, underscoring India’s growing capacity.
Evidence
“And he’s the one who’s involved into many designs, a large scale data centers and gigawatt designs” [59].
Major discussion point
Infrastructure and Data Center Strategies for AI
Topics
Artificial intelligence | Environmental impacts
Peter Panfil
Speech speed
139 words per minute
Speech length
2977 words
Speech time
1275 seconds
AI Factory Pod Design and Speed at Scale
Explanation
Peter outlines the AI‑factory concept where a single GPU pod design can be reused across multiple GPU generations, enabling rapid deployment and “speed at scale”.
Evidence
“I could show you a pod design that supports three generations of GPUs so this year, next year, next year after that three generations of GPUs just by changing the way those pods are populated on the compute side” [62]. “Now, we can take a GPU pod, design it once, build it many, and apply it to the GPU that we need from that generation” [63]. “Let’s start at the chip, define the most economical, most efficient, fastest from a compute perspective, and figure out how to deploy that as a pod, then replicate that pod, and achieve the speed” [64]. “We live now in the GPU world of project scales of between four and six months” [122]. “But speed at scale” [123]. “It’s scale at speed” [129].
Major discussion point
Speed, Scale, and Future‑Proofing of AI Factories
Topics
Artificial intelligence | Capacity development | Environmental impacts
Prefabricated Systems and Reference Designs
Explanation
He emphasizes Vertiv’s use of prefabricated, purpose‑built systems and global reference designs to accelerate AI‑factory roll‑outs.
Evidence
“Reference designs” [84]. “One is, you know, as you keep going up on the scale, the prefab systems that Vertebra is developing are going to be absolutely critical because, like, when I was talking about the enterprise -level difficulties right now, all that can be solved with” [89]. “We worked closely with your team on reference designs” [90].
Major discussion point
Infrastructure and Data Center Strategies for AI
Topics
Artificial intelligence | The enabling environment for digital development
Jigar Halani
Speech speed
170 words per minute
Speech length
3536 words
Speech time
1246 seconds
AI for All and Sovereignty Building Blocks
Explanation
Jigar calls for AI to be inclusive (“AI for all”) and discusses the building blocks of AI factories while stressing the need for indigenization and data sovereignty.
Evidence
“and make it work for everyone, to make AI for all, how India wants to call it as, AI for all, is what I feel we should be discussing about here” [4]. “What is that is the building blocks of these AI factories and the sovereignty aspect of it?” [8]. “We want to push, you know, India ecosystem, be more indigenized as much as possible, be more self -reliant” [13]. “We are helping the nation with AI factories” [14]. “That means we will still allow these processing to happen outside of India, but at the same time, we will do the regulatory reasons of some of the verticals, say, fintech, say, healthcare, defence and so on” [23].
Major discussion point
AI Sovereignty and Indigenous Development in India
Topics
Artificial intelligence | Building confidence and security in the use of ICTs
IRCTC Ticket‑Bot Mitigation and Fraud Prevention
Explanation
He describes large‑scale AI use for government services, including handling 100 million requests per hour and preventing ticket‑booking bot fraud, saving millions of dollars daily.
Evidence
“we have only touched 1 ,000 and we are already hitting 100 million requests per hour right and this translates to in a simple term roughly about 2 million 2 megawatts of data center consumption per minute right in in 2 megawatt if I’m able to cater to 100 million requests a minute that’s massive” [98]. “A call per day, if government has been able to scale in the last one month to about 50 ,000 calls a day to citizens through a bot, which is talking in a local language, has been able to save a fraud work of around per day, and I’m talking about per day, in the range of a couple of millions of dollars” [102].
Major discussion point
Use Cases and Applications of AI in India
Topics
Artificial intelligence | Social and economic development
Audience
Speech speed
128 words per minute
Speech length
418 words
Speech time
195 seconds
AI Everywhere Observation
Explanation
Audience members repeatedly note that AI has become pervasive across sectors and daily life.
Evidence
“And in today’s era, it is totally about AI” [6]. “As I can see, everywhere is AI” [12]. “There is any chance where humans and AI both are in the same niche?” [39]. “So, as you also said that this is a… AI whereas everywhere industry and company and education in every sector using AI” [101].
Major discussion point
AI Inclusivity and Accessibility
Topics
Artificial intelligence | Closing all digital divides
Moderator
Speech speed
176 words per minute
Speech length
1315 words
Speech time
447 seconds
Vertiv Skill Development Programs
Explanation
The moderator outlines Vertiv’s initiatives to train engineers through collaborations with IIT Chennai and on‑site/off‑site programs, aiming to close the talent gap for AI data‑center operations.
Evidence
“And let me also tell you, on Vertiv’s side, whatever innovations we are doing in the US, we are real -time bringing to India, so that there’s no latency here, and absolutely whatever is going to happen in the US, we want to bring it to India” [26]. “That’s something which we have started with in collaboration with Indian Institute of Technology, Chennai, where Diploma and BTEC are graduate engineers” [35]. “We train them for managing how to manage operation and maintenance of data centers” [112]. “So at Vati, we realized this challenge much ahead of time, and we started with a lot skill development program” [114].
Major discussion point
Skill Development and Talent Pipeline
Topics
Capacity development | The enabling environment for digital development
Srikanth Cherukuri
Speech speed
171 words per minute
Speech length
2202 words
Speech time
770 seconds
AI Factory Build‑Out and High‑Density Racks
Explanation
Srikanth describes the worldwide AI‑factory rollout, noting that India must adopt gigawatt‑scale, high‑density, liquid‑cooled racks similar to the U.S. experience.
Evidence
“And we are in year two or year three of this AI factory build -out worldwide” [15]. “I think the next step in terms of where India is going in this AI factory build -out is if you look at the U.S., it’s a little further ahead in terms of gigawatt scale and high -density racks, deploying high -density racks, high -density liquid -cooled racks” [17]. “But if you were to build a purpose, of course, this is an opportunity for India, if you’re building an AI factory today, there is no reason why you can’t integrate telemetry from chip to chip” [19]. “Someone has to build out the data centers” [24]. “We know, that’s why we’ve redefined the data center as an AI factory, which is a fully integrated, you know, where you go from a chip design to system design to the liquid cooling design or the power design” [57].
Major discussion point
Infrastructure and Data Center Strategies for AI
Topics
Artificial intelligence | Environmental impacts
Future‑Proof Design with Row‑Level Density
Explanation
He argues for moving beyond rack‑level density to row‑level (data‑hole) design and using digital‑twin simulations to ensure AI factories remain adaptable across GPU generations.
Evidence
“we have to stop thinking about rack density we have to start thinking about row and data hole level density because how we almost are slowly retrofitting the entire footprint to match an AI factor design we will not be doing that generation to generation that’s just very expensive” [133]. “And we went… We went from very quickly from 10 megawatts to… Now we’re talking about gigawatts” [143].
Major discussion point
Speed, Scale, and Future‑Proofing of AI Factories
Topics
Artificial intelligence | Capacity development
Reference Design Consistency
Explanation
He highlights the importance of global reference designs that stay consistent across GPU generations, facilitating rapid deployment.
Evidence
“So they have mastered the concept of a reference design, a global reference design, where you once lock in that design, you generation -wise you stay in consistency” [92].
Major discussion point
Infrastructure and Data Center Strategies for AI
Topics
Artificial intelligence | The enabling environment for digital development
Sanjay Kumar Sainani
Speech speed
170 words per minute
Speech length
2078 words
Speech time
733 seconds
Energy Consumption and PUE Critique
Explanation
Sanjay points out that PUE is often misused, emphasizing that true energy efficiency must consider seasonal temperature swings and holistic cooling strategies.
Evidence
“We have a 30 megawatt data center” [61]. “I think the word PUE is, I don’t know if this is the right word, but probably a very abused word in the industry” [168]. “but you increase the temperature so your electrical load reduces, I mean your cooling load reduces, you suddenly have a better calculation” [169]. “I have a 5 megawatt power, 5 megawatt cooling capacity” [172]. “But we haven’t, even if we have to do what Sanjay just suggested, which is optimize that, there is no automated way of doing that because the chip -level telemetry doesn’t talk to the data center -level telemetry” [179]. “And that optimizes for, you know, the water temperatures that we’re recommending, the port sizes that we’re recommending, the redundancy that we’re recommending, the integration between telemetry that we’re recommending” [180].
Major discussion point
Energy Efficiency and Sustainability
Topics
Environmental impacts | Artificial intelligence
High‑Power Rack Designs and Future Capacity
Explanation
He discusses the move toward 200‑kilowatt‑plus racks and the vision of megawatt‑scale racks that could power large portions of India’s digital services.
Evidence
“And as we announced it in January, we are now talking about 240, 230, 210 kilowatt per rack, which means this size hall could probably run a great portion of India with so many services that is probably never imagined before” [184]. “And while I don’t want to give away a bit too much, but one megawatt rack is not too far away” [186].
Major discussion point
Energy Efficiency and Sustainability
Topics
Environmental impacts | Artificial intelligence
Agreements
Agreement points
AI infrastructure design should start with compute requirements rather than traditional grid-first approach
Speakers
– Srirang Deshpande
– Peter Panfil
– Srikanth Cherukuri
Arguments
Data centers must transition from ‘outside-in’ to ‘inside-out’ approach, starting with GPU requirements rather than traditional infrastructure planning
Infrastructure design should start at the chip level and work outward to grid, not traditional grid-to-chip thinking
Future-proof design requires thinking in terms of pods and bounding boxes rather than individual racks, accommodating multiple GPU generations
Summary
All speakers agree that the traditional approach of designing data centers from grid infrastructure inward is outdated for AI applications. They advocate for starting with GPU and compute requirements first, then building supporting infrastructure around those needs using standardized pods and reference designs.
Topics
Artificial intelligence | The enabling environment for digital development
Speed of deployment is critical for AI infrastructure success
Speakers
– Peter Panfil
– Sanjay Kumar Sainani
– Srikanth Cherukuri
Arguments
Project timelines have compressed from 18 months in cloud world to 4-6 months in GPU world, requiring faster capacity building
Speed to token is critical because expensive GPU investments ($2 billion worth) need quick monetization to achieve return on capital
Reference designs and prefabricated systems enable faster deployment by moving testing and integration off-site
Summary
There is strong consensus that rapid deployment is essential for AI infrastructure, with speakers agreeing that traditional 18-month project timelines are inadequate for GPU-based systems that require 4-6 month deployment cycles to achieve proper return on investment.
Topics
Artificial intelligence | Financial mechanisms | The enabling environment for digital development
India has significant potential for AI sovereignty and development
Speakers
– Ankush Sabharwal
– Jigar Halani
– Nitin Gupta
Arguments
India will become the hub of AI development for the world within months due to its aspirational nature and readiness to adopt new technology
India generates 20% of world’s data but has only 3% of global data center capacity, creating opportunity for sovereign data processing
Sovereignty and innovation must run together, not as competing choices, with Google building data centers in India while providing indigenous solutions for critical data
Summary
All speakers agree that India is well-positioned for AI leadership, citing the country’s large data generation, consumer base, and growing infrastructure capabilities. They see sovereignty and innovation as complementary rather than competing priorities.
Topics
Artificial intelligence | Data governance | The enabling environment for digital development
Energy efficiency should be maximized from source to compute
Speakers
– Peter Panfil
– Jigar Halani
– Sanjay Kumar Sainani
Arguments
Every watt saved at the source eliminates need for generation, distribution, and heat rejection, maximizing compute efficiency
India’s advantage in green energy generation (50% renewable) provides strong foundation for sustainable AI infrastructure
PUE optimization should focus on thermal and load cycle management rather than simple temperature adjustments that may increase overall power consumption
Summary
Speakers agree that energy efficiency is crucial for AI infrastructure, emphasizing the importance of minimizing waste from source to compute and leveraging India’s renewable energy capabilities for sustainable AI development.
Topics
Environmental impacts | Artificial intelligence
Skill development is essential for scaling AI infrastructure
Speakers
– Audience
– Moderator
– Srikanth Cherukuri
Arguments
Scaling infrastructure requires specialized skills for operating advanced cooling systems and high-density environments that cannot be easily taught in traditional schools
Vertiv has initiated 8-12 week training programs with IIT Chennai for data center operations and maintenance, with programs available online
Prefabricated systems and standardized designs help address skill gaps by moving complex integration work to controlled factory environments
Summary
There is consensus that the rapid scaling of AI infrastructure requires specialized skills that are not readily available, and that both training programs and prefabricated systems are needed to address this challenge.
Topics
Capacity development | Artificial intelligence
Similar viewpoints
Both speakers see India’s role as a major data consumer driving the need for local processing infrastructure, with Halani predicting 10-12 gigawatts capacity and Panfil supporting data sovereignty from an efficiency perspective.
Speakers
– Jigar Halani
– Peter Panfil
Arguments
India will cross 10-12 gigawatts of AI infrastructure in next three years, driven primarily by inference workloads as a consumer nation
The most efficient place to process data is at its source, making sovereign data centers logical from efficiency perspective
Topics
Artificial intelligence | Data governance | The enabling environment for digital development
Both speakers advocate for practical, domain-specific AI applications that solve real-world problems by working with organizations that understand their specific challenges, rather than pursuing generic solutions.
Speakers
– Ankush Sabharwal
– Sudeesh VC Nambiar
Arguments
Bharat GPT focuses on enterprise solutions with ‘AI with purpose and trust,’ working with partners who understand their domains rather than trying to solve all problems
IRCTC uses advanced AI solutions to combat automated booking tools during peak tatkal periods, combining indigenous startups with global technology
Topics
Artificial intelligence | Social and economic development | The digital economy
Both speakers emphasize that India can leverage global experience to accelerate AI infrastructure deployment while avoiding costly mistakes, focusing on rapid time-to-value for expensive GPU investments.
Speakers
– Srikanth Cherukuri
– Sanjay Kumar Sainani
Arguments
India can benefit from lessons learned in other regions over the past 18 months, avoiding mistakes and accelerating deployment
Speed to token is critical because expensive GPU investments ($2 billion worth) need quick monetization to achieve return on capital
Topics
Artificial intelligence | Financial mechanisms | The enabling environment for digital development
Unexpected consensus
Data processing sovereignty as efficiency optimization rather than protectionism
Speakers
– Peter Panfil
– Jigar Halani
– Nitin Gupta
Arguments
The most efficient place to process data is at its source, making sovereign data centers logical from efficiency perspective
India generates 20% of world’s data but has only 3% of global data center capacity, creating opportunity for sovereign data processing
Sovereignty and innovation must run together, not as competing choices, with Google building data centers in India while providing indigenous solutions for critical data
Explanation
Unexpectedly, speakers from both global technology companies (Google, Vertiv) and Indian AI initiatives agree that data sovereignty makes business and technical sense from an efficiency perspective, not just regulatory compliance. This consensus suggests that sovereignty and global collaboration can coexist.
Topics
Data governance | Artificial intelligence | The enabling environment for digital development
Prefabrication and standardization as solutions to multiple challenges
Speakers
– Srikanth Cherukuri
– Sanjay Kumar Sainani
– Peter Panfil
Arguments
Reference designs and prefabricated systems enable faster deployment by moving testing and integration off-site
Speed to token is critical because expensive GPU investments ($2 billion worth) need quick monetization to achieve return on capital
Project timelines have compressed from 18 months in cloud world to 4-6 months in GPU world, requiring faster capacity building
Explanation
There is unexpected consensus that prefabrication addresses multiple challenges simultaneously – speed, skill gaps, quality control, and cost efficiency. This represents a significant shift from traditional custom data center construction approaches.
Topics
Artificial intelligence | Capacity development | The enabling environment for digital development
Overall assessment
Summary
The discussion reveals strong consensus on fundamental shifts needed for AI infrastructure: chip-first design, rapid deployment timelines, energy efficiency optimization, and the importance of skill development. Speakers agree that India has significant potential for AI leadership through a combination of sovereignty and global collaboration.
Consensus level
High level of consensus across technical, business, and policy dimensions. The agreement spans both global technology providers and Indian stakeholders, suggesting alignment between commercial interests and national development goals. This consensus indicates a mature understanding of AI infrastructure requirements and India’s strategic position in the global AI ecosystem.
Differences
Different viewpoints
Timeline for India becoming AI development hub
Speakers
– Ankush Sabharwal
Arguments
India will become the hub of AI development for the world within months due to its aspirational nature and readiness to adopt new technology
Summary
Sabharwal makes an optimistic claim about India becoming the global AI hub ‘within months, not years,’ but this timeline appears overly ambitious compared to the infrastructure scaling challenges discussed by other speakers who talk about 3-5 year horizons for gigawatt-scale deployments
Topics
Artificial intelligence | The enabling environment for digital development
Approach to AI model development – consumer vs enterprise focus
Speakers
– Ankush Sabharwal
Arguments
Bharat GPT focuses on enterprise solutions with ‘AI with purpose and trust,’ working with partners who understand their domains rather than trying to solve all problems
Summary
Sabharwal explicitly states that Bharat GPT is ‘not ready for consumers yet’ and focuses on enterprise partnerships, which contrasts with the broader consumer-focused applications discussed by other speakers like government services for citizens
Topics
Artificial intelligence | The digital economy
Infrastructure capacity projections for India
Speakers
– Jigar Halani
Arguments
India will cross 10-12 gigawatts of AI infrastructure in next three years, driven primarily by inference workloads as a consumer nation
Summary
Halani projects 10-12 gigawatts versus industry reports of 5-6 gigawatts, showing disagreement with mainstream industry projections about India’s AI infrastructure scaling
Topics
Artificial intelligence | The enabling environment for digital development
Unexpected differences
Role of traditional PUE metrics in AI infrastructure
Speakers
– Sanjay Kumar Sainani
Arguments
PUE optimization should focus on thermal and load cycle management rather than simple temperature adjustments that may increase overall power consumption
Explanation
Sainani’s criticism of PUE as an ‘abused word’ in the industry is unexpected, as PUE is typically considered a standard efficiency metric. His argument that improving PUE numbers can actually increase total power consumption challenges conventional wisdom about data center efficiency measurement
Topics
Environmental impacts | Artificial intelligence
Consumer vs infrastructure readiness assessment
Speakers
– Jigar Halani
Arguments
India generates 20% of world’s data but has only 3% of global data center capacity, creating opportunity for sovereign data processing
Explanation
While Halani presents India as a major consumer nation ready for AI scaling, the infrastructure gap he identifies (20% data generation vs 3% capacity) suggests India may not be as ready as other speakers suggest, creating an unexpected tension between consumption capability and infrastructure readiness
Topics
Artificial intelligence | The enabling environment for digital development | Data governance
Overall assessment
Summary
The discussion shows relatively low levels of direct disagreement among speakers, with most conflicts being implicit rather than explicit. The main areas of tension involve timelines for AI development, focus on enterprise vs consumer markets, and infrastructure scaling projections. Most speakers are aligned on the fundamental importance of AI sovereignty, energy efficiency, and rapid deployment, but differ on specific approaches and timelines.
Disagreement level
Low to moderate disagreement level with significant implications for strategic planning. The optimistic timelines and projections by some speakers may not align with the technical and infrastructure realities discussed by others, potentially leading to unrealistic expectations for India’s AI development trajectory. The focus on different market segments (enterprise vs consumer) could result in fragmented development approaches rather than coordinated national AI strategy.
Partial agreements
Partial agreements
Both speakers agree on the importance of data sovereignty and local processing, but they approach it differently – Gupta emphasizes balancing sovereignty with innovation through hybrid solutions, while Panfil focuses purely on efficiency arguments for local processing
Speakers
– Nitin Gupta
– Peter Panfil
Arguments
Sovereignty and innovation must run together, not as competing choices, with Google building data centers in India while providing indigenous solutions for critical data
The most efficient place to process data is at its source, making sovereign data centers logical from efficiency perspective
Topics
Data governance | Artificial intelligence | The enabling environment for digital development
All speakers agree on the critical need for speed in AI infrastructure deployment, but they emphasize different approaches – Panfil focuses on design philosophy changes, Cherukuri on prefabrication and reference designs, and Sainani on financial imperatives driving speed requirements
Speakers
– Peter Panfil
– Srikanth Cherukuri
– Sanjay Kumar Sainani
Arguments
Project timelines have compressed from 18 months in cloud world to 4-6 months in GPU world, requiring faster capacity building
Reference designs and prefabricated systems enable faster deployment by moving testing and integration off-site
Speed to token is critical because expensive GPU investments ($2 billion worth) need quick monetization to achieve return on capital
Topics
Artificial intelligence | The enabling environment for digital development | Financial mechanisms
Both speakers agree on the importance of energy efficiency in AI infrastructure, but they focus on different aspects – Panfil emphasizes source-to-GPU power delivery efficiency, while Sainani focuses on proper PUE measurement and thermal cycle management
Speakers
– Peter Panfil
– Sanjay Kumar Sainani
Arguments
Every watt saved at the source eliminates need for generation, distribution, and heat rejection, maximizing compute efficiency
PUE optimization should focus on thermal and load cycle management rather than simple temperature adjustments that may increase overall power consumption
Topics
Environmental impacts | Artificial intelligence
Similar viewpoints
Both speakers see India’s role as a major data consumer driving the need for local processing infrastructure, with Halani predicting 10-12 gigawatts capacity and Panfil supporting data sovereignty from an efficiency perspective.
Speakers
– Jigar Halani
– Peter Panfil
Arguments
India will cross 10-12 gigawatts of AI infrastructure in next three years, driven primarily by inference workloads as a consumer nation
The most efficient place to process data is at its source, making sovereign data centers logical from efficiency perspective
Topics
Artificial intelligence | Data governance | The enabling environment for digital development
Both speakers advocate for practical, domain-specific AI applications that solve real-world problems by working with organizations that understand their specific challenges, rather than pursuing generic solutions.
Speakers
– Ankush Sabharwal
– Sudeesh VC Nambiar
Arguments
Bharat GPT focuses on enterprise solutions with ‘AI with purpose and trust,’ working with partners who understand their domains rather than trying to solve all problems
IRCTC uses advanced AI solutions to combat automated booking tools during peak tatkal periods, combining indigenous startups with global technology
Topics
Artificial intelligence | Social and economic development | The digital economy
Both speakers emphasize that India can leverage global experience to accelerate AI infrastructure deployment while avoiding costly mistakes, focusing on rapid time-to-value for expensive GPU investments.
Speakers
– Srikanth Cherukuri
– Sanjay Kumar Sainani
Arguments
India can benefit from lessons learned in other regions over the past 18 months, avoiding mistakes and accelerating deployment
Speed to token is critical because expensive GPU investments ($2 billion worth) need quick monetization to achieve return on capital
Topics
Artificial intelligence | Financial mechanisms | The enabling environment for digital development
Takeaways
Key takeaways
India is positioned to become a global AI hub within months, leveraging its aspirational population and technology adoption readiness
AI infrastructure design must shift from traditional ‘grid-to-chip’ thinking to ‘chip-to-grid’ approach, starting with GPU requirements
India will require 10-12 gigawatts of AI infrastructure capacity within 3 years, driven primarily by inference workloads as the country is fundamentally a consumer nation
Rack densities have evolved dramatically from 10kW to 130+kW currently, with future generations reaching 240+kW and potentially 1 megawatt per rack
AI sovereignty requires processing data at its source for maximum efficiency, with India generating 20% of world’s data but having only 3% of global data center capacity
Real-world AI applications are already delivering significant economic benefits, such as preventing millions of dollars in fraud daily through government applications
Speed-to-token is critical for ROI, as expensive GPU investments (often $2 billion worth) require rapid monetization through compressed 4-6 month deployment timelines
Future-proof design requires thinking in terms of pods and bounding boxes that can accommodate multiple GPU generations rather than individual rack optimization
Energy efficiency optimization should focus on thermal and load cycle management with integrated telemetry from chip to data center level
Prefabricated systems and reference designs are essential for scaling at speed while addressing skill development challenges
Resolutions and action items
Vertiv has established 8-12 week training programs with IIT Chennai for data center operations and maintenance, with online enrollment available
NVIDIA and Vertiv are providing reference designs and prefabricated solutions to accelerate deployment and reduce complexity
Government of India has launched multiple AI applications including fraud detection systems and multilingual translation services (Bhashini)
Google announced free JEE mock exams on Gemini platform to improve AI accessibility for students
NVIDIA has open-sourced control plane technology to Indian government and companies for local inferencing capabilities
Industry partners are implementing NVIDIA-ready data center certification programs for faster deployment
Unresolved issues
Specific timeline and investment details for India’s semiconductor manufacturing capabilities (Semicon mission) remain unclear
Workforce scaling challenges persist, particularly for specialized skills in liquid cooling and high-density operations that cannot be easily taught in traditional educational settings
Regulatory framework details for data sovereignty requirements under DPDP law implementation are not fully defined
Long-term sustainability of rapid infrastructure scaling given the pace of technological change and equipment obsolescence cycles
Integration challenges between existing traditional data centers and new AI factory requirements for retrofit scenarios
Specific mechanisms for knowledge transfer from global deployments to Indian market to avoid repeating mistakes
Suggested compromises
Balancing sovereignty with innovation by allowing continued processing outside India while building local capabilities for critical sectors
Hybrid approach combining global technology expertise with indigenous solutions and local manufacturing partnerships
Phased deployment strategy using reference designs that can accommodate multiple GPU generations to balance current needs with future requirements
Mixed-mode facilities supporting both traditional CPU workloads and new GPU-intensive AI workloads during transition period
Collaborative approach between global companies and Indian startups/government for technology transfer and capability building
Thought provoking comments
I think India being so aspirational and ready to adopt new technology for the welfare of themselves and the welfare of the businesses, I think we would be the hub of AI development for the world. You will start seeing that happening in a few months, not years.
Speaker
Ankush Sabharwal
Reason
This bold prediction reframes India’s position from a consumer of AI technology to a global hub for AI development, challenging the conventional narrative of India as primarily a market rather than an innovation center.
Impact
This comment set an optimistic and ambitious tone for the entire discussion, prompting other speakers to explore India’s infrastructure readiness and capabilities. It shifted the conversation from theoretical possibilities to concrete timelines and actionable steps.
We cannot learn, right? And if you say, hey, I can create travel AI solutions, very, very difficult, right? So they know travel, they know railways. So it would be, I think, much better to work with them, learn from them.
Speaker
Ankush Sabharwal
Reason
This insight challenges the typical tech industry approach of building universal solutions, instead advocating for domain-specific partnerships where enterprises contribute their deep sector knowledge while AI companies provide the technical infrastructure.
Impact
This comment fundamentally shifted the discussion from technology-first to problem-first thinking, influencing how other speakers discussed AI implementation and the importance of understanding specific use cases before building solutions.
We used to live in the cloud world at project scales of 18 months. We live now in the GPU world of project scales of between four and six months. So a dramatic compression of schedules, a dramatic increase in capacity.
Speaker
Peter Panfil
Reason
This observation reveals a fundamental shift in infrastructure deployment timelines, highlighting how AI demands are forcing the industry to completely rethink traditional project management and deployment strategies.
Impact
This comment introduced urgency into the discussion and helped explain why traditional data center approaches are inadequate for AI infrastructure. It led to deeper exploration of prefabrication, reference designs, and the need for speed at scale.
We are a consumer country. We have always been in the mode of first to consume, then to build. And thereby, we are the largest chat GPT consumer base for the globe.
Speaker
Jigar Halani
Reason
This honest assessment of India’s position provides crucial context for understanding the scale of opportunity – if India is the largest consumer but processes this data elsewhere, bringing that compute capacity domestically represents enormous potential.
Impact
This comment provided the foundation for discussing data sovereignty and the economic opportunity of processing India’s own data domestically. It helped quantify the potential market size and justified the ambitious infrastructure projections.
Having a node down means few hundred dollars getting downtime. A GPU node down translates to few thousands of dollars going into a downtime… if a node fails, you start from the checkpoint… For eight hours of, say, 4,000 GPUs of time multiplied by that much is what you have lost the compute time… Hundreds of thousands of dollars. Real money.
Speaker
Jigar Halani
Reason
This stark comparison between traditional computing and AI infrastructure costs reveals why reliability and reference designs are not just technical preferences but economic necessities, fundamentally changing how infrastructure must be approached.
Impact
This comment elevated the discussion from technical specifications to business imperatives, explaining why cutting corners in AI infrastructure is economically devastating and why following proven reference designs is crucial.
The most efficient and effective place to process data is at the source of the data… I see a world where the data gets generated, it gets processed as closely and as quickly after the generation of the data.
Speaker
Peter Panfil
Reason
This reframes data sovereignty from a protectionist concept to an efficiency optimization, suggesting that local processing isn’t just about control but about optimal resource utilization and performance.
Impact
This comment shifted the sovereignty discussion from regulatory compliance to technical and economic optimization, providing a more compelling business case for local AI infrastructure development.
We generate 20% of world’s data, and what data center capacity the country has today is 3% of the world’s data center capacity… We still have a long, long way to go in building the large-scale data centers just to make sure our own data we process it by ourselves.
Speaker
Jigar Halani
Reason
This data point starkly illustrates the massive infrastructure gap between India’s data generation and processing capacity, providing concrete justification for the ambitious gigawatt-scale projections discussed throughout the session.
Impact
This statistic provided quantitative backing for all the ambitious infrastructure projections and helped the audience understand the scale of opportunity and necessity for rapid AI infrastructure development in India.
Overall assessment
These key comments fundamentally shaped the discussion by establishing India’s unique position as both a massive consumer and potential producer of AI capabilities. The conversation evolved from abstract possibilities to concrete business cases, driven by insights about economic imperatives (downtime costs), market opportunities (data generation vs. processing capacity), and strategic advantages (domain expertise partnerships). The speakers successfully built a compelling narrative that India’s AI infrastructure development isn’t just aspirational but economically necessary and technically feasible, with the discussion moving from ‘whether’ to ‘how’ and ‘how fast’ throughout the session.
Follow-up questions
How can India accelerate the development of indigenous AI chips to achieve complete sovereignty in the AI stack?
Speaker
Jigar Halani
Explanation
Jigar mentioned that while India is progressing well in 4 out of 5 layers of the AI stack (energy, infrastructure, models, applications), the missing piece is indigenous chip manufacturing, which is crucial for complete AI sovereignty
What specific strategies can be implemented to bridge the digital divide and make AI more accessible to underprivileged and rural populations in India?
Speaker
Akanksha Swarup
Explanation
This was raised as a concern about inclusivity in AI adoption, with only a brief mention of Google’s JEE mock exams initiative, but requiring deeper exploration of comprehensive solutions
How can the integration of chip-level telemetry with data center-level telemetry be implemented to optimize energy efficiency automatically?
Speaker
Srikanth Cherukuri
Explanation
This was identified as a key opportunity for AI factories to achieve better energy optimization through automated systems rather than manual control room operations
What are the detailed technical specifications and implementation guidelines for NVIDIA-ready data center certification programs?
Speaker
Moderator
Explanation
The certification program was mentioned but not elaborated upon, requiring more detailed information about requirements, processes, and benefits
How can India scale talent development programs to meet the exponential growth in AI infrastructure requirements?
Speaker
Dal Bhanushali (Audience)
Explanation
With capacity doubling every year, there’s a critical need for comprehensive skill development programs beyond the current 8-12 week programs mentioned
What are the specific economic impact measurements and ROI calculations for AI implementations in government services like fraud detection and language translation?
Speaker
Jigar Halani
Explanation
While examples were given of millions of dollars saved daily through AI fraud detection, detailed economic impact studies and ROI frameworks need further research
How can prefabricated AI factory systems be standardized and scaled for rapid deployment across different geographical and regulatory environments?
Speaker
Srikanth Cherukuri
Explanation
The concept of prefab systems was discussed as crucial for scaling, but detailed implementation strategies and standardization approaches require further development
What are the optimal cooling technologies and PUE optimization strategies for different climate zones across India?
Speaker
Sanjay Kumar Sainani
Explanation
While thermal management across India’s diverse climate was discussed, specific regional optimization strategies and technology selections need detailed research
How can the transition from CPU-based to GPU-based critical infrastructure be systematically managed for enterprise customers?
Speaker
Peter Panfil
Explanation
The complexity of this transition was acknowledged but detailed methodologies and best practices for enterprise-level transitions require further exploration
Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.
Related event

