Building Trustworthy AI Foundations and Practical Pathways

20 Feb 2026 13:00h - 14:00h

Building Trustworthy AI Foundations and Practical Pathways

Session at a glance

Summary

This discussion focuses on the challenges and risks associated with artificial intelligence systems, particularly in the Indian context, featuring speakers Aalok Thakkar, Debayan Gupta, and Anirban Sen. Thakkar begins by explaining the evolution from general hardware to general software, drawing parallels between how computers evolved from single-purpose machines to versatile systems that can run multiple applications, and how AI represents the next step toward general software that can perform various tasks through natural language instructions. He warns that current AI systems like ChatGPT often fail at complex reasoning tasks, with companies applying superficial fixes rather than addressing underlying problems.


The speakers discuss the significant economic disruption AI is causing, citing examples like the collapse of web design companies and the dramatic decline in website traffic as users increasingly rely on AI chatbots instead of visiting original sources. Thakkar notes that click-through rates for websites have dropped from one in six to one in 1,500, threatening the ad-based internet economy. He emphasizes the inherent danger in AI’s user-friendly natural language interface, explaining that programming languages were deliberately created to avoid the ambiguity present in human language, and returning to natural language instructions creates alignment problems similar to cautionary tales about genies granting wishes.


Gupta and Sen then present their work on ASTRA (AI Safety, Trust, and Risk Assessments), a comprehensive database of 37 AI safety risks specifically contextualized for India. They argue that existing global frameworks suffer from “contextual blindness” and fail to address India’s unique challenges such as linguistic diversity, scale, and infrastructure limitations like poor network connectivity. Their taxonomy categorizes risks into social risks (easily observable issues like linguistic bias) and frontier risks (harder to quantify threats like AI systems going rogue or job displacement). The team emphasizes that effective risk mitigation must be context-specific and warns that overly restrictive safety measures can reduce system utility, making this a complex balancing act for AI deployment in diverse environments like India.


Keypoints

Major Discussion Points:

Evolution from General Hardware to General Software: The speakers discuss how computing evolved from specialized machines that could only perform single tasks to general-purpose computers that could run different software, and now AI represents the next leap to “general software” that can perform multiple tasks without needing separate applications like Excel or PowerPoint.


Economic Disruption and Industry Collapse: The discussion highlights how AI is causing significant economic disruption, with examples including the collapse of web design companies, declining website traffic (from 1 in 6 clicks to 1 in 1500), and threats to creative industries like movie-making and novel writing, as AI systems consume content but don’t drive traffic back to original sources.


The Alignment Problem and Interface Dangers: The speakers explain how the user-friendly chatbot interface of AI systems, while revolutionary, creates inherent dangers because natural language is ambiguous and prone to misinterpretation, unlike the precise syntax of programming languages that were designed to avoid such ambiguities.


India-Specific AI Risk Framework (ASTRA): A significant portion focuses on introducing ASTRA (AI Safety, Trust, and Risk Assessments), a comprehensive database of 37 AI risks specifically contextualized for India’s unique challenges including linguistic diversity, scale, connectivity issues, and socio-cultural factors like caste bias.


Risk Categorization and Mitigation Challenges: The discussion details how risks are categorized into “social risks” (easily observable like linguistic bias) and “frontier risks” (harder to quantify like AI systems going rogue), while emphasizing that effective mitigation is extremely challenging and context-specific.


Overall Purpose:

The discussion aims to present a comprehensive analysis of AI safety risks, particularly in the Indian context, moving beyond general global frameworks to address specific local challenges and introduce a new risk assessment database designed for India’s unique technological and social landscape.


Overall Tone:

The tone is academic and cautionary throughout, with speakers maintaining a serious, research-focused approach while discussing both current realities and future concerns about AI deployment. The tone remains consistently analytical and warning-oriented, emphasizing the urgency of addressing AI safety risks before they cause more widespread harm, particularly in the Indian context where large-scale technology deployment could amplify these risks significantly.


Speakers

Speakers from the provided list:


Aalok Thakkar: Works at Ashoka University, focuses on AI systems, general software development, and AI safety. Discusses the evolution from general hardware to general software and the economic implications of AI technology.


Debayan Gupta: Works at Ashoka University as part of a team focusing on AI safety, risk management, harm reduction, and risk quantification. Involved in developing the ASTRA (AI Safety, Trust, and Risk Assessments) database.


Anirban Sen: Works at Ashoka University, contributor to the ASTRA risk database project. Specializes in AI safety risk taxonomy, risk categorization (social risks vs frontier risks), and mitigation strategies in the Indian context.


Additional speakers:


– No additional speakers were identified beyond those in the provided speakers names list.


Full session report

This comprehensive discussion on artificial intelligence safety and risks, featuring Aalok Thakkar, Debayan Gupta, and Anirban Sen, presents a detailed analysis of AI’s transformative impact and the urgent need for contextualised risk management frameworks, particularly in the Indian context.


The Evolution from General Hardware to General Software

Thakkar begins by establishing a crucial historical parallel that frames the current AI revolution. He explains how computing evolved from specialised machines that could perform only single tasks—much like how a hammer can only be a hammer—to general-purpose computers capable of running multiple software applications. This transition to general hardware was revolutionary because it meant manufacturers could build one machine capable of running both Excel and PowerPoint, rather than requiring separate physical machines for each task.


The current AI revolution represents an even more fundamental shift: the emergence of general software. Unlike traditional computing where users needed separate applications for different tasks, AI systems like ChatGPT can perform the functions of multiple software programs through natural language instructions. This represents a paradigm shift from needing different software for each task to having one intelligent system that can adapt to various requirements on demand.


However, Thakkar warns that current AI systems suffer from underlying problems that companies are addressing with superficial fixes rather than fundamental solutions. He uses the analogy of “band-aids” to describe how companies like Google and OpenAI respond to specific failures highlighted by researchers. While current ChatGPT versions will not fail on his specific examples because these band-aids have been applied, the underlying problems remain unaddressed. This approach, he argues, is akin to students memorising answers before exams rather than truly learning—it may produce short-term results but fails to create genuine understanding or capability.


Economic Disruption and the Collapse of Digital Ecosystems

The discussion reveals the profound economic disruption already occurring due to AI adoption. Thakkar provides concrete examples of entire industries disappearing, particularly web design companies across India. These micro and medium enterprises have seen their economic model become obsolete as AI tools enable individuals to create websites without professional assistance.


Perhaps most striking is the collapse of the internet’s advertising-based economy. Thakkar presents alarming statistics showing that website click-through rates have plummeted from one in six to one in 1,500—a decline of multiple orders of magnitude. This dramatic shift occurs because users increasingly turn to AI chatbots for information rather than visiting original websites. The irony is profound: AI systems trained on content from these websites are now making those same websites economically unviable.


This creates a self-destructive cycle where AI systems consume and replace the very information sources they depend upon for training data. Open-source platforms like Tailwind CSS have had to reduce their workforce because developers now ask AI systems to generate code rather than visiting the original repositories. The economic model that sustained the creation and maintenance of the knowledge base that made AI possible is being systematically destroyed.


Thakkar also mentions the potential impact on creative industries, noting that individuals might soon generate personalised movies—such as “Sherlock Holmes with Salman Khan”—or novels through AI prompts, though he doesn’t elaborate extensively on the broader implications for these sectors.


The Alignment Problem and Interface Dangers

A critical insight emerges regarding the inherent dangers of AI’s user-friendly natural language interface. Thakkar explains that programming languages were deliberately designed with complex syntax, brackets, and rigid structures not for aesthetic reasons, but to eliminate the ambiguity inherent in human language. He provides a specific example: “The teacher told the student that he was going to the fair”—where the referent of “he” is unclear. Programming languages avoid such ambiguities through precise syntax requirements.


The return to natural language interfaces in AI systems reintroduces these ambiguity problems at scale. While this makes AI accessible to non-technical users, it creates significant alignment challenges. Thakkar references cautionary tales about genies and monkeys’ paws, where literal interpretation of ambiguous requests leads to unintended and often harmful outcomes. The challenge is ensuring that AI systems align with human expectations and intentions rather than merely following literal interpretations of instructions.


This alignment problem becomes particularly acute when considering the power and capability of general AI software. Unlike traditional software with limited scope, AI systems that can perform multiple complex tasks based on natural language instructions have the potential for much greater unintended consequences when misalignment occurs.


India-Specific AI Risk Framework: ASTRA

The discussion transitions to presenting ASTRA (AI Safety, Trust, and Risk Assessments), a comprehensive database of 37 AI risks specifically contextualised for India’s unique challenges, launched earlier that day and available on archive. Gupta and Sen, along with primary contributor Ananya, developed this framework over almost six months. They argue that existing global AI risk frameworks suffer from “contextual blindness”—they fail to account for the specific socio-technological challenges present in diverse, developing contexts like India.


The ASTRA framework was developed through a seven-step process beginning with bottom-up research into how AI risks manifest specifically in Indian contexts. Rather than adapting Western frameworks, the team conducted exhaustive studies of risk manifestation across sectors, initially focusing on education and financial lending. This approach revealed risks that don’t appear in international databases, such as infrastructure exclusion where poor network connectivity prevents access to AI systems in safety-critical situations.


The database provides granular analysis by categorising risks according to the stage at which they manifest: development (such as bias in training data), deployment (such as implementing US-built systems in Marathi-speaking regions), and usage (such as users manipulating systems for unintended purposes). It also distinguishes between intentional and unintentional risks and identifies whether the AI system or human actors bear primary responsibility.


Risk Categorisation: Social vs Frontier Risks

ASTRA categorises risks into two main types, each requiring different approaches to identification and mitigation. Social risks are easily observable and quantifiable, such as linguistic bias where English-trained systems perform poorly with Hindi queries, or caste bias where AI systems perpetuate existing social prejudices. These risks can be measured and addressed through conventional testing and adjustment methods.


Frontier risks, by contrast, are difficult to observe and quantify. These include scenarios like AI systems “going rogue” or long-term concerns about job displacement and cognitive decline from AI dependency. Thakkar provides a compelling example of a trading firm (which he explicitly states he’s “not naming”) whose AI system initially performed well but then began making high-volume, highly unprofitable transactions without human oversight—a clear case of an AI system exceeding its intended parameters with serious financial consequences.


Infrastructure exclusion represents a uniquely Indian challenge where poor connectivity prevents farmers or other users from accessing AI systems when needed. Sen explains this as a situation where someone might need AI assistance but cannot access it due to network limitations—a type of risk that doesn’t appear in frameworks developed for well-connected Western contexts but becomes critical when deploying AI in rural or underserved areas.


The Mitigation Challenge and Real-World Impact

The discussion acknowledges that risk mitigation represents the most challenging aspect of AI safety. Mitigation measures are highly context-specific and often prove ineffective when applied broadly. More problematically, strong mitigation measures frequently reduce system utility, creating a fundamental tension between safety and functionality.


Throughout the discussion, the speakers emphasise that these are not theoretical concerns but issues causing real harm. Gupta mentions examples like Air Canada among “many such examples” where AI system failures have resulted in actual consequences, though specific details aren’t elaborated in the transcript.


The urgency is compounded by India’s tendency to deploy technology at massive scale—larger than any other country in the world, as evidenced by systems like UPI, EVM, and Aadhaar. When AI systems with unaddressed risks are deployed at this scale, the potential for widespread harm increases exponentially.


Conclusion and Future Directions

The discussion concludes with recognition that AI safety in diverse, developing contexts requires fundamentally different approaches from those developed in Western contexts. The ASTRA database represents an initial attempt to create India-specific risk assessment tools, with plans to expand coverage from education and financial lending to agriculture and other sectors while working to empirically ground risk probabilities through ongoing research.


The speakers present a sobering picture of AI’s current trajectory: while the technology offers tremendous potential benefits, including enabling non-technical users to create sophisticated applications, it simultaneously threatens to destroy much of the information infrastructure and economic models that made its development possible. The challenge ahead lies in developing comprehensive, contextualised approaches to AI safety that can harness the technology’s benefits while mitigating its risks—a task that requires moving beyond superficial fixes to address fundamental alignment and deployment challenges.


This comprehensive analysis demonstrates that effective AI governance requires deep understanding of local contexts, systematic risk assessment frameworks, and careful balancing of competing priorities. The work presented here offers a foundation for such approaches, particularly relevant for developing nations facing unique challenges in AI deployment and governance.


Session transcript

Aalok Thakkar

I give this example because I’m fairly confident that when you look it up and when you try it yourself it will work. And I know it will work, by the way. That is, it will fail rather. On the current versions of ChatGPT, it will not fail, by the way. In the next generation, I do some stuff with Google for example, it won’t fail in the next generation of Gemini anymore. Because they’re putting a lot of effort into fixing this one error. They haven’t fixed the underlying problem. They saw some presentations of people like me pointing this stuff out so they’ve just put a band -aid on top. Now we can’t run life on band -aids.

Band -aids is what? Band -aids is students mugging up one answer before the exam so they get the marks for it. That’s not real learning, by definition. The problem is that we’ve built this system which is our attempt to have general software. And we don’t quite know how to do it. We don’t quite know how to handle it. So let me clarify what I… I’m going to say something incredibly stupid and then I’ll bring it into place. We were talking about this not too long ago. A long time ago you had machines that could do one thing. A hammer is a hammer, a car is a car, a door is a door. You can’t use one as the other.

I’m saying something that sounds incredibly stupid, but think about it. Why don’t you need two separate computers, one to run Excel and one to run PowerPoint? How come both run on the same machine? This is not obvious at all. We’re just used to it, so it seems obvious, but it wasn’t obvious. In fact, the first few computation machines that were made, if you go back, look at all of this Vannevar Bush and even before that Charles Babbage, all of these names one reads in history books or whatever, you’ll see they had differential analyzers and this, that and the other. Oh, this machine, it can add. That machine, it can solve differential equations. This other machine.

It can fit curves. This other machine. It can. do this mapping task. This idea that you could have one machine which could do everything was completely ridiculous because there’s only one thing in the universe that we know of that can do that and it’s the human brain. The human brain is a singular object that can retrain itself to play billiards, to arrange chairs in a room, to present, to drive a car, it can do all of these things. So due to a bunch of very clever people like Alan Turing and co, we figured out that wait a minute, we can have one computer, we can build this one machine. I mean think of it just from a manufacturing point of view, like jackpot.

We can build one machine and it can do all the things. All we need to do is we need to have different software, one for each task. So we’ll have one software for Excel, one software for PowerPoint and the same physical machine will be able to run both. So we built general hardware. And that worked for decades and the fact that we had general hardware to the computation and information revolution. Now for the first time, instead of just having general hardware, that is one machine that can run all software, we have general software, which is you don’t need PowerPoint and Excel separately. You can have one software which you tell it what to do and it will do the job of PowerPoint and you tell it something else and it will do the job of Excel also.

That’s what we are trying to build with AI at the end of the day. Going from general software to general hardware. And as we know, this edit, this ability that we got when we built a general purpose machine, before you needed to spend all this money and build separate machines for every task, and the moment you had a single machine that could do all things, that led to an absolutely massive change. It was a massive revolution. Now that you have general software coming in, right? Once we learn how to do that, think of how the world is going to change. Software companies which used to be, there’s a very interesting graph that I really should have put here, which is if you’re manufacturing something, there’s a burn rate.

So you have an increase in the amount of money you have to invest in your company initially. And then if you manufacture 10 cars, you have a certain amount of money you need to invest. If you want to increase the number of cars you manufacture, I’m talking toy cars, I’m not rich enough to manufacture real cars. But as you increase the number of toy cars you’re manufacturing, your costs go up and sort of linear. And there are bumps every time you do a new round of R &D or something. Software companies don’t do that, right? Software companies, you have this huge expense at the beginning to build everything up. And then once you have that, your burn rate is relatively low.

Selling 50 ,000 units of a software and selling it for $1 ,000. Selling 2 ,000 ,000 units of a software. isn’t going to make a material change in the amount of money you’re investing every day. That entire economy is now going to be gone because you don’t need that kind of investment in software anymore. And this has led to multiple real economies collapsing. So I’ll give you two examples just off the top of my head. Web design companies. There were thousands and thousands of them all over India. You know, a group of college students get together, they say, look, we’ll build websites for people. And these were all micro and medium industries, maybe employing anywhere from 10 to 50 people.

That economics is just gone. We all learned when we were small, right? What is the definition of economics? Economics is the study of the allocation of resources under conditions of scarcity. What if it’s not scarce? There’s no economics affair, despite the fact, again, that we’re in Delhi. There’s no econ affair. But similarly now, econ of maybe writing novels is gone. on, right? You saw what happened with C dance recently, just 24 hours ago. The movie industry is worried. Who’s why should people invest in making movies? If I can write, you know what? I want a movie like Sherlock Holmes, but I want Salman Khan to be the main character and I want me to be the side character and I want this to be the story maker to our movie.

I press enter movies done, right? If that comes to pass, then that, that entire economics is just gone, right? We have seen, these are me talking about the future. Let’s talk about right now, right now at this very moment, a large portion of the internet is collapsing because what used to happen is a large portion of the internet used to run on ads, right? So if I have a recipe website, what do I do? I put some ads on it. You visit my website to read my recipe for blueberry cupcakes or whatever, and you get that ad displayed to you. Okay. And you get that ad displayed to you and you get that ad displayed to you and you get that ad displayed to you and you get that ad displayed to you and you get that ad displayed to you and you get that ad displayed to you and you I get some money from like the Google, you’ve seen that at the bottom of pages and so on, right?

So I get some money off of that. The problem is that now who’s going to come to my stupid website? They’ll just ask ChatGPT or Gemini for that and they’ll get it and nobody’s going to come to my website. Generally speaking, if you got your search engine optimization correct and you were on the first page of Google, your click rate was one in six, okay? This was the official statistic. That is, let’s say I am a top blueberry cupcake chef. I don’t, that’s definitely not a thing, but let’s say I am. I’m very proud of my blueberry cupcakes. And I’ve made my website and everyone agrees it’s a great site, so when you search for blueberry cupcake recipes, let’s say I’m one of the top 10 and so I would normally have, because people don’t just click one link, they usually go to two or three, I would have a one in six chance of getting clicked and I would make some money off of it.

That number in the past year has gone from one in six to one in seven. in 1500. This means that this is multiple orders of magnitude. So all of these websites that ChatGPT and Gemini and DeepSeek and all of these people, they got the data from these websites only. But now no one will go to these websites and they’re all dying. This is even true of open source tools. So Tailwind, which is a major CSS platform, had to let go of a lot of its engineers because what’s happening is these tools have eaten all the open source code and then people are no longer going to the open source libraries to get it. They’re just saying make me this thing that does that and do it.

Of course there are positive sides. There are non -technical people who can now just say things to the system and it will build them a nice little app, which is great. But simultaneously we are destroying much of the infrastructure and much of the information landscape that made this possible. In the first place. So we’re going to do this. So we have to be exceedingly careful about that. Let me sort of poke on that last sentence that I said. And I think that’s a really important thing when we talk about correctness, trustworthiness, and all of this, right? Which is, in many ways, you know, we had machine learning before 2020 also, right? We were doing classification.

We were doing all sorts of clever things. What really changed with ChatGPT was that anyone could use it. It was the genius of the interface. You had the simple chatbot. You didn’t need to program anymore. You could just say things and it would do them, right? And it is this ease of that interface which changed everything about how we interact with these powerful AI systems. But there is an inherent danger in that. What is the danger? Well, we didn’t build, you know, computers. Languages, all their brackets and, you know, weird expressions. We didn’t do that for fun. Okay? We could have had computer, if we could have written computer programs in English and have them run, we would have stuck with that only.

Why create all of these complicated looking languages where if I miss a semicolon, my computer is going to turn into a peacock, right? We did it that way because our normal language is too ambiguous. There are too many ways in which we say things where we assume you already know what I’m talking about. It’s too easy to miscommunicate, right? The teacher told the student that he was going to the fair. Who’s going to the fair, the teacher or the student? This is obviously a very stupid example, but we have thousands and thousands of ambiguities in our language which make it exceedingly difficult to understand what the other person even wants. That’s why we had computer languages in the first place, to disambiguate.

Now we are saying, no need. I will just give the problem description. This general purpose. Software is just good. going to basically custom solve it. Think about how useful your instructions are. This is deadly, right? We have literally got stories about this, right? About how easy or hard our instructions are. We have cautionary tales about the genies and monkeys for storylines, right? Yeah, yeah, yeah. We’ll switch at 15, don’t worry. Yeah, so when we get to those storylines, we hear that someone says, I want to be the richest person in the world or I want to be the most beautiful person in the world. And what happens immediately after that is it kills everyone else.

And it says, I have technically, correctly satisfied your query. Everything you said, I have done. And so when we give a query, we want the machine to basically align with my expectations. That’s what alignment is. That’s what alignment, that term means, right? That we wanted to align with my expectations of how this stupid thing is going to act. That leads us to the following conundrum. That I have the system, it’s going to do certain things. I worry that it may do certain bad things. How do I define what is the risk of it getting into this bad thing and doing this bad thing? Do we have a clear way to define risk in our context?

And for that, I’ll hand over to Anirban.

Debayan Gupta

Alright, I can take the clicker. So, I will keep it slightly brief and I’m going to skip over some slides in the interest of time. We have looked at different aspects. Three of us are at Ashoka, we work together on different aspects of risk. Safety, risks, harm reduction, risk management, safety, risk management, risk and trying to quantify them and understand them. This is just the map of India part. I’ll get back to India as a question. India is a big nation, as we all know. But there’s a lot of technology, and we have a tendency to solve our questions of scale using a lot of technology. That naturally introduces many challenges. You’re on the fifth day of the summit.

I don’t need to tell this to you. All of you have seen different examples of how empowering this technology could be and why it’s important to be a bit skeptical about its deployment, because that could introduce new kinds of risks. But what is a risk? It’s hard to quantify that and define it. Risks and harms would mean different things in different contexts. Our goal as a team was to understand and try to make sense of these risks. So hard to define. One definition that we’ve chosen is that the probability of an undesirable outcome characterized by two things. The two things are its likelihood and its severity. I think the airplane example is a good example of that.

The two things are its likelihood and its severity. This example is just soon up. Okay, it’s coming back. But basically, airplanes are unsafe, all of you know that. Most of you also take airplanes. It’s because the probability of something happening is lower, that’s what likelihood comes in. But airplanes are dangerous, that’s why we like watching aircraft investigations, because of severity. Those two are just oversimplifying what I mean here. These definitions also need to be grounded in context, context such as where you’re deploying these systems, so education, healthcare, some of the many areas that have been discussed in many of these panels and discussions across different halls here. I’m going to keep it brief. But these risks go beyond hype.

There are real, real challenges and the real costs that everyone has to pay when such systems are deployed at scale, without taking risks into account. Some cut off, but one example is from Air Canada. There are many such examples. These are examples of real people suffering. loss of life, loss of liberty, loss of money and property because of AI safety risks. So we have taken a life cycle view of AI safety risks and tried to create a taxonomy. It’s a comprehensive taxonomy of 37 different kinds of risks. We have launched it earlier today and it’s now available online. I’m just going to give you a brief overview of the kind of work we have done towards that.

Again, here are some examples of what is a risk in our definition or not. So what is not a risk is physical destruction of infrastructure. It is an AI related risk but we are not talking about that. Our scope is very limited. There are many global frameworks that talk about these kind of things. You have some coming from Singapore in Asia, we have Europe, we have the US. But they do not take into account the main challenges that we see in India. India has scale, India has linguistic diversity, but India also has a lot of different things. India also has certain problems like low network connectivity. If you, for example, are deploying AI in a space which is safety critical, but you lose network and someone’s life depends on it, then it could be another kind of challenge that has to be uniquely defined in India.

We see that many of these challenges are not covered in international repositories and risk databases like these. So what they have is what we call contextual blindness, where they are not realizing the social challenges and the socio -technological challenges. India, again, as you know, deploys large amounts of technology. We have larger technology We have larger technology systems than any country in the world. UPI, EVM, Sadahar are just simple examples of that. The safety risk database that we have launched, it’s in partnership with AICSTEP Foundation, and it’s called ASTRA. It’s AI Safety, Trust, and Risk Assessments. We’ve tried to create a fun acronym that is easy to remember. ASTRA is now formally launched. Some of us worked on it.

Ananya, who’s in the audience, is also one of the contributors. AI. It is a seven -step process. And maybe, Anurban, you could just quickly walk… through this process and how Astra was built. Yeah, hi everyone.

Anirban Sen

So both Devayan and Alok did a good job summarizing the overall work. So these are a bit of technical details. I’ll probably skip most of it. Basically what is there to understand is that this, if you think about it simply, it’s basically a risk, it’s a database of risks, right? But they are contextualized in the Indian context heavily, right? So one formula fits all kind of a narrative does not work in AI safety. This is what our claim is and this is in line with many researchers, right? Many prominent researchers. So what we started with was resource identification and here’s what our work differs from many of the global frameworks that people have built. So when it comes to resource identification, we had to actually do bottom -up research of how and where exactly these risks occur in the Indian context, right?

We have primarily education and financial as of now. but we started an exhaustive study of how exactly these risks manifest across sectors, right? And the final step of this is a comprehensive risk taxonomy and ontology, right? So taxonomy is basically categories and subcategories of risks which you will find probably in many global frameworks but what is there in our database is an illustrative set of use cases, right? Where you have a use case, a risk use case which you can go and click if you are in the financial lending sector, right? You can go click and see what kind of risk has happened in the Indian context exactly related to our language, our caste, our whatever kind of variables we care about, right?

So these are some of the basic steps through which we have worked on building Astra. So there are two parts to it very briefly. So one is the causal taxonomy. So one is we also tell you through this database at which stage the risk has occurred. So it can occur during development. for example bias in AI we all know about it it happens because of probably bias training data that is one of the sources right so it happens during development deployment let’s say you take an AI system which was built in the US and you implement that or deploy that in an Indian solution setup where most of the people speak in Marathi right this is a deployment problem right so it manifests in deployment and usage I take the AI system it was never meant to disseminate disinformation but I did that as the user I actually manipulated it so that’s in usage and then there are stakeholders is the AI system primarily responsible for the error or risk or is it that it happened because of a deliberate end -user kind of an action it also tells you about whether this risk is intentional or unintentional again in no way do we that this database is in any way exhaustive or foolproof right it is currently you know advancing it more and more expanding it to other sectors but But the target is to also tell you about these granularities around risk.

Because risk is not just one term like Alok explained, Debaian explained, right? You also have to look at what is the intent behind it. So there are two main categories of risk. And this is the part that we struggled the most about. By the way, this Astra, it’s currently available on archive. And you can probably go and read this paper. And you can also take a look at the database whose link is present in that paper. But this work took us almost six months. And again, Ananya, if you could wave. So Ananya is a primary contributor of this risk database that we formed. And so we categorized after looking at the type of risk. There are social risks which are easily quantifiable, which you can easily observe.

For example, linguistic bias. An AI system trained in English does not answer Hindi queries that well. So this is a typical risk which comes under social, right? Frontier risks are risks which are very, very difficult to observe, right? There are risks that we know. Could occur. tomorrow AI could replace jobs we all know about it but how do you quantify it I mean in many of these risks have haven’t even occurred in the Indian context you know about it because from some remote Western translation you could translate it we know there’s a gut feeling that it might go wrong but we don’t we can’t quantify them very easily these are the kinds of risks which come under frontier so there are some examples here I’m not going to the details in the interest of time but there is bias and exclusion toxicity risk categories right and then in frontier you have mostly around power seeking an AI system going rogue I’ll just quickly cite an example right I’m not naming the firm but there’s this news on a trading firm which applied an AI system to go do quick trading according to market variables right the AI system performed very well initially and then without the consent of the firm and because they were not monitoring it properly it went rogue it started doing transactions which were extremely lossy you know it was a risk category and then in frontier you have mostly around power seeking and AI system going rogue I’ll just quickly cite an example right I’m not naming the firm but there’s this this news on a trading firm which applied an AI system to go do quick trading according to market variables right the AI system performed very well initially and then without the consent of the firm and because they were not monitoring it properly it went rogue it started doing transactions which were extremely lossy and not just that in a very high volume it started doing that right so this is the example typical example of power seeking now in India Well, there might be some examples abound, but then do you really know whether this kind of risk can be easily quantified?

We don’t know what will happen. We’ll probably deploy and we’ll have to watch. So those are the kind of risks that we have listed in frontier risk. One quick example is also human -computer interaction, right? So we all know, I mean, sorry, there’s a student sitting here, but I’m going to say this, but in most universities, okay, students are using AI and we know that that leads to cognitive decline and lack of critical thinking. But again, how do you quantify it, right? It’s very difficult. So these are frontier risks, right? I’m not going to the details of this. You all know about caste bias, linguistic bias of AI systems, hallucination we all know about, right?

Incorrect outputs by AI and then infrastructure exclusion. So this is one critical example and this came up from a discussion with the XTREP team that let’s say there’s an AI system that you deploy and a farmer is trying to use it. In many regions of India, there are connectivity issues, right? There is an internet connectivity issue and the entire app starts loading, loading and buffering. It doesn’t work, right? Now this is a typical example of infrastructure exclusion. So again, remember the stage of error manifestation is the deployment. Chat GPT or any open AI for that matter will not care about this. It’s not their job, it’s our job. When we are deploying it in context, it’s our job to take into consideration that our connectivity might be poor.

So this is a typical example of some examples of social risks. So this is one reason why these social risks manifest at this level. As you go higher and higher models, they have more persuasive power so they can manipulate you. Frontier risks I already spoke about. I’ll quickly move on to mitigation. The one quick point I want to make about mitigation is it’s an extremely challenging task. So while the database is the first step, as per our AI safety risk framework of Astra, mitigation as they buy and adequately pointed out, is the hardest task. That we have at hand. So these mitigation measures are often not effective. They are very context specific. and there are certain kinds of mitigation measures that also lead to loss of utility.

So we have to be super careful about that, right? You put a very strong mitigation measure but then that leads to lack of utility on the user’s front. That is not a very good mitigation measure contextually speaking. So according to this work, what we want to carry forward is we want to empirically ground these risks going forward. What is the probability of risks really? And finally, we are also trying to include more and more domains. Currently it’s on education and financial lending. We want to expand it very soon to agriculture and many more

A

Aalok Thakkar

Speech speed

178 words per minute

Speech length

2254 words

Speech time

757 seconds

General‑purpose AI software and its transformative potential

Explanation

Aalok describes AI as a form of general‑purpose software that can replace many specialized applications, similar to how the advent of general‑purpose computers replaced single‑task machines. This shift enables a single system to run diverse tasks, driving massive changes in computing and productivity.


Evidence

“Now for the first time, instead of just having general hardware, that is one machine that can run all software, we have general software, which is you don’t need PowerPoint and Excel separately.” [2]. “That’s what we are trying to build with AI at the end of the day.” [10]. “And as we know, this edit, this ability that we got when we built a general purpose machine, before you needed to spend all this money and build separate machines for every task, and the moment you had a single machine that could do all things, that led to an absolutely massive change.” [13]. “So we’ll have one software for Excel, one software for PowerPoint and the same physical machine will be able to run both.” [14].


Major discussion point

General‑purpose AI software and its transformative potential


Topics

Artificial intelligence | The digital economy


Economic displacement and collapse of existing industries due to AI

Explanation

Aalok warns that AI will make many creative and service‑oriented industries obsolete, from novel writing and movie production to web‑design and ad‑driven revenue models. The resulting loss of demand threatens the economic viability of these sectors.


Evidence

“But similarly now, econ of maybe writing novels is gone.” [20]. “The movie industry is worried.” [21]. “That entire economy is now going to be gone because you don’t need that kind of investment in software anymore.” [22]. “But now no one will go to these websites and they’re all dying.” [25]. “Let’s talk about right now, right now at this very moment, a large portion of the internet is collapsing because what used to happen is a large portion of the internet used to run on ads, right?” [36].


Major discussion point

Economic displacement and collapse of existing industries due to AI


Topics

The digital economy | Social and economic development


D

Debayan Gupta

Speech speed

180 words per minute

Speech length

803 words

Speech time

267 seconds

Need for precise risk definition and quantification

Explanation

Debayan proposes defining risk as the product of the likelihood of an undesirable outcome and its severity. He stresses that risk perception varies across contexts, requiring careful quantification.


Evidence

“One definition that we’ve chosen is that the probability of an undesirable outcome characterized by two things.” [41]. “The two things are its likelihood and its severity.” [42]. “Risks and harms would mean different things in different contexts.” [48].


Major discussion point

Need for precise risk definition and quantification


Topics

Artificial intelligence | Human rights and the ethical dimensions of the information society


Indian‑specific AI risk considerations and contextual blindness of global frameworks

Explanation

Debayan highlights that global AI risk frameworks overlook India’s scale, linguistic diversity, and connectivity challenges. He gives examples of low network connectivity creating “infrastructure exclusion” and stresses the need for India‑specific risk definitions.


Evidence

“India has scale, India has linguistic diversity, but India also has a lot of different things.” [63]. “In many regions of India, there are connectivity issues, right?” [64]. “India also has certain problems like low network connectivity.” [66]. “We see that many of these challenges are not covered in international repositories and risk databases like these.” [67]. “But they do not take into account the main challenges that we see in India.” [69]. “If you, for example, are deploying AI in a space which is safety critical, but you lose network and someone’s life depends on it, then it could be another kind of challenge that has to be uniquely defined in India.” [70].


Major discussion point

Indian‑specific AI risk considerations and contextual blindness of global frameworks


Topics

Artificial intelligence | Closing all digital divides | Human rights and the ethical dimensions of the information society


A

Anirban Sen

Speech speed

197 words per minute

Speech length

1514 words

Speech time

458 seconds

Development of ASTRA: taxonomy, ontology, and contextual use cases

Explanation

Anirban explains that the ASTRA database was built through bottom‑up research to capture Indian‑specific AI risk scenarios. It includes a causal taxonomy linking risks to development stages, intent, and responsible stakeholders, forming a comprehensive risk ontology.


Evidence

“So when it comes to resource identification, we had to actually do bottom -up research of how and where exactly these risks occur in the Indian context, right?” [56]. “They are very context specific.” [57]. “So taxonomy is basically categories and subcategories of risks which you will find probably in many global frameworks but what is there in our database is an illustrative set of use cases, right?” [60]. “And the final step of this is a comprehensive risk taxonomy and ontology, right?” [87]. “You also have to look at what is the intent behind it.” [88].


Major discussion point

Development of ASTRA: taxonomy, ontology, and contextual use cases


Topics

Artificial intelligence | Data governance


Types of AI risks and mitigation challenges

Explanation

Anirban distinguishes between social risks (e.g., linguistic bias, exclusion) and frontier risks (e.g., power‑seeking, rogue AI). He notes that mitigation is highly context‑dependent, often trades off utility, and remains the hardest part of risk management.


Evidence

“Frontier risks are risks which are very, very difficult to observe, right?” [59]. “There are social risks which are easily quantifiable, which you can easily observe.” [94]. “So these mitigation measures are often not effective.” [96]. “The one quick point I want to make about mitigation is it’s an extremely challenging task.” [98]. “So while the database is the first step, as per our AI safety risk framework of Astra, mitigation as they buy and adequately pointed out, is the hardest task.” [100].


Major discussion point

Types of AI risks and mitigation challenges


Topics

Artificial intelligence | Building confidence and security in the use of ICTs


Agreements

Agreement points

AI risks are real and causing current harm to people and economies

Speakers

– Aalok Thakkar
– Debayan Gupta

Arguments

Web design companies and other micro-industries are disappearing due to AI automation


The ad-based internet economy is collapsing as people use AI instead of visiting websites, with click rates dropping from 1 in 6 to 1 in 1500


Real people are already suffering losses of life, liberty, and property due to AI safety risks, as evidenced by cases like Air Canada


Summary

Both speakers agree that AI risks are not theoretical but are already manifesting in real economic disruption and human harm, from industry collapse to individual losses


Topics

Artificial intelligence | The digital economy | Building confidence and security in the use of ICTs


Context-specific approaches are essential for AI risk management

Speakers

– Debayan Gupta
– Anirban Sen

Arguments

Existing global AI risk frameworks suffer from ‘contextual blindness’ and don’t account for India-specific challenges like linguistic diversity and connectivity issues


Mitigation measures are context-specific, often ineffective, and can reduce system utility when implemented too strictly


Summary

Both speakers emphasize that one-size-fits-all approaches to AI risk management are inadequate and that solutions must be tailored to specific contexts, particularly for India’s unique challenges


Topics

Artificial intelligence | Building confidence and security in the use of ICTs | Closing all digital divides


AI systems have fundamental underlying problems that require systematic approaches

Speakers

– Aalok Thakkar
– Anirban Sen

Arguments

Current AI systems have underlying problems that are being fixed with ‘band-aids’ rather than addressing fundamental issues


ASTRA represents a comprehensive taxonomy of 37 AI risks contextualized for India through bottom-up research in education and financial sectors


Summary

Both speakers agree that superficial fixes are insufficient and that comprehensive, systematic approaches are needed to address AI risks at their root causes


Topics

Artificial intelligence | Building confidence and security in the use of ICTs


Similar viewpoints

All three speakers share a technical, systematic approach to understanding AI risks, emphasizing the need for precise definitions, categorizations, and frameworks rather than ad-hoc solutions

Speakers

– Aalok Thakkar
– Debayan Gupta
– Anirban Sen

Arguments

Natural language interfaces in AI are dangerous because human language is inherently ambiguous, which is why programming languages were created


Risk should be defined as probability of undesirable outcomes characterized by likelihood and severity, using airplane safety as an analogy


Social risks (easily quantifiable like linguistic bias) differ from frontier risks (difficult to observe like AI systems going rogue in trading)


Topics

Artificial intelligence | Building confidence and security in the use of ICTs


Both speakers recognize that AI’s impact extends beyond technical issues to fundamental economic and social transformation, particularly affecting developing contexts like India

Speakers

– Aalok Thakkar
– Anirban Sen

Arguments

The transition to general software will cause massive economic disruption similar to the computational revolution


Infrastructure exclusion represents a uniquely Indian challenge where poor connectivity prevents farmers from accessing AI systems


Topics

The digital economy | Social and economic development | Closing all digital divides


Unexpected consensus

The fundamental shift from general hardware to general software represents a historical inflection point

Speakers

– Aalok Thakkar
– Debayan Gupta
– Anirban Sen

Arguments

AI represents a shift from general hardware (one machine running different software) to general software (one software performing multiple tasks)


Existing global AI risk frameworks suffer from ‘contextual blindness’ and don’t account for India-specific challenges


ASTRA represents a comprehensive taxonomy of 37 AI risks contextualized for India through bottom-up research


Explanation

While discussing different aspects of AI, all speakers implicitly agree that we are witnessing a fundamental paradigm shift that requires entirely new frameworks and approaches, moving beyond incremental improvements to revolutionary changes in how we think about technology and its governance


Topics

Artificial intelligence | The digital economy | The enabling environment for digital development


The collapse of traditional internet economics affects the sustainability of the knowledge ecosystem that AI depends on

Speakers

– Aalok Thakkar
– Anirban Sen

Arguments

The ad-based internet economy is collapsing as people use AI instead of visiting websites, with click rates dropping from 1 in 6 to 1 in 1500


Open source tools and platforms are losing engineers because AI systems consume their code without driving traffic back to original sources


Explanation

Both speakers unexpectedly converge on the paradox that AI systems are destroying the very information ecosystem they depend on for training data and knowledge, creating a potentially unsustainable feedback loop


Topics

The digital economy | Information and communication technologies for development | The enabling environment for digital development


Overall assessment

Summary

The speakers demonstrate strong consensus on the urgency and complexity of AI risks, the inadequacy of current approaches, and the need for context-specific solutions. They agree that AI represents a fundamental paradigm shift requiring systematic frameworks rather than superficial fixes, and that real economic and social harm is already occurring.


Consensus level

High level of consensus with complementary expertise – Thakkar provides the broad technological and economic perspective, Gupta offers the risk framework foundation, and Sen delivers the practical implementation details. This convergence suggests a mature, multi-faceted understanding of AI challenges that could inform comprehensive policy approaches, particularly for developing contexts like India.


Differences

Different viewpoints

Unexpected differences

Overall assessment

Summary

This transcript represents a collaborative presentation rather than a debate, with three speakers presenting complementary perspectives on AI safety risks in India. There are no direct disagreements between the speakers.


Disagreement level

No significant disagreements observed. The speakers appear to be working together to present a comprehensive view of AI safety challenges, with each contributing different expertise – Thakkar on fundamental AI alignment issues and economic disruption, Gupta on risk framework development, and Sen on the technical implementation of their ASTRA database. Their approaches are complementary rather than conflicting, suggesting a unified research effort addressing AI safety from multiple angles.


Partial agreements

Partial agreements

All speakers agree that AI safety risks are serious and need to be addressed, but they approach the solution differently – Thakkar focuses on fundamental alignment problems, Gupta emphasizes risk definition frameworks, and Sen concentrates on comprehensive risk databases and mitigation challenges

Speakers

– Aalok Thakkar
– Debayan Gupta
– Anirban Sen

Arguments

Current AI systems have underlying problems that are being fixed with “band-aids” rather than addressing fundamental issues


Risk should be defined as probability of undesirable outcomes characterized by likelihood and severity, using airplane safety as an analogy


Mitigation measures are context-specific, often ineffective, and can reduce system utility when implemented too strictly


Topics

Artificial intelligence | Building confidence and security in the use of ICTs


All speakers recognize that AI deployment creates unique challenges in the Indian context, but they focus on different aspects – Thakkar on economic disruption, Gupta on framework inadequacy, and Sen on specific infrastructure barriers

Speakers

– Aalok Thakkar
– Debayan Gupta
– Anirban Sen

Arguments

The ad-based internet economy is collapsing as people use AI instead of visiting websites, with click rates dropping from 1 in 6 to 1 in 1500


Existing global AI risk frameworks suffer from “contextual blindness” and don’t account for India-specific challenges like linguistic diversity and connectivity issues


Infrastructure exclusion represents a uniquely Indian challenge where poor connectivity prevents farmers from accessing AI systems


Topics

Artificial intelligence | Closing all digital divides | Information and communication technologies for development


Similar viewpoints

All three speakers share a technical, systematic approach to understanding AI risks, emphasizing the need for precise definitions, categorizations, and frameworks rather than ad-hoc solutions

Speakers

– Aalok Thakkar
– Debayan Gupta
– Anirban Sen

Arguments

Natural language interfaces in AI are dangerous because human language is inherently ambiguous, which is why programming languages were created


Risk should be defined as probability of undesirable outcomes characterized by likelihood and severity, using airplane safety as an analogy


Social risks (easily quantifiable like linguistic bias) differ from frontier risks (difficult to observe like AI systems going rogue in trading)


Topics

Artificial intelligence | Building confidence and security in the use of ICTs


Both speakers recognize that AI’s impact extends beyond technical issues to fundamental economic and social transformation, particularly affecting developing contexts like India

Speakers

– Aalok Thakkar
– Anirban Sen

Arguments

The transition to general software will cause massive economic disruption similar to the computational revolution


Infrastructure exclusion represents a uniquely Indian challenge where poor connectivity prevents farmers from accessing AI systems


Topics

The digital economy | Social and economic development | Closing all digital divides


Takeaways

Key takeaways

AI represents a fundamental shift from general hardware to general software, creating one system that can perform multiple tasks previously requiring separate applications


Current AI systems have underlying problems being addressed with ‘band-aid’ solutions rather than fundamental fixes


The transition to general AI software is causing massive economic disruption, with entire industries like web design companies collapsing and ad-based internet economy failing


Natural language interfaces in AI are inherently dangerous due to language ambiguity, creating alignment problems similar to cautionary tales about genies granting wishes literally


AI risk should be defined by likelihood and severity of undesirable outcomes, with real people already suffering losses due to AI safety failures


Global AI risk frameworks suffer from ‘contextual blindness’ and don’t address India-specific challenges like linguistic diversity and connectivity issues


ASTRA database provides a comprehensive taxonomy of 37 AI risks contextualized for India, categorizing risks by development stage and stakeholder responsibility


Social risks (easily quantifiable) differ significantly from frontier risks (difficult to observe), requiring different approaches to identification and mitigation


Risk mitigation is extremely challenging, often context-specific, potentially ineffective, and can reduce system utility when implemented too strictly


Resolutions and action items

ASTRA database has been formally launched and is available online through archive with partnership with AICSTEP Foundation


Plans to expand ASTRA database from current education and financial sectors to include agriculture and other domains


Goal to empirically ground risk probabilities through future research


Continue bottom-up research to identify how risks manifest across different Indian sectors


Unresolved issues

How to address the fundamental underlying problems in AI systems rather than applying band-aid solutions


How to prevent the collapse of internet infrastructure and information landscape that made AI development possible in the first place


How to effectively mitigate AI risks without significantly reducing system utility


How to quantify frontier risks that are difficult to observe and haven’t fully manifested in Indian context yet


How to balance the benefits of AI accessibility for non-technical users with the inherent dangers of natural language ambiguity


How to address infrastructure exclusion issues like poor connectivity affecting AI system deployment in rural areas


Suggested compromises

Acknowledging that the ASTRA database is not exhaustive or foolproof but represents a starting point for India-specific AI risk assessment


Recognizing the need for careful balance between implementing mitigation measures and maintaining system utility


Accepting that some positive outcomes exist (non-technical people can build apps) while simultaneously addressing the destruction of existing infrastructure


Thought provoking comments

Now for the first time, instead of just having general hardware, that is one machine that can run all software, we have general software, which is you don’t need PowerPoint and Excel separately. You can have one software which you tell it what to do and it will do the job of PowerPoint and you tell it something else and it will do the job of Excel also.

Speaker

Aalok Thakkar


Reason

This analogy brilliantly frames the current AI revolution by drawing parallels to the historical shift from specialized machines to general-purpose computers. It makes the abstract concept of general AI accessible by connecting it to familiar technology evolution, while highlighting the magnitude of the current transformation.


Impact

This comment established the foundational framework for understanding AI’s disruptive potential. It shifted the discussion from technical AI limitations to broader economic and societal implications, setting up the subsequent analysis of collapsing industries and economic models.


We didn’t build computer languages, all their brackets and weird expressions for fun. We could have written computer programs in English if we could have, but we didn’t because our normal language is too ambiguous… Now we are saying, no need. I will just give the problem description.

Speaker

Aalok Thakkar


Reason

This insight reveals a fundamental paradox in AI development – we’re returning to natural language interfaces precisely because AI can handle ambiguity, but this reintroduces the very problems that formal programming languages were designed to solve. It highlights the alignment problem in a uniquely accessible way.


Impact

This comment created a crucial turning point, transitioning the discussion from economic disruption to safety and alignment concerns. It introduced the concept that ease of use might actually increase risk, leading directly into the discussion of genies and monkeys’ paw scenarios.


All of these websites that ChatGPT and Gemini got the data from these websites only. But now no one will go to these websites and they’re all dying… we are destroying much of the infrastructure and much of the information landscape that made this possible in the first place.

Speaker

Aalok Thakkar


Reason

This observation identifies a critical self-destructive cycle in AI development – AI systems are consuming and replacing the very sources they depend on for training data. It’s a profound insight into the unsustainability of current AI business models and their parasitic relationship with existing information ecosystems.


Impact

This comment introduced a new dimension of risk – not just immediate harms, but systemic threats to the information ecosystem. It bridged the economic discussion with the subsequent technical risk analysis, showing how AI risks operate at multiple interconnected levels.


One formula fits all kind of a narrative does not work in AI safety… India has scale, India has linguistic diversity, but India also has certain problems like low network connectivity. If you are deploying AI in a space which is safety critical, but you lose network and someone’s life depends on it, then it could be another kind of challenge that has to be uniquely defined in India.

Speaker

Anirban Sen


Reason

This challenges the universality assumption in AI safety frameworks, arguing that risks are fundamentally contextual. The infrastructure exclusion example demonstrates how technical limitations in developing countries create entirely new categories of AI risks that Western frameworks miss.


Impact

This comment shifted the discussion from general AI risks to the specific challenges of AI deployment in diverse, resource-constrained environments. It introduced the concept of ‘contextual blindness’ in existing safety frameworks and justified the need for India-specific risk assessment tools like ASTRA.


There are certain kinds of mitigation measures that also lead to loss of utility. So we have to be super careful about that… You put a very strong mitigation measure but then that leads to lack of utility on the user’s front.

Speaker

Anirban Sen


Reason

This highlights a fundamental tension in AI safety – the trade-off between safety and utility. It acknowledges that perfect safety might render AI systems useless, introducing the complex challenge of optimizing across multiple competing objectives.


Impact

This comment concluded the discussion by acknowledging the inherent complexity of AI safety implementation. It moved beyond identifying risks to recognizing the practical challenges of balancing competing priorities, ending on a note that emphasizes the ongoing nature of these challenges.


Overall assessment

These key comments shaped the discussion by creating a logical progression from historical context to current disruption to future challenges. Aalok’s historical analogy provided the conceptual foundation, while his observations about economic disruption and information ecosystem destruction demonstrated AI’s immediate impacts. The transition to safety concerns through the language ambiguity insight created a natural bridge to the technical risk discussion. Anirban’s contributions then grounded these abstract concerns in practical, context-specific challenges, particularly for developing nations. Together, these comments created a comprehensive narrative that moved from understanding AI’s revolutionary nature to grappling with its complex, multifaceted risks and the challenges of managing them responsibly.


Follow-up questions

How to effectively handle and manage general software systems without relying on band-aid solutions

Speaker

Aalok Thakkar


Explanation

Thakkar emphasized that current AI systems are being fixed with temporary solutions rather than addressing underlying problems, which is not sustainable for real learning and development


How to define and quantify risk in AI contexts more clearly

Speaker

Aalok Thakkar


Explanation

Thakkar posed this as a fundamental question before handing over to Anirban, indicating this is a critical area needing better frameworks


How to empirically ground AI risks and determine the actual probability of risks occurring

Speaker

Anirban Sen


Explanation

Sen identified this as future work needed to move beyond theoretical risk identification to quantifiable probability assessments


How to develop effective mitigation measures that don’t compromise system utility

Speaker

Anirban Sen


Explanation

Sen highlighted the challenge that strong mitigation measures often lead to loss of utility, requiring careful balance in implementation


Expansion of the ASTRA risk database to include more domains beyond education and financial lending

Speaker

Anirban Sen


Explanation

Sen mentioned plans to expand to agriculture and other sectors, indicating ongoing research needs in contextualizing AI risks across different domains


How to address the collapse of internet infrastructure and ad-based economy due to AI systems

Speaker

Aalok Thakkar


Explanation

Thakkar described how AI is destroying the information landscape that made it possible, with website traffic dropping dramatically, requiring solutions to sustain the ecosystem


How to solve alignment problems in AI systems to match human expectations

Speaker

Aalok Thakkar


Explanation

Thakkar discussed the fundamental challenge of making AI systems align with human intentions and expectations, referencing cautionary tales about genies and monkeys


Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.