Steering the future of AI

8 Jul 2025 16:45h - 17:15h

Session at a glance

Summary

This discussion features Yann LeCun, Meta’s Chief AI Scientist and one of the “godfathers of AI,” in conversation with Nicholas Thompson from The Atlantic about steering the future of artificial intelligence. LeCun clarifies his controversial stance that large language models (LLMs) are a “dead end,” explaining that while LLMs are useful tools, they are insufficient for achieving human-level AI or artificial superintelligence. He argues that LLMs lack four essential capabilities for true intelligence: understanding the physical world, persistent memory, reasoning, and planning. LeCun identifies two fundamental limitations of current LLMs: they use a fixed amount of computation for every response regardless of complexity, and they cannot engage in the kind of prolonged thinking that humans use for difficult problems.


As an alternative, LeCun advocates for JEPA (Joint Embedding Predictive Architecture), which would allow AI systems to understand the world through abstract representations rather than trying to predict every pixel detail in video. He expresses optimism that human-level AI could be achieved within 5-10 years, though likely through architectures beyond current LLMs. Regarding AI safety, LeCun believes future AI systems will be designed to be objective-driven with built-in guardrails, making them safe by construction, comparing the risk level to airplane safety through careful engineering and regulation.


LeCun strongly advocates for open-source AI development, arguing it leads to faster progress, better safety through transparency, and prevents dangerous concentration of AI power in a few companies. He envisions a future with a small number of dominant open-source AI platforms, similar to how Linux dominates computing infrastructure today, potentially requiring international cooperation to train models on distributed data while preserving sovereignty.


Keypoints

**Major Discussion Points:**


– **Limitations of Large Language Models (LLMs):** LeCun argues that while LLMs are useful tools, they are insufficient for achieving human-level AI because they lack essential capabilities like planning, reasoning, persistent memory, and understanding of the physical world. He clarifies that he considers them a “dead end” only in the context of reaching artificial superintelligence, not as useful technology.


– **JEPA Architecture as an Alternative:** LeCun proposes Joint Embedding Predictive Architecture (JEPA) as a superior approach to current LLMs. This architecture would make predictions in abstract representation space rather than at the pixel level, allowing AI systems to understand the physical world more effectively by eliminating unpredictable details.


– **AI Safety and Design Philosophy:** LeCun expresses confidence that AI systems can be made safe by design, comparing the risk level to airplane safety. He advocates for objective-driven systems with built-in guardrails, arguing that the likelihood of AI causing major harm is extremely small when properly engineered.


– **Open Source vs. Closed AI Development:** LeCun strongly advocates for open-source AI development, arguing it leads to faster progress, better safety through more scrutiny, and prevents dangerous concentration of AI power. He discusses how Meta’s open-source Llama models have accelerated global AI development, including in China with DeepSeek.


– **Future AI Landscape and Global Cooperation:** LeCun envisions a future dominated by 1-2 open-source AI platforms (similar to Linux in operating systems) and proposes international partnerships for training foundation models where countries contribute data and computation while maintaining data sovereignty.


**Overall Purpose:**


The discussion aimed to explore LeCun’s perspective on the current state and future direction of AI development, particularly his contrarian views on LLMs, his advocacy for open-source development, and his optimistic stance on AI safety compared to other prominent AI researchers.


**Overall Tone:**


The conversation maintained a consistently intellectual and engaging tone throughout. LeCun was confident and articulate in defending his positions, while Thompson served as an effective moderator by asking probing questions that challenged LeCun’s views. The tone was respectful but direct, with Thompson pushing back on potential contradictions in LeCun’s arguments, particularly around open-source development and geopolitical implications. The audience appeared engaged and supportive, as evidenced by positive reactions to certain points about data sovereignty.


Speakers

– **LJ Rich**: Role/title not specified, appears to be involved in conference organization and introductions


– **Yann LeCun**: Chief AI Scientist at Meta, known as one of the “godfathers of AI” for helping create the architecture that led to current AI systems, conducts research at FAIR (Fundamental AI Research lab at Meta)


– **Nicholas Thompson**: Moderator from The Atlantic


Additional speakers:


None – all speakers in the transcript are included in the provided speakers names list.


Full session report

# Discussion Report: Yann LeCun on the Future of Artificial Intelligence


## Introduction and Context


This discussion featured Yann LeCun, Meta’s Chief AI Scientist, in conversation with Nicholas Thompson from The Atlantic, moderated by LJ Rich. LeCun, who described his role as “I have the best job in the company because I have nobody reporting to me. I don’t run any organization,” focused on his views regarding current AI limitations and future development paths.


## LeCun’s Position on Large Language Models


### Clarification on the “Dead End” Statement


LeCun clarified his widely reported criticism of large language models: “I didn’t say they [LLMs] were a dead end. I said they were a dead end if you are interested in reaching human-level AI or artificial superintelligence… They’re certainly a useful tool… But if we want to reach the type of intelligence that we observe in humans and animals, we are going to have to invent other methods.”


### Four Missing Capabilities for Human-Level AI


LeCun identified four critical capabilities that current LLMs lack:


1. **Understanding the physical world**: “Intelligence needs to be grounded in some reality,” noting that “a four-year-old child has gotten as much information through vision as the biggest LLMs today that are trained on all the publicly available texts on the internet.”


2. **Persistent memory**: Current LLMs cannot maintain long-term memory across interactions.


3. **Reasoning capabilities**: While LLMs can simulate reasoning, they lack deep reasoning abilities.


4. **Planning abilities**: LLMs cannot engage in strategic planning.


### Architectural Limitations


LeCun highlighted fundamental problems with current LLM architecture:


**Fixed Computational Capacity**: “The way an LLM produces an answer is by running through a fixed number of layers in the neural net and then producing a token… And the amount of computation that goes into producing this word is constant… when we have a hard question to answer, we tend to think about it for a long time, right?”


**Language vs. Reality**: LeCun argued that “language is simple. It’s simple because it’s discrete, and it’s discrete because it’s composed of discrete words and there is only a finite number of them,” contrasting this with the complexity of understanding the physical world.


## Alternative Architecture: Joint Embedding Predictive Architecture (JEPA)


LeCun advocated for Joint Embedding Predictive Architecture (JEPA) as an alternative approach. This architecture would make predictions in abstract representation space rather than attempting to predict every detail in video or text. The key advantage, according to LeCun, is that it eliminates unpredictable details by working with abstract representations, allowing AI systems to understand the physical world more effectively.


## AI Safety Philosophy


### Safety by Design Approach


LeCun expressed confidence in AI safety through careful design: “AI systems will be designed with objectives and guardrails, making them safe by construction similar to how jetliners are engineered for safety.” He argued that “the likelihood of AI causing harm is extremely small, comparable to jetliner accidents.”


Thompson challenged this aviation analogy by pointing out “the difference in regulatory approaches between highly regulated aviation industry and current minimal AI regulation.”


### Unresolved Safety Questions


Despite his optimism, LeCun acknowledged: “How do we design or train objectives and guardrails so that AI systems do not do things that are intrinsically dangerous?” He admitted this was “not a solved problem.”


## Democratic and Cultural Risks


LeCun identified concentration of power as a primary threat: “The biggest danger is that there is a future in which every single one of our interactions with the digital world will be mediated by AI assistance… We need to have a very wide diversity of AI assistance that gives people choice about what type of bias, linguistic capabilities, cultural knowledge and cultural value systems… for the same reason we need social diversity in the press.”


## Open Source Development Advocacy


### Arguments for Open Source


LeCun strongly advocated for open-source AI development, arguing that:


– Open source platforms are historically safer due to more people examining them


– Open research accelerates progress by enabling broader contribution


– It prevents concentration of power in a few companies


He noted that “FAIR has open-sourced around 1,000 research projects over 11 years.”


### Challenges and Responses


Thompson raised concerns about national security implications, specifically asking about “models like Llama being used by DeepSeek.” LeCun responded by highlighting the international nature of AI development: “Lama the first version of Lama was built by a team of 13 people, 12 of whom were in Paris… it’s not an American model, it’s a French model… ideas can pop up from anywhere.”


## Future Predictions and International Cooperation


### Timeline for Human-Level AI


LeCun provided a specific timeline: “The absolute shortest time I can imagine for at least being on a good path to reaching human-level intelligence is no less than five years. It could be 10, it could be more.”


### Market Structure and Cooperation


LeCun predicted a future with “2-3 dominant foundation models similar to operating system distribution, with 1-2 being open source.” He proposed that “Future foundation models should be trained through international partnerships preserving data sovereignty… each country or region will provide computation and its own data but will preserve its own data.”


### AI Assistants and “Artificial People”


When asked about Meta potentially creating “artificial people,” LeCun clarified that the focus would be on AI assistants: diverse AI systems that could serve different cultural and linguistic needs while preventing any single company from controlling all digital interactions.


## Key Areas of Disagreement


The main tensions in the discussion centered on:


1. **Regulatory approaches**: LeCun favored industry self-regulation through design, while Thompson highlighted the need for more robust oversight


2. **Open source vs. national security**: Thompson questioned whether open access undermines Western technological advantages


3. **Evidence for safety claims**: Thompson consistently asked for empirical evidence supporting LeCun’s theoretical arguments about open source benefits


## Conclusion


The discussion revealed LeCun’s comprehensive critique of current AI approaches while proposing specific alternatives through JEPA architecture and open-source development. While acknowledging significant unresolved technical and safety challenges, LeCun expressed optimism about achieving human-level AI within 5-10 years through international cooperation and diverse, democratically-oriented AI development.


The conversation highlighted ongoing debates about the best paths forward for AI development, particularly regarding the balance between innovation, safety, and democratic values in an increasingly global technological landscape.


Session transcript

LJ Rich: covering from seeing him speak at World AI Con. So, Jann and you will also be in really good hands with our fabulous moderator from the Atlantic. It’s Nicholas Thompson. Everybody is ready now, so it’s a pleasure to introduce the next session on how we can steer the future of AI. Thank you.


Nicholas Thompson: All right, Jann, you ready to be information-dense? That was a good introduction. How are you? I’m pretty good. I’m as ready as I’ll ever get. Wonderful. Jann is the chief AI scientist at Meta. He is also known as one of the godfathers of AI because he helped create the architecture that led to all the AI that we have today. And he is incredibly interesting and fun to talk to. And given that last panel, I should verify that man is real, and you can tell because there are no Pikachus behind him. So, let’s get cracking. Jann, first thing I want to ask you is, you have often said that large language models are a dead end. That is quite contrary to what many other people on this stage have been saying all day, many other people in the world, many other people investing billions of dollars. Tell me whether you still believe they’re a dead end and why.


Yann LeCun: Okay, it’s a bit of a fake news due to the soundbite habit. I didn’t say they were a dead end. I said they were a dead end if you are interested in reaching human-level AI or artificial superintelligence, if you want to call it this way, in the sense that they’re not sufficient. They’re certainly a useful tool. There’s a lot of people working on LLMs. I use them. Everybody now in the world uses them for something. There’s a lot of computer technology, including AI technology, that is very useful, yet is not necessarily a good path towards human-level intelligence. So, LLMs are very limited in some ways. That doesn’t make them useless. It doesn’t mean that we shouldn’t be working on them. In fact, we are, and many of my colleagues are working on it. It’s something we should do. But if we want to reach the type of intelligence that we observe in humans and animals, we are going to have to invent other methods to be the basis of the architecture of the systems. Admittedly, LLMs will still have a role to play because they’re extremely good at transforming thoughts into language. So, we’re still going to use them for that. But what about thoughts? How do we elaborate thoughts? How do we understand the physical world? How do we have persistent memory? How do we reason? How do we plan? Those are four essential things for intelligence that LLMs currently are not really capable of doing to the extent that we’d like. Part of the reason I ask you is because you are so optimistic about AI, and in some ways, your optimism seems to stem from a different belief that some of your colleagues have, and your belief that large language models under current architecture will never become all-powerful or incapable of becoming all-powerful. Is that a correct assessment of your views? Yeah, it’s a correct assessment. I’m certainly optimistic in the sense that I have no doubt that at some time in the future, we’ll be able to produce systems that match or surpass human intelligence in pretty much all domains, and that may arrive fairly soon, possibly within the next decade. The absolute shortest time I can imagine for at least being on a good path to reaching human-level intelligence is no less than five years. It could be 10, it could be more. It’s probably more because it’s almost always harder than we think. There are obstacles on the way that we’re not seeing yet, but I’m certainly optimistic that we will reach that goal. It’s just that LLMs, as they are practiced today, do not have the essential characteristics that are necessary, at least that’s my opinion, and that of an increasingly large number of my colleagues,


Nicholas Thompson: I’m happy to say. And they don’t have the capacity. So large-language models, it is impossible for large-language models to reach human intelligence because they do not have the capacity for planning, they do not have some fundamental capabilities that humans have,


Yann LeCun: and not only that, you think they will never get there. Well, something will get there, and at this point, I think we will not be able to call them LLMs anymore. So here is the two things that I think are essential that LLMs really are not capable of. The first one is the type of inference that they produce. The way an LLM produces an answer is by running through a fixed number of layers in the neural net and then producing a token, which is a word, essentially. And the amount of computation that goes into producing this word is constant. So it’s kind of a problem because when we have a hard question to answer, we tend to think about it for a long time, right? And so if you ask a system, an LLM, to answer by yes or no, the amount of computation you can devote to the answer is fixed. And that doesn’t make any sense because if the answer is simple, of course you can answer it directly, but if it’s complicated, you need to think about it. So the process of thinking about something, generally in humans at least, involves kind of manipulating some mental model of the situation, if you want, and sort of imagining scenarios. And that takes place all in your head. It’s not something you have to write down on paper unless you do mathematics, which is a special case, or perhaps coding. But most of that process takes place in your head using your kind of abstract mental model of the situation you’re thinking of. And you do this for reasoning, for planning, for all kinds of tasks, for everything that we do in our daily lives for which we’re not kind of capable of just taking an action subconsciously without thinking about it, which is what psychologists call the system two type of thinking. So the first characteristic an AI system should have if you wanted to reach human intelligence is this ability to kind of think for a long time and basically produce an answer by search, searching for an answer, not just computing it, but searching for it. This is a very sort of intrinsically different process. That’s the first thing. The second thing is, I’m of the opinion, as many of my colleagues, that intelligence needs to be grounded in some reality. It doesn’t need to be a physical reality. It could be virtual, it could be a situation, but some underlying reality in the sense that the information you get about this underlying reality is much higher bandwidth than what you get through purely text, which is human produced. Okay, so we get a huge amount of information through vision, through touch, through audition. And a four year old child has gotten as much information through vision as the biggest LLMs today that are trained on all the publicly available texts on the internet. And so that tells you, you know, we’re not gonna get to human level AI without having systems that understand the physical world. And here is the problem, with the current architectures that we have, which are basically trained to produce the next tokens, and those tokens have to be discrete, essentially, we cannot use this architecture to get a system to understand the world. We have to use a different architecture. I’ve been advocating for one called JEPA, that means Joint Embedding Predictive Architecture. We explained just a few weeks ago, released an example of this. This is a different architecture. It’s not LLS, it’s completely different. Why don’t you explain the JEPA architecture for the room and explain why that will lead to a more profound form of intelligence than the large language model path that we’re pursuing, so many companies are pursuing. Okay, so really something that may be shocking for a lot of people, language is simple. It’s simple because it’s discrete, and it’s discrete because it’s composed of discrete words and there is only a finite number of them. And so to some extent, it’s relatively simple to kind of predict the sequence of words that will follow, which is what LLMs do. Now the real world, as it turns out, is much more complicated. And of course, the idea that we’re going to train an AI system to understand the physical world by predicting what’s happening in a video is an old idea. I’ve been working on this for almost 20 years now, and my colleagues and I admit I’ve been working on this for the better part of the last 10 years. It turns out it doesn’t work. If you train a system to try to predict all the details of what happens in the video, you don’t get much kind of useful, the kind of representation of the world the system produces are not that useful. And it’s because predicting all the details of what happens in the video is essentially impossible. You know, I’m not seeing the room that with the audience at the moment, I can infer some characteristics of it at the abstract level. I can’t tell how many people are there, but I’m guessing most of the seats are occupied and there is, you know, equal proportion of men and women and, you know, the age distribution. There’s actually several thousand people and they’re all kind of dancing somewhat wildly but mutely, but go on. Okay. Well, that violates my prediction. I guess my distribution of what was happening was wrong. So anyway, there is no way I can predict, you know, the details of what everybody looks like and, you know, what the texture of the carpet is or whatever. Like, you know, there’s a lot of things in sort of video prediction that you just cannot predict. And if you ask a neural net, if you train a neural net to predict all those details, it basically gives up. I mean, it kind of, it can’t really solve the problem. And so it really doesn’t do a good job. Now, of course, if you want to train a system to actually produce cute looking video, you can do that. All right. And we can do that at the moment. But that’s a very far cry from actually understanding the sort of underlying physics if you want or the structure of what goes on. So the idea of JPEG is you have a system that looks at video and you may train it to predict what’s going to happen in a video in time or in space. But instead of making the prediction at the pixel level, you make it in an abstract representation level. So the system is trained to compute a representation, basically a neural net that computes an abstract representation of the world. And it makes prediction in that representation space where a lot of the details that are not predictable have been eliminated, essentially. All right. So you have a different architecture for how we should learn about the world.


Nicholas Thompson: And your view is that it will make AI closer to human intelligence, more powerful, down the line maybe we’ll be using JEP architecture, maybe we’ll be using something else. Let me ask you another question. So one of the exercises I did, initially we were going to have you and Jeff Hinton on stage to debate each other about the risks of AI. It didn’t work out. And instead you’re on today, Jeff is on tomorrow. And I was trying to think about what the fundamental differences are between you and Jeff. One, of course, is your view that there are limits to AI’s intelligence, whereas Jeff has a different view. The second, which I think is really important, is your view that it’s highly unlikely that AI would ever have bad intentions. It’s highly unlikely that any AI system, whether a large language model, whether something else, would ever act against the interests of its creator. Explain why you’re confident in that view and why it gives you confidence that we’re not going to head to a bad scenario with AI. Okay. What I’m confident about is that we will design an AI system once we figure out how to potentially


Yann LeCun: reach human level intelligence or something approaching it. And by the way, this is not going to be an event. It’s going to be kind of progressive innovations. And we’ll start with something at the level of a rat or a cat or something like this before we get to humans. So we’re going to have some time to figure out how to make those systems safe. But what I’m convinced of is that we’ll design those systems so that they are objective driven. You give them an objective, they accomplish it. And you give them guardrails. And by construction, they will have to abide by those guardrails and produce actions, sequences of actions or whatever, that are safe. Now we’ll design them for safety. It will be very different in the design from current LLMs that really don’t have any objectives. The objective of an LLM is specified by the prompt, and it’s a very weak specification because it entirely depends on the type of training the LLM has gone through. In the architecture I’m proposing, there is an objective that needs to be satisfied and the system can do nothing else but accomplish this objective subject to the guardrails. So that’s kind of an architecture that would be safe by construction. That leaves the question of how do we design or train those objectives and those guardrails so that the system does not actually do things that are intrinsically dangerous. And it’s not a solved problem, so I don’t want to minimize the complexity of this. But it’s not an infeasible problem either. The likelihood that some AI system will do something deleterious or dangerous to at least some humans is about the same as the likelihood that the next jetliner you fly on will blow up, which is an extremely, extremely small probability. And it’s very small because the people designing it have been working on the reliability of jet engines and airplanes for decades and kind of fine-tuning them. It’s going to be the same thing. AI systems, there’s going to be a design that makes them safe, and then within this, you’ll have to actually find the details of how you make them safe. Wait, Jan, let’s stay on this metaphor for a second because with jetliners, it’s one of the most regulated industries in the world. Before Boeing wants to build a new engine, they have to go through a million specs. There’s regulation to the left, there’s regulation to the right. Before the jetliner goes off, you check everything. The pilot makes you deplane if there’s extra moisture on the windshield. With AI, we’re building things that are so much more powerful, and there is no regulatory structure right now or minimal regulatory structure right now. We’re charging ahead. I’m not saying that’s a bad thing. My job is just the moderator. But isn’t there quite a difference between the way we regulate the thing on the left and the thing on the right? No. If you want to deploy an AI-based system to do a medical diagnosis or a medical image analysis or a driving assistance system for your car, it also has to go through careful testing and basically clinical trial for medical systems and authorization of deployment by government agencies. I mean, I think it’s pretty much the same. Now, there is no such regulation for chatbots because what they can do at the moment is… I mean, the risk is tiny and extremely… The risk of major harm is small. I mean, people have been sort of playing around this for the last two years. There were a lot of discussions as to whether LLMs are the type that we know today. Because of concerns of this type, it turns out LLMs have been around for a long time. They’re increasing in power, but not by huge amounts. And the world has not ended. And in fact, not much bad things have happened because of them.


Nicholas Thompson: Let’s talk about one of the architectural things that you’ve been a very vocal, powerful… Proponent of which is open source. And one of the arguments you’ve made from the very beginning is that open source architecture will lead to More safety if you have more Companies more countries more individuals involved in decisions. You’ll get much safer outcomes Is there any empirical evidence two years in that that is true or that it is false. There’s two. Yeah There’s two factors. The first thing is that historically it is the case that open source


Yann LeCun: Platforms and AI is going to be a platform, right? It’s going to be some sort of common platform that a lot of people are going to use and build upon So it is the case that open source platform software Are safer and more portable More flexible. I mean there’s all kinds of advantages and the reason why for example the entire Software architecture of the Internet and the mobile communication network is open source Is because it’s safer more portable, you know There’s more eyeballs looking at it and you know bugs are fixed faster and everything like that There is a pressure from the market for this to happen So that’s the evidence now for for AI systems is a lot of academic studies of safety or studies in general of safety that are enabled by the fact that a number of AI systems are open source. It’s very difficult to assess the the safety Of a system if you don’t have access If you can’t kind of, you know, run it locally kind of modify it in various ways So yes Now there is a more important factor though, because I think the biggest danger of AI going forward is not At least in the short term not whether you know AI systems are going to you know, do bad things or take over the world The biggest danger is that there is a future in which every single one of our interactions with the digital world Will be mediated by AI assistance Okay, so our entire digital diet will come from AI systems It’s already the case that a lot of our digital layer comes from machine learning systems But, you know, they do ranking and filtering and you know all that stuff right and content moderation But in the future, you know, our entire digital diet will come from from AI system. We need to have a very wide Diversity of AI assistance that gives people choice about what type of bias linguistic Capabilities cultural Knowledge and you know cultural value systems, etc Our political bias those assistants have for the same reason we need social diversity in the press If our AI system come from a handful of companies on the west coast of the US or China First of all, a lot of Countries in the world a lot of governments in the world would just not accept that because they will think it’s a danger to their Culture and democracy and so we need high diversity and that’s only enabled by


LJ Rich: Platforms this I mean, this is an extraordinarily important argument. It’s an extremely interesting argument Let me ask you a question. You may not be able to fully answer I know that you are the chief research scientist at Metta You are not in charge of Metta’s product roadmap But Metta has also talked about creating or Mark Zuckerberg on the Dworkish podcast talked about creating You know fully fake individuals because people have a demand for more friends than they currently have so to populate the social platforms that Metta runs with individuals Do you influence product decisions like that with the philosophy? You just gave like if you’re going to make people make sure there’s a diversity of people from different backgrounds with open source


Yann LeCun: Does your philosophy go through the rest of the company or is that something you just believe in the products that you’re building? Okay, so so I don’t build products, right? So Because it’s very confusing I’m chief AI scientist. Okay, and I have the best job in the company because I have nobody reporting to me I don’t run any organization Okay, I I conduct research projects on my own With people who want to work for me. They work with me because they want to know because I’m their boss and I try to Inspire people doing research to kind of work on projects and topics that I think are promising and interesting like for example This JEPA idea I was talking about and and things like persistent memory and reasoning and planning of various types So this is really the mission of FAIR, which is the fundamental AI research lab at Metta Okay And this is really long term. So the horizon of projects we work on is You know on average maybe two to three years sometimes a little shorter Sometimes quite a bit longer. It could be ten years Now I Also have you know a bit of input on strategy about about AI not at the product level but at the kind of more Long-term strategy and perhaps how to approach certain types of market So I was certainly a vocal advocate within the company. I was not the only one for open sourcing Lama for example the Lama series and and various other things in fact FAIR in its 11 year history has open source on the order of 1,000 Research projects and kind of various packages. So it’s actually in the DNA of the company really Open sourcing is is really a practice that preceded my arrival at Metta. You know, it’s not it’s not recent so For reasons of you know discussions of safeties and liability and things like this. We have to do this carefully But but certainly the strategy of open source is something that I’m personally extremely happy that Metta has embraced Yeah Now the you know to the idea that you know, should we build artificial people? No, the the idea is that AI systems are going to amplify human intelligence now the best way you can amplify human intelligence is by providing a system that People are used to interact with okay, which basically are other humans, right? So the best way to kind of Help people in their daily lives is to provide them with assistance that basically behave a little bit like humans or not You know frustrating to talk to because they don’t understand what you’re telling them but in that future, you know, we’ll all of us will behave like Like a boss of a manager of a team of potentially super intelligent people who would be working for us and Assisting us, you know in our daily lives That’s making us more Intelligent more rational Perhaps even wiser we can always move. Yeah I I asked that question in part guy I just think it’s such an interesting it’s such an interesting moment where the company that is building AI that will shape so much of society has a lead AI scientist who has These amazing but also quite different from much of the rest of world views about open source It’ll be extremely interesting to watch. I don’t think anybody knows how it plays out let me ask you another question on open source, which is Your models and the models built by llama Were used as part of the architecture in deep seek which in some way is a great validation of what you’ve been saying and It’s a way that AI has spread around the world to many people in America’s national security this was Not what they were looking for


Nicholas Thompson: Explain your views on that Well, so the magic of open research And open source is that you accelerate progress because you get more people involved and they have different ideas and Every idea every good idea kind of gets picked up by everybody else. So


Yann LeCun: So yes deep seek learn from llama. I mean they learn from you know, the whole history of AI before and but then vice versa when when DeepSea came out and and other open source model in China and by the way China is very aggressive about open sourcing their model which I think is interesting and a good thing we learn from them too we we learn you know a few tricks from from DeepSea and Quen and and and the other models and it’s not just you know my colleagues and I in research it’s you know everyone in the in the industry and so the big question is not you know you know who who has the best ideas because if those ideas are shared then they profit everybody the question is who is most capable of basically building on top of those ideas and making them useful for for people and then what matters is the speed at which the information circulates so you you might think that closing up research in the West in the Western world okay like let’s say you know some catastrophic scenario of banning open source or something like this which maybe you know stop geopolitical rivals from getting those ideas but you know they will seep anyway at some point it will just be with it with a delay the problem is we’ll be slowing us down ourselves we’ll we’ll be progressing slower and so basically the geopolitical rivals would be more you know able to kind of keep up because we’re not going to go as fast so it’s a bit like shooting yourself in the foot right I mean what you we have to do is be the the best country region industry whatever you want to call it at picking up on innovations and then pushing them to the market to become the leader and that goes through fast communication and information exchange but wait yeah let me I feel like they’re two kind of competing ideas and the answer you just gave there one is that it sounds like you think you want the West to be leading in AI and if that’s the case and if that is the goal surely it would be better for China not to be able to just go online and look at the structure of llama or perhaps even get weights that were you know leaked online


Nicholas Thompson: shouldn’t there be restrictions on you know a model as powerful as llama and the access that folks around the rest of the world can have okay you’re gonna be careful what you wish for here because first of all I’m not saying what I wish for I’m just asking whether there’s a contradiction I mean I feel like you can say it doesn’t really matter whether the West is ahead or behind because the world will share in this technology it’s a human technology it’s kind of part of the theme of this event in any event but I feel like if you’re gonna make the position if you’re gonna argue the position that the West should have a lead then when you also make this second position which would run counter to what you said that’s all I’m asking okay so first of all even if


Yann LeCun: you manage to completely isolate intellectually China from the rest of the world China has excellent resources in terms of engineers and scientists and they’re not going to be far behind in fact they may come up with ideas that the West does not come up with and so in some respect they might actually be ahead which is why DeepSea was such a shock for a lot of people in kind of the Silicon Valley bubble if you want but likely okay but let me let me give you a bit of information here Lama the first version of Lama was built by a team of 13 people 12 of whom were in Paris okay it’s not an American model it’s a French model now Lama 2 then was okay Lama 2 was built on top of this and that required a big team of engineers and scientists and you know spread over you know Paris New York Menlo Park Seattle and you know various London and various other places okay when you scale up when you pick it up and you try to make it real it becomes a bigger team DeepSea was also you know a small team 10-15 people in China most projects start this way same with PyTorch which you know Meta also open-sourced and actually transferred the thing Gemini produced or Gemini produced by Google a lot of it you know is developed in London so it’s not an American thing it’s not like you know the the the US is alone in this thing ideas can pop up from anywhere and we should have a system by which the best ideas circulate but then the the big question is how fast can you capitalize on those ideas to kind of put things out the door and China is not as efficient as as the US and Silicon Valley in particular are doing this let


Nicholas Thompson: me ask you what one quick question because we’re almost out of time here Yam which relates to this in five years maybe in three years how many foundation LLMs do you think the world will be using well there’ll be a small number of large winners as tends to happen in tech or will there be an infinite or not infinite but extremely large number of open source models will be like it is today there are there some very big closed source models and some open source models what’s it gonna be I think it’s gonna be very much like the situation in operating systems today you’re gonna have like two or three one with one of which maybe one or two will be open source and dominant and then two


Yann LeCun: or three would be proprietary and you know we’ll kind of cater to kind of you know niche markets and this is a situation you you you see today in operating systems almost every single computer in the world today runs Linux okay except for a few a few desktop and a few iPhones and really frankly that’s a small number of computers compared to all the the servers and you know the embedded devices and everything in your car and your phone so it’s gonna be the same situation you’re gonna have a one or two dominant open source platforms AI platforms foundation models on top of which the entire industry would be built and then perhaps it would be space for a few non open source ones that are more proprietary the reason I’m saying this is because the future I’m envisioning is one in which you know LLMs or their descendants will basically constitute a repository of all human knowledge and culture and if you want this to happen you need access to data that currently reside perhaps in digital form perhaps not but in various regions of the world various countries and those countries are not ready to just share the data unless they get something out of it and they want AI sovereignty so the only way out of this that I see is some sort of international partnership in which future foundation models will be trained in a distributed fashion so we’ll have one common big you know model repository of all human knowledge but each country or region will provide computation and its own data but will preserve its own data so the the way the global system will be trained is not by exchanging data but by exchanging parameter vectors basically so we can arrive at a consensus model without every region giving up their data and therefore kind of you know preserving some level of


Nicholas Thompson: sovereignty all right this will have to be open source yeah and I don’t think you heard and you can’t see the room but you got some hoots when you said data sovereignty which tells you a little bit about this conference Jan you said you have one of the best jobs in the world you definitely one of the most interesting jobs in the world I think he’s one of the most interesting minds in AI thank you so much Yann LeCun it’s always a pleasure to talk oh my goodness another amazing conversation thank you very much to Yann LeCun and Nicholas Thompson for a mind-bending discussion


Y

Yann LeCun

Speech speed

167 words per minute

Speech length

4451 words

Speech time

1591 seconds

LLMs are not a dead end but insufficient for reaching human-level AI due to lack of planning, reasoning, persistent memory, and world understanding capabilities

Explanation

LeCun clarifies that he didn’t say LLMs were a dead end, but rather that they are insufficient for reaching human-level AI or artificial superintelligence. He argues that while LLMs are useful tools, they lack four essential characteristics for intelligence: the ability to understand the physical world, persistent memory, reasoning, and planning capabilities.


Evidence

LLMs are good at transforming thoughts into language but cannot elaborate thoughts, understand the physical world, maintain persistent memory, reason, or plan effectively


Major discussion point

Limitations and Future of Large Language Models (LLMs)


Topics

Economic | Legal and regulatory


LLMs have fixed computational capacity per token generation, unlike human thinking which varies time based on problem complexity

Explanation

LeCun explains that LLMs produce answers by running through a fixed number of neural network layers with constant computation per token, which doesn’t make sense because complex problems require more thinking time. He argues that humans engage in ‘system two’ thinking, manipulating mental models and imagining scenarios, which takes variable time depending on problem complexity.


Evidence

When humans have hard questions, they think longer; LLMs use the same computational resources whether answering simple yes/no questions or complex problems


Major discussion point

Limitations and Future of Large Language Models (LLMs)


Topics

Economic | Infrastructure


Current LLM architecture cannot lead to artificial superintelligence under existing frameworks

Explanation

LeCun argues that something will eventually reach human-level intelligence, but at that point it won’t be called LLMs anymore. He believes the current architecture of predicting discrete tokens cannot achieve the type of intelligence observed in humans and animals, requiring entirely different methods and architectures.


Evidence

LLMs lack the ability to think for extended periods and search for answers rather than just computing them; they cannot understand the physical world through current discrete token prediction methods


Major discussion point

Limitations and Future of Large Language Models (LLMs)


Topics

Economic | Infrastructure


JEPA (Joint Embedding Predictive Architecture) offers a different approach by making predictions in abstract representation space rather than pixel level

Explanation

LeCun proposes JEPA as an alternative architecture that trains systems to predict what happens in video at an abstract representation level rather than trying to predict all pixel-level details. This approach eliminates unpredictable details while maintaining understanding of underlying structure and physics.


Evidence

Traditional video prediction fails because predicting all details is impossible; JEPA computes abstract representations where unpredictable details are eliminated while maintaining useful understanding


Major discussion point

Alternative AI Architectures and Approaches


Topics

Infrastructure | Economic


Intelligence needs to be grounded in reality with high-bandwidth information from vision, touch, and audition, not just text

Explanation

LeCun argues that intelligence must be grounded in some underlying reality, whether physical or virtual, because text alone provides insufficient information bandwidth. He notes that a four-year-old child receives as much visual information as the largest LLMs trained on all publicly available internet text.


Evidence

A four-year-old gets as much information through vision as the biggest LLMs trained on all public internet text; humans get huge amounts of information through vision, touch, and hearing


Major discussion point

Alternative AI Architectures and Approaches


Topics

Infrastructure | Sociocultural


Future AI systems should be objective-driven with built-in guardrails for safety by construction

Explanation

LeCun envisions AI systems designed with specific objectives and guardrails that they must abide by, producing only safe sequences of actions. This differs from current LLMs where objectives are weakly specified through prompts and depend entirely on training, making future systems safer by architectural design.


Evidence

Current LLMs have objectives specified only by prompts and depend on training; proposed architecture would have built-in objectives and guardrails that systems cannot violate


Major discussion point

AI Safety and Risk Assessment


Topics

Legal and regulatory | Cybersecurity


Agreed with

– Nicholas Thompson

Agreed on

AI systems require careful design and safety considerations


AI systems will be designed with objectives and guardrails, making them safe by construction similar to how jetliners are engineered for safety

Explanation

LeCun compares AI safety to aviation safety, arguing that the likelihood of AI systems causing harm will be extremely small, similar to jetliner accidents. He believes this will be achieved through careful design and decades of fine-tuning, just as has been done with jet engines and airplanes.


Evidence

Jetliner safety achieved through decades of work on reliability; AI systems will undergo similar design processes for safety


Major discussion point

AI Safety and Risk Assessment


Topics

Legal and regulatory | Cybersecurity


Agreed with

– Nicholas Thompson

Agreed on

AI systems require careful design and safety considerations


Disagreed with

– Nicholas Thompson

Disagreed on

Regulatory approach to AI safety


The likelihood of AI causing harm is extremely small, comparable to jetliner accidents

Explanation

LeCun argues that the probability of AI systems doing something dangerous to humans is about the same as a jetliner blowing up – an extremely small probability. He emphasizes that this low risk will be achieved through careful design work similar to aviation engineering.


Evidence

People have been using LLMs for two years without major incidents; aviation achieves safety through careful engineering over decades


Major discussion point

AI Safety and Risk Assessment


Topics

Legal and regulatory | Cybersecurity


Current LLMs pose minimal risk as evidenced by two years of widespread use without major incidents

Explanation

LeCun points out that LLMs have been available and increasingly powerful for some time, with people using them extensively for two years, yet no major harmful incidents have occurred. He argues this demonstrates that current LLMs present minimal risk of significant harm.


Evidence

LLMs have been around for a long time with increasing power; people have been using them for two years without the world ending or major bad incidents


Major discussion point

AI Safety and Risk Assessment


Topics

Legal and regulatory | Cybersecurity


Open source platforms are historically safer, more portable, and flexible due to more people examining and improving them

Explanation

LeCun argues that open source platforms have historically proven safer and more reliable because more people can examine the code, bugs are fixed faster, and there are various other advantages. He notes that the entire software architecture of the internet and mobile networks is open source for these safety and portability reasons.


Evidence

Internet and mobile communication network software architecture is entirely open source; more eyeballs looking at code means bugs are fixed faster


Major discussion point

Open Source AI Development


Topics

Infrastructure | Legal and regulatory


Disagreed with

– Nicholas Thompson

Disagreed on

Evidence for open source safety benefits


Diversity of AI assistants is crucial to prevent a handful of companies from controlling all digital interactions and cultural biases

Explanation

LeCun warns that in the future, all digital interactions will be mediated by AI assistants, making diversity crucial to prevent a small number of companies from controlling cultural values, political biases, and linguistic capabilities. He compares this need to the importance of diversity in the press and argues that many countries would reject AI systems from a handful of US or Chinese companies.


Evidence

Future digital diet will come entirely from AI systems; need for diversity similar to press diversity; countries won’t accept AI systems that threaten their culture and democracy


Major discussion point

Open Source AI Development


Topics

Sociocultural | Human rights


Open research accelerates progress by enabling more people to contribute different ideas and learn from each other

Explanation

LeCun explains that open research and open source accelerate progress because more people get involved with different ideas, and every good idea gets picked up by everyone else. He notes that when DeepSeek learned from Llama, the reverse also happened – Western researchers learned tricks from DeepSeek and other Chinese models.


Evidence

DeepSeek learned from Llama, but Western researchers also learned from DeepSeek and Quen; China is aggressive about open sourcing models; ideas profit everybody when shared


Major discussion point

Open Source AI Development


Topics

Economic | Infrastructure


Agreed with

– Nicholas Thompson

Agreed on

Current AI development involves global collaboration and knowledge sharing


Restricting open source would slow down Western progress more than it would hinder geopolitical rivals

Explanation

LeCun argues that closing up research in the West or banning open source would be like shooting yourself in the foot. While it might delay rivals from getting ideas, it would slow down Western progress more significantly, allowing geopolitical rivals to keep up more easily because the West wouldn’t advance as quickly.


Evidence

Ideas will seep through anyway with delay; restricting open source slows down domestic progress more than it hinders rivals


Major discussion point

Geopolitical Implications of AI Development


Topics

Economic | Legal and regulatory


Disagreed with

– Nicholas Thompson

Disagreed on

Open source vs national security concerns


China has excellent resources and may develop ideas the West doesn’t, as demonstrated by DeepSeek’s impact

Explanation

LeCun points out that even if China were completely isolated intellectually, they have excellent engineering and scientific resources and might come up with ideas the West doesn’t develop. He cites DeepSeek as an example that shocked many in Silicon Valley, demonstrating China’s capability to innovate independently.


Evidence

DeepSeek was a shock to Silicon Valley; first Llama was built by 13 people, 12 of whom were in Paris; DeepSeek was also a small team of 10-15 people in China


Major discussion point

Geopolitical Implications of AI Development


Topics

Economic | Infrastructure


Future foundation models should be trained through international partnerships preserving data sovereignty

Explanation

LeCun envisions international partnerships where future foundation models are trained in a distributed fashion, with each country providing computation and data while preserving sovereignty. The global system would be trained by exchanging parameter vectors rather than raw data, achieving consensus models without countries giving up their data.


Evidence

Countries want AI sovereignty and won’t share data unless they get something out of it; training can happen by exchanging parameter vectors instead of data


Major discussion point

Geopolitical Implications of AI Development


Topics

Legal and regulatory | Human rights


Serves as chief AI scientist conducting long-term research projects without managing products or people

Explanation

LeCun describes his role as chief AI scientist at Meta, emphasizing that he doesn’t build products or manage people directly. Instead, he conducts research projects and tries to inspire people to work on promising topics like JEPA, persistent memory, and reasoning, with project horizons averaging 2-3 years but sometimes extending to 10 years.


Evidence

Has nobody reporting to him; works on long-term projects like JEPA; project horizons average 2-3 years, sometimes up to 10 years


Major discussion point

Meta’s AI Strategy and LeCun’s Role


Topics

Economic | Infrastructure


FAIR has open-sourced around 1,000 research projects over 11 years, reflecting company DNA

Explanation

LeCun notes that Meta’s Fundamental AI Research (FAIR) lab has open-sourced approximately 1,000 research projects and packages over its 11-year history. He emphasizes that open sourcing is in the company’s DNA and predated his arrival, though they must do it carefully for safety and liability reasons.


Evidence

FAIR has open-sourced around 1,000 research projects over 11 years; open sourcing practice preceded LeCun’s arrival at Meta


Major discussion point

Meta’s AI Strategy and LeCun’s Role


Topics

Economic | Legal and regulatory


Advocates for AI systems that amplify human intelligence by providing human-like assistance

Explanation

LeCun argues that the best way to amplify human intelligence is by providing AI systems that people are used to interacting with – essentially other humans. He envisions a future where people will manage teams of potentially super-intelligent AI assistants, making humans more intelligent, rational, and potentially wiser.


Evidence

Best way to amplify intelligence is through systems that behave like humans; future humans will be like managers of super-intelligent teams working for them


Major discussion point

Meta’s AI Strategy and LeCun’s Role


Topics

Economic | Sociocultural


Expects 2-3 dominant foundation models similar to operating system distribution, with 1-2 being open source

Explanation

LeCun predicts the AI landscape will resemble today’s operating system market, with 2-3 dominant models where 1-2 are open source and dominant, while 2-3 proprietary ones serve niche markets. He notes that almost every computer runs Linux except for some desktops and iPhones, which represent a small fraction of all computing devices.


Evidence

Almost every computer runs Linux except desktops and iPhones; servers, embedded devices, cars, and phones mostly use open source operating systems


Major discussion point

Future AI Landscape Predictions


Topics

Economic | Infrastructure


Predicts AI will constitute a repository of all human knowledge requiring international cooperation

Explanation

LeCun envisions future LLMs or their descendants becoming repositories of all human knowledge and culture. He argues this will require access to data from various regions and countries, necessitating international partnerships where countries contribute data and computation while preserving sovereignty through distributed training methods.


Evidence

Future foundation models need data from various regions; countries want AI sovereignty; solution requires international partnerships with distributed training


Major discussion point

Future AI Landscape Predictions


Topics

Sociocultural | Legal and regulatory


Human-level AI may be achievable within 5-10 years, starting with animal-level intelligence

Explanation

LeCun expresses optimism that human-level AI will be achieved, possibly within the next decade, though he acknowledges it could take longer due to unforeseen obstacles. He emphasizes this will be a progressive process, starting with rat or cat-level intelligence before reaching human levels, providing time to develop safety measures.


Evidence

Absolute shortest time for good path to human-level AI is 5 years; could be 10 or more; will start with animal-level intelligence before human level


Major discussion point

Future AI Landscape Predictions


Topics

Economic | Infrastructure


N

Nicholas Thompson

Speech speed

178 words per minute

Speech length

938 words

Speech time

314 seconds

Questions whether LeCun still believes LLMs are a dead end given widespread investment and adoption

Explanation

Thompson challenges LeCun’s previous statements about LLMs being a dead end, noting that this view contradicts what many other speakers, people worldwide, and investors putting billions of dollars believe. He asks LeCun to clarify and defend his position given the widespread adoption and investment in LLM technology.


Evidence

Many other people on stage, worldwide, and investors are putting billions into LLMs; contradicts widespread belief in LLM potential


Major discussion point

Limitations and Future of Large Language Models (LLMs)


Topics

Economic | Infrastructure


Asks for explanation of JEPA architecture and why it leads to more profound intelligence

Explanation

Thompson requests that LeCun explain the JEPA (Joint Embedding Predictive Architecture) to the audience and clarify why this alternative approach will lead to more profound forms of intelligence compared to the large language model path that many companies are currently pursuing.


Major discussion point

Alternative AI Architectures and Approaches


Topics

Infrastructure | Economic


Points out the difference in regulatory approaches between highly regulated aviation industry and current minimal AI regulation

Explanation

Thompson challenges LeCun’s aviation safety analogy by highlighting that the aviation industry is one of the most regulated in the world, with extensive specifications, regulations, and safety checks. He contrasts this with AI development, which currently has minimal regulatory structure while building potentially more powerful systems.


Evidence

Aviation requires million specs before building engines; extensive regulation and safety checks; AI has minimal regulatory structure while building more powerful systems


Major discussion point

AI Safety and Risk Assessment


Topics

Legal and regulatory | Cybersecurity


Agreed with

– Yann LeCun

Agreed on

AI systems require careful design and safety considerations


Disagreed with

– Yann LeCun

Disagreed on

Regulatory approach to AI safety


Questions LeCun’s confidence that AI systems will never act against their creators’ interests

Explanation

Thompson identifies a key difference between LeCun and other AI researchers like Jeff Hinton, specifically LeCun’s belief that it’s highly unlikely AI systems would ever have bad intentions or act against their creators’ interests. He asks LeCun to explain the basis for this confidence and how it informs his optimistic view of AI safety.


Evidence

Contrasts LeCun’s views with Jeff Hinton’s different perspective on AI risks and capabilities


Major discussion point

AI Safety and Risk Assessment


Topics

Legal and regulatory | Cybersecurity


Challenges whether open source truly leads to better safety outcomes and asks for empirical evidence

Explanation

Thompson questions LeCun’s argument that open source architecture leads to better safety outcomes, asking for empirical evidence after two years of implementation. He wants to know if there’s concrete proof that having more companies, countries, and individuals involved in AI development actually produces safer results.


Evidence

Two years of open source AI development provides a timeframe for evaluating safety claims


Major discussion point

Open Source AI Development


Topics

Legal and regulatory | Infrastructure


Disagreed with

– Yann LeCun

Disagreed on

Evidence for open source safety benefits


Questions the contradiction between wanting Western AI leadership while supporting open access to models like Llama

Explanation

Thompson identifies what he sees as competing ideas in LeCun’s position – wanting the West to lead in AI while simultaneously supporting open source access that allows countries like China to benefit from models like Llama. He asks whether there’s a contradiction between these two positions and suggests restrictions might be warranted if Western leadership is the goal.


Evidence

China’s use of Llama architecture in DeepSeek; national security concerns in America about open access to powerful models


Major discussion point

Open Source AI Development


Topics

Economic | Legal and regulatory


Disagreed with

– Yann LeCun

Disagreed on

Open source vs national security concerns


Raises concerns about models like Llama being used by DeepSeek and implications for national security

Explanation

Thompson points out that LeCun’s models and Llama were used in DeepSeek’s architecture, which while validating LeCun’s open source philosophy, created concerns among America’s national security establishment. He frames this as an example of the tension between open source benefits and national security considerations.


Evidence

DeepSeek used Llama architecture; American national security establishment had concerns about this development


Major discussion point

Geopolitical Implications of AI Development


Topics

Cybersecurity | Legal and regulatory


Agreed with

– Yann LeCun

Agreed on

Current AI development involves global collaboration and knowledge sharing


Explores how LeCun’s philosophy on diversity and open source influences Meta’s product decisions

Explanation

Thompson asks whether LeCun’s philosophy about ensuring diversity in AI systems and open source development influences Meta’s product decisions, specifically referencing Mark Zuckerberg’s discussion about creating artificial individuals to populate social platforms. He wants to understand the connection between LeCun’s research philosophy and company products.


Evidence

Mark Zuckerberg discussed creating fake individuals on the Dworkish podcast; Meta’s plans to populate social platforms with AI-generated people


Major discussion point

Meta’s AI Strategy and LeCun’s Role


Topics

Sociocultural | Economic


Asks about the future structure of foundation model market and competition

Explanation

Thompson inquires about LeCun’s prediction for the foundation model landscape in 3-5 years, asking whether there will be a small number of large winners as typically happens in tech, or if there will be an extremely large number of open source models coexisting with some big closed source models as seen today.


Evidence

Current market has some very big closed source models and some open source models; tech industry tendency toward few large winners


Major discussion point

Future AI Landscape Predictions


Topics

Economic | Infrastructure


L

LJ Rich

Speech speed

191 words per minute

Speech length

181 words

Speech time

56 seconds

Introduces the session on steering AI’s future and presents LeCun as Meta’s chief AI scientist

Explanation

Rich introduces the session topic of how to steer the future of AI and presents Yann LeCun as Meta’s chief AI scientist. She also introduces Nicholas Thompson from The Atlantic as the moderator and notes LeCun’s reputation as one of the godfathers of AI for helping create the architecture behind current AI systems.


Evidence

LeCun helped create the architecture that led to all current AI; he’s known as one of the godfathers of AI; Nicholas Thompson is from The Atlantic


Major discussion point

Conference Context and Introductions


Topics

Economic | Infrastructure


Provides closing remarks thanking the speakers for their discussion

Explanation

Rich concludes the session by thanking both Yann LeCun and Nicholas Thompson for what she describes as an amazing and mind-bending conversation. She expresses appreciation for the quality and depth of their discussion about AI’s future.


Evidence

Describes the conversation as amazing and mind-bending


Major discussion point

Conference Context and Introductions


Topics

Economic | Infrastructure


Agreements

Agreement points

AI systems require careful design and safety considerations

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Future AI systems should be objective-driven with built-in guardrails for safety by construction


AI systems will be designed with objectives and guardrails, making them safe by construction similar to how jetliners are engineered for safety


Points out the difference in regulatory approaches between highly regulated aviation industry and current minimal AI regulation


Summary

Both speakers acknowledge that AI systems need careful safety design, though Thompson raises concerns about the current lack of regulation compared to other industries like aviation


Topics

Legal and regulatory | Cybersecurity


Current AI development involves global collaboration and knowledge sharing

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Open research accelerates progress by enabling more people to contribute different ideas and learn from each other


Raises concerns about models like Llama being used by DeepSeek and implications for national security


Summary

Both speakers recognize that AI development is inherently global and collaborative, with ideas and models being shared across borders, though they may differ on the implications


Topics

Economic | Infrastructure


Similar viewpoints

Both speakers engage seriously with the timeline and feasibility of achieving human-level AI, with Thompson acknowledging the significant investment and interest in this goal

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Human-level AI may be achievable within 5-10 years, starting with animal-level intelligence


Questions whether LeCun still believes LLMs are a dead end given widespread investment and adoption


Topics

Economic | Infrastructure


Both speakers recognize the importance of diversity in AI systems and the potential risks of concentration of power in a few companies

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Diversity of AI assistants is crucial to prevent a handful of companies from controlling all digital interactions and cultural biases


Explores how LeCun’s philosophy on diversity and open source influences Meta’s product decisions


Topics

Sociocultural | Human rights


Unexpected consensus

Need for international cooperation in AI development

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Future foundation models should be trained through international partnerships preserving data sovereignty


Raises concerns about models like Llama being used by DeepSeek and implications for national security


Explanation

Despite Thompson raising national security concerns, both speakers seem to accept that international collaboration in AI is inevitable and necessary, with LeCun proposing structured partnerships that preserve sovereignty


Topics

Legal and regulatory | Human rights


Recognition of China’s AI capabilities and innovation

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

China has excellent resources and may develop ideas the West doesn’t, as demonstrated by DeepSeek’s impact


Raises concerns about models like Llama being used by DeepSeek and implications for national security


Explanation

Both speakers acknowledge China’s significant AI capabilities and innovation potential, even when discussing competitive concerns, showing unexpected consensus on China’s technical competence


Topics

Economic | Infrastructure


Overall assessment

Summary

The speakers show substantial agreement on the technical challenges and future directions of AI development, the importance of safety considerations, the need for diversity in AI systems, and the reality of global collaboration in AI research. They also both recognize the significant capabilities of international competitors like China.


Consensus level

High level of consensus on technical and structural issues, with disagreements mainly on policy approaches rather than fundamental assessments. This suggests that while there may be debates about regulation and open source policies, there is broad agreement among experts about the technical trajectory and challenges of AI development.


Differences

Different viewpoints

Regulatory approach to AI safety

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

AI systems will be designed with objectives and guardrails, making them safe by construction similar to how jetliners are engineered for safety


Points out the difference in regulatory approaches between highly regulated aviation industry and current minimal AI regulation


Summary

LeCun argues that AI safety can be achieved through careful design similar to aviation, while Thompson challenges this analogy by highlighting that aviation is heavily regulated whereas AI development currently has minimal regulatory oversight despite building more powerful systems.


Topics

Legal and regulatory | Cybersecurity


Open source vs national security concerns

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Restricting open source would slow down Western progress more than it would hinder geopolitical rivals


Questions the contradiction between wanting Western AI leadership while supporting open access to models like Llama


Summary

LeCun believes restricting open source would harm Western progress more than help, while Thompson questions whether supporting open access undermines Western AI leadership, especially given national security concerns about models like Llama being used by competitors like DeepSeek.


Topics

Economic | Legal and regulatory


Evidence for open source safety benefits

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Open source platforms are historically safer, more portable, and flexible due to more people examining and improving them


Challenges whether open source truly leads to better safety outcomes and asks for empirical evidence


Summary

LeCun argues that open source is historically safer based on general software principles, while Thompson specifically asks for empirical evidence that this applies to AI systems after two years of implementation.


Topics

Legal and regulatory | Infrastructure


Unexpected differences

Fundamental disagreement on current AI regulation adequacy

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Current LLMs pose minimal risk as evidenced by two years of widespread use without major incidents


Points out the difference in regulatory approaches between highly regulated aviation industry and current minimal AI regulation


Explanation

This disagreement is unexpected because both speakers generally support AI development, yet they have fundamentally different views on whether current regulatory approaches are adequate. LeCun sees current evidence as proof that minimal regulation is sufficient, while Thompson argues that more powerful systems require more robust regulatory frameworks.


Topics

Legal and regulatory | Cybersecurity


Overall assessment

Summary

The main disagreements center around regulatory approaches to AI safety, the balance between open source benefits and national security concerns, and the adequacy of current oversight mechanisms. While both speakers support AI development, they differ significantly on risk assessment and appropriate safeguards.


Disagreement level

Moderate to significant disagreement with important implications for AI governance. The disagreements reflect broader tensions in the AI community between innovation acceleration and risk management, between global collaboration and national competitiveness, and between industry self-regulation and government oversight. These disagreements could influence policy decisions about AI regulation, international cooperation, and open source development practices.


Partial agreements

Partial agreements

Similar viewpoints

Both speakers engage seriously with the timeline and feasibility of achieving human-level AI, with Thompson acknowledging the significant investment and interest in this goal

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Human-level AI may be achievable within 5-10 years, starting with animal-level intelligence


Questions whether LeCun still believes LLMs are a dead end given widespread investment and adoption


Topics

Economic | Infrastructure


Both speakers recognize the importance of diversity in AI systems and the potential risks of concentration of power in a few companies

Speakers

– Yann LeCun
– Nicholas Thompson

Arguments

Diversity of AI assistants is crucial to prevent a handful of companies from controlling all digital interactions and cultural biases


Explores how LeCun’s philosophy on diversity and open source influences Meta’s product decisions


Topics

Sociocultural | Human rights


Takeaways

Key takeaways

Large Language Models (LLMs) are useful tools but insufficient for achieving human-level AI due to fundamental limitations in planning, reasoning, persistent memory, and world understanding


Alternative architectures like JEPA (Joint Embedding Predictive Architecture) are needed that can make predictions in abstract representation space rather than at pixel level


AI safety can be achieved through systems designed with built-in objectives and guardrails, making them safe by construction similar to aviation engineering


Open source AI development accelerates progress, increases safety through broader scrutiny, and prevents concentration of power in a few companies


Future AI landscape will likely have 2-3 dominant foundation models, with 1-2 being open source, similar to current operating system distribution


International cooperation and data sovereignty will be crucial for training future foundation models that represent global human knowledge


Human-level AI may be achievable within 5-10 years, progressing through animal-level intelligence first


Restricting open source AI development would slow Western progress more than it would hinder geopolitical rivals


Resolutions and action items

N


o


n


e


i


d


e


n


t


i


f


i


e


d


Unresolved issues

How to design and train objectives and guardrails for AI systems to prevent intrinsically dangerous behavior – acknowledged as complex and unsolved


Whether open source AI development truly leads to better safety outcomes – empirical evidence still being gathered


The apparent contradiction between wanting Western AI leadership while supporting open access to powerful models


How to balance national security concerns with the benefits of open source AI development


The technical details of implementing international partnerships for distributed AI training while preserving data sovereignty


How Meta’s AI research philosophy translates into actual product decisions and implementations


Suggested compromises

International partnerships for training foundation models where countries provide computation and data but preserve data sovereignty through parameter vector exchange rather than raw data sharing


A future AI ecosystem with both open source dominant platforms and proprietary niche models, similar to current operating system distribution


Careful but not restrictive approach to open sourcing AI models, balancing safety considerations with the benefits of open development


Thought provoking comments

I didn’t say they [LLMs] were a dead end. I said they were a dead end if you are interested in reaching human-level AI or artificial superintelligence… They’re certainly a useful tool… But if we want to reach the type of intelligence that we observe in humans and animals, we are going to have to invent other methods.

Speaker

Yann LeCun


Reason

This comment is deeply insightful because it reframes the entire AI debate by distinguishing between utility and sufficiency. LeCun clarifies that his criticism isn’t about LLMs being useless, but about their fundamental limitations in achieving human-level intelligence. This nuanced position challenges both AI optimists who see LLMs as the path to AGI and pessimists who might dismiss them entirely.


Impact

This comment set the foundational tone for the entire discussion, moving it away from sensationalized ‘dead end’ interpretations toward a more sophisticated analysis of AI architectures. It allowed Thompson to probe deeper into LeCun’s specific technical objections and alternative proposals, structuring the conversation around the core question of what constitutes a path to human-level AI.


The way an LLM produces an answer is by running through a fixed number of layers in the neural net and then producing a token… And the amount of computation that goes into producing this word is constant… when we have a hard question to answer, we tend to think about it for a long time, right?

Speaker

Yann LeCun


Reason

This is a profound technical insight that makes abstract AI limitations concrete and relatable. By connecting the fixed computational architecture of LLMs to human cognitive processes, LeCun reveals a fundamental mismatch between current AI and human intelligence. The observation that humans allocate variable thinking time to problems of different complexity exposes a core architectural limitation.


Impact

This comment shifted the discussion from general criticisms to specific technical limitations, providing a concrete foundation for understanding why LLMs might be insufficient for human-level AI. It led directly into LeCun’s explanation of ‘system two thinking’ and the need for AI systems that can ‘search’ for answers rather than just compute them, deepening the technical sophistication of the conversation.


Intelligence needs to be grounded in some reality… a four year old child has gotten as much information through vision as the biggest LLMs today that are trained on all the publicly available texts on the internet.

Speaker

Yann LeCun


Reason

This comparison is strikingly thought-provoking because it challenges our assumptions about the scale and nature of learning. The idea that a four-year-old’s visual experience equals the entire internet’s text content reframes our understanding of information density and learning efficiency. It suggests that text-based training, no matter how massive, may be fundamentally insufficient.


Impact

This insight pivoted the conversation toward embodied cognition and multimodal learning, setting up LeCun’s introduction of the JEPA architecture. It provided compelling justification for why alternative approaches are needed and helped the audience understand why purely text-based systems might hit fundamental limits.


The biggest danger is that there is a future in which every single one of our interactions with the digital world will be mediated by AI assistance… We need to have a very wide diversity of AI assistance that gives people choice about what type of bias, linguistic capabilities, cultural knowledge and cultural value systems… for the same reason we need social diversity in the press.

Speaker

Yann LeCun


Reason

This comment is exceptionally insightful because it reframes AI safety from technical risks to democratic and cultural risks. LeCun identifies a more immediate and arguably more important danger than AI takeover: the concentration of AI systems in few hands could create unprecedented control over human information consumption and cultural expression. The analogy to press diversity makes this abstract concern tangible.


Impact

This comment fundamentally shifted the safety discussion away from science fiction scenarios toward concrete democratic concerns. It provided the intellectual foundation for LeCun’s open source advocacy and led to a more nuanced exploration of geopolitical implications. The conversation moved from technical safety to societal governance of AI systems.


Lama the first version of Lama was built by a team of 13 people, 12 of whom were in Paris… it’s not an American model, it’s a French model… ideas can pop up from anywhere and we should have a system by which the best ideas circulate.

Speaker

Yann LeCun


Reason

This revelation is thought-provoking because it challenges nationalist narratives about AI development while providing concrete evidence for international collaboration. By revealing the French origins of Llama, LeCun demonstrates that AI innovation is inherently global, making arguments for technological isolation seem both impractical and counterproductive.


Impact

This comment effectively countered Thompson’s probing about national security concerns and technological competition. It shifted the discussion from zero-sum geopolitical thinking toward a more collaborative vision of AI development, supporting LeCun’s arguments about open source benefits and international cooperation.


The future I’m envisioning is one in which LLMs or their descendants will basically constitute a repository of all human knowledge and culture… each country or region will provide computation and its own data but will preserve its own data… we can arrive at a consensus model without every region giving up their data.

Speaker

Yann LeCun


Reason

This vision is remarkably thought-provoking because it proposes a technical solution to the fundamental tension between global AI development and data sovereignty. The idea of distributed training with shared parameters but protected data offers a potential path forward for international AI cooperation while respecting national concerns about data control.


Impact

This comment provided a compelling conclusion to the discussion by offering a concrete vision for how AI development could proceed globally while addressing sovereignty concerns. It synthesized many of the conversation’s themes – open source, international cooperation, and democratic access to AI – into a practical framework for future development.


Overall assessment

These key comments shaped the discussion by systematically building a comprehensive alternative vision for AI development. LeCun’s insights moved the conversation through several levels: from technical architecture (why LLMs are limited) to safety philosophy (democratic rather than existential risks) to geopolitical strategy (open source and international cooperation). Each comment built upon previous ones, creating a coherent argument that challenged mainstream AI discourse on multiple fronts. The discussion evolved from defensive clarifications about LLM limitations to a proactive vision for global, democratic AI development. Thompson’s skilled questioning helped draw out these insights, but LeCun’s comments consistently elevated the conversation beyond typical AI hype or fear-mongering toward substantive technical and philosophical analysis. The overall effect was to present a mature, nuanced perspective that acknowledged both AI’s potential and its current limitations while proposing concrete paths forward.


Follow-up questions

How do we design or train objectives and guardrails so that AI systems do not do things that are intrinsically dangerous?

Speaker

Yann LeCun


Explanation

LeCun acknowledged this as ‘not a solved problem’ and emphasized its complexity while discussing AI safety architecture, indicating it requires further research and development.


How do we elaborate thoughts in AI systems?

Speaker

Yann LeCun


Explanation

This was mentioned as one of four essential capabilities for intelligence that current LLMs lack, requiring new methods and architectures to achieve human-level AI.


How do we enable AI systems to understand the physical world?

Speaker

Yann LeCun


Explanation

LeCun identified this as a fundamental limitation of current LLMs and a necessary component for reaching human-level intelligence, requiring different architectural approaches.


How do we implement persistent memory in AI systems?

Speaker

Yann LeCun


Explanation

This was listed as one of the four essential capabilities missing from current LLMs that would be necessary for human-level intelligence.


How do we enable AI systems to reason effectively?

Speaker

Yann LeCun


Explanation

Reasoning was identified as a critical capability that LLMs currently cannot perform to the desired extent for achieving human-level intelligence.


How do we enable AI systems to plan effectively?

Speaker

Yann LeCun


Explanation

Planning was mentioned as one of the four essential missing capabilities in current LLMs that would need to be developed for human-level AI.


What obstacles on the path to human-level AI are we not seeing yet?

Speaker

Yann LeCun


Explanation

LeCun mentioned that reaching human-level AI is ‘almost always harder than we think’ and there are likely unforeseen obstacles that need to be identified and addressed.


How can we implement international partnerships for distributed training of foundation models while preserving data sovereignty?

Speaker

Yann LeCun


Explanation

LeCun proposed a complex system where countries could contribute data and computation without sharing raw data, but the technical and political implementation details require further development.


What is the empirical evidence for open source leading to safer AI outcomes?

Speaker

Nicholas Thompson


Explanation

Thompson specifically asked whether there was empirical evidence supporting LeCun’s claims about open source safety benefits, indicating a need for more concrete data and research on this topic.


How will the JEPA architecture specifically lead to more profound intelligence than LLMs?

Speaker

Nicholas Thompson


Explanation

While LeCun explained the concept, the practical implementation and validation of JEPA’s superiority over current architectures requires further research and development.


Disclaimer: This is not an official session record. DiploAI generates these resources from audiovisual recordings, and they are presented as-is, including potential errors. Due to logistical challenges, such as discrepancies in audio/video or transcripts, names may be misspelled. We strive for accuracy to the best of our ability.