Gemini AI can now summarise videos in Google Drive

Google is expanding Gemini AI’s capabilities in Drive by enabling it to analyse video files and respond to user questions or generate concise summaries.

The new feature aims to save users time by providing quick insights from lengthy content such as meetings, classes or announcements, instead of requiring them to watch the entire video. Until now, Gemini could only summarise documents and PDFs stored in Drive.

According to a blog post published on 28 May 2025, the feature will support prompts like ‘Summarise the video’ or ‘List action items from the meeting.’ Users can access Gemini’s functionality either through Drive’s overlay previewer or a standalone viewer in a separate browser tab.

However, captions must be enabled within the user’s domain for the feature to work properly.

The update is being gradually rolled out and is expected to be available to all eligible users by 19 June. At the moment, it is limited to English and accessible only to users of Google Workspace and Google One AI Premium, or those with Gemini Business or Enterprise add-ons.

For administrators, smart features and personalisation settings must be activated to grant access.

To use the new function, users can double-click on a video file in Drive and select the ‘Ask Gemini’ option marked by a star icon in the top right corner. Google says the upgrade reflects a broader effort to integrate AI seamlessly into everyday workflows by making content easier to navigate and understand.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Meta faces backlash over open source AI claims

Meta is under renewed scrutiny for what critics describe as ‘open washing’ after sponsoring a Linux Foundation whitepaper on the benefits of open source AI.

The paper highlights how open models help reduce enterprise costs—claiming companies using proprietary AI tools spend over three times more. However, Meta’s involvement has raised questions, as its Llama AI models are presented as open source despite industry experts insisting otherwise.

Amanda Brock, head of OpenUK, argues that Llama does not meet accepted definitions of open source due to licensing terms that restrict commercial use.

She referenced the Open Source Initiative’s (OSI) standards, which Llama fails to meet, pointing to the presence of commercial limitations that contradict open source principles. Brock noted that open source should allow unrestricted use, which Llama’s license does not support.

Meta has long branded its Llama models as open source, but the OSI and other stakeholders have repeatedly pushed back, stating that the company’s licensing undermines the very foundation of open access.

While Brock acknowledged Meta’s contribution to the broader open source conversation, she also warned that such mislabelling could have serious consequences—especially as lawmakers and regulators increasingly reference open source in crafting AI legislation.

Other firms have faced similar allegations, including Databricks with its DBRX model in 2024, which was also criticised for failing to meet OSI standards. As the AI sector continues to evolve, the line between truly open and merely accessible models remains a point of growing tension.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Anthropic CEO warns of mass job losses from AI

Just one week after releasing its most advanced AI models to date — Opus 4 and Sonnet 4 — Anthropic CEO Dario Amodei warned in an interview with Axios that AI could soon reshape the job market in alarming ways.

AI, he said, may be responsible for eliminating up to half of all entry-level white-collar roles within the next one to five years, potentially driving unemployment as high as 10% to 20%.

Amodei’s goal in speaking publicly is to help workers prepare and to urge both AI companies and governments to be more transparent about coming changes. ‘Most of them [workers] are unaware that this is about to happen,’ he told Axios. ‘It sounds crazy, and people just don’t believe it.’

According to Amodei, the shift from AI augmenting jobs to fully automating them could begin as soon as two years from now. He highlighted how widespread displacement may threaten democratic stability and deepen inequality, as large groups of people lose the ability to generate economic value.

Despite these warnings, Amodei explained that competitive pressures prevent developers from slowing down. Regulatory caution in the US, he suggested, would only result in countries like China advancing more rapidly.

Still, not all implications are negative. Amodei pointed to major breakthroughs in other areas, such as healthcare, as part of the broader impact of AI.

‘Cancer is cured, the economy grows at 10% a year, the budget is balanced — and 20% of people don’t have jobs,’ he said.

To prepare society, Amodei called for increased public awareness, encouraging individuals to reconsider career paths and avoid the most automation-prone fields.

He referenced the Anthropic Economic Index, which monitors how AI affects different occupations. At its launch in February, the index showed that 57% of AI use cases still supported human tasks rather than replacing them.

However, during a press-only session at Code with Claude, Amodei noted that augmentation is likely to be a short-term strategy. He described a ‘rising waterline’ — the gradual shift from assistance to full replacement — which may soon outpace efforts to retain human roles.

‘When I think about how to make things more augmentative, that is a strategy for the short and the medium term — in the long term, we are all going to have to contend with the idea that everything humans do is eventually going to be done by AI systems. That is a constant. That will happen,’ he said.

His other recommendations included boosting AI literacy and equipping public officials with a deeper understanding of superintelligent systems, so they can begin forming policy for a radically transformed economy.

While Amodei’s outlook may sound daunting, it echoes a pattern seen throughout history: every major technological disruption brings workforce upheaval. Though some roles vanish, others emerge. Several studies suggest AI may even highlight the continued relevance of distinctively human skills.

Regardless of the outcome, one thing remains clear — learning to work with AI has never been more important.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

New York Times partners with Amazon on AI integration

The New York Times Company and Amazon have signed a multi-year licensing agreement that will allow Amazon to integrate editorial content from The New York Times, NYT Cooking, and The Athletic into a range of its AI-powered services, the companies announced Wednesday.

Under the deal, Amazon will use licensed content for real-time display in consumer-facing products such as Alexa, as well as for training its proprietary foundation models. The agreement marks an expansion of the firms’ existing partnership.

‘The agreement expands the companies’ existing relationship, and will deliver additional value to Amazon customers while bringing Times journalism to broader audiences,’ the companies said in a joint statement.

According to the announcement, the licensing terms include ‘real-time display of summaries and short excerpts of Times content within Amazon products and services’ alongside permission to use the content in AI model development. Amazon platforms will also feature direct links to full Times articles.

Both companies described the partnership as a reflection of a shared commitment to delivering global news and information across Amazon’s AI ecosystem. Financial details of the agreement were not made public.

The announcement comes amid growing industry debate about the role of journalistic material in training AI systems.

By entering a formal licensing arrangement, The New York Times positions itself as one of the first major media outlets to publicly align with a technology company for AI-related content use.

The companies have yet to name additional Amazon products that will feature Times content, and no timeline has been disclosed for the rollout of the new integrations.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Croatia urged to embed human rights into AI law

Politiscope recently held an event at the Croatian Journalists’ Association to highlight the human rights risks of AI.

As Croatia begins drafting a national law to implement the EU AI Act, the event aimed to push for stronger protections and transparency instead of relying on vague promises of innovation.

Croatia’s working group is still forming key elements of the law, such as who will enforce it, making it an important moment for public input.

Experts warned that AI systems could increase surveillance, discrimination, and exclusion. Speakers presented troubling examples, including inaccurate biometric tools and algorithms that deny benefits or profile individuals unfairly.

Campaigners from across Europe, including EDRi, showcased how civil society has already stopped invasive AI tools in places like the Netherlands and Serbia. They argued that ‘values’ embedded in corporate AI systems often lack accountability and harm marginalised groups instead of protecting them.

Rather than presenting AI as a distant threat or a miracle cure, the event focused on current harms and the urgent need for safeguards. Speakers called for a public register of AI use in state institutions, a ban on biometric surveillance in public, and full civil society participation in shaping AI rules.

A panel urged Croatia to go beyond the EU Act’s baseline by embracing more transparent and citizen-led approaches.

Despite having submitted recommendations, Politiscope and other civil society organisations remain excluded from the working group drafting the law. While business groups and unions often gain access through social dialogue rules, CSOs are still sidelined.

Politiscope continues to demand an open and inclusive legislative process, arguing that democratic oversight is essential for AI to serve people instead of controlling them.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Gen Z turns to AI to plan their next holiday

AI is quickly becoming an essential tool for Gen Z travellers, with 45% either already using it or planning to use it this year to help organise their holidays.

According to a survey by Multitrip.com, AI is now a key influence for trip inspiration and destination selection, especially for younger generations looking to personalise their travel experiences.

Among Gen Z respondents, a quarter have already used AI to choose where to go—an increase from 20% last year—while interest among Millennials and Gen X is also growing.

Millennials’ use rose to 22%, and Gen X saw a notable jump from 4% to 10%. Baby Boomers, however, remain largely unconvinced, with only 3% reporting past use and 84% saying they have no plans to try it.

Social media continues to play a major role in travel decisions, especially for younger people, while TV and film remain popular sources of inspiration across all generations.

AI is not only helping travellers choose where to go, but also shaping itineraries, with nearly one in four Gen Z users already relying on it to structure their trips.

As AI technology becomes more user-friendly and accessible, its role in trip planning is expected to grow. However, experts still recommend traditional travel essentials, including insurance.

Multitrip.com’s Christian Bennett reminds travellers that no matter how a trip is planned, protection from unexpected events remains crucial.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Startup Odyssey unveils streaming AI-powered 3D video demo

Odyssey, a startup founded by self-driving technology pioneers Oliver Cameron and Jeff Hawke, has unveiled a new AI model capable of generating interactive, 3D-like video experiences in real time.

The system streams video frames every 40 milliseconds and lets users explore scenes with basic controls, much like moving around in a video game.

The model is powered by a newly developed ‘world model’ designed to predict what comes next in a virtual environment based on prior actions and context.

While the current demo, available online, still features blurred and unstable visuals, Odyssey claims it can maintain coherent video streams for over five minutes. Streaming performance reaches up to 30 frames per second using Nvidia H100 GPUs, costing $1–2 per user-hour.

Odyssey is positioning its platform as the future of interactive media, suggesting it could transform entertainment, education, advertising, and training.

Unlike some AI developers facing criticism for displacing creative workers, Odyssey pledges to work alongside artists and provide tools compatible with software like Unreal Engine, Blender, and After Effects.

To support development, the company built a 360-degree backpack-mounted camera to collect real-world footage, aiming to produce more realistic outputs than models trained solely on public datasets.

Backed by $27 million in funding and supported by Pixar co-founder Ed Catmull, Odyssey plans to significantly improve model stability and expand its action capabilities in the near future.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

ASEAN, the GCC, and China unite to advance digital innovation

The Association of Southeast Asian Nations (ASEAN), the Gulf Cooperation Council (GCC), and China convened their inaugural trilateral summit in Kuala Lumpur with a shared vision to strengthen regional cooperation and integration.

The three parties committed to deepening economic ties through sustainable development, free trade, and enhanced collaboration across key sectors such as the digital and green economy, technology, and finance.

A major focus of the summit was boosting connectivity, notably through infrastructure development and expanded cooperation under China’s Belt and Road Initiative.

Key policy areas include the promotion of AI, the advancement of financial technology (fintech), and the development of digital skills to prepare their populations for the evolving digital economy.

They also expressed a shared interest in supporting clean energy innovation, highlighting the role of technology in driving sustainable and inclusive energy transitions.

These technology-focused policies aim to build resilient, future-ready economies by enhancing cross-border collaboration in emerging digital industries and strengthening regional capabilities in science, technology, and innovation.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

DeepSeek updates its R1 AI model

Chinese AI startup DeepSeek has rolled out an updated version of its R1 reasoning model on the popular developer platform Hugging Face. The release was quietly announced via a WeChat post and marks a continuation of DeepSeek’s effort to challenge leading players in the AI field, including US-based OpenAI.

The new R1 model is a minor update, released under the permissive MIT license, allowing commercial use. However, the Hugging Face repository only includes the model’s configuration files and internal weights, with no public documentation describing the model’s capabilities.

https://twitter.com/deepseek_ai/status/1881318130334814301

These ‘weights’ — the parameters that determine how an AI model behaves — total a massive 685 billion, making the model too large to run on standard consumer hardware without special adaptations. DeepSeek gained attention earlier this year when the original R1 demonstrated competitive performance against leading models from OpenAI.

That advancement hasn’t gone unnoticed by international observers; US regulators have expressed concern over the potential national security risks posed by DeepSeek’s technologies. Despite the controversy, the company continues to make bold moves on the global AI stage.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

Google Drive adds AI video summaries

Google Drive is gaining a new AI-powered tool that allows Workspace users to summarise and interact with video content using Gemini, Google’s generative AI assistant.

Instead of manually skipping through videos, users can now click the ‘Ask Gemini’ button to get instant summaries, key highlights, or action items from uploaded recordings.

The tool builds on Gemini 2.5 Pro’s strong video analysis capabilities, which recently scored 84.8% on the VideoMME benchmark. Gemini’s side panel, already used for summarising documents and folders, can now handle natural language prompts like ‘Summarise this video’ or ‘List key points from this meeting’.

However, the feature only works in English and requires captions to be enabled by the Workspace admin.

Google is rolling out the feature across various Workspace plans, including Business Standard and Enterprise tiers, with access available through Drive’s overlay preview or a new browser tab.

Instead of switching between windows or scrubbing through videos, users can now save time by letting Gemini handle the heavy lifting.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!