Japan to boost spending on semiconductors and AI

Japan’s Ministry of Economy, Trade and Industry is set to significantly increase funding for advanced semiconductors and AI in the coming fiscal year.

Spending on chips and AI is expected to nearly quadruple to ¥1.23 trillion ($7.9 billion), accounting for the majority of the ministry’s ¥3.07 trillion budget, a 50% increase from last year. The budget, approved by Prime Minister Sanae Takaichi’s Cabinet, will be debated in parliament early next year.

The funding boost reflects Japan’s push to strengthen its position in frontier technologies amid global competition with the US and China. The government will fund most of the additional support through regular budgets, ensuring more stable backing for semiconductor and AI development.

Key initiatives include ¥150 billion for chip venture Rapidus and ¥387.3 billion for domestic foundation AI models, data infrastructure, and ‘physical AI’ for robotics and machinery control.

The budget also allocates ¥5 billion for critical minerals and ¥122 billion for decarbonisation, including next-generation nuclear power. Special bonds worth ¥1.78 trillion will also support Japanese investment in the US, reinforcing the trade agreement between the two countries.

The increase in funding demonstrates Japan’s strategic focus on achieving technological self-sufficiency and enhancing global competitiveness in emerging industries, thereby ensuring long-term support for innovation and critical infrastructure.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Groq partners with Nvidia to expand inference technology

Groq has signed a non-exclusive licensing agreement with Nvidia to share its inference technology, aiming to make high-performance, cost-efficient AI processing more widely accessible.

Groq’s founder, Jonathan Ross, president Sunny Madra, and other team members will join Nvidia to help develop and scale the licensed technology. Despite the collaboration, Groq will remain an independent company, with Simon Edwards taking over as Chief Executive Officer.

Operations of GroqCloud will continue without interruption, ensuring ongoing services for existing customers. The agreement highlights a growing trend of partnerships in the AI sector, combining innovation with broader access to advanced processing capabilities.

The partnership could speed up AI inference adoption, offering companies more scalable and cost-effective options for deploying AI workloads. Analysts suggest such collaborations are likely to drive competition and innovation in the rapidly evolving AI hardware and software market.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

SK hynix urges government to ease fair trade regulations

SK Hynix has urged the South Korean government to relax fair trade rules to allow the creation of a special-purpose company for raising funds for significant investments. The move comes as the semiconductor firm faces high capital demands amid the global AI boom.

Currently, SK hynix, a second-tier subsidiary of SK Group through SK Square, must retain full ownership when establishing third-tier subsidiaries. The government pledged to cut the ownership requirement to 50 percent, giving chipmakers more flexibility in funding projects.

The company highlighted the rising costs of advanced facilities, noting that a cleanroom at the Yongin semiconductor cluster in 2019 required 7.5 trillion won ($5.14 billion), while the new M15X fabrication plant in 2025 cost around 20 trillion won.

The size and long-term nature of modern semiconductor investments increasingly strain existing methods for raising funds.

SK hynix said letting subsidiaries partner with external investors would ease financial pressure and improve corporate health. The company added that regulatory flexibility is crucial for sustaining investment and competitiveness in a sector marked by high volatility.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

‘All is fair in RAM and war’: RAM price crisis in 2025 explained

If you are piecing together a new workstation or gaming rig, or just hunting for extra RAM or SSD storage, you have stumbled into the worst possible moment. With GPU prices already sky-high, the recent surge in RAM and storage costs has hit consumers hard, leaving wallets lighter and sparking fresh worries about where the tech market is headed.

On the surface, the culprit behind these soaring prices is a sudden RAM shortage. Prices for 32GB and 64GB sticks have skyrocketed by as much as 600 percent, shelves are emptying fast, and the balance between supply and demand has completely unraveled.

But blaming the sky-high prices on empty shelves only tells part of the story. Why has affordable RAM vanished? How long will this chaos last? And most intriguingly, what role does AI play in this pricing storm?

Tracing the causes of RAM pricing spikes

The US tariffs imposed on China on 1 August 2025, played a substantial role in the increase in DRAM prices. Global imports of various goods have become more costly, investments and workforce onboarding have been halted, and many businesses relying on imports have adopted a ‘wait-and-see’ approach to how they will do business going forward.

However, the worst was yet to come. On 3 December, Micron, one of the world’s leading manufacturers of data storage and computer memory components, announced its withdrawal from the RAM consumer market, citing a ‘surge in demand for memory and storage’ driven by supply shortages of memory and storage for AI data centres.

With Micron out of the picture, we are left with only two global consumer RAM and high-bandwidth memory (HBM) manufacturers: Samsung and SK Hynix. While there are countless RAM brands on the market, with Corsair, Kingston, and Crucial leading the charge, all of them rely on the three aforementioned suppliers for memory chips.

Micron’s exit was likely met with obscured glee by Samsung and SK Hynix of South Korea, who seized the opportunity to take over Crucial’s surrendered territory and set the stage for their DRAM/HBM supply duel. The latter supplier was quick to announce the completion of its M15X semiconductor fabrication plant (fab), but warned that RAM supply constraints are likely to last until 2028 at the earliest.

Amid the ruckus, rumours surfaced that Samsung would be sunsetting its SATA SSD production, which the company quickly extinguished. On the contrary, the Korean giant announced its intention to dethrone SK Hynix as the top global RAM provider, with more than 80 percent of its projected profits coming directly from Samsung Electronics.

Despite their established market shares, both enterprises were caught off guard when their main rival threw in the towel, and their production facilities are unable, at current capacity, to accommodate the resulting market void. It is nigh certain that the manufacturers will use their newly gained market dominance to their advantage, setting prices based on their profit margins and customers’ growing demand for their products. In a nutshell, they have the baton, and we must play to their tune.

AI infrastructure and the reallocation of RAM supply

Micron, deeming commodity RAM a manufacturing inconvenience, made a move that was anything but rash. In October, Samsung and SK Hynix joined forces with OpenAI to supply the AI giant with a monthly batch of 900,000 DRAM wafers. OpenAI’s push to enhance its AI infrastructure and development was presumably seen by Micron as a gauntlet thrown by its competitors, and Crucial’s parent company took no time in allocating its forces to a newly opened front.

Lured by lucrative, long-term, high-volume contracts, all three memory suppliers saw AI as an opportunity to open new income streams that would not dry up for years to come. While fears of the AI bubble bursting are omnipresent and tangible, neither Samsung, SK Hynix, nor Micron are overly concerned about what the future holds for LLMs and AGI, as long as they continue to get their RAM money’s worth (literally).

AI has expanded across multiple industries, and three competitors judged Q4 2025 the opportune time to put all their RAM eggs in one basket. AI as a business model has yet to reach profitability, but corporate investors poured more than USD 250 billion into AI in 2024 alone. Predictions for 2025 have surpassed the USD 500 billion mark, but financiers will inevitably grow more selective as the AI startup herd thins and predicted cash cows fail to deliver future profits.

To justify massive funding rounds, OpenAI, Microsoft, Google, and other major AI players need to keep their LLMs in a perpetual growth cycle by constantly expanding their memory capacity. A hyperscale AI data centre can contain tens of thousands to hundreds of thousands of GPUs, each with up to 180 gigabytes of VRAM. Multiply that by 1,134, the current number of hyperscale data centres, and it is easy to see why Micron was eager to ditch the standard consumer market for more bankable opportunities.

The high demand for RAM has changed the ways manufacturers view risk and opportunity. AI infrastructure brings more volume, predictability, and stable contracts than consumer markets, especially during uncertain times and price swings. Even if some areas of AI do not meet long-term hopes, the need for memory in the near and medium term is built into data centre growth plans. For memory makers, shifting capacity to AI is a practical response to current market incentives, not just a risky bet on a single trend.

The aftermath of the RAM scarcity

The sudden price inflation and undersupply of RAM have affected more than just consumers building high-end gaming PCs and upgrading laptops. Memory components are critical to all types of devices, thereby affecting the prices of smartphones, tablets, TVs, game consoles, and many other IoT devices. To mitigate production costs and maintain profit margins, device manufacturers are tempted to offer their products with less RAM, resulting in substandard performance at the same price.

Businesses that rely on servers, cloud services, or data processing are also expected to get caught in the RAM crossfire. Higher IT costs are predicted to slow down software upgrades, digital services, and cybersecurity improvements. Every SaaS company, small or large, risks having its platforms overloaded or its customers’ data compromised.

Public institutions, such as schools, hospitals, and government agencies, will also have to bend backwards to cover higher hardware costs due to more expensive RAM. Operating on fixed budgets allows only so much wiggle room to purchase the required software and hardware, likely leading to delays in public digital projects and the continued use of outdated electronic equipment.

Man putting up missing posters with a picture of RAM memory sticks on them.

Rising memory costs also influence innovation and competition. When basic components become more expensive, it is harder for new companies to enter the market or scale up their services. This can favour large, well-funded firms and reduce diversity in the tech ecosystem. Finally, higher RAM prices can indirectly affect digital access and inclusion. More expensive devices and services make it harder for individuals and communities to afford modern technology, widening existing digital divides.

In short, when RAM becomes scarce or expensive, the effects extend far beyond memory pricing, influencing how digital services are accessed, deployed, and maintained across the economy. While continued investment in more capable AI models is a legitimate technological goal, it also raises a practical tension.

Advanced systems deliver limited value if the devices and infrastructure most people rely on lack the memory capacity required to run them efficiently. The challenge of delivering advanced AI models and AI-powered apps to subpar devices is one that AI developers will have to take into account moving forward. After all, what good is a state-of-the-art LLM if a run-of-the-mill PC or smartphone lacks the RAM to handle it?

The road ahead for RAM supply and pricing

As mentioned earlier, some memory component manufacturers predict that the RAM shortage will remain a burr under consumers’ saddles for at least a few years. Pompous predictions of the AI bubble’s imminent bursting have mostly ended up in the ‘I’ll believe it when I see it’ archive section, across the hall from the ‘NFTs are the future of digital ownership’ district.

Should investments continue to fill the budgets of OpenAI, Perplexity, Anthropic, and the rest, they will have the resources to reinforce their R&D departments, acquire the necessary memory components, and further develop their digital infrastructure. In the long run, the technology powering AI models may become more sophisticated to the point where energy demands reach a plateau. In that case, opportunities for expansion would be limitless.

Even though one of the biggest RAM manufacturers has fully shifted to making AI infrastructure components, there is still a gap large enough to be filled by small- and medium-sized producers. Companies such as Nanya Technology from Taiwan or US-based Virtium hold a tenth of the overall market share, but they have been given the opportunity to carry Micron’s torch and maintain competitiveness in their own capacities.

The current RAM price crisis is not caused by a single event, but by the way new technologies are changing the foundations of the digital economy. As AI infrastructure takes up more of the global memory supply, higher prices and limited availability are likely to continue across consumer, business, and public-sector markets. How governments, manufacturers, and buyers respond will shape not only the cost of hardware but also how accessible and resilient digital systems remain.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

AI growth changes the cycle for memory chip manufacturers

The growing demand for AI is reshaping the fortunes of the memory chip industry, according to leading manufacturers, who argue that the scale of AI investment is altering the sector’s typical boom-and-bust pattern.

The technology is creating more structural demand, rather than the sharp cyclical spikes that previously defined the market.

AI workloads depend heavily on robust memory systems, particularly as companies expand data centre capacity worldwide. Major chipmakers now expect steadier growth because AI models require vast data handling rather than one-off hardware surges.

Analysts suggest it could reduce the volatility that has often led to painful downturns for the industry.

Additionally, some reports claim that Japanese technology group Rakuten is prioritising low-cost AI development to improve profitability across its businesses.

Its AI leadership stresses the need to deploy systems that maximise margins instead of simply chasing capability for its own sake.

The developments underscore how AI is not only transforming software and services but also reshaping the economics of the hardware required to power them, from memory chips to cloud infrastructure on a global scale.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

ByteDance prepares major AI investment for 2026

ByteDance plans a major jump in AI spending next year as global chip access remains uncertain. The firm is preparing heavier investment in processors and infrastructure to support demanding models across its apps and cloud platforms.

The company is budgeting nearly nine billion pounds for AI chips despite strict US export rules. A potential trial purchase of Nvidia H200 hardware could expand its computing capacity if wider access is approved for Chinese firms.

Rivals in the US continue to outspend ByteDance, with large tech groups pouring hundreds of billions into data centres. Chinese platforms face tighter limits and are developing models that run efficiently with fewer resources.

ByteDance’s consumer AI ecosystem keeps accelerating, led by its Doubao chatbot and growing cloud business. Private ownership gives the firm flexibility to invest aggressively while placing AI at the heart of its long-term strategy.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Nvidia seeks China market access as US eases AI chip restrictions

The US tech giant NVIDIA has largely remained shut out of China’s market for advanced AI chips, as US export controls have restricted sales due to national security concerns.

High-performance processors such as the H100 and H200 were barred, forcing NVIDIA to develop downgraded alternatives tailored for Chinese customers instead of flagship products.

A shift in policy emerged after President Donald Trump announced that H200 chip sales to China could proceed following a licensing review and a proposed 25% fee. The decision reopened a limited pathway for exporting advanced US AI hardware, subject to regulatory approval in both Washington and Beijing.

If authorised, the H200 shipments would represent the most powerful US-made AI chips permitted in China since restrictions were introduced. The move could help NVIDIA monetise existing H200 inventory while easing pressure on its China business as it transitions towards newer Blackwell chips.

Strategically, the decision may slow China’s push for AI chip self-sufficiency, as domestic alternatives still lag behind NVIDIA’s technology.

At the same time, the policy highlights a transactional approach to export controls, raising uncertainty over long-term US efforts to contain China’s technological rise.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Civil servants and AI will work together in 2050

Public administrations worldwide are facing unprecedented change as AI reshapes automation, procurement, and decision-making. Governments must stay flexible, open, and resilient, preparing for multiple futures with foresight, continuous learning, and adaptability.

During World Futures Day, experts from the SPARK-AI Alliance and representatives from governments, academia, and the private sector explored four potential scenarios for public service in 2050.

Scenarios ranged from human-centred administrations that reinforce trust, to algorithmic bureaucracies focused on oversight, agentic administrations with semi-autonomous AI actors, and data-eroded futures that require renewed governance of poor-quality data.

Key insights highlighted the growing importance of anticipatory capacity, positioning AI as a ‘co-worker’ rather than a replacement, and emphasising the need to safeguard public trust.

Civil servants will increasingly focus on ethical reasoning, interpretation of automated processes, and cross-disciplinary collaboration, supported by robust accountability and transparent data governance.

The SPARK-AI Alliance has launched a Working Group on the Future of Work in the Public Sector to help governments anticipate and prepare for change. Its focus will be on building resilient public administrations, evolving civil-service roles, and maintaining trust in AI-enabled governance.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

Amazon considers 10 billion investment in OpenAI

Amazon is reportedly considering a $10 billion investment in OpenAI, highlighting its growing focus on the generative AI market. The investment follows OpenAI’s October restructuring, giving it more flexibility to raise funds and form new tech partnerships.

OpenAI has recently secured major infrastructure agreements, including a $38 billion cloud computing deal with Amazon Web Services (AWS). Deals with Nvidia, AMD, and Broadcom boost OpenAI’s access to computing power for its AI development.

Amazon has invested $8 billion in Anthropic and continues developing AI hardware through AWS’s Inferentia and Trainium chips. The move into OpenAI reflects Amazon’s strategy to expand its influence across the AI sector.

OpenAI’s prior $13 billion Microsoft exclusivity has ended, enabling it to pursue new partnerships. The combination of fresh funding, cloud capacity, and hardware support positions OpenAI for continued growth in the AI industry.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot

The limits of raw computing power in AI

As the global race for AI accelerates, a growing number of experts are questioning whether simply adding more computing power still delivers meaningful results. In a recent blog post, digital policy expert Jovan Kurbalija argues that AI development is approaching a critical plateau, where massive investments in hardware produce only marginal gains in performance.

Despite the dominance of advanced GPUs and ever-larger data centres, improvements in accuracy and reasoning among leading models are slowing, exposing what he describes as an emerging ‘AI Pareto paradox’.

According to Kurbalija, the imbalance is striking: around 80% of AI investment is currently spent on computing infrastructure, yet it accounts for only a fraction of real-world impact. As hardware becomes cheaper and more widely available, he suggests it is no longer the decisive factor.

Instead, the next phase of AI progress will depend on how effectively organisations integrate human knowledge, skills, and processes into AI systems.

That shift places people, not machines, at the centre of AI transformation. Kurbalija highlights the limits of traditional training approaches and points to new models of learning that focus on hands-on development and deep understanding of data.

Building a simple AI tool may now take minutes, but turning it into a reliable, high-precision system requires sustained human effort, from refining data to rethinking internal workflows.

Looking ahead to 2026, the message is clear. Success in AI will not be defined by who owns the most powerful chips, but by who invests most wisely in people.

As Kurbalija concludes, organisations that treat AI as a skill to be cultivated, rather than a product to be purchased, are far more likely to see lasting benefits from the technology.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!