An AI factory brings Nvidia compute into Samsung’s fabs

Nvidia and Samsung outlined a semiconductor AI factory that embeds accelerated computing into production. Over 50,000 GPUs will drive digital twins, predictive maintenance, and real-time optimisation. Partners present the project as a template for autonomous fabs.

The alliance spans design and manufacturing. Samsung uses CUDA-X and EDA tools to speed simulation and verification. Integrating cuLitho into OPC reports roughly twentyfold gains in computational lithography at advanced nodes.

Factory planning and logistics run on Omniverse digital twins and RTX PRO servers. Unified analytics support anomaly detection, capacity planning, and flow balancing. Managers expect shorter ramps and smoother changeovers with higher equipment effectiveness.

Robotics and edge AI extend intelligence to the line. Isaac Sim, Cosmos models, and Jetson Thor target safe collaboration, faster task retargeting, and teleoperation. Samsung’s in-house models enable multilingual assistance and on-site decision support.

A decades-long Nvidia–Samsung relationship underpins the effort, from NV1 DRAM to HBM3E and HBM4. Work continues on memory, modules, and foundry services, plus AI-RAN research with networks in South Korea and academia linking factory intelligence with next-gen connectivity.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

CXMT launches LPDDR5X chips as China advances in semiconductor race

ChangXin Memory Technologies has begun mass production of LPDDR5X chips, marking a major milestone in China’s effort to strengthen its position in the global semiconductor market.

The Hefei-based manufacturer, preparing for a Shanghai stock listing, said its new DRAM generation will support faster data transfer and lower power use across mobile devices and AI systems.

The LPDDR5X range includes chips with speeds of up to 10,667 Mbps, positioning CXMT as a growing competitor to industry leaders such as Samsung, SK Hynix and Micron.

Earlier LPDDR5 versions launched in 2023 had already helped the firm progress towards advanced 16-nanometre manufacturing, narrowing the technological gap with global rivals.

Industry data indicate a rising global demand for memory chips, driven by AI applications and high-bandwidth computing. Additionally, DRAM revenue increased 17.1 percent in the second quarter, reaching US$31.6 billion.

CXMT’s expansion comes as it targets a Shanghai IPO valued at around 300 billion yuan, highlighting both investor interest and the ambition of China to achieve greater chip self-sufficiency.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Trainium2 power surges as AWS’s Project Rainier enters service for Anthropic

Anthropic and AWS switched on Project Rainier, a vast Trainium2 cluster spanning multiple US sites to accelerate Claude’s evolution.

Project Rainier is now fully operational, less than a year after its announcement. AWS engineered an EC2 UltraCluster of Trainium2 UltraServers to deliver massive training capacity. Anthropic says it offers more than five times the compute used for prior Claude models.

UltraServers bind four Trainium2 servers with high-speed NeuronLinks so 64 chips act as one. Tens of thousands of networks are connected through Elastic Fabric Adapter across buildings. The design reduces latency within racks while preserving flexible scale across data centres.

Anthropic is already training and serving Claude on Rainier across the US and plans to exceed one million Trainium2 chips by year’s end. More computing should raise model accuracy, speed evaluations, and shorten iteration cycles for new frontier releases.

AWS controls the stack from chip to data centre for reliability and efficiency. Teams tune power delivery, cooling, and software orchestration. New sites add water-wise cooling, contributing to the company’s renewable energy and net-zero goals.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Humanoid robots set to power Foxconn’s new Nvidia server plant in Houston

Foxconn will add humanoid robots to a new Houston plant building Nvidia AI servers from early 2026. Announced at Nvidia’s developer conference, the move deepens their partnership and positions the site as a US showcase for AI-driven manufacturing.

Humanoid systems based on Nvidia’s Isaac GR00T N are built to perceive parts, adapt on the line, and work with people. Unlike fixed industrial arms, they handle delicate assembly and switch tasks via software updates. Goals include flexible throughput, faster retooling, and fewer stoppages.

AI models are trained in simulation using digital twins and reinforcement learning to improve accuracy and safety. On the line, robots self-tune as analytics predict maintenance and balance workloads, unlocking gains across logistics, assembly, testing, and quality control.

Texas, US, offers proximity to a growing semiconductor and AI cluster, as well as policy support for domestic capacity. Foxconn also plans expansions in Wisconsin and California to meet global demand for AI servers. Scaling output should ease supply pressures around Nvidia-class compute in data centres.

Job roles will shift as routine tasks automate and oversight becomes data-driven. Human workers focus on design, line configuration, and AI supervision, with safety gates for collaboration. Analysts see a template for Industry 4.0 factories running near-continuously with rapid changeovers.

Would you like to learn more about AI, tech, and digital diplomacy? If so, ask our Diplo chatbot!

Nvidia and Deutsche Telekom plan €1 billion AI data centre in Germany

Plans are being rolled out for a €1 billion data centre in Germany to bolster Europe’s AI infrastructure, with Nvidia and Deutsche Telekom set to co-fund the project.

The facility is expected to serve enterprise customers, including SAP SE, Europe’s largest software company, and to deploy around 10,000 advanced chips known as graphics processing units (GPUs).

While significant for Europe, the build is modest compared with gigawatt-scale sites elsewhere, highlighting the region’s push to catch up with US and Chinese capacity.

An announcement is anticipated next month in Berlin alongside senior industry and government figures, with Munich identified as the planned location.

The move aligns with the EU efforts to expand AI compute, including the €200 billion initiative announced in February to grow capacity over the next five to seven years.

Would you like to learn more about AI, tech and digital diplomacyIf so, ask our Diplo chatbot!

AI200 and AI250 set a rack-scale inference push from Qualcomm

Qualcomm unveiled AI200 and AI250 data-centre accelerators aimed at high-throughput, low-TCO generative AI inference. AI200 targets rack-level deployment with high performance per pound per watt and 768 GB LPDDR per card for large models.

AI250 introduces a near-memory architecture that boosts adequate memory bandwidth by over tenfold while lowering power draw. Qualcomm pitches the design for disaggregated serving, improving hardware utilisation across large fleets.

Both arrive as full racks with direct liquid cooling, PCIe for scale-up, Ethernet for scale-out, and confidential computing. Qualcomm quotes around 160 kW per rack for thermally efficient, dense inference.

A hyperscaler-grade software stack spans apps to system software with one-click onboarding of Hugging Face models. Support covers leading frameworks, inference engines, and optimisation techniques to simplify secure, scalable deployments.

Commercial timing splits the roadmap: AI200 in 2026 and AI250 in 2027. Qualcomm commits to an annual cadence for data-centre inference, aiming to lead in performance, energy efficiency, and total cost of ownership.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

AMD powers US AI factory supercomputers for national research

The US Department of Energy and AMD are joining forces to expand America’s AI and scientific computing power through two new supercomputers at Oak Ridge National Laboratory.

Named Lux and Discovery, the systems will drive the country’s sovereign AI strategy, combining public and private investment worth around $1 billion to strengthen research, innovation, and security infrastructure.

Lux, arriving in 2026, will become the nation’s first dedicated AI factory for science.

Built with AMD’s EPYC CPUs and Instinct GPUs alongside Oracle and HPE technologies, Lux will accelerate research across materials, medicine, and advanced manufacturing, supporting the US AI Action Plan and boosting the Department of Energy’s AI capacity.

Discovery, set for deployment in 2028, will deepen collaboration between the DOE, AMD, and HPE. Powered by AMD’s next-generation ‘Venice’ CPUs and MI430X GPUs, Discovery will train and deploy AI models on secure US-built systems, protecting national data and competitiveness.

It aims to deliver faster energy, biology, and national security breakthroughs while maintaining high efficiency and open standards.

AMD’s CEO, Dr Lisa Su, said the collaboration represents the best public-private partnerships, advancing the nation’s foundation for science and innovation.

US Energy Secretary Chris Wright described the initiative as proof that America leads when government and industry work together toward shared AI and scientific goals.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Qualcomm and HUMAIN power Saudi Arabia’s AI transformation

HUMAIN and Qualcomm Technologies have launched a collaboration to deploy advanced AI infrastructure in Saudi Arabia, aiming to position the Kingdom as a global hub for AI.

Announced ahead of the Future Investment Initiative conference, the project will deliver the world’s first fully optimised edge-to-cloud AI system, expanding Saudi Arabia’s regional and global inferencing services capabilities.

In 2026, HUMAIN plans to deploy 200 megawatts of Qualcomm’s AI200 and AI250 rack solutions to power large-scale AI inference services.

The partnership combines HUMAIN’s regional infrastructure and full AI stack with Qualcomm’s semiconductor expertise, creating a model for nations seeking to develop sovereign AI ecosystems.

However, the initiative will also integrate HUMAIN’s Saudi-developed ALLaM models with Qualcomm’s AI platforms, offering enterprise and government customers tailor-made solutions for industry-specific needs.

The collaboration supports Saudi Arabia’s strategy to drive economic growth through AI and semiconductor innovation, reinforcing its ambition to lead the next wave of global intelligent computing.

Qualcomm’s CEO Cristiano Amon said the partnership would help the Kingdom build a technology ecosystem to accelerate its AI ambitions.

HUMAIN CEO Tareq Amin added that combining local insight with Qualcomm’s product leadership will establish Saudi Arabia as a key player in global AI and semiconductor development.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

Gigawatt-scale AI marks Anthropic’s next compute leap

Anthropic will massively expand on Google Cloud, planning to deploy up to 1 million TPUs and bring well over a gigawatt online in 2026. The multiyear investment totals tens of billions to accelerate research and product development.

Google Cloud CEO Thomas Kurian said Anthropic’s move reflects TPUs’ price-performance and efficiency, citing ongoing innovations and the seventh-generation ‘Ironwood’ TPU. Google will add capacity and drive further efficiency across its accelerator portfolio.

Anthropic now serves over 300,000 business customers, with large accounts up nearly sevenfold year over year. Added compute will meet demand while enabling deeper testing, alignment research, and responsible deployment at a global scale.

CFO Krishna Rao said the expansion keeps Claude at the frontier for Fortune 500s and AI-native startups alike. Increased capacity ensures reliability as usage and mission-critical workloads grow rapidly.

Anthropic’s diversified strategy spans Google TPUs, Amazon Trainium, and NVIDIA GPUs. It remains committed to Amazon as its primary training partner, including Project Rainier’s vast US clusters, and will continue investing to advance model capabilities.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!

South Korea moves to lead the AI era with OpenAI’s economic blueprint

Poised to become a global AI powerhouse, South Korea has the right foundations in place: advanced semiconductor production, robust digital infrastructure, and a highly skilled workforce.

OpenAI’s new Economic Blueprint for Korea sets out how the nation can turn those strengths into broad, inclusive growth through scaled and trusted AI adoption.

The blueprint builds on South Korea’s growing momentum in frontier technology.

Following OpenAI’s first Asia–Pacific country partnership, initiatives such as Stargate with Samsung and SK aim to expand advanced memory supply and explore next-generation AI data centres alongside the Ministry of Science and ICT.

A new OpenAI office in Seoul, along with collaboration with Seoul National University, further signals the country’s commitment to becoming an AI hub.

A strategy that rests on two complementary paths: building sovereign AI capabilities in infrastructure, data governance, and GPU supply, while also deepening cooperation with frontier developers like OpenAI.

The aim is to enhance operational maturity and cost efficiency across key industries, including semiconductors, shipbuilding, healthcare, and education.

By combining domestic expertise with global partnerships, South Korea could boost productivity, improve welfare services, and foster regional growth beyond Seoul. With decisive action, the nation stands ready to transform from a fast adopter into a global standard-setter for safe, scalable AI systems.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!