South Korean firm unveils faster AI data centre architecture with CXL-over-Xlink

Panmnesia’s CXL-over-Xlink model uses sub-100ns latency links and multi-level CXL switching to power scalable, efficient AI data centre operations without costly hardware upgrades.

Korean CXL specialist Panmnesia reveals new AI data centre model combining shared memory pools and scalable fabrics to reduce latency and improve efficiency across large-scale AI workloads.

South Korean company Panmnesia has introduced a new architecture for AI data centres aimed at improving speed and efficiency.

Instead of using only PCIe or RDMA-based systems, its CXL-over-Xlink approach combines Compute Express Link (CXL) with fast accelerator links such as UALink and NVLink.

The company claims this design can deliver up to 5.3 times faster AI training and reduce inference latency sixfold. By allowing CPUs and GPUs to access large shared memory pools via the CXL fabric, AI workloads are no longer restricted by the fixed memory limits inside each GPU.

It will enable data centres to scale compute and memory independently, adapting to changing workload demands without hardware overprovisioning.

Panmnesia’s system also reduces communication overhead using accelerator-optimised links for CXL traffic, helping maintain high throughput with sub-100ns latency.

The architecture incorporates a hierarchical memory model blending local high-bandwidth memory with pooled CXL memory, alongside scalable CXL 3.1 switches that connect hundreds of devices efficiently without bottlenecks.

Would you like to learn more about AI, tech and digital diplomacy? If so, ask our Diplo chatbot!