MIT researchers developed Attention Matching, a KV cache compaction technique that compresses LLM memory by 50x in seconds — ...
AI infrastructure can't evolve as fast as model innovation. Memory architecture is one of the few levers capable of accelerating deployment cycles. Enter SOCAMM2 ...
LLC, positioned between external memory and internal subsystems, stores frequently accessed data close to compute resources.
At the Huawei Product & Solution Launch during MWC Barcelona 2026, Yuan Yuan, President of Huawei Data Storage Product Line, officially launched Huawei's AI Data Platform. The platform integrates ...
When we talk about the cost of AI infrastructure, the focus is usually on Nvidia and GPUs -- but memory is an increasingly important part of the picture.
Micron Technology, Inc. (Nasdaq: MU) today extended its leadership in low-power server memory by shipping customer samples of the industry’s highest-capacity LPDRAM module - 256GB SOCAMM2. Enabled by ...
As the demand for real-time data processing escalates, the technology behind Compute Express Link, known as CXL, is emerging as a critical solution for modern data centers. CXL memory is one solution ...
At MWC Barcelona 2026 the president of Huawei Data Storage Product Line shared Huawei's key insights and innovations ...
The purpose of this application note is to familiarize the reader with the Level 1 (L1) CPU cache implementation in the PIC32MZ device family by bringing awareness to the hazards that can occur in a ...