Menu

Memcon 2024: Memory Technologies Are a Key Component to Scale AI

Memcon 2024: Memory Technologies Are a Key Component to Scale AI

In the second year of the Memcon conference, it was all memory and how to feed the “beast,” the GPU. In 2023, the focus was on CXL technology, how and where it would be adopted. Because of the explosion of generative AI, the commentary on types of memory has shifted to high-bandwidth memory (HBM) to address the bandwidth needed to maximize the GPU performance. Overall, there was a general agreement, CXL, while it addressed the capacity, it cannot deliver on the bandwidth needed for the training of generative AI applications. Rather we would see CXL memory modes applied to inference or more likely In Memory Database applications, such as SAP Hana.

In a very related note to generative AI and HBM, the community addressed the pressing issues of designing for the scale that this new AI will require. Those areas are advanced packaging for memory, for instance scaling HBM to 4, 8, 16+ layers. The question is how many layers before the signaling or mechanical design become issues? The second area is scalable networks. Is it PCIe or IB? These are big bets that organizations will need to make in their deployments. The trend seems to be Ethernet, but the high-end systems will probably cling to IB.

The next issue was cooling and energy discussed throughout the two days. The more processors, cores, and memory, the hotter these systems become. The major research labs have turned to liquid cooling. Expect this to become the norm for AI systems as they grow, at least self-contained systems with liquid. Perhaps we will be back with water cooled facilities in the near future.

Methods and techniques being deployed to overcome the memory constraints were also presented. Tejas Chopra of Netflix discussed the gyrations the company’s data scientists and programmers deployed, which spanned model pruning, efficient mini-batch selection, data quantization, and paging. Asked if this goes away with new memory offerings, he stated no. It just advances the capabilities of the current environment. Some of these methods were echoed by Shell (the energy company) and how it broke data into cubes to fit into memory or using more checkpointing and compression.

So what about CXL? Samsung was a major sponsor for the event and used the forum to launch its CXL Memory Module box CMM-B with 2 TB of memory for memory hungry databases and their HBM3M, 12-stack offering. There was some question as to when we are going to see this in real deployment—an understandable position given the 5 years we have been talking about CXL. Well, we are getting there. Very exciting were the partnerships with VMware and Red Hat in regard to joint development.

VMware will be releasing support for Samsung’s CMM-H, which is CXL 2.0 pooled memory. In a release planned for later in 2024, vSphere will support tiered memory. This will enable disaggregated memory to feed core capacity. The result: increased VM density per core, more memory for database-driven apps such as SAP Hana and cluster-wide memory for large-scale environments. Given all the noise on VMware licensing, this might give a bit of relief depending on how the next release is priced.

Red Hat and Samsung had previously announced the qualification of DRAM Memory Module (CMM-D) for pooling of memory with RHEL 9.3.

We are making progress with the memory constraints but do not expect workarounds to disappear. Rather, the applications will continue to consume whatever we can feed them. So back to work memory engineers!

Disclosure: The Futurum Group is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of The Futurum Group as a whole.

Other Insights from The Futurum Group:

The Six Five Talk Samsung’s Memory Tech Day

Memory Market: Call it a Comeback?

Marvell Industry Analyst Day 2023: Accelerated Computing Takes Off

Author Information

Camberley Bates

Now retired, Camberley brought over 25 years of executive experience leading sales and marketing teams at Fortune 500 firms. Before joining The Futurum Group, she led the Evaluator Group, an information technology analyst firm as Managing Director.

Her career spanned all elements of sales and marketing including a 360-degree view of addressing challenges and delivering solutions was achieved from crossing the boundary of sales and channel engagement with large enterprise vendors and her own 100-person IT services firm.

Camberley provided Global 250 startups with go-to-market strategies, creating a new market category “MAID” as Vice President of Marketing at COPAN and led a worldwide marketing team including channels as a VP at VERITAS. At GE Access, a $2B distribution company, she served as VP of a new division and succeeded in growing the company from $14 to $500 million and built a successful 100-person IT services firm. Camberley began her career at IBM in sales and management.

She holds a Bachelor of Science in International Business from California State University – Long Beach and executive certificates from Wellesley and Wharton School of Business.

Related Insights
Arm Q3 FY 2026 Earnings Highlight AI-Driven Royalty Momentum
February 6, 2026

Arm Q3 FY 2026 Earnings Highlight AI-Driven Royalty Momentum

Futurum Research analyzes Arm’s Q3 FY 2026 results, highlighting CPU-led AI inference momentum, CSS-driven royalty leverage, and diversification across data center, edge, and automotive, with guidance pointing to continued growth....
Qualcomm Q1 FY 2026 Earnings Record Revenue, Memory Headwinds
February 6, 2026

Qualcomm Q1 FY 2026 Earnings: Record Revenue, Memory Headwinds

Futurum Research analyzes Qualcomm’s Q1 FY 2026 earnings, highlighting AI-native device momentum, Snapdragon X PCs, and automotive SDV traction amid near-term handset build constraints from industry-wide memory tightness....
Alphabet Q4 FY 2025 Highlights Cloud Acceleration and Enterprise AI Momentum
February 6, 2026

Alphabet Q4 FY 2025 Highlights Cloud Acceleration and Enterprise AI Momentum

Nick Patience, VP and AI Practice Lead at Futurum analyzes Alphabet’s Q4 FY 2025 results, highlighting AI-driven momentum across Cloud and Search, Gemini scale, and 2026 capex priorities to expand...
Amazon CES 2026 Do Ring, Fire TV, and Alexa+ Add Up to One Strategy
February 5, 2026

Amazon CES 2026: Do Ring, Fire TV, and Alexa+ Add Up to One Strategy?

Olivier Blanchard, Research Director at The Futurum Group, examines Amazon’s CES 2026 announcements across Ring, Fire TV, and Alexa+, focusing on AI-powered security, faster interfaces, and expanded assistant access across...
Is 2026 the Turning Point for Industrial-Scale Agentic AI?
February 5, 2026

Is 2026 the Turning Point for Industrial-Scale Agentic AI?

VP and Practice Lead Fernando Montenegro shares insights from the Cisco AI Summit 2026, where leaders from the major AI ecosystem providers gathered to discuss bridging the AI ROI gap...
NXP Q4 FY 2025: Auto Stabilises, Edge AI Platforms Gain Traction
February 5, 2026

NXP Q4 FY 2025: Auto Stabilises, Edge AI Platforms Gain Traction

Futurum Research analyzes NXP’s Q4 FY 2025 earnings, highlighting SDV design wins, edge AI platform traction, and portfolio focus, with guidance pointing to steady margins and disciplined channel management into...

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.