Menu

AMD: Our GPUs Running LLMs

AMD: Our GPUs Running LLMs

The News: On September 26, large language model (LLM) fine-tuning specialist Lamini revealed in a blog post the availability of its LLM Superstation, a graphics processing unit (GPU) compute platform powered by AMD GPUs that is optimized to run private enterprise LLMs. Here are the key details:

  • The solution combines Lamini’s enterprise LLM infrastructure with AMD Instinct MI210 and MI250 accelerators.
  • Superstation can run Llama 2 70B out of the box. Lamini claims the setup costs 10x less than Amazon Web Services (AWS). (Note that it is unclear what is being compared.)
  • LLM Superstation is available now, both in the cloud and on premises. Lamini highlights this availability compared with the current 52-week lead time for NVIDIA H100s.
  • Lamini has been running its fine-tuned LLMs on AMD Instinct GPUs “for the past year.”
  • In a comparison of AMD’s current accelerators versus NVIDIA, the Instinct m1250x is comparable in compute power and memory to NVIDIA A100 but is not nearly as powerful as the NVIDIA H100.
  • Lamini ran tests that show AMD’s ROCm software “provides a solid foundation for high-performance applications like fine-tuning LLMs.”

Read Lamini’s post about the AMD Superstation here.

Read Lamini’s NVIDIA-AMD Oven-Grill post on X here.

AMD: Our GPUs Running LLMs

Analyst Take: LLM workloads are not the exclusive purview of NVIDIA. AMD’s Lamini initiative has impact on the AI compute market. Here is my take.

Comparing Apples to Oranges

The news that the AMD-Lamini LLM Superstation can run LLMs is welcome, but you have to run Lamini fine-tuning, which is designed to reduce the workload – it is unclear whether the Superstation would be able to run an LLM without fine-tuning it. Lamini uses the example of Llama 2 70B as the LLM used, which is an open source model. Will the setup run other LLMs?

The AMD GPUs compare well with NVIDIA’s A100s but are not nearly as powerful as NVIDIA’s H100s. Not apples to oranges.

On the plus side, some developers feel an equalizer for AMD has been an improvement in its GPU software. In Y combinator’s Hacker News, one commentor posted this about the Lamini-AMD initiative:

“The hard part about using any AI Chips other than NVIDIA has been software. ROCm is finally at the point where it can train and deploy LLMs like Llama 2 in production.”

Lamini CTO Greg Diamos, who was an early architect of CUDA at NVIDIA said, “Using Lamini software, ROCm has achieved software parity with CUDA for LLMs. We chose the Instinct MI250 as the foundation for Lamini because it runs the biggest models that our customers demand and integrates fine-tuning optimizations. We use the large HBM capacity (128 GB) on MI250 to run bigger models with lower software complexity than clusters of A100s.”

Availability

The backlog of NVIDIA H100s, Lamini says a 52-week lead time, is an extremely compelling argument for enterprises to consider the AMD-Lamini Superstation. Perhaps it is not as powerful and might not offer the broadest range of LLM options, but it is a system that can be operationalized right now, and with the pace of AI innovation, that is a critical consideration.

Seeking Options To Reduce AI Compute

Along with availability, perhaps the winds of AI compute are set to change anyway. The extraordinarily large AI compute loads necessary to run training and inference for giant LLMs is probably not economically viable or sustainable. The trend has been toward “smaller” LLMs with the fine-tuning being championed by Lamini and others. AI workloads might not require the horsepower NVIDIA’s H100s provide.

Conclusions

Despite some limitations, the AMD-Lamini Superstation is a viable option for enterprises to consider for deploying LLMs. Savvy enterprises that cannot wait a calendar year to run AI workloads will be test-driving the system.

Disclosure: The Futurum Group is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of The Futurum Group as a whole.

Other insights from The Futurum Group:

AMD and Hugging Face Team Up to Democratize AI Compute – Shrewd Alliance Could Lead to AI Compute Competition, Lower AI Costs

AMD Revenue Hits $5.4 Billion in Q2, Down 18% YoY, But Beats Estimates

Hybrid Cloud Journey: How Nutanix, AMD and HPE Power Modern Apps | Futurum Tech Webcast

Author Information

Based in Tampa, Florida, Mark is a veteran market research analyst with 25 years of experience interpreting technology business and holds a Bachelor of Science from the University of Florida.

Related Insights
CIO Take Smartsheet's Intelligent Work Management as a Strategic Execution Platform
December 22, 2025

CIO Take: Smartsheet’s Intelligent Work Management as a Strategic Execution Platform

Dion Hinchcliffe analyzes Smartsheet’s Intelligent Work Management announcements from a CIO lens—what’s real about agentic AI for execution at scale, what’s risky, and what to validate before standardizing....
Will Zoho’s Embedded AI Enterprise Spend and Billing Solutions Drive Growth
December 22, 2025

Will Zoho’s Embedded AI Enterprise Spend and Billing Solutions Drive Growth?

Keith Kirkpatrick, Research Director with Futurum, shares his insights on Zoho’s latest finance-focused releases, Zoho Spend and Zoho Billing Enterprise Edition, further underscoring Zoho’s drive to illustrate its enterprise-focused capabilities....
Will IFS’ Acquisition of Softeon Help Attract New Supply Chain Customers
December 19, 2025

Will IFS’ Acquisition of Softeon Help Attract New Supply Chain Customers?

Keith Kirkpatrick, Research Director at Futurum, shares his insights into IFS’ acquisition of WMS provider Softeon, and provides his assessment on the impact to IFS’s market position and the overall...
NVIDIA Bolsters AI/HPC Ecosystem with Nemotron 3 Models and SchedMD Buy
December 16, 2025

NVIDIA Bolsters AI/HPC Ecosystem with Nemotron 3 Models and SchedMD Buy

Nick Patience, AI Platforms Practice Lead at Futurum, shares his insights on NVIDIA's release of its Nemotron 3 family of open-source models and the acquisition of SchedMD, the developer of...
Will a Digital Adoption Platform Become a Must-Have App in 2026?
December 15, 2025

Will a DAP Become the Must-Have Software App in 2026?

Keith Kirkpatrick, Research Director with Futurum, covers WalkMe’s 2025 Analyst Day, and discusses the company’s key pillars for driving success with enterprise software in an AI- and agentic-dominated world heading...
Broadcom Q4 FY 2025 Earnings AI And Software Drive Beat
December 15, 2025

Broadcom Q4 FY 2025 Earnings: AI And Software Drive Beat

Futurum Research analyzes Broadcom’s Q4 FY 2025 results, highlighting accelerating AI semiconductor momentum, Ethernet AI switching backlog, and VMware Cloud Foundation gains, alongside system-level deliveries....

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.