Menu

AMD: Our GPUs Running LLMs

AMD: Our GPUs Running LLMs

The News: On September 26, large language model (LLM) fine-tuning specialist Lamini revealed in a blog post the availability of its LLM Superstation, a graphics processing unit (GPU) compute platform powered by AMD GPUs that is optimized to run private enterprise LLMs. Here are the key details:

  • The solution combines Lamini’s enterprise LLM infrastructure with AMD Instinct MI210 and MI250 accelerators.
  • Superstation can run Llama 2 70B out of the box. Lamini claims the setup costs 10x less than Amazon Web Services (AWS). (Note that it is unclear what is being compared.)
  • LLM Superstation is available now, both in the cloud and on premises. Lamini highlights this availability compared with the current 52-week lead time for NVIDIA H100s.
  • Lamini has been running its fine-tuned LLMs on AMD Instinct GPUs “for the past year.”
  • In a comparison of AMD’s current accelerators versus NVIDIA, the Instinct m1250x is comparable in compute power and memory to NVIDIA A100 but is not nearly as powerful as the NVIDIA H100.
  • Lamini ran tests that show AMD’s ROCm software “provides a solid foundation for high-performance applications like fine-tuning LLMs.”

Read Lamini’s post about the AMD Superstation here.

Read Lamini’s NVIDIA-AMD Oven-Grill post on X here.

AMD: Our GPUs Running LLMs

Analyst Take: LLM workloads are not the exclusive purview of NVIDIA. AMD’s Lamini initiative has impact on the AI compute market. Here is my take.

Comparing Apples to Oranges

The news that the AMD-Lamini LLM Superstation can run LLMs is welcome, but you have to run Lamini fine-tuning, which is designed to reduce the workload – it is unclear whether the Superstation would be able to run an LLM without fine-tuning it. Lamini uses the example of Llama 2 70B as the LLM used, which is an open source model. Will the setup run other LLMs?

The AMD GPUs compare well with NVIDIA’s A100s but are not nearly as powerful as NVIDIA’s H100s. Not apples to oranges.

On the plus side, some developers feel an equalizer for AMD has been an improvement in its GPU software. In Y combinator’s Hacker News, one commentor posted this about the Lamini-AMD initiative:

“The hard part about using any AI Chips other than NVIDIA has been software. ROCm is finally at the point where it can train and deploy LLMs like Llama 2 in production.”

Lamini CTO Greg Diamos, who was an early architect of CUDA at NVIDIA said, “Using Lamini software, ROCm has achieved software parity with CUDA for LLMs. We chose the Instinct MI250 as the foundation for Lamini because it runs the biggest models that our customers demand and integrates fine-tuning optimizations. We use the large HBM capacity (128 GB) on MI250 to run bigger models with lower software complexity than clusters of A100s.”

Availability

The backlog of NVIDIA H100s, Lamini says a 52-week lead time, is an extremely compelling argument for enterprises to consider the AMD-Lamini Superstation. Perhaps it is not as powerful and might not offer the broadest range of LLM options, but it is a system that can be operationalized right now, and with the pace of AI innovation, that is a critical consideration.

Seeking Options To Reduce AI Compute

Along with availability, perhaps the winds of AI compute are set to change anyway. The extraordinarily large AI compute loads necessary to run training and inference for giant LLMs is probably not economically viable or sustainable. The trend has been toward “smaller” LLMs with the fine-tuning being championed by Lamini and others. AI workloads might not require the horsepower NVIDIA’s H100s provide.

Conclusions

Despite some limitations, the AMD-Lamini Superstation is a viable option for enterprises to consider for deploying LLMs. Savvy enterprises that cannot wait a calendar year to run AI workloads will be test-driving the system.

Disclosure: The Futurum Group is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of The Futurum Group as a whole.

Other insights from The Futurum Group:

AMD and Hugging Face Team Up to Democratize AI Compute – Shrewd Alliance Could Lead to AI Compute Competition, Lower AI Costs

AMD Revenue Hits $5.4 Billion in Q2, Down 18% YoY, But Beats Estimates

Hybrid Cloud Journey: How Nutanix, AMD and HPE Power Modern Apps | Futurum Tech Webcast

Author Information

Based in Tampa, Florida, Mark is a veteran market research analyst with 25 years of experience interpreting technology business and holds a Bachelor of Science from the University of Florida.

Related Insights
Will Supermicro's Legal Crisis Shift Server Market Share to New Dell and HPE GPU Platforms?
March 27, 2026

Will Supermicro’s Legal Crisis Shift Server Market Share to New Dell and HPE GPU Platforms?

Brendan Burke, Research Director at Futurum, shares insights on how Supermicro's export crisis creates a GPU allocation opening for Dell and HPE, reshaping the AI server competitive landscape post-NVIDIA GTC...
Is Workflow AI Now Native After Microsoft Embeds Copilot in Power Platform?
March 27, 2026

Is Workflow AI Now Native After Microsoft Embeds Copilot in Power Platform?

Keith Kirkpatrick, Research Director at Futurum, examines Microsoft Copilot integration within Power Platform and how agentic workflows and process mining reshape enterprise application execution and decision-making....
Infosys Bets on P&C Insurance Depth With Stratus Acquisition
March 27, 2026

Infosys Bets on P&C Insurance Depth With Stratus Acquisition

Infosys agreed to acquire Stratus, a New Jersey tech firm specializing in property and casualty insurance, signaling a strategic shift toward domain-specific AI in regulated industries....
Can Palo Alto's Agentic NetOps Actually Kill the NOC Ticket?
March 27, 2026

Can Palo Alto’s Agentic NetOps Actually Kill the NOC Ticket?

Palo Alto Networks' agentic AI Troubleshooting Agent in Prisma SD-WAN enables autonomous network remediation, reducing resolution time from hours to minutes, marking AI-native networking's transition to production....
Infosys Bets on Anthropic to Survive the Automation Wave It Helped Build
March 27, 2026

Infosys Bets on Anthropic to Survive the Automation Wave It Helped Build

Infosys expands its Anthropic partnership to develop enterprise AI agents, signaling that its labor arbitrage model faces disruption and reflecting an urgent pivot toward AI-first service delivery....
Red Piranha's Global InfoSec Win: Can Smaller Vendors Break the Cybersecurity Stalemate?
March 26, 2026

Red Piranha’s Global InfoSec Win: Can Smaller Vendors Break the Cybersecurity Stalemate?

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.