Menu

Large Language Model AI Needs to Be Invisible and Cheaper

Large Language Model AI Needs to Be Invisible and Cheaper

There is no doubt that large language model (LLM) AI is revolutionizing the ability of a computer to augment or replace human effort. The challenge is taking that revolution and gaining business value from LLMs. ChatGPT’s attention-grabbing ability to write college-level essays and Sora’s ability to generate life-like videos differ from using LLMs within a business application. LLMs must integrate easily into the tools that build business applications to deliver widespread business value from LLMs. The cost of running LLM inference in business applications must be controlled to allow maximum value.

Many organizations are trying to integrate LLMs into their applications and finding that months of work are required to gain any value from an LLM, let alone transform their applications. The primary issue is that the organization’s data and business processes must be integrated with the foundation LLM. The currently available technologies are a collection of incredibly powerful science projects that require significant tuning to each use case within an organization. The mix of projects is natural in the early stages of a new application class, and more will spring up over a few years. The usual maturity curve will apply. Over time, a few projects will rise to the top as the most useful, and these will become easier to implement. In the same way, machine learning (ML)-based AI for video has become a core component of some applications, and we will see LLMs become a feature rather than a product.

The cost to run LLM-based inference is a barrier to some use cases; LLM inference is resource-intensive and often requires GPUs installed in application servers. The high cost means that LLMs are only used where there is a high return. For broader use, the cost must come down. We are already seeing the use of quantization to reduce resource requirements, and as LLM sizes increase, we will need more techniques to reduce resource use. One development is that Intel has added a matrix math accelerator to the latest Xeon Scalable CPUs, reducing the need for GPUs to deliver business value from LLM inference performance.

I hope we see another seismic shift in AI, and LLMs become more applicable because they are easier and cheaper to integrate into business applications. I doubt we will see the future of intelligent assistance robots, flying cars, and unlimited leisure. But it would be nice if LLMs could make everyday applications more straightforward to use and more insight driven.

Disclosure: The Futurum Group is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of The Futurum Group as a whole.

Other Insights from The Futurum Group:

HPE Infuses GenAI LLMs to Uplift HPE Aruba Networking Central AIOps

AI Field Day: Nature Fresh Farms Profits by Machine Learning, Not LLMs

Why the Launch of LLM Gemini Will Underpin Google Revenue

Author Information

Alastair has made a twenty-year career out of helping people understand complex IT infrastructure and how to build solutions that fulfil business needs. Much of his career has included teaching official training courses for vendors, including HPE, VMware, and AWS. Alastair has written hundreds of analyst articles and papers exploring products and topics around on-premises infrastructure and virtualization and getting the most out of public cloud and hybrid infrastructure. Alastair has also been involved in community-driven, practitioner-led education through the vBrownBag podcast and the vBrownBag TechTalks.

Related Insights
Glean Doubles ARR to $200M. Can Its Knowledge Graph Beat Copilot
April 3, 2026

Glean Doubles ARR to $200M. Can Its Knowledge Graph Beat Copilot?

Nick Patience, VP & Practice Lead at Futurum, examines Glean's platform evolution from enterprise search to agentic AI, as it doubles ARR to $200M and battles Microsoft 365 Copilot for...
HP IQ Finally Brings Useful On-Device AI To Workspaces
April 3, 2026

HP IQ Finally Brings Useful On-Device AI To Workspaces

Olivier Blanchard, Research Director at Futurum, shares insights on HP IQ, HP’s workplace intelligence layer combining on-device AI, proximity-based connectivity, and IT control across devices and workflows....
CrowdStrike Deepens Agentic SOC Strategy Across Partners, Services, and Devices
April 1, 2026

CrowdStrike Deepens Agentic SOC Strategy Across Partners, Services, and Devices

Fernando Montenegro, VP & Practice Lead for Cybersecurity & Resilience at Futurum, examines CrowdStrike’s agentic SOC expansion across partners, IBM, and Intel, and what it means for security execution and...
LevelBlue–SentinelOne Partnership: Does Unified Security Improve Outcomes?
April 1, 2026

LevelBlue–SentinelOne Partnership: Does Unified Security Improve Outcomes?

Fernando Montenegro, VP & Practice Lead for Cybersecurity & Resilience at Futurum, analyzes the LevelBlue SentinelOne partnership and its focus on integrating threat intelligence, AI detection, and response to improve...
Palo Alto Bets on Agentic Endpoints Before Anyone Else Does
April 1, 2026

Palo Alto Bets on Agentic Endpoints Before Anyone Else Does

Palo Alto Networks bets big on AI agent security through Koi acquisition, with CEO Nikesh Arora backing it with a $10M stock purchase, positioning the company as first to formally...
IBM Pushes Voice AI Into watsonx. Can It Drive Enterprise Adoption
March 31, 2026

IBM Pushes Voice AI Into watsonx. Can It Drive Enterprise Adoption?

Nick Patience, VP and AI Practice Lead at Futurum, examines IBM ElevenLabs voice AI integration and how it enables multilingual, secure, voice-first AI agents within watsonx Orchestrate for enterprise workflows....

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.