Generative AI Investment Accelerating: $1.3 Billion for LLM Inflection

Generative AI Investment Accelerating: $1.3 Billion for LLM Inflection

The News: Inflection announced on June 29 they have raised a new round of funding to the tune of $1.3 billion. Inflection is a large language model (LLM) founded by the co-founder of DeepMind, Mustafa Suleyman and LinkedIn co-founder Reed Hoffman. The company, launched approximately 1 year ago, is now valued at $4 billion. Investors include Microsoft, NVIDIA, Bill Gates, and former Google CEO Eric Schmidt.

Forbes reports that Inflection will install the largest GPU cluster for AI apps in the world – 23,000 NVIDIA H100s. Inflection’s chatbot is called “Pi”, and the company’s ambition is to create a more personal, emotional interface/LLM. Read the full Press Release about the investment round on Inflection’s website.

Generative AI Investment Accelerating: $1.3 Billion for LLM Inflection

Analyst Take: The $10 billion investment by Microsoft in Open AI aside (note: that’s a bit misleading, since it is not all cash to OpenAI, and a big chunk of the “investment” is providing Azure compute), the Inflection funding marks a significant milestone in the generative AI landscape. What will the Inflection investment mean to the generative AI market?

Short Term: More Investment in LLMs, GenAI Model Management, LLM Management

At this early stage of the generative AI market, investments are being made in the more foundational elements of the generative AI stack – AI compute and AI tools/platforms. This makes perfect sense since generative AI use cases are extremely formative. We are in the innovation and experimentation stage of the market. As enterprises contemplate the generative AI opportunity, many are thinking about how to create a competitive edge, and how to build their own IP, with generative AI.

AI cloud compute is established with cloud providers AWS, Microsoft Azure, Google Cloud, and to an extent, Oracle. The investment opportunity lies in the next layer – LLMs, generative AI model management, and LLM management. This group of companies provides the foundational tools enterprises need to experiment with and build generative AI applications. LLM options are exploding, and their value propositions are being refined so rapidly it is difficult to keep up with the market. The primary driver in this competitive space will be how to present an LLM that meets enterprise-grade requirements for scalable applications – security, accuracy, privacy, etc.

LLMs and other foundational models (diffusion, etc.) are proving to be a bit messy and not necessarily standalone, plug-and-play platforms. Consequently, a range of ancillary services has emerged for generative AI model management and LLM management, such as Trustwise, Whylabs, Galileo AI, OctoML, and Anyscale. These companies help enterprises scale AI compute, tune models, and help tackle hallucination, among other functions. Investment in these companies will grow significantly in the short term.

Short Term: More Investment in Data Management

Generative AI has sparked a renewed interest by enterprises in leveraging proprietary data. As such, data management is even more critical (See Databrick’s MosaicML Acquisition, LakehouseIQ Launch, Data + AI Summit Show Gen AI Savvy). Investments in companies like Databricks, Snowflake, MongoDB, SingleStore, and narrower data management specialists like LlamaIndex and Datasaur will grow significantly in the short term.

Longer Term: 2024 Investment in GenAI Applications and Specific Use Cases Accelerates

Generative AI use cases are formative, and with the exception of automated code development, there are few proven use cases and consequently, limited specific proven applications. While generative AI application startups are proliferating, significant investment in such companies will lag the more foundational layers of the generative AI stack, with acceleration in this top layer coming sometime in 2024.

Disclosure: The Futurum Group is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of The Futurum Group as a whole.

Other insights from The Futurum Group:

UK AI Regulations Criticized: A Cautionary Tale for AI Safety

New Generative AI-Powered Capabilities in Oracle Fusion Cloud HCM Announced

Improving Contact Center Experiences via NLP, NLU, and Analytics-Focused AI

Author Information

Mark comes to The Futurum Group from Omdia’s Artificial Intelligence practice, where his focus was on natural language and AI use cases.

Previously, Mark worked as a consultant and analyst providing custom and syndicated qualitative market analysis with an emphasis on mobile technology and identifying trends and opportunities for companies like Syniverse and ABI Research. He has been cited by international media outlets including CNBC, The Wall Street Journal, Bloomberg Businessweek, and CNET. Based in Tampa, Florida, Mark is a veteran market research analyst with 25 years of experience interpreting technology business and holds a Bachelor of Science from the University of Florida.

SHARE:

Latest Insights:

NVIDIA Expands Isaac Platform With New Tools and Systems To Accelerate Physical AI Development
Olivier Blanchard, Research Director at Futurum, shares insights on how NVIDIA’s Isaac GR00T platform and Blackwell systems aim to accelerate humanoid robotics through simulation, synthetic data, and integrated infrastructure.
Plug and Play Taps Synopsys to Expand Semiconductor Ecosystem and Support Startup Chip Design
Olivier Blanchard, Research Director at Futurum, explores how Synopsys’ collaboration with Plug and Play could reduce chip design costs and accelerate time-to-market for startups entering the semiconductor space.
Google Cloud Introduces Enhanced Data Boundary Controls, New Security Testing, and Localised Partnerships To Meet Rising Regional Demands for Data Governance and Operational Autonomy
Daniel Newman, CEO at Futurum, shares insights on Google Cloud’s sovereign cloud expansion, highlighting how new data controls and local partnerships aim to address EU regulatory concerns and customer sovereignty demands.
VMware-Led Software Sales and AI Chips Powered 20% YoY Topline Growth in Q2 FY 2025
Olivier Blanchard and Daniel Newman, at Futurum, explore Broadcom’s strong Q2 FY2025 earnings driven by robust growth in AI networking, VMware-led infrastructure software, and strategic acceleration of custom AI chip deployments.

Book a Demo

Thank you, we received your request, a member of our team will be in contact with you.