Search
The Futurum Group's Statement on Israel

Is ChatGPT’s OpenAI Looking to Make Its Own AI Chips?

Is ChatGPT’s OpenAI Looking to Make Its Own AI Chips?

Is ChatGPT’s OpenAI Looking to Make Its Own AI Chips?

Of all the AI chip strategies from all the generative AI startups in the world, ChatGPT’s Open AI is reportedly contemplating its own move to enter the AI chip-making business. According to a recent report by Reuters, OpenAI, the creator of ChatGPT, is exploring the idea of either building its own chips, acquiring an existing chip maker, or expanding its pool of chip suppliers beyond NVIDIA, which is its lone supplier today. The Reuters report said the company has been discussing various options since 2022 due to shortages of the kinds of AI chips that are needed in its work. No decision has been made by OpenAI on the matter, according to Reuters, but this possibility certainly raises some interesting issues.

Is ChatGPT’s OpenAI Looking to Make Its Own AI Chips?

I get it that ChatGPT’s OpenAI might want more control of the availability and pricing of AI chips, which are critical to the company’s operations and business. That makes perfect sense. So, the Reuters story is right on point and relates an intriguing storyline that is apparently unfolding in the executive offices and hallways of OpenAI.

At first glance, it looks like creating its own AI chips could be a good idea for OpenAI to consider. Sure, build your own chips so you do not have to rely on anyone else; you can produce and secure the chip supplies that you need to serve yourself and keep your company ahead of your competitors.

On second glance, though, you must think of the significant ramifications of making such a move―and they are not small. To start, there are the immense costs of taking on the manufacturing of AI chips. You think paying someone else for their chips is expensive? Then start looking at what it is going to cost you to design your own chips and then build your own chip-making facilities or line up a fab that might have the capacity to make them for you. And what is it going to cost you to develop a roadmap of new and better chips on a neverending schedule into the future? And as if that is not enough, what about your own supply chain worries about keeping the chips flowing and selling? There are an awful lot of zeroes in the price tags for such operations, even if you decide to go out and acquire an existing chip maker or hire a fab to make them for you.

Meanwhile, let us say that even with all these challenges you still decide to pursue the idea. Where does that leave you?

Well, none of these complex processes will happen quickly, so your new AI chips will just start coming to market years after you begin all this complex work. And that means that your competitors will have been working on their core technologies all that time―while you were just getting things off the ground. And in that timespan, those competitors will be upgrading and replacing their products with faster chips rather than spending their time and money just getting started. It seems to me that you might need a very long time to catch up and make it worthwhile, even if you could catch up.

You know, all of this makes my head spin. It makes me think of carmakers and the similar decisions they must make each year to introduce new car and truck models. Stamping machines that make body panels, engine production lines and casting systems that must be changed out, and a million other decisions and production steps are affected. And sometimes, by the time their new vehicle models are out after a few years, the market might have changed. Oops. OpenAI might not want to get into that situation at all.

My Bottom Line on the ChatGPT OpenAI Chip-Making Rumor

First, let us remember that so far, this story is just that, a rumor from an article from Reuters. Maybe this will not come to fruition. Maybe it is just the company venting in the marketplace.
However, maybe the idea for Open AI to produce its own chips is not so crazy.

Maybe OpenAI could pull off this idea with the continuing help of Microsoft and other financial backers that might envision real benefits from such an uphill battle. Yes, I might see it as more posturing and ego than as a smart strategy at this point, but maybe OpenAI foresees something I am missing in my own crystal ball.

Certainly, AI chips are never going to be inexpensive to develop. There must be a very attractive business reason to jump into that market, given all its risks. And maybe OpenAI has such a reason in mind that is on the company’s corporate radar and that none of us yet understand.

Or perhaps all this talk is just to stir the field and make some noise. OpenAI is not shy about making noise in the AI marketplace, so that is also possible.

For now, we will have to wait and see what happens. As technology watchers like me have followed the swift global expansion of generative AI and AI in the marketplace in the past few years, it would not have struck me that producing one’s own AI chips would be part of the strategy.

But as I think about it, OpenAI has never looked at things the way other companies look at things. With new ideas, visions, technologies, and directions, ChatGPT’s OpenAI could well be looking at solving its own IT challenges by using what could be yet another all-new approach, by producing its own AI chips. It will be fascinating to watch how this goes and to learn whether the rumors are true or not.

Other insights from The Futurum Group:

The Ramifications of ChatGPT Going Realtime Web

OpenAI ChatGPT Enterprise: A Tall Order

Google Cloud’s TPU v5e Accelerates the AI Compute War

Author Information

Todd joined The Futurum Group as an Analyst after over 20 years as a technology journalist covering such topic areas as artificial intelligence (AI), deep learning (DL), machine learning (ML), open source and Linux, high-performance computing, supercomputers, cloud computing, virtualization, containers and microservices, IT security and more.

Prior to his work with The Futurum Group, Todd previously served as managing editor of EnterpriseAI.news from 2020 through 2022 where he worked to drive coverage of AI use and innovation in the enterprise. He also served in the past as a staff writer for Computerworld and eWEEK and freelanced for a wide range of tech websites, including TechRepublic, Channel Futures and Channel Partners, Computerworld, PC World, Data Center Knowledge, IT Pro Today, Linux.com and The Linux Foundation.

Todd holds a Bachelor’s Degree in Journalism from the University of Wisconsin-Madison. A Philadelphia native, he lives in Lancaster County, Pa., and spends his spare time tinkering with his vintage Mazda Miata convertible and collecting toy taxis from around the world.

SHARE:

Latest Insights:

Shopping Muse Uses Generative AI to Help Shoppers Find Just What They Need—Even Without the Exact Words to Describe It
Sherril Hanson, Senior Analyst at The Futurum Group, breaks down Dynamic Yield by Mastercard’s new personal shopping assistant solution, Shopping Muse, that can provide a personalized digital shopping experience.
On this episode of The Six Five – On The Road, hosts Daniel Newman and Patrick Moorhead welcome Chetan Kapoor, Director at AWS EC2 for a conversation on AWS Generative AI Infrastructure announced at AWS re:Invent.
A Deep Dive into Databricks’ RAG Tool Suite
Steven Dickens, VP and Practice Leader at The Futurum Group, shares his insights on Databricks' RAG suite and Vector Search service, which are reshaping AI application development.

Latest Research:

In our latest Market Insight Report, ‘Tis the Season: CX “Must Haves” for High-Volume Travel Periods, we dive into the five must haves for high-volume travel periods, and what companies can do to keep the relationship as positive as possible for their customers with supporting CX technologies and strategies.
In our latest Research Brief, Delivering Personalized Outreach via Dynamics 365 Customer Insights, completed in partnership with Microsoft, The Futurum Group covers demand drivers for greater personalization and discusses the ways in which organizations must engage with customers to create personalized, curated journeys that meet their needs and exceed their expectations.
In our latest Research Brief, Enhancing Sales and Service Efficiency via Microsoft Sales Copilot, completed in partnership with Microsoft, The Futurum Group discusses Microsoft Sales Copilot, designed to allow the automated generation of content, the surfacing of insights, and the creation of recommended actions, all without switching applications or interrupting a typical workflow.