The News: IBM announced on July 11 that it has begun to roll out IBM watsonx, the company’s enterprise-ready AI and data platform that is designed to help organizations accelerate and scale AI. Previewed 2 months ago at IBM THINK, watsonx encompasses the watsonx.ai studio for new foundation models (FMs), generative AI and machine learning (ML), and the watsonx.data fit-for-purpose data store, built on an open lakehouse architecture, which are now available. The watsonx.governance toolkit to help enable AI workflows to be built with responsibility, transparency, and explainability is coming later in 2023. You can read the original Press Release at this link.
watsonx.ai Leverages Foundation Models to Accelerate AI Application Development
Analyst Take: IBM announced that it is making several watsonx products generally available to enterprise customers, with the goal of helping organizations accelerate the development of AI technology, which can be then deployed at scale. watsonx.ai, the enterprise studio targeted at AI builders and developers, is among the applications that are now available, and supports the training, validation, tuning, and deployment of traditional ML and generative AI capabilities, powered by FMs.
Pre-Trained Models from IBM and Hugging Face for AI Development Tasks
Today in watsonx.ai, AI builders can leverage pre-trained models from IBM and from the Hugging Face community for a range of AI development tasks, including natural language processing (NLP)-type tasks, such as question answering, content generation and summarization, text classification, and extraction. Further, watsonx.ai includes several additional capabilities designed to enable the creation and use of generative AI-based applications, including:
- Retrieval-Augmented Generation (RAG), which allows the creation of a chatbot or a question-answering feature grounded on specific content
- Summarization, which allows the transformation of text into overviews with key points
- Insight extraction, which analyzes existing unstructured text and then surfacing insights in specialized domain areas
- Content generation, which generates text for a specific purpose
- Named Entity Recognition, which identifies and extracts essential and specific information from unstructured text
- Classification, which allows the reading and classification of written input with zero examples
In future releases, IBM says it will provide access to a larger variety of IBM-trained proprietary foundation models for efficient domain and task specialization.
Foundation Models Allow Enterprises to Jump-Start the Development Process
The use of foundation models allows organizations to create AI-driven applications more quickly, compared with the traditional method of training each model to handle a specific task. This is because the process starts with a model that has already been pre-trained. With traditional model development, the process begins with the initial collection of data, followed by additional curation, cleaning, and labeling of that data to create a specific task. To handle a different task, the same process needed to be run again, resulting in a significant time and resource allocation, given the need to label anywhere from a thousand to even millions of data points to ensure the model can be properly trained.
Automating Prompt Engineering Can Speed Up the Tuning Process
The use of FMs allows organizations to accelerate application development, which could be used right out of the box. To provide more domain functionality, the model can (and should be) tweaked or tuned to suit a specific use case with as little as a hundred or a thousand additional data labels, which is far less time-consuming or expensive than developing another model from scratch. This enables faster time-to-value and in many cases, the realization of a better ROI.
To tune a model, model builders often use prompt engineering, a process by which prompts are used to refine the model to format or structure its output based on the input provided via a prompt. A prompt engineer would input text with examples of how the model should respond, based on certain inputs, and over time, the model will be tuned to respond in the desired way.
However, for large enterprises with hundreds of tasks or applications, tuning is a better approach. Model tuning involves the automation of the creation of prompt, so that the prompt engineering process is automatic, enabling a more cost-effective way to further refine the model’s output as the number of tasks, functions, or variables increase. The watsonx.ai Studio includes the Prompt Lab, in which users can experiment with zero-shot, one-shot, or few-shot prompting to support a range of NLP-type tasks including question answering, content generation and summarization, text classification, and extraction.
Later in the year, future watsonx.ai releases will include capabilities for prompt tuning and fine-tuning FMs as part of its Tuning Studio to help tune FMs with labeled data for better performance and accuracy.
Disclosure: The Futurum Group is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.
Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of The Futurum Group as a whole.
Other insights from The Futurum Group:
IBM watsonx.data is Now Generally Available as Part of Major watsonx Announcement
IBM Brings Advanced Feature Set to New FlashSystem 5045
IBM’s Opportunity for Generative AI in the Enterprise
Author Information
Keith has over 25 years of experience in research, marketing, and consulting-based fields.
He has authored in-depth reports and market forecast studies covering artificial intelligence, biometrics, data analytics, robotics, high performance computing, and quantum computing, with a specific focus on the use of these technologies within large enterprise organizations and SMBs. He has also established strong working relationships with the international technology vendor community and is a frequent speaker at industry conferences and events.
In his career as a financial and technology journalist he has written for national and trade publications, including BusinessWeek, CNBC.com, Investment Dealers’ Digest, The Red Herring, The Communications of the ACM, and Mobile Computing & Communications, among others.
He is a member of the Association of Independent Information Professionals (AIIP).
Keith holds dual Bachelor of Arts degrees in Magazine Journalism and Sociology from Syracuse University.