Evaluating OpenText Aviator: The Emergence of Enterprise AI Platforms

Evaluating OpenText Aviator: The Emergence of Enterprise AI Platforms

The News: OpenText launches Aviator solution set, which represents the company’s vision and direction for AI. Read the full press release on the OpenText website.

Evaluating OpenText Aviator: The Emergence of Enterprise AI Platforms

Analyst Take: Mark Barrenchea the CEO and CTO of OpenText, took the stage earlier this week to announce the company’s pivot to embracing AI across its entire portfolio with a suite of Aviator solutions. The recent announcement marks a pivotal moment in enterprise AI development and a key rallying cry for the company as it looks to cement its position in the market post the acquisition of Micro Focus.

The new releases are laden with a multitude of generative AI features aimed at operational efficiency, robust data management, and heightened security. This release is especially timely given that consensus forecasts put AI IT spending to surpass $300 billion by the middle of this decade. But as organizations grapple with rapidly evolving business landscapes, how does OpenText’s Aviator stack up against other prominent AI services such as Google Vertex AI and Amazon Web Services (AWS) Bedrock?

OpenText Aviator: An Overview

OpenText Aviator represents a comprehensive approach to enterprise AI, encompassing features that cater to a broad range of organizational needs. Aviator intends to be the central nervous system for information-driven decision-making within an organization, from IT operations to DevOps, from content management to cybersecurity. Of particular interest is the platform’s ability to overlay large language models (LLMs) on private, secured data, thus enabling a stack-based or modular approach to practical AI.

Competitive Landscape

Google Vertex AI and Amazon AWS Bedrock are significant players in the enterprise AI landscape, each with distinct advantages and limitations. Google Vertex AI prioritizes automated machine learning (autoML) capabilities and simplifies the deployment process, aiming for ease of use so that developers can focus more on fine-tuning models rather than managing infrastructure. Its feature set is expansive, supporting tasks such as model training, evaluation, and large-scale prediction, although its offerings are mainly tailored for ML practitioners. In contrast, AWS Bedrock is designed for seamless integration within the broader AWS ecosystem, providing a highly extensible set of foundational AI services that enable organizations to develop complex AI applications with relative ease. However, unlike Vertex AI, AWS Bedrock is more infrastructure-focused and generally requires a certain level of technical expertise for optimal utilization.

Incorporating Microsoft’s association with OpenAI provides a third vantage point to the enterprise AI landscape, offering an intriguing counterbalance to Vertex AI and AWS Bedrock. Microsoft has made significant investments in OpenAI and has been a key partner in deploying models such as GPT-3, which leverages deep learning techniques to create one of the most sophisticated natural language models to date. Similar to Google Vertex, Microsoft’s involvement with OpenAI shows a keen interest in democratizing AI and ML capabilities. However, what sets this partnership apart is the focus on LLMs that can process and generate human-like text based on the data they are trained on, thereby making them highly applicable for tasks such as conversational agents, automated content generation, and complex data analysis.

Although Vertex AI excels in providing a streamlined ML workflow and AWS Bedrock shines in its extensible foundational services within its own ecosystem, Microsoft’s affiliation with OpenAI potentially bridges the gap between technical AI capabilities and real-world applications. The focus here is on a model that does not just execute tasks based on algorithms but also “understands” and “generates” human-like text, making it more versatile in addressing business problems that require a level of semantic understanding.

Moreover, the OpenAI-Microsoft alliance benefits from Microsoft’s well-established enterprise software ecosystem, which includes products such as Azure, Microsoft 365, and Dynamics 365. This collaboration could offer a more seamless integration experience for organizations already invested in Microsoft’s suite of services, thus lowering the barrier to entry.

In summary, while Google Vertex AI is engineered for ML practitioners focusing on autoML and deployment efficiency, and AWS Bedrock is designed for infrastructure-centric, technically skilled professionals, Microsoft’s relationship with OpenAI represents a synthesis of technical prowess with a keen eye toward versatile, real-world applications. The collaboration aims to merge cutting-edge AI capabilities with the demands of today’s business challenges, within an established enterprise ecosystem.

Key areas of differentiation include:

  • Flexibility and Adaptability: OpenText Aviator offers a wide range of use-case scenarios from IT operations to content management, thus showing higher adaptability for diverse business needs. Google Vertex AI and AWS Bedrock are generally more specialized, focusing mainly on ML and AI infrastructure.
  • Data Security: OpenText appears to strongly emphasize data security, enabling organizations to keep their data in house. This approach contrasts with Google Vertex AI and AWS Bedrock, where cloud-native solutions could potentially raise data sovereignty and privacy concerns.
  • Integration with Existing Systems: AWS Bedrock and Google Vertex AI require organizations to adapt to their respective ecosystems, potentially leading to compatibility issues with existing systems. OpenText Aviator, however, focuses on seamless integration across multiple clouds and knowledge bases.
  • User Experience (UX): Both Vertex AI and AWS Bedrock cater to a more technically skilled audience. With its intuitive chat interfaces and natural language queries, OpenText Aviator is designed for end users, facilitating a more user-friendly experience.
  • Cost-Efficiency: Although direct cost comparisons are difficult due to varying pricing models, OpenText’s focus on enhancing operational efficiency and reducing service management costs could benefit enterprises in the long term.

Looking Ahead

OpenText’s Aviator appears poised to offer a full-stack, versatile AI solution with a unique emphasis on data security and broad adaptability. Although Google Vertex AI and AWS Bedrock also offer compelling solutions, they are often more specialized and could require a certain level of technical expertise for full utility. Organizations considering a foray into enterprise AI would do well to comprehensively evaluate their specific needs, available skill sets, and long-term strategic objectives before choosing a platform. OpenText Aviator, with its user-centric design and focus on security, could be a strong contender for those looking for an all-encompassing enterprise AI solution.

Disclosure: The Futurum Group is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Analysis and opinions expressed herein are specific to the analyst individually and data and other information that might have been provided for validation, not those of The Futurum Group as a whole.

Other insights from The Futurum Group:

OpenText Uses its OpenText World EMEA Event to Break Cover

OpenText Aviator Delivers Generative AI Use Cases Beyond CX

OpenText Reports Strong Q4 and FY 2023 Earnings, Driven by Cloud and ARR Growth

Author Information

Steven engages with the world’s largest technology brands to explore new operating models and how they drive innovation and competitive edge.

Related Insights
Is Anthropic’s $100 Billion Pact for AWS Silicon a Bargain in a Supply-Constrained Market?
April 23, 2026

Is Anthropic’s $100 Billion Pact for AWS Silicon a Bargain in a Supply-Constrained Market?

Brendan Burke, Research Director at Futurum, examines how Anthropic's $100 billion decade-long commitment to AWS Trainium and Graviton reshapes frontier AI infrastructure economics and supply dynamics....
ChatGPT Images 2.0 Raises the Stakes in Enterprise AI—But Will Reliability Keep Pace?
April 23, 2026

ChatGPT Images 2.0 Raises the Stakes in Enterprise AI—But Will Reliability Keep Pace?

OpenAI's ChatGPT Images 2.0 intensifies competition with Microsoft and Google, but enterprise adoption hinges on reliability. Futurum Group's Decision Maker Survey reveals 55% cite AI agent hallucination management as the...
Qodo Hands PR-Agent to the Community: Will Open Governance Accelerate AI Code Review?
April 23, 2026

Qodo Hands PR-Agent to the Community: Will Open Governance Accelerate AI Code Review?

Qodo's transfer of PR-Agent to community ownership marks a pivotal test for open-source AI against proprietary competitors demanding transparency and rapid innovation....
Qualcomm’s Snapdragon Wear Elite Redefines the AI Wearable Stakes—But Who Wins the Wrist War?
April 22, 2026

Qualcomm’s Snapdragon Wear Elite Redefines the AI Wearable Stakes—But Who Wins the Wrist War?

Qualcomm's Snapdragon Wear Elite marks a turning point in wearable AI, delivering a dedicated neural processing unit for on-device intelligence, privacy, and real-time voice interactions—positioning the company against Apple and...
VAST Data Valuation Triples. Can a Unified Platform Scale AI Globally?
April 22, 2026

VAST Data Valuation Triples. Can a Unified Platform Scale AI Globally?

Brad Shimmin, Vice President & Practice Lead at Futurum, analyzes VAST Data valuation and its AI operating system strategy, questioning whether unified infrastructure can scale amid persistent market fragmentation....
Cerebras S-1 Teardown: Is the $23B Wafer-Scale IPO the End of GPU Homogeneity?
April 22, 2026

Cerebras S-1 Teardown: Is the $23B Wafer-Scale IPO the End of GPU Homogeneity?

Brendan Burke, Research Director at Futurum, examines Cerebras Systems' S-1 filing and $23B valuation, dissecting the $20B OpenAI deal, 86% UAE revenue concentration, and whether wafer-scale silicon can survive the...

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.