With Growth of In-Production Kubernetes, Container Platform Providers Add New Capabilities

In-Production Kubernetes Use Expanding

Kubernetes-managed containers have escaped the confines of application development and pilot projects and are expanding rapidly in full production.  EG’s recent survey of container management revealed that Kubernetes is in production at over 50% of surveyed customers, with 60% of the customers running more than five workloads (applications), and 70% already running multiple clusters.   Within a year, 55% of customers expect to be running six or more clusters, driving higher requirements in global observability and management tools…. and challenges for IT Operations Executives.

Some customers are avoiding the IT Operations challenge by leveraging managed container services from providers like AWS, GCP, Azure and VMware.  But for customers seeking more flexibility, Container Management Platform providers are reacting to this race to productization with a range of new capabilities to support self-management of production scale out.  Recent announcements by Red Hat and D2IQ announcement offer some key examples.

Vendors Address the Scale Issue

To address rapidly scaling workloads, Red Hat (in Open Shift 4.11) has recently made autoscaling of application workloads easier and more flexible. Customers can now define custom metrics for scaling – leveraging an Open Shift automated operator to scale their application workloads without need for operator intervention.

D2IQ’s enhancements are focused on scaling out clusters.  DKP now includes what D2IQ calls “federated application management” for fleets of clusters.   Application lifecycle management – the ability to define the desired state of an application and then manage updates, revisions etc. over time – is a necessity for container management.  The ability to manage application updates and configuration across a multi-cluster environment (including large fleets) has become a must-have feature for any customer intending to scale-out.  DKP’s latest enhancement takes the capability to a new level, with the ability for the customer to define a desired state for a configuration of multiple applications running together.  DKP then performs the application lifecycle management as a single integrated activity, eliminating the need for the customer to define and manage the desired state of each application individually.  Customers using DKP 3.4 can define and execute application lifecycle management across multi-cluster (and multi-cloud) environments. This is a particularly valuable capability for customers deploying large numbers of clusters with similar container contents (e.g., an edge-based application), and/or where rolling updates are required to manage certain clusters as a group.

The company also announced a significant new capability in global observability.  As customer production environments scale (in both volume and complexity), administrators need automation to help maintain availability and performance.   DKP Insights, a predictive analytics tool that detects anomalies in Kubernetes clusters or workload configurations, has historically offered global observability, running an analytics engine on each Kubernetes cluster, and then accumulating information for reporting, alerting, and troubleshooting on the DKP management console.  But with this latest announcement, Insights also automatically checks workload configurations against preset definitions of best practices and suggests improvements to the administrator.  The admin can then drill down on the anomaly to get additional information, a root cause analysis, and suggested solutions.

D2IQ product executives noted that with this management and engine model, the “suggested improvements” could be enhanced to automatically enable deployment by the management system and application by the engines, in a continuous AIOps model.  However, as customers may not be ready to adopt an operating model which is fully automated, a logical interim step would be to write and package the YAML code necessary to deploy the suggestion, allowing for a quick review and approve “push button” deployment to alleviate DevOps toil.

Enterprise customers who see multi-cluster in their future will want to keep an eye on these efforts to see what develops.   Improvements in global observability and management – including automated operations – will be required to support the massive Kubernetes environments of the future.

Related Insights
How Big A Role Will Commvault Play In Securing Agentic AI?
April 17, 2026

How Big A Role Will Commvault Play In Securing Agentic AI?

Fernando Montenegro and Brad Shimmin, VPs at Futurum, analyze Commvault's new offerings—Data Activate, AI Protect, and AI Studio—and their strategic role in securing enterprise agentic AI ecosystems against rising competition....
Can Starburst's AIDA Crack the Enterprise AI Data Access Problem?
April 17, 2026

Can Starburst’s AIDA Crack the Enterprise AI Data Access Problem?

Starburst's AIDA represents a fundamental shift in how enterprises approach AI data access. Rather than centralizing data, agentic AI systems reason across distributed sources, addressing accuracy concerns and accelerating AI...
ClickHouse Builds a CLI to Make its Databases Agent-Native
April 15, 2026

ClickHouse Builds a CLI to Make its Databases Agent-Native

Brad Shimmin, VP and Practice Lead at Futurum, explores Cloudera’s latest hybrid data platform updates. He analyzes how the Hybrid Multi-Cloud Fabric and Iceberg REST Catalog are designed to reduce...
Neo4j's Context Gap
April 14, 2026

Does Neo4j’s Context Gap Thesis Expose Enterprise AI’s Biggest Blind Spot?

Neo4j's latest analysis exposes a critical flaw in enterprise AI: the neglect of structural, relational context. Discover why graph databases are positioned as the missing memory layer for agentic AI...
Wasabi Acquires Lyve Cloud. Does This Strengthen Its Storage Position?
April 14, 2026

Wasabi Acquires Lyve Cloud. Does This Strengthen Its Storage Position?

Alex Smith and Brad Shimmin from Futurum examine the Wasabi Lyve Cloud acquisition and whether the deal meaningfully shifts cloud storage dynamics or primarily consolidates customers, capacity, and ecosystem integrations....
Oracle Redefines Mission-Critical Tiers as AI Workloads Demand Always-On Data
April 14, 2026

Oracle Redefines Mission-Critical Tiers as AI Workloads Demand Always-On Data

Brad Shimmin, Research Director at Futurum, explores Oracle's new Diamond-Grade resilience and how sub-millisecond latency and post-quantum security are redefining mission-critical data for the age of agentic AI....

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.