Menu

NVIDIA GTC 2021: Cloudera and NVIDIA Expand Partnership, Look to RAPIDly Advance Data Scientist Adoption of GPUs

The News: Cloudera Machine Learning and NVIDIA are now offering the RAPIDS Edition Machine Learning (ML) Runtime. ML Runtimes are developed to be secure, customizable, and containerized working environments. The RAPIDS Edition Runtime is built on top of community-built RAPIDS docker images, potentially enabling data scientists to get up and running on GPUs with the single click of a button, with all the resources and libraries they need. Cloudera Machine Learning (CML) is one of the Data Services available in the Cloudera Data Platform.

CML is developed to offer the functionality needed from a data science platform, including scalable compute resources and access to preferred tools, along with the capability of being managed, governed, and secured by Cloudera’s Shared Data Experience, or SDX. NVIDIA RAPIDS is a suite of software libraries that enables users to run data science workflows entirely on GPUs. RAPIDS relies on NVIDIA CUDA primitives for low-level compute optimization and exposes performance gains through user-friendly Python interfaces. Read more on the Cloudera Blog here.

NVIDIA GTC 2021: Cloudera and NVIDIA Expand Partnership, Look to RAPIDly Advance Data Scientist Adoption of GPUs

Analyst Take: Cloudera and NVIDIA continue to strengthen their alliance with Cloudera unveiling tighter integration of Cloudera Data Platform components with NVIDIA GPUs ahead of the NVIDIA GTC 2021 event. The alliance is now taking the initiative to remove key barriers to broader data scientist adoption of GPUs for workloads beyond deep learning. The main barriers identified include the time needed to configure an environment with GPUs and the time required to refactor CPU code.

Already data science requires proficiency in programming, math, communication, statistics, and specialty knowledge. As such, data scientists look to avoid learning an assemblage of new libraries as well as taking on the major task of learning a new programming language. To address this challenge, RAPIDS supports Python interfaces including NVIDIA RAPIDS.ai libraries, which offer near-identical syntax replicas of popular CPU-based Python data science libraries such as Pandas and Scikit-Learn that are designed enable data scientists to instead run with GPU based Python libraries such as cuDF for dataframes and cuML for ML.

I view the augmented NVIDIA RADIDS.ai libraries as critical to streamlining and assuring ease of use for data scientists to accelerate the configuration of any large data workload environment using GPUs, as well as refactoring CPU code. For instance, Cloudera is asserting that runtimes can decrease up to 98% using cuDF and cuML code.

Additionally, data science productivity can be increased through the SDX-enabled visibility, security, and governance capabilities applied to the entire data science lifecycle. Data scientists can use CDP tooling to potentially accelerate ML, agile experimentation, and advanced analytics applications by large differentials at lower costs with GPU parallelization. As such, I believe Cloudera and NVIDIA can deliver a definitive competitive advantage in boosting time-to-insight and return on investment over CPU-only computing architectures, especially when scaling out for large processes and large cloud data workloads.

Key Takeaways on NVIDIA and Cloudera Expanded CML and RAPIDS Collaboration

I anticipate that the new Cloudera NVIDIA partnership push can broaden data science community consideration and adoption of running more data science pipelines on NVIDIA GPU infrastructure to improve their data-driven operations, especially for massive cloud data workloads. Through CDP, data scientists can accelerate their toolchains with only minimal or modest code changes required. Together, CML and RAPIDS can diminish the impediments that have kept both new and experienced data scientists from using GPUs, thus potentially improving their code performance without requiring additional training in new languages, libraries, or frameworks.

Disclosure: Futurum Research is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Other insights from Futurum Research:

NVIDIA Continues its Record Breaking Run With a Huge Q2 Result

Cloudera Will Go Private in a Private Equity Deal Valued at $5.3 Billion

NVIDIA Receives Public Vote of Confidence on its $40 Billion Arm Deal

Image Credit: Mar Tech Series

Author Information

Ron is an experienced, customer-focused research expert and analyst, with over 20 years of experience in the digital and IT transformation markets, working with businesses to drive consistent revenue and sales growth.

Ron holds a Master of Arts in Public Policy from University of Nevada — Las Vegas and a Bachelor of Arts in political science/government from William and Mary.

Related Insights
Yann LeCun’s AMI Raises $1BN Seed Round - Is the World Model Era Finally Here
March 13, 2026

Yann LeCun’s AMI Raises $1BN Seed Round – Is the World Model Era Finally Here?

Nick Patience, VP & AI Platforms Practice Lead at Futurum, examines AMI Labs' $1.03B seed round - Europe's largest - and what it means for the world model era, sovereign...
Domo Q4 FY 2026 Earnings Show Record Billings And Profitability Gains
March 13, 2026

Domo Q4 FY 2026 Earnings Show Record Billings And Profitability Gains

Brad Shimmin, Vice President & Practice Lead Futurum, analyzes Domo’s Q4 FY 2026 results, focusing on record billings, improving retention, and AI-led workflow automation strategy as the company pushes consumption...
Nebius Designs the Agentic Era of AI Cloud Platforms with NVIDIA Investment
March 13, 2026

Nebius Designs the Agentic Era of AI Cloud Platforms with NVIDIA Investment

Brendan Burke, Research Director at Futurum, examines NVIDIA’s $2 billion investment in Nebius and its implications for AI cloud infrastructure, including the push toward AI factories, inference platforms, and large-scale...
Synopsys Converge – Is the New Synopsys Ready to Own Multi-Physics Design
March 13, 2026

Synopsys Converge – Is the New Synopsys Ready to Own Multi-Physics Design?

Brendan Burke, Research Director at Futurum, shares his insights on Synopsys Converge 2026, where the "New Synopsys" unveiled Multiphysics-Fusion, AgentEngineer, and Ansys 2026 R1, among other announcements aimed at the...
Oracle Q3 FY 2026 Earnings Driven by OCI AI Infrastructure Demand
March 13, 2026

Oracle Q3 FY 2026 Earnings Driven by OCI AI Infrastructure Demand

Futurum Research analyzes Oracle’s Q3 FY 2026 earnings, focusing on OCI AI infrastructure momentum, sovereign cloud positioning, and Fusion’s embedded AI agents as the company scales capacity and backlog....
Enterprise Connect 2026 - How Will AI’s Emergence Impact CCaaS Vendors
March 13, 2026

Enterprise Connect 2026 — How Will AI’s Emergence Impact CCaaS Vendors?

Keith Kirkpatrick, VP & Research Director with Futurum, covers Enterprise Connect 2026, and shares his insights on the evolution of the CX market and its impact on CCaaS vendors....

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.