Menu

NVIDIA GTC 2021: Cloudera and NVIDIA Expand Partnership, Look to RAPIDly Advance Data Scientist Adoption of GPUs

The News: Cloudera Machine Learning and NVIDIA are now offering the RAPIDS Edition Machine Learning (ML) Runtime. ML Runtimes are developed to be secure, customizable, and containerized working environments. The RAPIDS Edition Runtime is built on top of community-built RAPIDS docker images, potentially enabling data scientists to get up and running on GPUs with the single click of a button, with all the resources and libraries they need. Cloudera Machine Learning (CML) is one of the Data Services available in the Cloudera Data Platform.

CML is developed to offer the functionality needed from a data science platform, including scalable compute resources and access to preferred tools, along with the capability of being managed, governed, and secured by Cloudera’s Shared Data Experience, or SDX. NVIDIA RAPIDS is a suite of software libraries that enables users to run data science workflows entirely on GPUs. RAPIDS relies on NVIDIA CUDA primitives for low-level compute optimization and exposes performance gains through user-friendly Python interfaces. Read more on the Cloudera Blog here.

NVIDIA GTC 2021: Cloudera and NVIDIA Expand Partnership, Look to RAPIDly Advance Data Scientist Adoption of GPUs

Analyst Take: Cloudera and NVIDIA continue to strengthen their alliance with Cloudera unveiling tighter integration of Cloudera Data Platform components with NVIDIA GPUs ahead of the NVIDIA GTC 2021 event. The alliance is now taking the initiative to remove key barriers to broader data scientist adoption of GPUs for workloads beyond deep learning. The main barriers identified include the time needed to configure an environment with GPUs and the time required to refactor CPU code.

Already data science requires proficiency in programming, math, communication, statistics, and specialty knowledge. As such, data scientists look to avoid learning an assemblage of new libraries as well as taking on the major task of learning a new programming language. To address this challenge, RAPIDS supports Python interfaces including NVIDIA RAPIDS.ai libraries, which offer near-identical syntax replicas of popular CPU-based Python data science libraries such as Pandas and Scikit-Learn that are designed enable data scientists to instead run with GPU based Python libraries such as cuDF for dataframes and cuML for ML.

I view the augmented NVIDIA RADIDS.ai libraries as critical to streamlining and assuring ease of use for data scientists to accelerate the configuration of any large data workload environment using GPUs, as well as refactoring CPU code. For instance, Cloudera is asserting that runtimes can decrease up to 98% using cuDF and cuML code.

Additionally, data science productivity can be increased through the SDX-enabled visibility, security, and governance capabilities applied to the entire data science lifecycle. Data scientists can use CDP tooling to potentially accelerate ML, agile experimentation, and advanced analytics applications by large differentials at lower costs with GPU parallelization. As such, I believe Cloudera and NVIDIA can deliver a definitive competitive advantage in boosting time-to-insight and return on investment over CPU-only computing architectures, especially when scaling out for large processes and large cloud data workloads.

Key Takeaways on NVIDIA and Cloudera Expanded CML and RAPIDS Collaboration

I anticipate that the new Cloudera NVIDIA partnership push can broaden data science community consideration and adoption of running more data science pipelines on NVIDIA GPU infrastructure to improve their data-driven operations, especially for massive cloud data workloads. Through CDP, data scientists can accelerate their toolchains with only minimal or modest code changes required. Together, CML and RAPIDS can diminish the impediments that have kept both new and experienced data scientists from using GPUs, thus potentially improving their code performance without requiring additional training in new languages, libraries, or frameworks.

Disclosure: Futurum Research is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Other insights from Futurum Research:

NVIDIA Continues its Record Breaking Run With a Huge Q2 Result

Cloudera Will Go Private in a Private Equity Deal Valued at $5.3 Billion

NVIDIA Receives Public Vote of Confidence on its $40 Billion Arm Deal

Image Credit: Mar Tech Series

Author Information

Ron is an experienced, customer-focused research expert and analyst, with over 20 years of experience in the digital and IT transformation markets, working with businesses to drive consistent revenue and sales growth.

Ron holds a Master of Arts in Public Policy from University of Nevada — Las Vegas and a Bachelor of Arts in political science/government from William and Mary.

Related Insights
Is SaaS Facing a Threat from AI Automation
February 10, 2026

Is SaaS Facing a Threat from AI Automation?

Keith Kirkpatrick, VP & Research Director at Futurum, shares his insights around the recent selloff of major SaaS vendors, fueled by fears of an AI-native software takeover, and provides his...
Qualtrics Names Jason Maynard CEO; Can Execution Match AI Ambition
February 10, 2026

Qualtrics Names Jason Maynard CEO; Can Execution Match AI Ambition?

Keith Kirkpatrick, VP & Research Director at The Futurum Group, examines Qualtrics’ appointment of Jason Maynard as CEO and what tighter execution, AI adoption, and product focus could mean for...
Truth or Dare What Can Claude Agent Teams And Developers Create Today
February 10, 2026

Truth or Dare: What Can Claude Agent Teams And Developers Create Today?

Mitch Ashley, VP and Practice Lead, Software Lifecycle Engineering at Futurum, examines what Anthropic’s Claude agent teams reveal about AI-driven software development today, separating experience reports from externally credible capability...
Google Adds Deeper Context and Control for Agentic Developer Workflows
February 10, 2026

Google Adds Deeper Context and Control for Agentic Developer Workflows

Mitch Ashley, VP and Practice Lead, Software Lifecycle Engineering at Futurum, examines how Google’s Developer Knowledge API and Gemini CLI hooks externalize agent context and governance, shaping production-ready AI development...
Texas Instruments Buys Silicon Labs To Fuel Edge AI Scale
February 10, 2026

Texas Instruments Buys Silicon Labs To Fuel Edge AI Scale

Brendan Burke, Research Director at Futurum, examines Texas Instruments’ acquisition of Silicon Labs, assessing how manufacturing integration, portfolio scale, and cost discipline could reshape embedded wireless connectivity markets....
OpenAI Frontier Close the Enterprise AI Opportunity Gap—or Widen It
February 9, 2026

OpenAI Frontier: Close the Enterprise AI Opportunity Gap—or Widen It?

Futurum Research Analysts Mitch Ashley, Keith Kirkpatrick, Fernando Montenegro, Nick Patience, and Brad Shimmin examine OpenAI Frontier and whether enterprise AI agents can finally move from pilots to production. The...

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.