Menu

NVIDIA GTC 2021: Cloudera and NVIDIA Expand Partnership, Look to RAPIDly Advance Data Scientist Adoption of GPUs

The News: Cloudera Machine Learning and NVIDIA are now offering the RAPIDS Edition Machine Learning (ML) Runtime. ML Runtimes are developed to be secure, customizable, and containerized working environments. The RAPIDS Edition Runtime is built on top of community-built RAPIDS docker images, potentially enabling data scientists to get up and running on GPUs with the single click of a button, with all the resources and libraries they need. Cloudera Machine Learning (CML) is one of the Data Services available in the Cloudera Data Platform.

CML is developed to offer the functionality needed from a data science platform, including scalable compute resources and access to preferred tools, along with the capability of being managed, governed, and secured by Cloudera’s Shared Data Experience, or SDX. NVIDIA RAPIDS is a suite of software libraries that enables users to run data science workflows entirely on GPUs. RAPIDS relies on NVIDIA CUDA primitives for low-level compute optimization and exposes performance gains through user-friendly Python interfaces. Read more on the Cloudera Blog here.

NVIDIA GTC 2021: Cloudera and NVIDIA Expand Partnership, Look to RAPIDly Advance Data Scientist Adoption of GPUs

Analyst Take: Cloudera and NVIDIA continue to strengthen their alliance with Cloudera unveiling tighter integration of Cloudera Data Platform components with NVIDIA GPUs ahead of the NVIDIA GTC 2021 event. The alliance is now taking the initiative to remove key barriers to broader data scientist adoption of GPUs for workloads beyond deep learning. The main barriers identified include the time needed to configure an environment with GPUs and the time required to refactor CPU code.

Already data science requires proficiency in programming, math, communication, statistics, and specialty knowledge. As such, data scientists look to avoid learning an assemblage of new libraries as well as taking on the major task of learning a new programming language. To address this challenge, RAPIDS supports Python interfaces including NVIDIA RAPIDS.ai libraries, which offer near-identical syntax replicas of popular CPU-based Python data science libraries such as Pandas and Scikit-Learn that are designed enable data scientists to instead run with GPU based Python libraries such as cuDF for dataframes and cuML for ML.

I view the augmented NVIDIA RADIDS.ai libraries as critical to streamlining and assuring ease of use for data scientists to accelerate the configuration of any large data workload environment using GPUs, as well as refactoring CPU code. For instance, Cloudera is asserting that runtimes can decrease up to 98% using cuDF and cuML code.

Additionally, data science productivity can be increased through the SDX-enabled visibility, security, and governance capabilities applied to the entire data science lifecycle. Data scientists can use CDP tooling to potentially accelerate ML, agile experimentation, and advanced analytics applications by large differentials at lower costs with GPU parallelization. As such, I believe Cloudera and NVIDIA can deliver a definitive competitive advantage in boosting time-to-insight and return on investment over CPU-only computing architectures, especially when scaling out for large processes and large cloud data workloads.

Key Takeaways on NVIDIA and Cloudera Expanded CML and RAPIDS Collaboration

I anticipate that the new Cloudera NVIDIA partnership push can broaden data science community consideration and adoption of running more data science pipelines on NVIDIA GPU infrastructure to improve their data-driven operations, especially for massive cloud data workloads. Through CDP, data scientists can accelerate their toolchains with only minimal or modest code changes required. Together, CML and RAPIDS can diminish the impediments that have kept both new and experienced data scientists from using GPUs, thus potentially improving their code performance without requiring additional training in new languages, libraries, or frameworks.

Disclosure: Futurum Research is a research and advisory firm that engages or has engaged in research, analysis, and advisory services with many technology companies, including those mentioned in this article. The author does not hold any equity positions with any company mentioned in this article.

Other insights from Futurum Research:

NVIDIA Continues its Record Breaking Run With a Huge Q2 Result

Cloudera Will Go Private in a Private Equity Deal Valued at $5.3 Billion

NVIDIA Receives Public Vote of Confidence on its $40 Billion Arm Deal

Image Credit: Mar Tech Series

Author Information

Ron is an experienced, customer-focused research expert and analyst, with over 20 years of experience in the digital and IT transformation markets, working with businesses to drive consistent revenue and sales growth.

Ron holds a Master of Arts in Public Policy from University of Nevada — Las Vegas and a Bachelor of Arts in political science/government from William and Mary.

Related Insights
March 11, 2026

AI Accelerators – Futurum Signal

The rapid acceleration of artificial intelligence is fundamentally reshaping the semiconductor and data center landscape. In our latest Futurum Signal Report: AI Accelerators, we examine how a new generation of...
OpenAI Acquires Promptfoo, Gaining 25% Foothold in Fortune 500 Enterprises
March 11, 2026

OpenAI Acquires Promptfoo, Gaining 25% Foothold in Fortune 500 Enterprises

Mitch Ashley, VP Practice Lead at Futurum, examines OpenAI's acquisition of Promptfoo and what it signals about the security and governance requirements blocking AI agents from enterprise production....
HPE Q1 FY 2026 Results Show Networking Strength, AI Backlog, and Higher Outlook
March 11, 2026

HPE Q1 FY 2026 Results Show Networking Strength, AI Backlog, and Higher Outlook

Futurum Research analyzes HPE’s Q1 FY 2026 earnings, focusing on networking-for-AI demand, memory-driven supply constraints, Juniper integration progress, and what the updated outlook implies for FY 2026 execution....
Can Applied Materials and Micron Crack the Materials Barrier Holding Back HBM?
March 11, 2026

Can Applied Materials and Micron Crack the Materials Barrier Holding Back HBM?

Brendan Burke, Research Director at Futurum, shares insights on the Applied Materials and Micron partnership to co-develop advanced materials engineering for next-generation memory scaling, and what it means for the...
Claude Marketplace Tests Whether Anthropic Can Win the Procurement Heart
March 11, 2026

Claude Marketplace Tests Whether Anthropic Can Win the Procurement Heart

Alex Smith, VP and Practice Lead at Futurum examines Anthropic’s Claude Marketplace and how commitment-based procurement and partner apps shift enterprise AI buying toward consolidated spend and workflow-specific tools....
CData Trades in Vibe-Coding for Industrial-Grade Enterprise AI Infrastructure
March 10, 2026

CData Trades in Vibe-Coding for Industrial-Grade Enterprise AI Infrastructure

Brad Shimmin, Vice President at Futurum, analyzes CData’s March 9th pivot toward enterprise AI infrastructure. CData is tackling the AI "accuracy gap" with a validated 98.5% query success rate and...

Book a Demo

Newsletter Sign-up Form

Get important insights straight to your inbox, receive first looks at eBooks, exclusive event invitations, custom content, and more. We promise not to spam you or sell your name to anyone. You can always unsubscribe at any time.

All fields are required






Thank you, we received your request, a member of our team will be in contact with you.