Current Status
Not Enrolled
Price
Free
Get Started

What you will learn

  • Understand the differences between processing data using CPU and GPU
  • Use cuDF as a replacement for pandas for GPU-accelerated processing
  • Implement codes using cuDF to manipulate DataFrames
  • Use cuPy as a replacement for numpy for GPU-accelerated processing
  • Use cuML as a replacement for scikit-learn for GPU-accelerated processing
  • Implement a complete machine learning project using cuDF and cuML
  • Compare the performance of classic Python libraries that run on the CPU with RAPIDS libraries that run on the GPU
  • Implement projects with DASK for parallel and distributed processing
  • Integrate DASK with cuDF and cuML for GPU performance

Requirements

  • Programming logic
  • Basic Python programming
  • Machine learning: basic understanding of the algorithm training process, as well as classification and regression techniques

Description

Data science and machine learning represent the largest computational sectors in the world, where modest improvements in the accuracy of analytical models can translate into billions of impact on the bottom line. Data scientists are constantly striving to train, evaluate, iterate, and optimize models to achieve highly accurate results and exceptional performance. With NVIDIA’s powerful RAPIDS platform, what used to take days can now be accomplished in a matter of minutes, making the construction and deployment of high-value models easier and more agile. In data science, additional computational power means faster and more effective insights. RAPIDS harnesses the power of NVIDIA CUDA to accelerate the entire data science model training workflow, running it on graphics processing units (GPUs).

In this course, you will learn everything you need to take your machine learning applications to the next level! Check out some of the topics that will be covered below:

  • Utilizing the cuDF, cuPy, and cuML libraries instead of Pandas, Numpy, and scikit-learn; ensuring that data is processed and machine learning algorithms are executed with high performance on the GPU.
  • Comparing the performance of classic Python libraries with RAPIDS. In some experiments conducted during the classes, we achieved acceleration rates exceeding 900x. This indicates that with certain databases and algorithms, RAPIDS can be 900 times faster!
  • Creating a complete, step-by-step machine learning project using RAPIDS, from data loading to predictions.
  • Using DASK for task parallelism on multiple GPUs or CPUs; integrated with RAPIDS for superior performance.

Throughout the course, we will use the Python programming language and the online Google Colab. This way, you don’t need to have a local GPU to follow the classes, as we will use the free hardware provided by Google.

Who this course is for

  • Data scientists and artificial intelligence professionals looking to enhance the performance of their applications
  • Professionals currently working or aspiring to work in the field of data science, particularly those seeking to improve their skills in machine learning model training and data analysis
  • Anyone interested in learning about machine learning, especially with a focus on high-performance implementations using GPUs
  • Professionals involved in the development and implementation of machine learning models
  • Undergraduate and graduate students studying subjects related to artificial intelligence

Ratings and Reviews

5.0
Avg. Rating
5 Ratings
5
5
4
0
3
0
2
0
1
0
What's your experience? We'd love to know!
Review posted on Udemy
Posted 1 month ago
by Asha Reilly

This is thoroughly explained as it is a beginner's course. To grasp concepts and techniques, the instructor pushes you to the next level.

×
Preview Image
Review posted on Udemy
Posted 1 month ago
by Jessica Howe

Excellent course! Many specific details. Also, the tutorial is very easy to understand. I appreciate your efforts!

×
Preview Image
Review posted on Udemy
Posted 1 month ago
by Sonja Lindholm

This is a great course and I highly recommend it. It will be extremely helpful in any AI Application Boost. The instructor is a highly knowledgeable.

×
Preview Image
Review posted on Udemy
Posted 1 month ago
by Gaspar Maheu

It was a really amazing experience, I learned a lot of new things for AI Application Boost with NVIDIA RAPIDS Acceleration. Thanks

×
Preview Image
Review posted on Udemy
Posted 1 month ago
by Miracle Mertz

Really amazing! I was able to learn all the basics on AI as well as advance techniques. It really helps to know more about AI.

×
Preview Image
Show more reviews
What's your experience? We'd love to know!
Scroll to Top