Introduction to Parallel Programming with CUDA

This course is part of GPU Programming Specialization

Instructor: Chancellor Thomas Pascale

What you'll learn

  •   Students will learn how to utilize the CUDA framework to write C/C++ software that runs on CPUs and Nvidia GPUs.
  •   Students will transform sequential CPU algorithms and programs into CUDA kernels that execute 100s to 1000s of times simultaneously on GPU hardware.
  • Skills you'll gain

  •   Algorithms
  •   Data Structures
  •   Distributed Computing
  •   Performance Tuning
  •   C (Programming Language)
  •   OS Process Management
  •   Hardware Architecture
  •   Debugging
  •   C++ (Programming Language)
  •   Computer Architecture
  •   Development Environment
  •   Performance Testing
  •   Program Development
  •   Data Storage
  •   System Programming
  •   Data Access
  • There are 5 modules in this course

    This course will help prepare students for developing code that can process large amounts of data in parallel on Graphics Processing Units (GPUs). It will learn on how to implement software that can solve complex problems with the leading consumer to enterprise-grade GPUs available using Nvidia CUDA. They will focus on the hardware and software capabilities, including the use of 100s to 1000s of threads and various forms of memory.

    Threads, Blocks and Grids

    Host and Global Memory

    Shared and Constant Memory

    Register Memory

    Explore more from Software Development

    ©2025  ementorhub.com. All rights reserved