Fundamentals of Accelerated Computing with CUDA C/C++

Europe/Berlin
Description

Together with Nvidia, we are offering a workshop on accelerated computing with CUDA in C/C++.

The Workshop will take place virtually. Participants will be provided with access to a suitable GPU setup.

Connection details for the workshop as well as required steps for access to the setup will be sent to the participants prior to the event.

Learning Objectives

By participating in this workshop, you’ll:

  • Write code to be executed by a GPU accelerator
  • Expose and express data and instruction-level parallelism in C/C++ applications using CUDA
  • Utilize CUDA-managed memory and optimize memory migration using asynchronous prefetching
  • Leverage command-line and visual profilers to guide your work
  • Utilize concurrent streams for instruction-level parallelism
  • Write GPU-accelerated CUDA C/C++ applications, or refactor existing CPU-only applications, using a profile-driven approach

Prerequisites

  • Basic C/C++ competency, including familiarity with variable types, loops, conditional statements, functions, and array manipulations
  • No previous knowledge of CUDA programming is assumed

 

This workshop is part of the Nvidia Deep Learning Institute, further details can be found at https://www.nvidia.com/en-us/training/instructor-led-workshops/fundamentals-of-accelerated-computing-with-cuda/

 

Registration
Registration
48 / 48
Surveys
Participants Feedback
    • 1
      Introduction

      Meet the instructor.
      Create an account at courses.nvidia.com/join

    • 2
      Accelerating Applications with CUDA C/C++

      Learn the essential syntax and concepts to be able to write GPU-enabled C/C++ applications with CUDA:

      Write, compile, and run GPU code.
      Control parallel thread hierarchy.
      Allocate and free memory for the GPU.

    • 11:15 AM
      Break
    • 3
      Managing Accelerated Application Memory with CUDA C/C++

      Learn the command-line profiler and CUDA-managed memory, focusing on observation-driven application improvements and a deep understanding of managed memory behavior:

      Profile CUDA code with the command-line profiler.
      Go deep on unified memory.
      Optimize unified memory management.

    • 2:15 PM
      Break
    • 4
      Asynchronous Streaming and Visual Profiling for Accelerated Applications with CUDA C/C++

      Identify opportunities for improved memory management and instruction-level parallelism:

      Profile CUDA code with the NVIDIA Visual Profiler.
      Use concurrent CUDA streams.

    • 5
      Final Review

      Review key learnings and wrap up questions.
      Complete the assessment to earn a certificate.
      Take the workshop survey.