What is HIP?#

HIP is a C++ runtime API and kernel language for AMD GPUs. It is part of AMD’s ROCm platform and lets developers create applications that run on heterogeneous systems, using CPUs and AMD GPUs from a single source code base.

HIP in an application.
  • HIP is a thin API with little or no performance impact over coding directly in AMD ROCm.

  • HIP enables coding in a single-source C++ programming language, including features such as templates, C++11 lambdas, classes, namespaces, and more.

  • Developers can tune for performance or handle tricky cases via HIP.

ROCm offers compilers (clang, hipcc), code profilers (rocprofv3), debugging tools (rocgdb), libraries and HIP with the runtime API and kernel language, to create heterogeneous applications running on both CPUs and GPUs. ROCm provides libraries like hipFFT and hipBLAS that provide API compatibility with equivalent NVIDIA CUDA libraries, making it easier to port existing NVIDIA CUDA applications. These libraries provide pointer-based memory interfaces and can be easily integrated into your applications.

GPU programmers with NVIDIA CUDA experience will find the HIP API straightforward. You can quickly port NVIDIA CUDA applications to run on AMD GPUs. The HIPify tools, based on the clang front-end and Perl language, can convert NVIDIA CUDA API calls into the corresponding HIP API calls. However, HIP is not intended to be a drop-in replacement for NVIDIA CUDA, and developers should expect to do some manual coding and performance tuning work to port existing projects to AMD GPUs as described in the HIP porting guide.

HIP provides two components: those that run on the CPU, also known as host system, and those that run on GPUs, also referred to as device. The host-based code is used to create device buffers, move data between the host application and a device, launch the device code (also known as kernel), manage streams and events, and perform synchronization. The kernel language provides a way to develop massively parallel programs that run on GPUs, and provides access to GPU specific hardware capabilities.

In summary, HIP simplifies porting NVIDIA CUDA applications to AMD GPUs, maintains performance, and provides a familiar C++ experience for GPU programming.

HIP components#

HIP consists of the following components. For information on the license associated with each component, see HIP licensing.

C++ runtime API#

HIP provides headers and a runtime library built on top of HIP-Clang compiler in the repository Compute Language Runtime (CLR). The HIP runtime implements HIP streams, events, and memory APIs, and is an object library that is linked with the application. The source code for all headers and the library implementation is available on GitHub.

For further details, check HIP Runtime API Reference.

Kernel language#

HIP provides a C++ syntax that is suitable for compiling most code that commonly appears in compute kernels (classes, namespaces, operator overloading, and templates). HIP also defines other language features that are designed to target accelerators, such as:

  • Short-vector headers that can serve on a host or device

  • Math functions that resemble those in math.h, which is included with standard C++ compilers

  • Built-in functions for accessing specific GPU hardware capabilities

For further details, check HIP C++ language extensions and Kernel language C++ support.