SAXPY - Hello, HIP#
This tutorial explains the basic concepts of the single-source Heterogeneous-computing Interface for Portability (HIP) programming model and the essential tooling around it. It also reviews some commonalities of heterogenous APIs in general. This topic assumes basic familiarity with the C/C++ compilation model and language.
Prerequisites#
To follow this tutorial, you’ll need installed drivers and a HIP compiler toolchain to compile your code. Because HIP for ROCm supports compiling and running on Linux and Windows with AMD and NVIDIA GPUs, the combination of install instructions is more than worth covering as part of this tutorial. For more information about installing HIP development packages, see Install HIP.
Heterogeneous programming#
Heterogeneous programming and offloading APIs are often mentioned together. Heterogeneous programming deals with devices of varying capabilities simultaneously. Offloading focuses on the “remote” and asynchronous aspects of computation. HIP encompasses both. It exposes GPGPU (general-purpose GPU) programming much like ordinary host-side CPU programming and lets you move data across various devices.
When programming in HIP (and other heterogenous APIs for that matter), remember that target devices are built for a specific purpose. They are designed with different tradeoffs than traditional CPUs and therefore have very different performance characteristics. Even subtle changes in code might adversely affect execution time.
Your first lines of HIP code#
First, let’s do the “Hello, World!” of GPGPU: SAXPY. Single-precision A times X Plus Y (SAXPY) is a mathematical acronym; a vector equation \(a\cdot x+y=z\) where \(a\in\mathbb{R}\) is a scalar and \(x,y,z\in\mathbb{V}\) are vector quantities of some large dimensionality. This vector space is defined over the set of reals. Practically speaking, you can compute this using a single for
loop over three arrays.
for (int i = 0 ; i < N ; ++i)
z[i] = a * x[i] + y[i];
In linear algebra libraries, such as BLAS (Basic Linear Algebra Subsystem) this operation is defined as AXPY “A times X Plus Y”. The “S” comes from single-precision, meaning that array element is float
-s (IEEE 754 binary32 representation).
To quickly get started, use the set of HIP samples from GitHub. With Git configured on your machine, open a command-line and navigate to your desired working directory, then run:
git clone https://github.com/amd/rocm-examples.git
A simple implementation of SAXPY resides in the HIP-Basic/saxpy/main.hip
file in this repository. The HIP code here mostly deals with where data has to be and when, and how devices transform this data. The first HIP calls deal with allocating device-side memory and copying data from host-side memory to device side in a C runtime-like fashion.
// Allocate and copy vectors to device memory.
float* d_x{};
float* d_y{};
HIP_CHECK(hipMalloc(&d_x, size_bytes));
HIP_CHECK(hipMalloc(&d_y, size_bytes));
HIP_CHECK(hipMemcpy(d_x, x.data(), size_bytes, hipMemcpyHostToDevice));
HIP_CHECK(hipMemcpy(d_y, y.data(), size_bytes, hipMemcpyHostToDevice));
HIP_CHECK
is a custom macro borrowed from the examples utilities which checks the error code returned by API functions for errors and reports them to the console. It is not essential to the API, but it is a good practice to check the error codes of the HIP APIs in case you pass on incorrect values to the API, or the API might be out of resources.
The code selects the device to allocate to and to copy to. Commands are issued to the HIP runtime per thread, and every thread has a device set as the target of commands. The default device is 0
, which is equivalent to calling hipSetDevice(0)
.
Launch the calculation on the device after the input data has been prepared.
__global__ void saxpy_kernel(const float a, const float* d_x, float* d_y, const unsigned int size)
{
// ...
}
int main()
{
// ...
// Launch the kernel on the default stream.
saxpy_kernel<<<dim3(grid_size), dim3(block_size), 0, hipStreamDefault>>>(a, d_x, d_y, size);
}
Analyze at the signature of the offloaded function:
__global__
instructs the compiler to generate code for this function as an entrypoint to a device program, such that it can be launched from the host.The function does not return anything, because there is no trivial way to construct a return channel of a parallel invocation. Device-side entrypoints may not return a value, their results should be communicated using output parameters.
Device-side functions are typically called compute kernels, or just kernels for short. This is to distinguish them from non-graphics-related graphics shaders, or just shaders for short.
Arguments are taken by value and all arguments shall be TriviallyCopyable, meaning they should be memcpy-friendly. (Imagine if they had custom copy constructors. Where would that logic execute? On the host? On the device?) Pointer arguments are pointers to device memory, one typically backed by VRAM.
We said that we’ll be computing \(a\cdot x+y=z\), however we only pass two pointers to the function. We’ll be canonically reusing one of the inputs as outputs.
This function is launched from the host using a language extension often called the triple chevron syntax. Inside the angle brackets, provide the following.
The amount of shared memory to allocate by the host
The device stream to enqueue the operation on
The block size and shared memory become important later in Reduction. For
now, a hardcoded 256
is a safe default for simple kernels such as this.
Following the triple chevron is ordinary function argument passing.
Look at how the kernel is implemented.
__global__ void saxpy_kernel(const float a, const float* d_x, float* d_y, const unsigned int size)
{
// Compute the current thread's index in the grid.
const unsigned int global_idx = blockIdx.x * blockDim.x + threadIdx.x;
// The grid can be larger than the number of items in the vectors. Avoid out-of-bounds addressing.
if(global_idx < size)
{
d_y[global_idx] = a * d_x[global_idx] + d_y[global_idx];
}
}
The unique linear index identifying the thread is computed from the block ID the thread is a member of, the block’s size and the ID of the thread within the block.
A check is made to avoid overindexing the input.
The useful part of the computation is carried out.
Retrieval of the result from the device is done much like input data copy. In this current step the results copied from device to host. The opposite direction of the input data copy:
HIP_CHECK(hipMemcpy(y.data(), d_y, size_bytes, hipMemcpyDeviceToHost));
Compiling on the command line#
Setting up the command line#
Strictly speaking there’s no such thing as “setting up the command-line for compilation” on Linux. To make invocations more terse, Linux and Windows example follow.
While distro maintainers might package ROCm so that it installs to system-default locations, AMD’s packages aren’t installed that way. They need to be added to the PATH by the user.
export PATH=/opt/rocm/bin:${PATH}
You should be able to call the compiler on the command line now:
amdclang++ --version
Note
Docker images distributed by AMD, such as rocm-terminal already have /opt/rocm/bin on the Path for convenience. This subtly affects CMake package detection logic of ROCm libraries.
Both distro maintainers and NVIDIA package CUDA so that nvcc
and related
tools are available on the command line by default. You can call the
compiler on the command line with:
nvcc --version
Windows compilers and command line tooling have traditionally relied on
extra environmental variables and PATH entries to function correctly.
Visual Studio refers to command lines with this setup as “Developer
Command Prompt” or “Developer PowerShell” for cmd.exe
and PowerShell
respectively.
The HIP SDK on Windows doesn’t include a complete toolchain. You will also need:
The Microsoft Windows SDK. It provides the import libs to crucial system libraries that all executables must link to and some auxiliary compiler tooling.
A Standard Template Library (STL). Installed as part of the Microsoft Visual C++ compiler (MSVC) or with Visual Studio.
If you don’t have a version of Visual Studio 2022 installed, for a minimal command line experience, install the Build Tools for Visual Studio 2022 with the Desktop Developemnt Workload. Under Individual Components select:
A version of the Windows SDK
“MSVC v143 - VS 2022 C++ x64/x86 build tools (Latest)”
“C++ CMake tools for Windows” (optional)
Note
The “C++ CMake tools for Windows” individual component is a convenience which
puts both cmake.exe
and ninja.exe
onto the PATH inside developer
command prompts. You can install these manually, but then you must manage
them manually.
Visual Studio 2017 and later are detectable as COM object instances via WMI. To setup a command line from any shell for the latest Visual Studio’s default Visual C++ toolset issue:
$InstallationPath = Get-CimInstance MSFT_VSInstance | Sort-Object -Property Version -Descending | Select-Object -First 1 -ExpandProperty InstallLocation
Import-Module $InstallationPath\Common7\Tools\Microsoft.VisualStudio.DevShell.dll
Enter-VsDevShell -InstallPath $InstallationPath -SkipAutomaticLocation -Arch amd64 -HostArch amd64 -DevCmdArguments '-no_logo'
$env:PATH = "${env:HIP_PATH}bin;${env:PATH}"
You should be able to call the compiler on the command line now:
clang++ --version
Windows compilers and command line tooling have traditionally relied on
extra environmental variables and PATH entries to function correctly.
Visual Studio refers to command lines with this setup as “Developer
Command Prompt” or “Developer PowerShell” for cmd.exe
and PowerShell
respectively.
The HIP and CUDA SDKs on Windows don’t include complete toolchains. You will also need:
The Microsoft Windows SDK. It provides the import libs to crucial system libraries that all executables must link to and some auxiliary compiler tooling.
A Standard Template Library (STL). Installed as part of the Microsoft Visual C++ compiler (MSVC) or with Visual Studio.
If you don’t have a version of Visual Studio 2022 installed, for a minimal command line experience, install the Build Tools for Visual Studio 2022 with the Desktop Developemnt Workload. Under Individual Components select:
A version of the Windows SDK
“MSVC v143 - VS 2022 C++ x64/x86 build tools (Latest)”
“C++ CMake tools for Windows” (optional)
Note
The “C++ CMake tools for Windows” individual component is a convenience which
puts both cmake.exe
and ninja.exe
onto the PATH inside developer
command prompts. You can install these manually, but then you must manage
them manually.
Visual Studio 2017 and later are detectable as COM object instances via WMI. To setup a command line from any shell for the latest Visual Studio’s default Visual C++ toolset issue:
$InstallationPath = Get-CimInstance MSFT_VSInstance | Sort-Object -Property Version -Descending | Select-Object -First 1 -ExpandProperty InstallLocation
Import-Module $InstallationPath\Common7\Tools\Microsoft.VisualStudio.DevShell.dll
Enter-VsDevShell -InstallPath $InstallationPath -SkipAutomaticLocation -Arch amd64 -HostArch amd64 -DevCmdArguments '-no_logo'
You should be able to call the compiler on the command line now:
nvcc --version
Invoking the compiler manually#
To compile and link a single-file application, use the following commands:
amdclang++ ./HIP-Basic/saxpy/main.hip -o saxpy -I ./Common -lamdhip64 -L /opt/rocm/lib -O2
nvcc ./HIP-Basic/saxpy/main.hip -o saxpy -I ./Common -I /opt/rocm/include -O2 -x cu
clang++ .\HIP-Basic\saxpy\main.hip -o saxpy.exe -I .\Common -lamdhip64 -L ${env:HIP_PATH}lib -O2
nvcc .\HIP-Basic\saxpy\main.hip -o saxpy.exe -I ${env:HIP_PATH}include -I .\Common -O2 -x cu
Depending on your computer, the resulting binary might or might not run. If not,
it typically complains about “Invalid device function”. That error
(corresponding to the hipErrorInvalidDeviceFunction
entry of hipError_t
)
means that the runtime could not find a device program binary of the
appropriate flavor embedded into the executable.
So far, the discussion has covered how data makes it from the host to the device and back. It has also discussed the device code as source, with the HIP runtime arguing that the correct binary to dispatch for execution. How can you find out what device binary flavors are embedded into the executable?
The utilities included with ROCm help significantly to inspect binary artifacts on disk. Add the ROCmCC installation folder to your PATH if you want to use these utilities (the utilities expect them to be on the PATH).
You can list embedded program binaries using roc-obj-ls
.
roc-obj-ls ./saxpy
It should return something like:
1 host-x86_64-unknown-linux file://./saxpy#offset=12288&size=0
1 hipv4-amdgcn-amd-amdhsa--gfx803 file://./saxpy#offset=12288&size=9760
The compiler embeds a version 4 code object (more on code
object versions)
and used the LLVM target triple amdgcn-amd-amdhsa–gfx803 (more on target triples). You can
extract that program object in a disassembled fashion for human consumption
via roc-obj
.
roc-obj -t gfx803 -d ./saxpy
This creates two files on disk and .s
extension is of most interest.
Opening this file or dumping it to the console using cat
lets find the disassembled binary of the SAXPY compute kernel, something
similar to:
Disassembly of section .text:
<_Z12saxpy_kernelfPKfPfj>:
s_load_dword s0, s[4:5], 0x2c // 000000001000: C0020002 0000002C
s_load_dword s1, s[4:5], 0x18 // 000000001008: C0020042 00000018
s_waitcnt lgkmcnt(0) // 000000001010: BF8C007F
s_and_b32 s0, s0, 0xffff // 000000001014: 8600FF00 0000FFFF
s_mul_i32 s6, s6, s0 // 00000000101C: 92060006
v_add_u32_e32 v0, vcc, s6, v0 // 000000001020: 32000006
v_cmp_gt_u32_e32 vcc, s1, v0 // 000000001024: 7D980001
s_and_saveexec_b64 s[0:1], vcc // 000000001028: BE80206A
s_cbranch_execz 22 // 00000000102C: BF880016 <_Z12saxpy_kernelfPKfPfj+0x88>
s_load_dwordx4 s[0:3], s[4:5], 0x8 // 000000001030: C00A0002 00000008
v_mov_b32_e32 v1, 0 // 000000001038: 7E020280
v_lshlrev_b64 v[0:1], 2, v[0:1] // 00000000103C: D28F0000 00020082
s_waitcnt lgkmcnt(0) // 000000001044: BF8C007F
v_mov_b32_e32 v3, s1 // 000000001048: 7E060201
v_add_u32_e32 v2, vcc, s0, v0 // 00000000104C: 32040000
v_addc_u32_e32 v3, vcc, v3, v1, vcc // 000000001050: 38060303
flat_load_dword v2, v[2:3] // 000000001054: DC500000 02000002
v_mov_b32_e32 v3, s3 // 00000000105C: 7E060203
v_add_u32_e32 v0, vcc, s2, v0 // 000000001060: 32000002
v_addc_u32_e32 v1, vcc, v3, v1, vcc // 000000001064: 38020303
flat_load_dword v3, v[0:1] // 000000001068: DC500000 03000000
s_load_dword s0, s[4:5], 0x0 // 000000001070: C0020002 00000000
s_waitcnt vmcnt(0) lgkmcnt(0) // 000000001078: BF8C0070
v_mac_f32_e32 v3, s0, v2 // 00000000107C: 2C060400
flat_store_dword v[0:1], v3 // 000000001080: DC700000 00000300
s_endpgm // 000000001088: BF810000
Alternatively, call the compiler with --save-temps
to dump all device
binary to disk in separate files.
amdclang++ ./HIP-Basic/saxpy/main.hip -o saxpy -I ./Common -lamdhip64 -L /opt/rocm/lib -O2 --save-temps
List all the temporaries created while compiling main.hip
with:
ls main-hip-amdgcn-amd-amdhsa-*
main-hip-amdgcn-amd-amdhsa-gfx803.bc
main-hip-amdgcn-amd-amdhsa-gfx803.cui
main-hip-amdgcn-amd-amdhsa-gfx803.o
main-hip-amdgcn-amd-amdhsa-gfx803.out
main-hip-amdgcn-amd-amdhsa-gfx803.out.resolution.txt
main-hip-amdgcn-amd-amdhsa-gfx803.s
Files with the .s
extension hold the disassembled contents of the binary.
The filename notes the graphics IPs used by the compiler. The contents of
this file are similar to what roc-obj
printed to the console.
Unlike HIP on AMD, when compiling using the NVIDIA support of HIP the resulting
binary will be a valid CUDA executable as far as the binary goes. Therefor
it’ll incorporate PTX ISA (Parallel Thread eXecution Instruction Set
Architecture) instead of AMDGPU binary. As s result, tooling shipping with the
CUDA SDK can be used to inspect which device ISA got compiled into a specific
executable. The tool most useful to us currently is cuobjdump
.
cuobjdump --list-ptx ./saxpy
Which will print something like:
PTX file 1: saxpy.1.sm_52.ptx
From this we can see that the saxpy kernel is stored as sm_52
, which shows
that a compute capability 5.2 ISA got embedded into the executable, so devices
which sport compute capability 5.2 or newer will be able to run this code.
The HIP SDK for Windows don’t yet sport the roc-*
set of utilities to work
with binary artifacts. To find out what binary formats are embedded into an
executable, one may use dumpbin
tool from the Windows SDK to obtain the
raw data of the .hip_fat
section of an executable. (This binary payload is
what gets parsed by the roc-*
set of utilities on Linux.) Skipping over the
reported header, the rendered raw data as ASCII has ~3 lines per entries.
Depending on how many binaries are embedded, you may need to alter the number
of rendered lines. An invocation such as:
dumpbin.exe /nologo /section:.hip_fat /rawdata:8 .\saxpy.exe | select -Skip 20 -First 12
The output may look like:
000000014004C000: 5F474E414C435F5F 5F44414F4C46464F __CLANG_OFFLOAD_
000000014004C010: 5F5F454C444E5542 0000000000000002 BUNDLE__........
000000014004C020: 0000000000001000 0000000000000000 ................
000000014004C030: 0000000000000019 3638782D74736F68 ........host-x86
000000014004C040: 6E6B6E752D34365F 756E696C2D6E776F _64-unknown-linu
000000014004C050: 0000000000100078 00000000000D9800 x...............
000000014004C060: 0000000000001F00 612D347670696800 .........hipv4-a
000000014004C070: 6D612D6E6367646D 617368646D612D64 mdgcn-amd-amdhsa
000000014004C080: 3630397866672D2D 0000000000000000 --gfx906........
000000014004C090: 0000000000000000 0000000000000000 ................
000000014004C0A0: 0000000000000000 0000000000000000 ................
000000014004C0B0: 0000000000000000 0000000000000000 ................
We can see that the compiler embedded a version 4 code object (more on code object versions) and used the LLVM target triple amdgcn-amd-amdhsa–gfx906 (more on target triples). Don’t be alarmed about linux showing up as a binary format, AMDGPU binaries uploaded to the GPU for execution are proper linux ELF binaries in their format.
Alternatively we can call the compiler with --save-temps
to dump all device
binary to disk in separate files.
clang++ .\HIP-Basic\saxpy\main.hip -o saxpy.exe -I .\Common -lamdhip64 -L ${env:HIP_PATH}lib -O2 --save-temps
Now we can list all the temporaries created while compiling main.hip
via
Get-ChildItem -Filter main-hip-* | select -Property Name
Name
----
main-hip-amdgcn-amd-amdhsa-gfx906.bc
main-hip-amdgcn-amd-amdhsa-gfx906.hipi
main-hip-amdgcn-amd-amdhsa-gfx906.o
main-hip-amdgcn-amd-amdhsa-gfx906.out
main-hip-amdgcn-amd-amdhsa-gfx906.out.resolution.txt
main-hip-amdgcn-amd-amdhsa-gfx906.s
Files with the .s
extension hold the disassembled contents of the binary and
the filename directly informs us of the graphics IPs used by the compiler.
Get-ChildItem main-hip-*.s | Get-Content
.text
.amdgcn_target "amdgcn-amd-amdhsa--gfx906"
.protected _Z12saxpy_kernelfPKfPfj ; -- Begin function _Z12saxpy_kernelfPKfPfj
.globl _Z12saxpy_kernelfPKfPfj
.p2align 8
.type _Z12saxpy_kernelfPKfPfj,@function
_Z12saxpy_kernelfPKfPfj: ; @_Z12saxpy_kernelfPKfPfj
; %bb.0:
s_load_dword s0, s[4:5], 0x4
s_load_dword s1, s[6:7], 0x18
s_waitcnt lgkmcnt(0)
s_and_b32 s0, s0, 0xffff
s_mul_i32 s8, s8, s0
v_add_u32_e32 v0, s8, v0
v_cmp_gt_u32_e32 vcc, s1, v0
s_and_saveexec_b64 s[0:1], vcc
s_cbranch_execz .LBB0_2
; %bb.1:
s_load_dwordx4 s[0:3], s[6:7], 0x8
v_mov_b32_e32 v1, 0
v_lshlrev_b64 v[0:1], 2, v[0:1]
s_waitcnt lgkmcnt(0)
v_mov_b32_e32 v3, s1
v_add_co_u32_e32 v2, vcc, s0, v0
v_addc_co_u32_e32 v3, vcc, v3, v1, vcc
global_load_dword v2, v[2:3], off
v_mov_b32_e32 v3, s3
v_add_co_u32_e32 v0, vcc, s2, v0
v_addc_co_u32_e32 v1, vcc, v3, v1, vcc
global_load_dword v3, v[0:1], off
s_load_dword s0, s[6:7], 0x0
s_waitcnt vmcnt(0) lgkmcnt(0)
v_fmac_f32_e32 v3, s0, v2
global_store_dword v[0:1], v3, off
.LBB0_2:
s_endpgm
...
Unlike HIP on AMD, when compiling using the NVIDIA support for HIP, the resulting
binary will be a valid CUDA executable. Therefore, it’ll incorporate PTX ISA
(Parallel Thread eXecution Instruction Set Architecture) instead of AMDGPU
binary. As a result, tooling included with the CUDA SDK can be used to
inspect which device ISA was compiled into a specific executable. The most
helpful to us currently is cuobjdump
.
cuobjdump.exe --list-ptx .\saxpy.exe
Which prints something like:
PTX file 1: saxpy.1.sm_52.ptx
This example shows that the SAXPY kernel is stored as sm_52
. It also shows
that a compute capability 5.2 ISA was embedded into the executable, so devices
that support compute capability 5.2 or newer will be able to run this code.
Now that you’ve found what binary got embedded into the executable, find which format our available devices use.
On Linux a utility called rocminfo
helps us list all the properties of the
devices available on the system, including which version of graphics IP
(gfxXYZ
) they employ. You can filter the output to have only these lines:
/opt/rocm/bin/rocminfo | grep gfx
Name: gfx906
Name: amdgcn-amd-amdhsa--gfx906:sramecc+:xnack-
Now that you know which graphics IPs our devices use, recompile your program with the appropriate parameters.
amdclang++ ./HIP-Basic/saxpy/main.hip -o saxpy -I ./Common -lamdhip64 -L /opt/rocm/lib -O2 --offload-arch=gfx906:sramecc+:xnack-
Now the sample will run.
./saxpy
Calculating y[i] = a * x[i] + y[i] over 1000000 elements.
First 10 elements of the results: [ 3, 5, 7, 9, 11, 13, 15, 17, 19, 21 ]
On Linux HIP with the NVIDIA back-end, the deviceQuery
CUDA SDK sample
can help us list all the properties of the devices available on the system,
including which version of compute capability a device sports.
<major>.<minor>
compute capability is passed to nvcc
on the
command-line as sm_<major><minor>
, for eg. 8.6
is sm_86
.
Because it’s not included as a binary, compile the matching example from ROCm.
nvcc ./HIP-Basic/device_query/main.cpp -o device_query -I ./Common -I /opt/rocm/include -O2
Filter the output to have only the lines of interest, for example:
./device_query | grep "major.minor"
major.minor: 8.6
major.minor: 7.0
Note
In addition to the nvcc
executable is another tool called __nvcc_device_query
which prints the SM Architecture numbers to standard out as a comma
separated list of numbers. The utility’s name suggests it’s not a user-facing
executable but is used by nvcc
to determine what devices are in the
system at hand.
Now that you know which graphics IPs our devices use, recompile your program with the appropriate parameters.
nvcc ./HIP-Basic/saxpy/main.hip -o saxpy -I ./Common -I /opt/rocm/include -O2 -x cu -arch=sm_70,sm_86
Note
If you want to portably target the development machine which is compiling, you
may specify -arch=native
instead.
Now the sample will run.
./saxpy
Calculating y[i] = a * x[i] + y[i] over 1000000 elements.
First 10 elements of the results: [ 3, 5, 7, 9, 11, 13, 15, 17, 19, 21 ]
On Windows, a utility called hipInfo.exe
helps us list all the properties
of the devices available on the system, including which version of graphics IP
(gfxXYZ
) they employ. Filter the output to have only these lines:
& ${env:HIP_PATH}bin\hipInfo.exe | Select-String gfx
gcnArchName: gfx1032
gcnArchName: gfx1035
Now that you know which graphics IPs our devices use, recompile your program with the appropriate parameters.
clang++ .\HIP-Basic\saxpy\main.hip -o saxpy.exe -I .\Common -lamdhip64 -L ${env:HIP_PATH}lib -O2 --offload-arch=gfx1032 --offload-arch=gfx1035
Now the sample will run.
.\saxpy.exe
Calculating y[i] = a * x[i] + y[i] over 1000000 elements.
First 10 elements of the results: [ 3, 5, 7, 9, 11, 13, 15, 17, 19, 21 ]
On Windows HIP with the NVIDIA back-end, the deviceQuery
CUDA SDK sample
can help us list all the properties of the devices available on the system,
including which version of compute capability a device sports.
<major>.<minor>
compute capability is passed to nvcc
on the
command-line as sm_<major><minor>
, for eg. 8.6
is sm_86
.
Because it’s not included as a binary, compile the matching example from ROCm.
nvcc .\HIP-Basic\device_query\main.cpp -o device_query.exe -I .\Common -I ${env:HIP_PATH}include -O2
Filter the output to have only the lines of interest, for example:
.\device_query.exe | Select-String "major.minor"
major.minor: 8.6
major.minor: 7.0
Note
Next to the nvcc
executable is another tool called __nvcc_device_query.exe
which simply prints the SM Architecture numbers to standard out as a comma
separated list of numbers. The naming of this utility suggests it’s not a user
facing executable but is used by nvcc
to determine what devices are in the
system at hand.
Now that you know which graphics IPs our devices use, recompile your program with the appropriate parameters.
nvcc .\HIP-Basic\saxpy\main.hip -o saxpy.exe -I ${env:HIP_PATH}include -I .\Common -O2 -x cu -arch=sm_70,sm_86
Note
If you want to portably target the development machine which is compiling, you
may specify -arch=native
instead.
Now the sample will run.
.\saxpy.exe
Calculating y[i] = a * x[i] + y[i] over 1000000 elements.
First 10 elements of the results: [ 3, 5, 7, 9, 11, 13, 15, 17, 19, 21 ]