// REQUIRES: amdgpu-registered-target // RUN: %clang_cc1 -fopenmp -x c++ -w -std=c++11 -triple x86_64-unknown-unknown -fopenmp-targets=amdgcn-amd-amdhsa -emit-llvm-bc %s -o %t-ppc-host.bc // RUN: %clang_cc1 -fopenmp -x c++ -w -std=c++11 -triple amdgcn-amd-amdhsa -fopenmp-targets=amdgcn-amd-amdhsa -emit-llvm %s -fopenmp-is-target-device -fopenmp-host-ir-file-path %t-ppc-host.bc -target-cpu gfx906 -o - | FileCheck %s // expected-no-diagnostics /*===-----------------------------------------------------------------------=== Inspired from SOLLVE tests: - 5.0/metadirective/test_metadirective_arch_is_nvidia.c ===------------------------------------------------------------------------===*/ #define N 1024 int metadirective1() { int v1[N], v2[N], v3[N]; int target_device_num, host_device_num, default_device; int errors = 0; #pragma omp target map(to:v1,v2) map(from:v3, target_device_num) device(default_device) { #pragma omp metadirective \ when(device={arch("amdgcn")}: teams distribute parallel for) \ default(parallel for) for (int i = 0; i < N; i++) { #pragma omp atomic write v3[i] = v1[i] * v2[i]; } } return errors; } // CHECK: define weak_odr protected amdgpu_kernel void @[[METADIRECTIVE:.+metadirective1[a-z0-9_]+]] // CHECK: entry: // CHECK: %{{[0-9]}} = call i32 @__kmpc_target_init // CHECK: user_code.entry: // CHECK: call void @[[METADIRECTIVE]]_omp_outlined // CHECK-NOT: call void @__kmpc_parallel_51 // CHECK: ret void // CHECK: define internal void @[[METADIRECTIVE]]_omp_outlined // CHECK: entry: // CHECK: call void @__kmpc_distribute_static_init // CHECK: omp.loop.exit: // CHECK: call void @__kmpc_distribute_static_fini // CHECK: define internal void @[[METADIRECTIVE]]_omp_outlined_omp_outlined // CHECK: entry: // CHECK: call void @__kmpc_for_static_init_4 // CHECK: omp.inner.for.body: // CHECK: store atomic {{.*}} monotonic // CHECK: omp.loop.exit: // CHECK-NEXT: call void @__kmpc_distribute_static_fini // CHECK-NEXT: ret void