103 lines
3.1 KiB
C
103 lines
3.1 KiB
C
|
// clang-format off
|
||
|
// RUN: %libomptarget-compile-generic
|
||
|
// RUN: env LIBOMPTARGET_INFO=16 \
|
||
|
// RUN: %libomptarget-run-generic 2>&1 | %fcheck-generic --check-prefix=DEFAULT
|
||
|
|
||
|
// UNSUPPORTED: nvptx64-nvidia-cuda
|
||
|
// UNSUPPORTED: nvptx64-nvidia-cuda-LTO
|
||
|
// UNSUPPORTED: aarch64-unknown-linux-gnu
|
||
|
// UNSUPPORTED: aarch64-unknown-linux-gnu-LTO
|
||
|
// UNSUPPORTED: x86_64-pc-linux-gnu
|
||
|
// UNSUPPORTED: x86_64-pc-linux-gnu-LTO
|
||
|
|
||
|
__attribute__((optnone)) int optnone() { return 1; }
|
||
|
|
||
|
int main() {
|
||
|
int N = optnone() * 4098 * 32;
|
||
|
|
||
|
// DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]]
|
||
|
#pragma omp target teams distribute parallel for simd
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]]
|
||
|
#pragma omp target teams distribute parallel for simd
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]]
|
||
|
#pragma omp target teams distribute parallel for simd
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]]
|
||
|
#pragma omp target
|
||
|
#pragma omp teams distribute parallel for
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 42 (MaxFlatWorkGroupSize: 1024
|
||
|
#pragma omp target thread_limit(optnone() * 42)
|
||
|
#pragma omp teams distribute parallel for
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 42 (MaxFlatWorkGroupSize: 42
|
||
|
#pragma omp target thread_limit(optnone() * 42) ompx_attribute(__attribute__((amdgpu_flat_work_group_size(42, 42))))
|
||
|
#pragma omp teams distribute parallel for
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 42 (MaxFlatWorkGroupSize: 42
|
||
|
#pragma omp target ompx_attribute(__attribute__((amdgpu_flat_work_group_size(42, 42))))
|
||
|
#pragma omp teams distribute parallel for
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: MaxFlatWorkGroupSize: 1024
|
||
|
#pragma omp target
|
||
|
#pragma omp teams distribute parallel for num_threads(optnone() * 42)
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: MaxFlatWorkGroupSize: 1024
|
||
|
#pragma omp target teams distribute parallel for thread_limit(optnone() * 42)
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: MaxFlatWorkGroupSize: 1024
|
||
|
#pragma omp target teams distribute parallel for num_threads(optnone() * 42)
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 9 (MaxFlatWorkGroupSize: 9
|
||
|
#pragma omp target
|
||
|
#pragma omp teams distribute parallel for num_threads(9)
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 4 (MaxFlatWorkGroupSize: 4
|
||
|
#pragma omp target thread_limit(4)
|
||
|
#pragma omp teams distribute parallel for
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 4 (MaxFlatWorkGroupSize: 4
|
||
|
#pragma omp target
|
||
|
#pragma omp teams distribute parallel for thread_limit(4)
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 9 (MaxFlatWorkGroupSize: 9
|
||
|
#pragma omp target teams distribute parallel for num_threads(9)
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
// DEFAULT: 4 (MaxFlatWorkGroupSize: 4
|
||
|
#pragma omp target teams distribute parallel for simd thread_limit(4)
|
||
|
for (int i = 0; i < N; ++i) {
|
||
|
optnone();
|
||
|
}
|
||
|
}
|
||
|
|