// clang-format off // RUN: %libomptarget-compile-generic // RUN: env LIBOMPTARGET_INFO=16 \ // RUN: %libomptarget-run-generic 2>&1 | %fcheck-generic --check-prefix=DEFAULT // UNSUPPORTED: nvptx64-nvidia-cuda // UNSUPPORTED: nvptx64-nvidia-cuda-LTO // UNSUPPORTED: aarch64-unknown-linux-gnu // UNSUPPORTED: aarch64-unknown-linux-gnu-LTO // UNSUPPORTED: x86_64-pc-linux-gnu // UNSUPPORTED: x86_64-pc-linux-gnu-LTO __attribute__((optnone)) int optnone() { return 1; } int main() { int N = optnone() * 4098 * 32; // DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]] #pragma omp target teams distribute parallel for simd for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]] #pragma omp target teams distribute parallel for simd for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]] #pragma omp target teams distribute parallel for simd for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: [[NT:(128|256)]] (MaxFlatWorkGroupSize: [[NT]] #pragma omp target #pragma omp teams distribute parallel for for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 42 (MaxFlatWorkGroupSize: 1024 #pragma omp target thread_limit(optnone() * 42) #pragma omp teams distribute parallel for for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 42 (MaxFlatWorkGroupSize: 42 #pragma omp target thread_limit(optnone() * 42) ompx_attribute(__attribute__((amdgpu_flat_work_group_size(42, 42)))) #pragma omp teams distribute parallel for for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 42 (MaxFlatWorkGroupSize: 42 #pragma omp target ompx_attribute(__attribute__((amdgpu_flat_work_group_size(42, 42)))) #pragma omp teams distribute parallel for for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: MaxFlatWorkGroupSize: 1024 #pragma omp target #pragma omp teams distribute parallel for num_threads(optnone() * 42) for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: MaxFlatWorkGroupSize: 1024 #pragma omp target teams distribute parallel for thread_limit(optnone() * 42) for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: MaxFlatWorkGroupSize: 1024 #pragma omp target teams distribute parallel for num_threads(optnone() * 42) for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 9 (MaxFlatWorkGroupSize: 9 #pragma omp target #pragma omp teams distribute parallel for num_threads(9) for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 4 (MaxFlatWorkGroupSize: 4 #pragma omp target thread_limit(4) #pragma omp teams distribute parallel for for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 4 (MaxFlatWorkGroupSize: 4 #pragma omp target #pragma omp teams distribute parallel for thread_limit(4) for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 9 (MaxFlatWorkGroupSize: 9 #pragma omp target teams distribute parallel for num_threads(9) for (int i = 0; i < N; ++i) { optnone(); } // DEFAULT: 4 (MaxFlatWorkGroupSize: 4 #pragma omp target teams distribute parallel for simd thread_limit(4) for (int i = 0; i < N; ++i) { optnone(); } }