// RUN: %clang_cc1 -w -triple x86_64-apple-darwin10 \ // RUN: -emit-llvm -o - %s | FileCheck %s --check-prefix=ALL --check-prefix=SSE // RUN: %clang_cc1 -w -triple i386-apple-darwin10 \ // RUN: -emit-llvm -o - %s | FileCheck %s --check-prefix=ALL --check-prefix=SSE // RUN: %clang_cc1 -w -triple x86_64-apple-darwin10 -target-feature +avx \ // RUN: -emit-llvm -o - %s | FileCheck %s --check-prefix=ALL --check-prefix=AVX // RUN: %clang_cc1 -w -triple i386-apple-darwin10 -target-feature +avx \ // RUN: -emit-llvm -o - %s | FileCheck %s --check-prefix=ALL --check-prefix=AVX // RUN: %clang_cc1 -w -triple x86_64-apple-darwin10 -target-feature +avx512f \ // RUN: -emit-llvm -o - %s | FileCheck %s --check-prefix=ALL --check-prefix=AVX512 // RUN: %clang_cc1 -w -triple i386-apple-darwin10 -target-feature +avx512f \ // RUN: -emit-llvm -o - %s | FileCheck %s --check-prefix=ALL --check-prefix=AVX512 // At or below target max alignment with no aligned attribute should align based // on the size of vector. double __attribute__((vector_size(16))) v1; // SSE: @v1 {{.*}}, align 16 // AVX: @v1 {{.*}}, align 16 // AVX512: @v1 {{.*}}, align 16 double __attribute__((vector_size(32))) v2; // SSE: @v2 {{.*}}, align 16 // AVX: @v2 {{.*}}, align 32 // AVX512: @v2 {{.*}}, align 32 typedef __attribute__((__ext_vector_type__(16))) _Bool v2b_type; v2b_type v2b; // ALL: @v2b {{.*}}, align 2 // Alignment above target max alignment with no aligned attribute should align // based on the target max. double __attribute__((vector_size(64))) v3; // SSE: @v3 {{.*}}, align 16 // AVX: @v3 {{.*}}, align 32 // AVX512: @v3 {{.*}}, align 64 double __attribute__((vector_size(1024))) v4; // SSE: @v4 {{.*}}, align 16 // AVX: @v4 {{.*}}, align 32 // AVX512: @v4 {{.*}}, align 64 typedef __attribute__((__ext_vector_type__(8192))) _Bool v4b_type; v4b_type v4b; // SSE: @v4b {{.*}}, align 16 // AVX: @v4b {{.*}}, align 32 // AVX512: @v4b {{.*}}, align 64 // Aliged attribute should always override. double __attribute__((vector_size(16), aligned(16))) v5; // ALL: @v5 {{.*}}, align 16 double __attribute__((vector_size(16), aligned(64))) v6; // ALL: @v6 {{.*}}, align 64 double __attribute__((vector_size(32), aligned(16))) v7; // ALL: @v7 {{.*}}, align 16 double __attribute__((vector_size(32), aligned(64))) v8; // ALL: @v8 {{.*}}, align 64 typedef __attribute__((ext_vector_type(256), aligned(128))) _Bool v8b_type; v8b_type v8b; // ALL: @v8b {{.*}}, align 128 // Check non-power of 2 widths. double __attribute__((vector_size(24))) v9; // SSE: @v9 {{.*}}, align 16 // AVX: @v9 {{.*}}, align 32 // AVX512: @v9 {{.*}}, align 32 double __attribute__((vector_size(40))) v10; // SSE: @v10 {{.*}}, align 16 // AVX: @v10 {{.*}}, align 32 // AVX512: @v10 {{.*}}, align 64 typedef __attribute__((ext_vector_type(248))) _Bool v10b_type; v10b_type v10b; // SSE: @v10b {{.*}}, align 16 // AVX: @v10b {{.*}}, align 32 // AVX512: @v10b {{.*}}, align 32 // Check non-power of 2 widths with aligned attribute. double __attribute__((vector_size(24), aligned(64))) v11; // ALL: @v11 {{.*}}, align 64 double __attribute__((vector_size(80), aligned(16))) v12; // ALL: @v12 {{.*}}, align 16 typedef __attribute__((ext_vector_type(248), aligned(4))) _Bool v12b_type; v12b_type v12b; // ALL: @v12b {{.*}}, align 4