; NOTE: Assertions have been autogenerated by utils/update_test_checks.py UTC_ARGS: --version 3 ; RUN: opt -safe-stack -S -mtriple=i386-pc-linux-gnu < %s -o - | FileCheck %s --check-prefix=I386 ; RUN: opt -safe-stack -S -mtriple=x86_64-pc-linux-gnu < %s -o - | FileCheck %s --check-prefix=X86-64 ; RUN: opt -passes=safe-stack -S -mtriple=i386-pc-linux-gnu < %s -o - | FileCheck %s --check-prefix=I386 ; RUN: opt -passes=safe-stack -S -mtriple=x86_64-pc-linux-gnu < %s -o - | FileCheck %s --check-prefix=X86-64 %struct.__jmp_buf_tag = type { [8 x i64], i32, %struct.__sigset_t } %struct.__sigset_t = type { [16 x i64] } @.str = private unnamed_addr constant [4 x i8] c"%s\0A\00", align 1 @buf = internal global [1 x %struct.__jmp_buf_tag] zeroinitializer, align 16 ; setjmp/longjmp test with dynamically sized array. ; Requires protector. define i32 @foo(i32 %size) nounwind uwtable safestack { ; I386-LABEL: define i32 @foo( ; I386-SAME: i32 [[SIZE:%.*]]) #[[ATTR0:[0-9]+]] { ; I386-NEXT: entry: ; I386-NEXT: [[UNSAFE_STACK_PTR:%.*]] = load ptr, ptr @__safestack_unsafe_stack_ptr, align 4 ; I386-NEXT: [[UNSAFE_STACK_DYNAMIC_PTR:%.*]] = alloca ptr, align 4 ; I386-NEXT: store ptr [[UNSAFE_STACK_PTR]], ptr [[UNSAFE_STACK_DYNAMIC_PTR]], align 4 ; I386-NEXT: [[TMP0:%.*]] = mul i32 [[SIZE]], 4 ; I386-NEXT: [[TMP1:%.*]] = load ptr, ptr @__safestack_unsafe_stack_ptr, align 4 ; I386-NEXT: [[TMP2:%.*]] = ptrtoint ptr [[TMP1]] to i32 ; I386-NEXT: [[TMP3:%.*]] = sub i32 [[TMP2]], [[TMP0]] ; I386-NEXT: [[TMP4:%.*]] = and i32 [[TMP3]], -16 ; I386-NEXT: [[A:%.*]] = inttoptr i32 [[TMP4]] to ptr ; I386-NEXT: store ptr [[A]], ptr @__safestack_unsafe_stack_ptr, align 4 ; I386-NEXT: store ptr [[A]], ptr [[UNSAFE_STACK_DYNAMIC_PTR]], align 4 ; I386-NEXT: [[CALL:%.*]] = call i32 @_setjmp(ptr @buf) #[[ATTR1:[0-9]+]] ; I386-NEXT: [[TMP5:%.*]] = load ptr, ptr [[UNSAFE_STACK_DYNAMIC_PTR]], align 4 ; I386-NEXT: store ptr [[TMP5]], ptr @__safestack_unsafe_stack_ptr, align 4 ; I386-NEXT: call void @funcall(ptr [[A]]) ; I386-NEXT: store ptr [[UNSAFE_STACK_PTR]], ptr @__safestack_unsafe_stack_ptr, align 4 ; I386-NEXT: ret i32 0 ; ; X86-64-LABEL: define i32 @foo( ; X86-64-SAME: i32 [[SIZE:%.*]]) #[[ATTR0:[0-9]+]] { ; X86-64-NEXT: entry: ; X86-64-NEXT: [[UNSAFE_STACK_PTR:%.*]] = load ptr, ptr @__safestack_unsafe_stack_ptr, align 8 ; X86-64-NEXT: [[UNSAFE_STACK_DYNAMIC_PTR:%.*]] = alloca ptr, align 8 ; X86-64-NEXT: store ptr [[UNSAFE_STACK_PTR]], ptr [[UNSAFE_STACK_DYNAMIC_PTR]], align 8 ; X86-64-NEXT: [[TMP0:%.*]] = zext i32 [[SIZE]] to i64 ; X86-64-NEXT: [[TMP1:%.*]] = mul i64 [[TMP0]], 4 ; X86-64-NEXT: [[TMP2:%.*]] = load ptr, ptr @__safestack_unsafe_stack_ptr, align 8 ; X86-64-NEXT: [[TMP3:%.*]] = ptrtoint ptr [[TMP2]] to i64 ; X86-64-NEXT: [[TMP4:%.*]] = sub i64 [[TMP3]], [[TMP1]] ; X86-64-NEXT: [[TMP5:%.*]] = and i64 [[TMP4]], -16 ; X86-64-NEXT: [[A:%.*]] = inttoptr i64 [[TMP5]] to ptr ; X86-64-NEXT: store ptr [[A]], ptr @__safestack_unsafe_stack_ptr, align 8 ; X86-64-NEXT: store ptr [[A]], ptr [[UNSAFE_STACK_DYNAMIC_PTR]], align 8 ; X86-64-NEXT: [[CALL:%.*]] = call i32 @_setjmp(ptr @buf) #[[ATTR1:[0-9]+]] ; X86-64-NEXT: [[TMP6:%.*]] = load ptr, ptr [[UNSAFE_STACK_DYNAMIC_PTR]], align 8 ; X86-64-NEXT: store ptr [[TMP6]], ptr @__safestack_unsafe_stack_ptr, align 8 ; X86-64-NEXT: call void @funcall(ptr [[A]]) ; X86-64-NEXT: store ptr [[UNSAFE_STACK_PTR]], ptr @__safestack_unsafe_stack_ptr, align 8 ; X86-64-NEXT: ret i32 0 ; entry: %a = alloca i32, i32 %size %call = call i32 @_setjmp(ptr @buf) returns_twice call void @funcall(ptr %a) ret i32 0 } declare i32 @_setjmp(ptr) declare void @funcall(ptr)