| ; RUN: opt < %s -passes='default<O2>' -S | FileCheck --check-prefixes=CHECK %s |
| target datalayout = "p:64:64:64" |
| |
| %swift.async_func_pointer = type <{ i32, i32 }> |
| %swift.context = type { %swift.context*, void (%swift.context*)* } |
| |
| @repoTU = global %swift.async_func_pointer <{ i32 trunc (i64 sub (i64 ptrtoint (void (%swift.context*)* @repo to i64), i64 ptrtoint (%swift.async_func_pointer* @repoTU to i64)) to i32), i32 16 }>, align 8 |
| |
| declare swifttailcc void @callee.0(%swift.context* swiftasync, i8*, i64, i64) |
| |
| define internal swifttailcc void @callee(i8* %0, i64 %1, i64 %2, %swift.context* %3) { |
| entry: |
| musttail call swifttailcc void @callee.0(%swift.context* swiftasync %3, i8* %0, i64 %1, i64 %2) |
| ret void |
| } |
| |
| define swifttailcc void @repo(%swift.context* swiftasync %0) { |
| entry: |
| %1 = alloca %swift.context*, align 8 |
| %2 = bitcast %swift.context* %0 to <{ %swift.context*, void (%swift.context*)* }>* |
| %3 = call token @llvm.coro.id.async(i32 16, i32 16, i32 0, i8* bitcast (%swift.async_func_pointer* @repoTU to i8*)) |
| %4 = call i8* @llvm.coro.begin(token %3, i8* null) |
| store %swift.context* %0, %swift.context** %1, align 8 |
| |
| ; This context.addr is the address in the frame of the first partial function after splitting. |
| %5 = call i8** @llvm.swift.async.context.addr() |
| store i8* null, i8** %5, align 8 |
| |
| %6 = call i8* @llvm.coro.async.resume() |
| %7 = call { i8* } (i32, i8*, i8*, ...) @llvm.coro.suspend.async.sl_p0i8s(i32 0, |
| i8* %6, |
| i8* bitcast (i8* (i8*)* @__swift_async_resume_get_context to i8*), |
| i8* bitcast (void (i8*, i64, i64, %swift.context*)* @callee to i8*), |
| i8* %6, i64 0, i64 0, %swift.context* %0) |
| %8 = load %swift.context*, %swift.context** %1, align 8 |
| %9 = bitcast %swift.context* %8 to <{ %swift.context*, void (%swift.context*)* }>* |
| %10 = getelementptr inbounds <{ %swift.context*, void (%swift.context*)* }>, <{ %swift.context*, void (%swift.context*)* }>* %9, i32 0, i32 1 |
| %11 = load void (%swift.context*)*, void (%swift.context*)** %10, align 8 |
| %12 = load %swift.context*, %swift.context** %1, align 8 |
| %13 = bitcast void (%swift.context*)* %11 to i8* |
| |
| ; This context.addr is the address in the frame of the second partial function after splitting. |
| ; It is not valid to CSE it with the previous call. |
| %14 = call i8** @llvm.swift.async.context.addr() |
| store i8* %13, i8** %14, align 8 |
| |
| %15 = call i1 (i8*, i1, ...) @llvm.coro.end.async(i8* %4, i1 false, void (i8*, %swift.context*)* @repo.0, i8* %13, %swift.context* %12) |
| unreachable |
| } |
| |
| ; Make sure we don't CSE the llvm.swift.async.context.addr calls |
| ; CHECK: define swifttailcc void @repo |
| ; CHECK: call i8** @llvm.swift.async.context.addr() |
| |
| ; CHECK: define {{.*}}swifttailcc void @repoTY0_ |
| ; CHECK: call i8** @llvm.swift.async.context.addr() |
| |
| define internal swifttailcc void @repo.0(i8* %0, %swift.context* %1) #1 { |
| entry: |
| %2 = bitcast i8* %0 to void (%swift.context*)* |
| musttail call swifttailcc void %2(%swift.context* swiftasync %1) |
| ret void |
| } |
| |
| define linkonce_odr hidden i8* @__swift_async_resume_get_context(i8* %0) #1 { |
| entry: |
| ret i8* %0 |
| } |
| |
| declare { i8* } @llvm.coro.suspend.async.sl_p0i8s(i32, i8*, i8*, ...) #1 |
| declare token @llvm.coro.id.async(i32, i32, i32, i8*) #1 |
| declare i8* @llvm.coro.begin(token, i8* writeonly) #1 |
| declare i1 @llvm.coro.end.async(i8*, i1, ...) #1 |
| declare i8* @llvm.coro.async.resume() #1 |
| declare i8** @llvm.swift.async.context.addr() #1 |
| |
| attributes #1 = { nounwind } |