aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--clang/docs/OpenMPSupport.rst2
-rw-r--r--clang/docs/ReleaseNotes.rst1
-rw-r--r--clang/lib/Sema/SemaOpenMP.cpp126
-rw-r--r--clang/test/OpenMP/target_default_codegen.cpp2020
-rw-r--r--clang/test/OpenMP/target_default_messages.cpp2
5 files changed, 2115 insertions, 36 deletions
diff --git a/clang/docs/OpenMPSupport.rst b/clang/docs/OpenMPSupport.rst
index 10a8d09..f7e6061 100644
--- a/clang/docs/OpenMPSupport.rst
+++ b/clang/docs/OpenMPSupport.rst
@@ -559,7 +559,7 @@ implementation.
+-------------------------------------------------------------+---------------------------+---------------------------+--------------------------------------------------------------------------+
| Clarifications to Fortran map semantics | :none:`unclaimed` | :none:`unclaimed` | |
+-------------------------------------------------------------+---------------------------+---------------------------+--------------------------------------------------------------------------+
-| default clause at target construct | :part:`In Progress` | :none:`unclaimed` | |
+| default clause at target construct | :good:`done` | :none:`unclaimed` | https://github.com/llvm/llvm-project/pull/162910 |
+-------------------------------------------------------------+---------------------------+---------------------------+--------------------------------------------------------------------------+
| ref count update use_device_{ptr, addr} | :none:`unclaimed` | :none:`unclaimed` | |
+-------------------------------------------------------------+---------------------------+---------------------------+--------------------------------------------------------------------------+
diff --git a/clang/docs/ReleaseNotes.rst b/clang/docs/ReleaseNotes.rst
index 5b95b44..88a05af 100644
--- a/clang/docs/ReleaseNotes.rst
+++ b/clang/docs/ReleaseNotes.rst
@@ -725,6 +725,7 @@ OpenMP Support
- Added support for 'omp fuse' directive.
- Updated parsing and semantic analysis support for ``nowait`` clause to accept
optional argument in OpenMP >= 60.
+- Added support for ``default`` clause on ``target`` directive.
Improvements
^^^^^^^^^^^^
diff --git a/clang/lib/Sema/SemaOpenMP.cpp b/clang/lib/Sema/SemaOpenMP.cpp
index 465dab2..2ab2fd1 100644
--- a/clang/lib/Sema/SemaOpenMP.cpp
+++ b/clang/lib/Sema/SemaOpenMP.cpp
@@ -17319,45 +17319,101 @@ OMPClause *SemaOpenMP::ActOnOpenMPDefaultClause(
<< getOpenMPClauseNameForDiag(OMPC_default);
return nullptr;
}
-
- switch (M) {
- case OMP_DEFAULT_none:
- DSAStack->setDefaultDSANone(MLoc);
- break;
- case OMP_DEFAULT_shared:
- DSAStack->setDefaultDSAShared(MLoc);
- break;
- case OMP_DEFAULT_firstprivate:
- DSAStack->setDefaultDSAFirstPrivate(MLoc);
- break;
- case OMP_DEFAULT_private:
- DSAStack->setDefaultDSAPrivate(MLoc);
- break;
- default:
- llvm_unreachable("DSA unexpected in OpenMP default clause");
- }
-
- switch (VCKind) {
- case OMPC_DEFAULT_VC_aggregate:
- DSAStack->setDefaultDSAVCAggregate(VCKindLoc);
- break;
- case OMPC_DEFAULT_VC_all:
- DSAStack->setDefaultDSAVCAll(VCKindLoc);
- break;
- case OMPC_DEFAULT_VC_allocatable:
- DSAStack->setDefaultDSAVCAllocatable(VCKindLoc);
- break;
- case OMPC_DEFAULT_VC_pointer:
- DSAStack->setDefaultDSAVCPointer(VCKindLoc);
- break;
- case OMPC_DEFAULT_VC_scalar:
- DSAStack->setDefaultDSAVCScalar(VCKindLoc);
- break;
- default:
+ if (VCKind == OMPC_DEFAULT_VC_unknown) {
Diag(VCKindLoc, diag::err_omp_default_vc)
<< getOpenMPSimpleClauseTypeName(OMPC_default, unsigned(M));
+ return nullptr;
}
+ bool IsTargetDefault =
+ getLangOpts().OpenMP >= 60 &&
+ isOpenMPTargetExecutionDirective(DSAStack->getCurrentDirective());
+
+ // OpenMP 6.0, page 224, lines 3-4 default Clause, Semantics
+ // If data-sharing-attribute is shared then the clause has no effect
+ // on a target construct;
+ if (IsTargetDefault && M == OMP_DEFAULT_shared)
+ return nullptr;
+
+ auto SetDefaultClauseAttrs = [&](llvm::omp::DefaultKind M,
+ OpenMPDefaultClauseVariableCategory VCKind) {
+ OpenMPDefaultmapClauseModifier DefMapMod;
+ OpenMPDefaultmapClauseKind DefMapKind;
+ // default data-sharing-attribute
+ switch (M) {
+ case OMP_DEFAULT_none:
+ if (IsTargetDefault)
+ DefMapMod = OMPC_DEFAULTMAP_MODIFIER_none;
+ else
+ DSAStack->setDefaultDSANone(MLoc);
+ break;
+ case OMP_DEFAULT_firstprivate:
+ if (IsTargetDefault)
+ DefMapMod = OMPC_DEFAULTMAP_MODIFIER_firstprivate;
+ else
+ DSAStack->setDefaultDSAFirstPrivate(MLoc);
+ break;
+ case OMP_DEFAULT_private:
+ if (IsTargetDefault)
+ DefMapMod = OMPC_DEFAULTMAP_MODIFIER_private;
+ else
+ DSAStack->setDefaultDSAPrivate(MLoc);
+ break;
+ case OMP_DEFAULT_shared:
+ assert(!IsTargetDefault && "DSA shared invalid with target directive");
+ DSAStack->setDefaultDSAShared(MLoc);
+ break;
+ default:
+ llvm_unreachable("unexpected DSA in OpenMP default clause");
+ }
+ // default variable-category
+ switch (VCKind) {
+ case OMPC_DEFAULT_VC_aggregate:
+ if (IsTargetDefault)
+ DefMapKind = OMPC_DEFAULTMAP_aggregate;
+ else
+ DSAStack->setDefaultDSAVCAggregate(VCKindLoc);
+ break;
+ case OMPC_DEFAULT_VC_pointer:
+ if (IsTargetDefault)
+ DefMapKind = OMPC_DEFAULTMAP_pointer;
+ else
+ DSAStack->setDefaultDSAVCPointer(VCKindLoc);
+ break;
+ case OMPC_DEFAULT_VC_scalar:
+ if (IsTargetDefault)
+ DefMapKind = OMPC_DEFAULTMAP_scalar;
+ else
+ DSAStack->setDefaultDSAVCScalar(VCKindLoc);
+ break;
+ case OMPC_DEFAULT_VC_all:
+ if (IsTargetDefault)
+ DefMapKind = OMPC_DEFAULTMAP_all;
+ else
+ DSAStack->setDefaultDSAVCAll(VCKindLoc);
+ break;
+ default:
+ llvm_unreachable("unexpected variable category in OpenMP default clause");
+ }
+ // OpenMP 6.0, page 224, lines 4-5 default Clause, Semantics
+ // otherwise, its effect on a target construct is equivalent to
+ // specifying the defaultmap clause with the same data-sharing-attribute
+ // and variable-category.
+ //
+ // If earlier than OpenMP 6.0, or not a target directive, the default DSA
+ // is/was set as before.
+ if (IsTargetDefault) {
+ if (DefMapKind == OMPC_DEFAULTMAP_all) {
+ DSAStack->setDefaultDMAAttr(DefMapMod, OMPC_DEFAULTMAP_aggregate, MLoc);
+ DSAStack->setDefaultDMAAttr(DefMapMod, OMPC_DEFAULTMAP_scalar, MLoc);
+ DSAStack->setDefaultDMAAttr(DefMapMod, OMPC_DEFAULTMAP_pointer, MLoc);
+ } else {
+ DSAStack->setDefaultDMAAttr(DefMapMod, DefMapKind, MLoc);
+ }
+ }
+ };
+
+ SetDefaultClauseAttrs(M, VCKind);
return new (getASTContext())
OMPDefaultClause(M, MLoc, VCKind, VCKindLoc, StartLoc, LParenLoc, EndLoc);
}
diff --git a/clang/test/OpenMP/target_default_codegen.cpp b/clang/test/OpenMP/target_default_codegen.cpp
new file mode 100644
index 0000000..eadd0e5
--- /dev/null
+++ b/clang/test/OpenMP/target_default_codegen.cpp
@@ -0,0 +1,2020 @@
+// NOTE: Assertions have been autogenerated by utils/update_cc_test_checks.py UTC_ARGS: --include-generated-funcs --replace-value-regex "__omp_offloading_[0-9a-z]+_[0-9a-z]+" "reduction_size[.].+[.]" "pl_cond[.].+[.|,]" --prefix-filecheck-ir-name _ --version 5
+// expected-no-diagnostics
+
+// RUN: %clang_cc1 -no-enable-noundef-analysis -verify -Wno-vla -fopenmp -fopenmp-version=60 -fopenmp-targets=powerpc64le-ibm-linux-gnu -x c++ -triple powerpc64le-unknown-unknown -emit-llvm %s -o - | FileCheck -allow-deprecated-dag-overlap %s --check-prefix CK-64
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp -fopenmp-version=60 -fopenmp-targets=powerpc64le-ibm-linux-gnu -x c++ -std=c++11 -triple powerpc64le-unknown-unknown -emit-pch -o %t %s
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp -fopenmp-version=60 -fopenmp-targets=powerpc64le-ibm-linux-gnu -x c++ -triple powerpc64le-unknown-unknown -std=c++11 -include-pch %t -verify -Wno-vla %s -emit-llvm -o - | FileCheck -allow-deprecated-dag-overlap %s --check-prefix CK-64
+// RUN: %clang_cc1 -no-enable-noundef-analysis -verify -Wno-vla -fopenmp -fopenmp-version=60 -fopenmp-targets=i386-pc-linux-gnu -x c++ -triple i386-unknown-unknown -emit-llvm %s -o - | FileCheck -allow-deprecated-dag-overlap %s --check-prefix CK-32
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp -fopenmp-version=60 -fopenmp-targets=i386-pc-linux-gnu -x c++ -std=c++11 -triple i386-unknown-unknown -emit-pch -o %t %s
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp -fopenmp-version=60 -fopenmp-targets=i386-pc-linux-gnu -x c++ -triple i386-unknown-unknown -std=c++11 -include-pch %t -verify -Wno-vla %s -emit-llvm -o - | FileCheck -allow-deprecated-dag-overlap %s --check-prefix CK-32
+
+// RUN: %clang_cc1 -no-enable-noundef-analysis -verify -Wno-vla -fopenmp-simd -fopenmp-version=60 -fopenmp-targets=powerpc64le-ibm-linux-gnu -x c++ -triple powerpc64le-unknown-unknown -emit-llvm %s -o - | FileCheck -allow-deprecated-dag-overlap --check-prefix SIMD-ONLY-64 %s
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp-simd -fopenmp-version=60 -fopenmp-targets=powerpc64le-ibm-linux-gnu -x c++ -std=c++11 -triple powerpc64le-unknown-unknown -emit-pch -o %t %s
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp-simd -fopenmp-version=60 -fopenmp-targets=powerpc64le-ibm-linux-gnu -x c++ -triple powerpc64le-unknown-unknown -std=c++11 -include-pch %t -verify -Wno-vla %s -emit-llvm -o - | FileCheck -allow-deprecated-dag-overlap --check-prefix SIMD-ONLY-64 %s
+// RUN: %clang_cc1 -no-enable-noundef-analysis -verify -Wno-vla -fopenmp-simd -fopenmp-version=60 -fopenmp-targets=i386-pc-linux-gnu -x c++ -triple i386-unknown-unknown -emit-llvm %s -o - | FileCheck -allow-deprecated-dag-overlap --check-prefix SIMD-ONLY-32 %s
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp-simd -fopenmp-version=60 -fopenmp-targets=i386-pc-linux-gnu -x c++ -std=c++11 -triple i386-unknown-unknown -emit-pch -o %t %s
+// RUN: %clang_cc1 -no-enable-noundef-analysis -fopenmp-simd -fopenmp-version=60 -fopenmp-targets=i386-pc-linux-gnu -x c++ -triple i386-unknown-unknown -std=c++11 -include-pch %t -verify -Wno-vla %s -emit-llvm -o - | FileCheck -allow-deprecated-dag-overlap --check-prefix SIMD-ONLY-32 %s
+
+#ifndef HEADER
+#define HEADER
+void foo1(int a) {
+ double d = (double)a;
+
+ #pragma omp target default(private: scalar)
+ {
+ d += 1.0;
+ }
+}
+
+void foo2() {
+ int pvtArr[10];
+
+ #pragma omp target default(private: aggregate)
+ {
+ pvtArr[5]++;
+ }
+}
+
+void foo3() {
+ int *pa;
+
+ #pragma omp target default(private: pointer)
+ {
+ pa[50]++;
+ }
+}
+
+// Specified variable-category doesn't apply to referenced variable, so
+// normal implicitly determined data-sharing applies.
+void foo4() {
+ int p;
+
+ #pragma omp target default(private: pointer)
+ {
+ p++;
+ }
+}
+
+// Verify default clause with variable-category 'all' is equivalent to no
+// variable-category. IR checks generated with 'all' but test runs without
+// variable-category.
+void foo5(int a) {
+ double d = (double)a;
+ int pvtArr[10];
+ int *pa;
+
+ #pragma omp target default(private)
+ {
+ d += 1.0;
+ pvtArr[5]++;
+ pa[50]++;
+ }
+}
+
+// Verify default clause with 'shared' DSA is ignored. This makes it
+// equivalent to target with no default clause. IR checks generated with
+// no default clause but test runs with default 'shared'.
+void foo6(int a) {
+ double d = (double)a;
+ int pvtArr[10];
+ int *pa;
+
+ #pragma omp target default(shared)
+ {
+ d += 1.0;
+ pvtArr[5]++;
+ pa[50]++;
+ }
+}
+
+// Verify default clause with 'firstprivate' DSA is equivalent to specifying
+// defaultmap with 'firstprivate'. IR checks generated with
+// defaultmap(firstprivate) but test runs with default(firstprivate).
+void foo7(int a) {
+ double d = (double)a;
+ int pvtArr[10];
+ int *pa;
+
+ #pragma omp target default(firstprivate)
+ {
+ d += 1.0;
+ pvtArr[5]++;
+ pa[50]++;
+ }
+}
+
+// Verify 'default' clause on a combined 'target' directive is equivalent to
+// specifying its constituent directives with 'default' clauses. IR checks
+// generated with constituent directives but test runs with combined
+// directive.
+void foo8() {
+ int x = 0;
+ #pragma omp target teams distribute parallel for default(firstprivate) firstprivate(x)
+ for (int i=0; i<10; i++)
+ x += 1;
+}
+#endif // HEADER
+// CK-64-LABEL: define dso_local void @_Z4foo1i(
+// CK-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0:[0-9]+]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-64-NEXT: [[D_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// CK-64-NEXT: store double [[TMP1]], ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = load i64, ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP3]], align 8
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP4]], align 8
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP5]], align 8
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP8]], align 4
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 1, ptr [[TMP9]], align 4
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP6]], ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP7]], ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes, ptr [[TMP13]], align 8
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP14]], align 8
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP15]], align 8
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 0, ptr [[TMP16]], align 8
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP17]], align 8
+// CK-64-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP18]], align 4
+// CK-64-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP19]], align 4
+// CK-64-NEXT: [[TMP20:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP20]], align 4
+// CK-64-NEXT: [[TMP21:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1:[0-9]+]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo1i_l23.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP22:%.*]] = icmp ne i32 [[TMP21]], 0
+// CK-64-NEXT: br i1 [[TMP22]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo1i_l23(i64 [[TMP2]]) #[[ATTR2:[0-9]+]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo1i_l23(
+// CK-64-SAME: i64 [[D:%.*]]) #[[ATTR1:[0-9]+]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[D_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[D1:%.*]] = alloca double, align 8
+// CK-64-NEXT: store i64 [[D]], ptr [[D_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load double, ptr [[D1]], align 8
+// CK-64-NEXT: [[ADD:%.*]] = fadd double [[TMP0]], 1.000000e+00
+// CK-64-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define dso_local void @_Z4foo2v(
+// CK-64-SAME: ) #[[ATTR0]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP0]], align 8
+// CK-64-NEXT: [[TMP1:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP1]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP2]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP5]], align 4
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 1, ptr [[TMP6]], align 4
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP3]], ptr [[TMP7]], align 8
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP4]], ptr [[TMP8]], align 8
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes.1, ptr [[TMP9]], align 8
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes.2, ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 0, ptr [[TMP13]], align 8
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP14]], align 8
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP15]], align 4
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP16]], align 4
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP17]], align 4
+// CK-64-NEXT: [[TMP18:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo2v_l32.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP19:%.*]] = icmp ne i32 [[TMP18]], 0
+// CK-64-NEXT: br i1 [[TMP19]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo2v_l32(ptr [[PVTARR]]) #[[ATTR2]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo2v_l32(
+// CK-64-SAME: ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[PVTARR1:%.*]] = alloca [10 x i32], align 4
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 8, !nonnull [[META19:![0-9]+]], !align [[META20:![0-9]+]]
+// CK-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR1]], i64 0, i64 5
+// CK-64-NEXT: [[TMP1:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP1]], 1
+// CK-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define dso_local void @_Z4foo3v(
+// CK-64-SAME: ) #[[ATTR0]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PA]], align 8
+// CK-64-NEXT: [[TMP1:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store ptr [[TMP0]], ptr [[TMP1]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store ptr [[TMP0]], ptr [[TMP2]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP3]], align 8
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP6]], align 4
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 1, ptr [[TMP7]], align 4
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP4]], ptr [[TMP8]], align 8
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP5]], ptr [[TMP9]], align 8
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes.3, ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes.4, ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP13]], align 8
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 0, ptr [[TMP14]], align 8
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP15]], align 8
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP16]], align 4
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP17]], align 4
+// CK-64-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP18]], align 4
+// CK-64-NEXT: [[TMP19:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo3v_l41.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP20:%.*]] = icmp ne i32 [[TMP19]], 0
+// CK-64-NEXT: br i1 [[TMP20]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo3v_l41(ptr [[TMP0]]) #[[ATTR2]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo3v_l41(
+// CK-64-SAME: ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[PA1:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PA1]], align 8
+// CK-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds i32, ptr [[TMP0]], i64 50
+// CK-64-NEXT: [[TMP1:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP1]], 1
+// CK-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define dso_local void @_Z4foo4v(
+// CK-64-SAME: ) #[[ATTR0]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[P:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[P_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[P]], align 4
+// CK-64-NEXT: store i32 [[TMP0]], ptr [[P_CASTED]], align 4
+// CK-64-NEXT: [[TMP1:%.*]] = load i64, ptr [[P_CASTED]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP1]], ptr [[TMP2]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP1]], ptr [[TMP3]], align 8
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP4]], align 8
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP7]], align 4
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 1, ptr [[TMP8]], align 4
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP5]], ptr [[TMP9]], align 8
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP6]], ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes.5, ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes.6, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP13]], align 8
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP14]], align 8
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 0, ptr [[TMP15]], align 8
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP16]], align 8
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP17]], align 4
+// CK-64-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP18]], align 4
+// CK-64-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP19]], align 4
+// CK-64-NEXT: [[TMP20:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo4v_l52.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP21:%.*]] = icmp ne i32 [[TMP20]], 0
+// CK-64-NEXT: br i1 [[TMP21]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo4v_l52(i64 [[TMP1]]) #[[ATTR2]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo4v_l52(
+// CK-64-SAME: i64 [[P:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[P_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: store i64 [[P]], ptr [[P_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[P_ADDR]], align 4
+// CK-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP0]], 1
+// CK-64-NEXT: store i32 [[INC]], ptr [[P_ADDR]], align 4
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define dso_local void @_Z4foo5i(
+// CK-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[D_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// CK-64-NEXT: store double [[TMP1]], ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = load i64, ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA]], align 8
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP4]], align 8
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP5]], align 8
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP6]], align 8
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 1
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP7]], align 8
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 1
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP8]], align 8
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 1
+// CK-64-NEXT: store ptr null, ptr [[TMP9]], align 8
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP3]], ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP3]], ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 2
+// CK-64-NEXT: store ptr null, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP15]], align 4
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 3, ptr [[TMP16]], align 4
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP13]], ptr [[TMP17]], align 8
+// CK-64-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP14]], ptr [[TMP18]], align 8
+// CK-64-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes.7, ptr [[TMP19]], align 8
+// CK-64-NEXT: [[TMP20:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes.8, ptr [[TMP20]], align 8
+// CK-64-NEXT: [[TMP21:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP21]], align 8
+// CK-64-NEXT: [[TMP22:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP22]], align 8
+// CK-64-NEXT: [[TMP23:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 0, ptr [[TMP23]], align 8
+// CK-64-NEXT: [[TMP24:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP24]], align 8
+// CK-64-NEXT: [[TMP25:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP25]], align 4
+// CK-64-NEXT: [[TMP26:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP26]], align 4
+// CK-64-NEXT: [[TMP27:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP27]], align 4
+// CK-64-NEXT: [[TMP28:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo5i_l66.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP29:%.*]] = icmp ne i32 [[TMP28]], 0
+// CK-64-NEXT: br i1 [[TMP29]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo5i_l66(i64 [[TMP2]], ptr [[PVTARR]], ptr [[TMP3]]) #[[ATTR2]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo5i_l66(
+// CK-64-SAME: i64 [[D:%.*]], ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]], ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[D_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[D1:%.*]] = alloca double, align 8
+// CK-64-NEXT: [[PVTARR2:%.*]] = alloca [10 x i32], align 4
+// CK-64-NEXT: [[PA3:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: store i64 [[D]], ptr [[D_ADDR]], align 8
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 8
+// CK-64-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 8, !nonnull [[META19]], !align [[META20]]
+// CK-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D1]], align 8
+// CK-64-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// CK-64-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// CK-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR2]], i64 0, i64 5
+// CK-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// CK-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA3]], align 8
+// CK-64-NEXT: [[ARRAYIDX4:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i64 50
+// CK-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX4]], align 4
+// CK-64-NEXT: [[INC5:%.*]] = add nsw i32 [[TMP4]], 1
+// CK-64-NEXT: store i32 [[INC5]], ptr [[ARRAYIDX4]], align 4
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define dso_local void @_Z4foo6i(
+// CK-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[D_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// CK-64-NEXT: store double [[TMP1]], ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = load i64, ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA]], align 8
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP4]], align 8
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP5]], align 8
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP6]], align 8
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 1
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP7]], align 8
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 1
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP8]], align 8
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 1
+// CK-64-NEXT: store ptr null, ptr [[TMP9]], align 8
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP3]], ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP3]], ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 2
+// CK-64-NEXT: store ptr null, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP15]], align 4
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 3, ptr [[TMP16]], align 4
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP13]], ptr [[TMP17]], align 8
+// CK-64-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP14]], ptr [[TMP18]], align 8
+// CK-64-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes.9, ptr [[TMP19]], align 8
+// CK-64-NEXT: [[TMP20:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes.10, ptr [[TMP20]], align 8
+// CK-64-NEXT: [[TMP21:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP21]], align 8
+// CK-64-NEXT: [[TMP22:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP22]], align 8
+// CK-64-NEXT: [[TMP23:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 0, ptr [[TMP23]], align 8
+// CK-64-NEXT: [[TMP24:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP24]], align 8
+// CK-64-NEXT: [[TMP25:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP25]], align 4
+// CK-64-NEXT: [[TMP26:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP26]], align 4
+// CK-64-NEXT: [[TMP27:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP27]], align 4
+// CK-64-NEXT: [[TMP28:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo6i_l82.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP29:%.*]] = icmp ne i32 [[TMP28]], 0
+// CK-64-NEXT: br i1 [[TMP29]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo6i_l82(i64 [[TMP2]], ptr [[PVTARR]], ptr [[TMP3]]) #[[ATTR2]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo6i_l82(
+// CK-64-SAME: i64 [[D:%.*]], ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]], ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[D_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: store i64 [[D]], ptr [[D_ADDR]], align 8
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 8
+// CK-64-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 8, !nonnull [[META19]], !align [[META20]]
+// CK-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D_ADDR]], align 8
+// CK-64-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// CK-64-NEXT: store double [[ADD]], ptr [[D_ADDR]], align 8
+// CK-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[TMP0]], i64 0, i64 5
+// CK-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// CK-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA_ADDR]], align 8
+// CK-64-NEXT: [[ARRAYIDX1:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i64 50
+// CK-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX1]], align 4
+// CK-64-NEXT: [[INC2:%.*]] = add nsw i32 [[TMP4]], 1
+// CK-64-NEXT: store i32 [[INC2]], ptr [[ARRAYIDX1]], align 4
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define dso_local void @_Z4foo7i(
+// CK-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[D_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [3 x ptr], align 8
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// CK-64-NEXT: store double [[TMP1]], ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = load i64, ptr [[D_CASTED]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA]], align 8
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP4]], align 8
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP2]], ptr [[TMP5]], align 8
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP6]], align 8
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 1
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP7]], align 8
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 1
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[TMP8]], align 8
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 1
+// CK-64-NEXT: store ptr null, ptr [[TMP9]], align 8
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP3]], ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP3]], ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 2
+// CK-64-NEXT: store ptr null, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP15]], align 4
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 3, ptr [[TMP16]], align 4
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP13]], ptr [[TMP17]], align 8
+// CK-64-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP14]], ptr [[TMP18]], align 8
+// CK-64-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes.11, ptr [[TMP19]], align 8
+// CK-64-NEXT: [[TMP20:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes.12, ptr [[TMP20]], align 8
+// CK-64-NEXT: [[TMP21:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP21]], align 8
+// CK-64-NEXT: [[TMP22:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP22]], align 8
+// CK-64-NEXT: [[TMP23:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 0, ptr [[TMP23]], align 8
+// CK-64-NEXT: [[TMP24:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP24]], align 8
+// CK-64-NEXT: [[TMP25:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP25]], align 4
+// CK-64-NEXT: [[TMP26:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP26]], align 4
+// CK-64-NEXT: [[TMP27:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP27]], align 4
+// CK-64-NEXT: [[TMP28:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo7i_l98.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP29:%.*]] = icmp ne i32 [[TMP28]], 0
+// CK-64-NEXT: br i1 [[TMP29]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo7i_l98(i64 [[TMP2]], ptr [[PVTARR]], ptr [[TMP3]]) #[[ATTR2]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo7i_l98(
+// CK-64-SAME: i64 [[D:%.*]], ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]], ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[D_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[PVTARR1:%.*]] = alloca [10 x i32], align 4
+// CK-64-NEXT: store i64 [[D]], ptr [[D_ADDR]], align 8
+// CK-64-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 8
+// CK-64-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 8, !nonnull [[META19]], !align [[META20]]
+// CK-64-NEXT: call void @llvm.memcpy.p0.p0.i64(ptr align 4 [[PVTARR1]], ptr align 4 [[TMP0]], i64 40, i1 false)
+// CK-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D_ADDR]], align 8
+// CK-64-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// CK-64-NEXT: store double [[ADD]], ptr [[D_ADDR]], align 8
+// CK-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR1]], i64 0, i64 5
+// CK-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// CK-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA_ADDR]], align 8
+// CK-64-NEXT: [[ARRAYIDX2:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i64 50
+// CK-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX2]], align 4
+// CK-64-NEXT: [[INC3:%.*]] = add nsw i32 [[TMP4]], 1
+// CK-64-NEXT: store i32 [[INC3]], ptr [[ARRAYIDX2]], align 4
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define dso_local void @_Z4foo8v(
+// CK-64-SAME: ) #[[ATTR0]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[X:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[X_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 8
+// CK-64-NEXT: [[TMP:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-64-NEXT: store i32 0, ptr [[X]], align 4
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[X]], align 4
+// CK-64-NEXT: store i32 [[TMP0]], ptr [[X_CASTED]], align 4
+// CK-64-NEXT: [[TMP1:%.*]] = load i64, ptr [[X_CASTED]], align 8
+// CK-64-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP1]], ptr [[TMP2]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: store i64 [[TMP1]], ptr [[TMP3]], align 8
+// CK-64-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i64 0, i64 0
+// CK-64-NEXT: store ptr null, ptr [[TMP4]], align 8
+// CK-64-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP6:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-64-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-64-NEXT: store i32 3, ptr [[TMP7]], align 4
+// CK-64-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-64-NEXT: store i32 1, ptr [[TMP8]], align 4
+// CK-64-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-64-NEXT: store ptr [[TMP5]], ptr [[TMP9]], align 8
+// CK-64-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-64-NEXT: store ptr [[TMP6]], ptr [[TMP10]], align 8
+// CK-64-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-64-NEXT: store ptr @.offload_sizes.13, ptr [[TMP11]], align 8
+// CK-64-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-64-NEXT: store ptr @.offload_maptypes.14, ptr [[TMP12]], align 8
+// CK-64-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-64-NEXT: store ptr null, ptr [[TMP13]], align 8
+// CK-64-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-64-NEXT: store ptr null, ptr [[TMP14]], align 8
+// CK-64-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-64-NEXT: store i64 10, ptr [[TMP15]], align 8
+// CK-64-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-64-NEXT: store i64 0, ptr [[TMP16]], align 8
+// CK-64-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP17]], align 4
+// CK-64-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-64-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP18]], align 4
+// CK-64-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-64-NEXT: store i32 0, ptr [[TMP19]], align 4
+// CK-64-NEXT: [[TMP20:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 0, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.region_id, ptr [[KERNEL_ARGS]])
+// CK-64-NEXT: [[TMP21:%.*]] = icmp ne i32 [[TMP20]], 0
+// CK-64-NEXT: br i1 [[TMP21]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-64: [[OMP_OFFLOAD_FAILED]]:
+// CK-64-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112(i64 [[TMP1]]) #[[ATTR2]]
+// CK-64-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-64: [[OMP_OFFLOAD_CONT]]:
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112(
+// CK-64-SAME: i64 [[X:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[X_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[X_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: store i64 [[X]], ptr [[X_ADDR]], align 8
+// CK-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[X_ADDR]], align 4
+// CK-64-NEXT: store i32 [[TMP0]], ptr [[X_CASTED]], align 4
+// CK-64-NEXT: [[TMP1:%.*]] = load i64, ptr [[X_CASTED]], align 8
+// CK-64-NEXT: call void (ptr, i32, ptr, ...) @__kmpc_fork_teams(ptr @[[GLOB1]], i32 1, ptr @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined, i64 [[TMP1]])
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined(
+// CK-64-SAME: ptr noalias [[DOTGLOBAL_TID_:%.*]], ptr noalias [[DOTBOUND_TID_:%.*]], i64 [[X:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[DOTGLOBAL_TID__ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[DOTBOUND_TID__ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[X_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOMP_IV:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[TMP:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_COMB_LB:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_COMB_UB:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_STRIDE:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_IS_LAST:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[I:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[X_CASTED:%.*]] = alloca i64, align 8
+// CK-64-NEXT: store ptr [[DOTGLOBAL_TID_]], ptr [[DOTGLOBAL_TID__ADDR]], align 8
+// CK-64-NEXT: store ptr [[DOTBOUND_TID_]], ptr [[DOTBOUND_TID__ADDR]], align 8
+// CK-64-NEXT: store i64 [[X]], ptr [[X_ADDR]], align 8
+// CK-64-NEXT: store i32 0, ptr [[DOTOMP_COMB_LB]], align 4
+// CK-64-NEXT: store i32 9, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-64-NEXT: store i32 1, ptr [[DOTOMP_STRIDE]], align 4
+// CK-64-NEXT: store i32 0, ptr [[DOTOMP_IS_LAST]], align 4
+// CK-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[DOTGLOBAL_TID__ADDR]], align 8
+// CK-64-NEXT: [[TMP1:%.*]] = load i32, ptr [[TMP0]], align 4
+// CK-64-NEXT: call void @__kmpc_for_static_init_4(ptr @[[GLOB2:[0-9]+]], i32 [[TMP1]], i32 92, ptr [[DOTOMP_IS_LAST]], ptr [[DOTOMP_COMB_LB]], ptr [[DOTOMP_COMB_UB]], ptr [[DOTOMP_STRIDE]], i32 1, i32 1)
+// CK-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-64-NEXT: [[CMP:%.*]] = icmp sgt i32 [[TMP2]], 9
+// CK-64-NEXT: br i1 [[CMP]], label %[[COND_TRUE:.*]], label %[[COND_FALSE:.*]]
+// CK-64: [[COND_TRUE]]:
+// CK-64-NEXT: br label %[[COND_END:.*]]
+// CK-64: [[COND_FALSE]]:
+// CK-64-NEXT: [[TMP3:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-64-NEXT: br label %[[COND_END]]
+// CK-64: [[COND_END]]:
+// CK-64-NEXT: [[COND:%.*]] = phi i32 [ 9, %[[COND_TRUE]] ], [ [[TMP3]], %[[COND_FALSE]] ]
+// CK-64-NEXT: store i32 [[COND]], ptr [[DOTOMP_COMB_UB]], align 4
+// CK-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[DOTOMP_COMB_LB]], align 4
+// CK-64-NEXT: store i32 [[TMP4]], ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: br label %[[OMP_INNER_FOR_COND:.*]]
+// CK-64: [[OMP_INNER_FOR_COND]]:
+// CK-64-NEXT: [[TMP5:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: [[TMP6:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-64-NEXT: [[CMP1:%.*]] = icmp sle i32 [[TMP5]], [[TMP6]]
+// CK-64-NEXT: br i1 [[CMP1]], label %[[OMP_INNER_FOR_BODY:.*]], label %[[OMP_INNER_FOR_END:.*]]
+// CK-64: [[OMP_INNER_FOR_BODY]]:
+// CK-64-NEXT: [[TMP7:%.*]] = load i32, ptr [[DOTOMP_COMB_LB]], align 4
+// CK-64-NEXT: [[TMP8:%.*]] = zext i32 [[TMP7]] to i64
+// CK-64-NEXT: [[TMP9:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-64-NEXT: [[TMP10:%.*]] = zext i32 [[TMP9]] to i64
+// CK-64-NEXT: [[TMP11:%.*]] = load i32, ptr [[X_ADDR]], align 4
+// CK-64-NEXT: store i32 [[TMP11]], ptr [[X_CASTED]], align 4
+// CK-64-NEXT: [[TMP12:%.*]] = load i64, ptr [[X_CASTED]], align 8
+// CK-64-NEXT: call void (ptr, i32, ptr, ...) @__kmpc_fork_call(ptr @[[GLOB1]], i32 3, ptr @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined.omp_outlined, i64 [[TMP8]], i64 [[TMP10]], i64 [[TMP12]])
+// CK-64-NEXT: br label %[[OMP_INNER_FOR_INC:.*]]
+// CK-64: [[OMP_INNER_FOR_INC]]:
+// CK-64-NEXT: [[TMP13:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: [[TMP14:%.*]] = load i32, ptr [[DOTOMP_STRIDE]], align 4
+// CK-64-NEXT: [[ADD:%.*]] = add nsw i32 [[TMP13]], [[TMP14]]
+// CK-64-NEXT: store i32 [[ADD]], ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: br label %[[OMP_INNER_FOR_COND]]
+// CK-64: [[OMP_INNER_FOR_END]]:
+// CK-64-NEXT: br label %[[OMP_LOOP_EXIT:.*]]
+// CK-64: [[OMP_LOOP_EXIT]]:
+// CK-64-NEXT: call void @__kmpc_for_static_fini(ptr @[[GLOB2]], i32 [[TMP1]])
+// CK-64-NEXT: ret void
+//
+//
+// CK-64-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined.omp_outlined(
+// CK-64-SAME: ptr noalias [[DOTGLOBAL_TID_:%.*]], ptr noalias [[DOTBOUND_TID_:%.*]], i64 [[DOTPREVIOUS_LB_:%.*]], i64 [[DOTPREVIOUS_UB_:%.*]], i64 [[X:%.*]]) #[[ATTR1]] {
+// CK-64-NEXT: [[ENTRY:.*:]]
+// CK-64-NEXT: [[DOTGLOBAL_TID__ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[DOTBOUND_TID__ADDR:%.*]] = alloca ptr, align 8
+// CK-64-NEXT: [[DOTPREVIOUS_LB__ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTPREVIOUS_UB__ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[X_ADDR:%.*]] = alloca i64, align 8
+// CK-64-NEXT: [[DOTOMP_IV:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[TMP:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_LB:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_UB:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_STRIDE:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[DOTOMP_IS_LAST:%.*]] = alloca i32, align 4
+// CK-64-NEXT: [[I:%.*]] = alloca i32, align 4
+// CK-64-NEXT: store ptr [[DOTGLOBAL_TID_]], ptr [[DOTGLOBAL_TID__ADDR]], align 8
+// CK-64-NEXT: store ptr [[DOTBOUND_TID_]], ptr [[DOTBOUND_TID__ADDR]], align 8
+// CK-64-NEXT: store i64 [[DOTPREVIOUS_LB_]], ptr [[DOTPREVIOUS_LB__ADDR]], align 8
+// CK-64-NEXT: store i64 [[DOTPREVIOUS_UB_]], ptr [[DOTPREVIOUS_UB__ADDR]], align 8
+// CK-64-NEXT: store i64 [[X]], ptr [[X_ADDR]], align 8
+// CK-64-NEXT: store i32 0, ptr [[DOTOMP_LB]], align 4
+// CK-64-NEXT: store i32 9, ptr [[DOTOMP_UB]], align 4
+// CK-64-NEXT: [[TMP0:%.*]] = load i64, ptr [[DOTPREVIOUS_LB__ADDR]], align 8
+// CK-64-NEXT: [[CONV:%.*]] = trunc i64 [[TMP0]] to i32
+// CK-64-NEXT: [[TMP1:%.*]] = load i64, ptr [[DOTPREVIOUS_UB__ADDR]], align 8
+// CK-64-NEXT: [[CONV1:%.*]] = trunc i64 [[TMP1]] to i32
+// CK-64-NEXT: store i32 [[CONV]], ptr [[DOTOMP_LB]], align 4
+// CK-64-NEXT: store i32 [[CONV1]], ptr [[DOTOMP_UB]], align 4
+// CK-64-NEXT: store i32 1, ptr [[DOTOMP_STRIDE]], align 4
+// CK-64-NEXT: store i32 0, ptr [[DOTOMP_IS_LAST]], align 4
+// CK-64-NEXT: [[TMP2:%.*]] = load ptr, ptr [[DOTGLOBAL_TID__ADDR]], align 8
+// CK-64-NEXT: [[TMP3:%.*]] = load i32, ptr [[TMP2]], align 4
+// CK-64-NEXT: call void @__kmpc_for_static_init_4(ptr @[[GLOB3:[0-9]+]], i32 [[TMP3]], i32 34, ptr [[DOTOMP_IS_LAST]], ptr [[DOTOMP_LB]], ptr [[DOTOMP_UB]], ptr [[DOTOMP_STRIDE]], i32 1, i32 1)
+// CK-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[DOTOMP_UB]], align 4
+// CK-64-NEXT: [[CMP:%.*]] = icmp sgt i32 [[TMP4]], 9
+// CK-64-NEXT: br i1 [[CMP]], label %[[COND_TRUE:.*]], label %[[COND_FALSE:.*]]
+// CK-64: [[COND_TRUE]]:
+// CK-64-NEXT: br label %[[COND_END:.*]]
+// CK-64: [[COND_FALSE]]:
+// CK-64-NEXT: [[TMP5:%.*]] = load i32, ptr [[DOTOMP_UB]], align 4
+// CK-64-NEXT: br label %[[COND_END]]
+// CK-64: [[COND_END]]:
+// CK-64-NEXT: [[COND:%.*]] = phi i32 [ 9, %[[COND_TRUE]] ], [ [[TMP5]], %[[COND_FALSE]] ]
+// CK-64-NEXT: store i32 [[COND]], ptr [[DOTOMP_UB]], align 4
+// CK-64-NEXT: [[TMP6:%.*]] = load i32, ptr [[DOTOMP_LB]], align 4
+// CK-64-NEXT: store i32 [[TMP6]], ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: br label %[[OMP_INNER_FOR_COND:.*]]
+// CK-64: [[OMP_INNER_FOR_COND]]:
+// CK-64-NEXT: [[TMP7:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: [[TMP8:%.*]] = load i32, ptr [[DOTOMP_UB]], align 4
+// CK-64-NEXT: [[CMP2:%.*]] = icmp sle i32 [[TMP7]], [[TMP8]]
+// CK-64-NEXT: br i1 [[CMP2]], label %[[OMP_INNER_FOR_BODY:.*]], label %[[OMP_INNER_FOR_END:.*]]
+// CK-64: [[OMP_INNER_FOR_BODY]]:
+// CK-64-NEXT: [[TMP9:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: [[MUL:%.*]] = mul nsw i32 [[TMP9]], 1
+// CK-64-NEXT: [[ADD:%.*]] = add nsw i32 0, [[MUL]]
+// CK-64-NEXT: store i32 [[ADD]], ptr [[I]], align 4
+// CK-64-NEXT: [[TMP10:%.*]] = load i32, ptr [[X_ADDR]], align 4
+// CK-64-NEXT: [[ADD3:%.*]] = add nsw i32 [[TMP10]], 1
+// CK-64-NEXT: store i32 [[ADD3]], ptr [[X_ADDR]], align 4
+// CK-64-NEXT: br label %[[OMP_BODY_CONTINUE:.*]]
+// CK-64: [[OMP_BODY_CONTINUE]]:
+// CK-64-NEXT: br label %[[OMP_INNER_FOR_INC:.*]]
+// CK-64: [[OMP_INNER_FOR_INC]]:
+// CK-64-NEXT: [[TMP11:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: [[ADD4:%.*]] = add nsw i32 [[TMP11]], 1
+// CK-64-NEXT: store i32 [[ADD4]], ptr [[DOTOMP_IV]], align 4
+// CK-64-NEXT: br label %[[OMP_INNER_FOR_COND]]
+// CK-64: [[OMP_INNER_FOR_END]]:
+// CK-64-NEXT: br label %[[OMP_LOOP_EXIT:.*]]
+// CK-64: [[OMP_LOOP_EXIT]]:
+// CK-64-NEXT: call void @__kmpc_for_static_fini(ptr @[[GLOB3]], i32 [[TMP3]])
+// CK-64-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo1i(
+// CK-32-SAME: i32 [[A:%.*]]) #[[ATTR0:[0-9]+]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-32-NEXT: [[TMP1:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP1]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP3]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP6]], align 4
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 1, ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP4]], ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP5]], ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes, ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes, ptr [[TMP11]], align 4
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP12]], align 4
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP13]], align 4
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 0, ptr [[TMP14]], align 8
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP15]], align 8
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP16]], align 4
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP18]], align 4
+// CK-32-NEXT: [[TMP19:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1:[0-9]+]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo1i_l23.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP20:%.*]] = icmp ne i32 [[TMP19]], 0
+// CK-32-NEXT: br i1 [[TMP20]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo1i_l23(ptr [[D]]) #[[ATTR2:[0-9]+]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo1i_l23(
+// CK-32-SAME: ptr nonnull align 4 dereferenceable(8) [[D:%.*]]) #[[ATTR1:[0-9]+]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[D_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[D1:%.*]] = alloca double, align 8
+// CK-32-NEXT: store ptr [[D]], ptr [[D_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[D_ADDR]], align 4, !nonnull [[META20:![0-9]+]], !align [[META21:![0-9]+]]
+// CK-32-NEXT: [[TMP1:%.*]] = load double, ptr [[D1]], align 8
+// CK-32-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// CK-32-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo2v(
+// CK-32-SAME: ) #[[ATTR0]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: [[TMP0:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP0]], align 4
+// CK-32-NEXT: [[TMP1:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP1]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP5]], align 4
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 1, ptr [[TMP6]], align 4
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP3]], ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP4]], ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes.1, ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes.2, ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP11]], align 4
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP12]], align 4
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 0, ptr [[TMP13]], align 8
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP14]], align 8
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP15]], align 4
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP16]], align 4
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo2v_l32.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP19:%.*]] = icmp ne i32 [[TMP18]], 0
+// CK-32-NEXT: br i1 [[TMP19]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo2v_l32(ptr [[PVTARR]]) #[[ATTR2]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo2v_l32(
+// CK-32-SAME: ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PVTARR1:%.*]] = alloca [10 x i32], align 4
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 4, !nonnull [[META20]], !align [[META21]]
+// CK-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR1]], i32 0, i32 5
+// CK-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP1]], 1
+// CK-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo3v(
+// CK-32-SAME: ) #[[ATTR0]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PA]], align 4
+// CK-32-NEXT: [[TMP1:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[TMP0]], ptr [[TMP1]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[TMP0]], ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP3]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP6]], align 4
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 1, ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP4]], ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP5]], ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes.3, ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes.4, ptr [[TMP11]], align 4
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP12]], align 4
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP13]], align 4
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 0, ptr [[TMP14]], align 8
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP15]], align 8
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP16]], align 4
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP18]], align 4
+// CK-32-NEXT: [[TMP19:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo3v_l41.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP20:%.*]] = icmp ne i32 [[TMP19]], 0
+// CK-32-NEXT: br i1 [[TMP20]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo3v_l41(ptr [[TMP0]]) #[[ATTR2]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo3v_l41(
+// CK-32-SAME: ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PA1:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PA1]], align 4
+// CK-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds i32, ptr [[TMP0]], i32 50
+// CK-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP1]], 1
+// CK-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo4v(
+// CK-32-SAME: ) #[[ATTR0]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[P:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[P_CASTED:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[P]], align 4
+// CK-32-NEXT: store i32 [[TMP0]], ptr [[P_CASTED]], align 4
+// CK-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[P_CASTED]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store i32 [[TMP1]], ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store i32 [[TMP1]], ptr [[TMP3]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP4]], align 4
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 1, ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP5]], ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP6]], ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes.5, ptr [[TMP11]], align 4
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes.6, ptr [[TMP12]], align 4
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP13]], align 4
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP14]], align 4
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 0, ptr [[TMP15]], align 8
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP16]], align 8
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP18]], align 4
+// CK-32-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP19]], align 4
+// CK-32-NEXT: [[TMP20:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo4v_l52.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP21:%.*]] = icmp ne i32 [[TMP20]], 0
+// CK-32-NEXT: br i1 [[TMP21]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo4v_l52(i32 [[TMP1]]) #[[ATTR2]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo4v_l52(
+// CK-32-SAME: i32 [[P:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[P_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: store i32 [[P]], ptr [[P_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[P_ADDR]], align 4
+// CK-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP0]], 1
+// CK-32-NEXT: store i32 [[INC]], ptr [[P_ADDR]], align 4
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo5i(
+// CK-32-SAME: i32 [[A:%.*]]) #[[ATTR0]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-32-NEXT: [[TMP1:%.*]] = load ptr, ptr [[PA]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP3]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP4]], align 4
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 1
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP5]], align 4
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 1
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP6]], align 4
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 1
+// CK-32-NEXT: store ptr null, ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP1]], ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP1]], ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 2
+// CK-32-NEXT: store ptr null, ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP13]], align 4
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 3, ptr [[TMP14]], align 4
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP11]], ptr [[TMP15]], align 4
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP12]], ptr [[TMP16]], align 4
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes.7, ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes.8, ptr [[TMP18]], align 4
+// CK-32-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP19]], align 4
+// CK-32-NEXT: [[TMP20:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP20]], align 4
+// CK-32-NEXT: [[TMP21:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 0, ptr [[TMP21]], align 8
+// CK-32-NEXT: [[TMP22:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP22]], align 8
+// CK-32-NEXT: [[TMP23:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP23]], align 4
+// CK-32-NEXT: [[TMP24:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP24]], align 4
+// CK-32-NEXT: [[TMP25:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP25]], align 4
+// CK-32-NEXT: [[TMP26:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo5i_l66.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP27:%.*]] = icmp ne i32 [[TMP26]], 0
+// CK-32-NEXT: br i1 [[TMP27]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo5i_l66(ptr [[D]], ptr [[PVTARR]], ptr [[TMP1]]) #[[ATTR2]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo5i_l66(
+// CK-32-SAME: ptr nonnull align 4 dereferenceable(8) [[D:%.*]], ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]], ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[D_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[D1:%.*]] = alloca double, align 8
+// CK-32-NEXT: [[PVTARR2:%.*]] = alloca [10 x i32], align 4
+// CK-32-NEXT: [[PA3:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: store ptr [[D]], ptr [[D_ADDR]], align 4
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 4
+// CK-32-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[D_ADDR]], align 4, !nonnull [[META20]], !align [[META21]]
+// CK-32-NEXT: [[TMP1:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 4, !nonnull [[META20]], !align [[META21]]
+// CK-32-NEXT: [[TMP2:%.*]] = load double, ptr [[D1]], align 8
+// CK-32-NEXT: [[ADD:%.*]] = fadd double [[TMP2]], 1.000000e+00
+// CK-32-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// CK-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR2]], i32 0, i32 5
+// CK-32-NEXT: [[TMP3:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP3]], 1
+// CK-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = load ptr, ptr [[PA3]], align 4
+// CK-32-NEXT: [[ARRAYIDX4:%.*]] = getelementptr inbounds i32, ptr [[TMP4]], i32 50
+// CK-32-NEXT: [[TMP5:%.*]] = load i32, ptr [[ARRAYIDX4]], align 4
+// CK-32-NEXT: [[INC5:%.*]] = add nsw i32 [[TMP5]], 1
+// CK-32-NEXT: store i32 [[INC5]], ptr [[ARRAYIDX4]], align 4
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo6i(
+// CK-32-SAME: i32 [[A:%.*]]) #[[ATTR0]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-32-NEXT: [[TMP1:%.*]] = load ptr, ptr [[PA]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP3]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP4]], align 4
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 1
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP5]], align 4
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 1
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP6]], align 4
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 1
+// CK-32-NEXT: store ptr null, ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP1]], ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP1]], ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 2
+// CK-32-NEXT: store ptr null, ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP13]], align 4
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 3, ptr [[TMP14]], align 4
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP11]], ptr [[TMP15]], align 4
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP12]], ptr [[TMP16]], align 4
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes.9, ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes.10, ptr [[TMP18]], align 4
+// CK-32-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP19]], align 4
+// CK-32-NEXT: [[TMP20:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP20]], align 4
+// CK-32-NEXT: [[TMP21:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 0, ptr [[TMP21]], align 8
+// CK-32-NEXT: [[TMP22:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP22]], align 8
+// CK-32-NEXT: [[TMP23:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP23]], align 4
+// CK-32-NEXT: [[TMP24:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP24]], align 4
+// CK-32-NEXT: [[TMP25:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP25]], align 4
+// CK-32-NEXT: [[TMP26:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo6i_l82.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP27:%.*]] = icmp ne i32 [[TMP26]], 0
+// CK-32-NEXT: br i1 [[TMP27]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo6i_l82(ptr [[D]], ptr [[PVTARR]], ptr [[TMP1]]) #[[ATTR2]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo6i_l82(
+// CK-32-SAME: ptr nonnull align 4 dereferenceable(8) [[D:%.*]], ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]], ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[D_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[D1:%.*]] = alloca double, align 8
+// CK-32-NEXT: store ptr [[D]], ptr [[D_ADDR]], align 4
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 4
+// CK-32-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[D_ADDR]], align 4, !nonnull [[META20]], !align [[META21]]
+// CK-32-NEXT: [[TMP1:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 4, !nonnull [[META20]], !align [[META21]]
+// CK-32-NEXT: [[TMP2:%.*]] = load double, ptr [[TMP0]], align 8
+// CK-32-NEXT: store double [[TMP2]], ptr [[D1]], align 8
+// CK-32-NEXT: [[TMP3:%.*]] = load double, ptr [[D1]], align 8
+// CK-32-NEXT: [[ADD:%.*]] = fadd double [[TMP3]], 1.000000e+00
+// CK-32-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// CK-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[TMP1]], i32 0, i32 5
+// CK-32-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP4]], 1
+// CK-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[TMP5:%.*]] = load ptr, ptr [[PA_ADDR]], align 4
+// CK-32-NEXT: [[ARRAYIDX2:%.*]] = getelementptr inbounds i32, ptr [[TMP5]], i32 50
+// CK-32-NEXT: [[TMP6:%.*]] = load i32, ptr [[ARRAYIDX2]], align 4
+// CK-32-NEXT: [[INC3:%.*]] = add nsw i32 [[TMP6]], 1
+// CK-32-NEXT: store i32 [[INC3]], ptr [[ARRAYIDX2]], align 4
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo7i(
+// CK-32-SAME: i32 [[A:%.*]]) #[[ATTR0]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[D:%.*]] = alloca double, align 8
+// CK-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// CK-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [3 x ptr], align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// CK-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// CK-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// CK-32-NEXT: [[TMP1:%.*]] = load ptr, ptr [[PA]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store ptr [[D]], ptr [[TMP3]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP4]], align 4
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 1
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP5]], align 4
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 1
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[TMP6]], align 4
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 1
+// CK-32-NEXT: store ptr null, ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP1]], ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP1]], ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 2
+// CK-32-NEXT: store ptr null, ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds [3 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP13]], align 4
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 3, ptr [[TMP14]], align 4
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP11]], ptr [[TMP15]], align 4
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP12]], ptr [[TMP16]], align 4
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes.11, ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes.12, ptr [[TMP18]], align 4
+// CK-32-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP19]], align 4
+// CK-32-NEXT: [[TMP20:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP20]], align 4
+// CK-32-NEXT: [[TMP21:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 0, ptr [[TMP21]], align 8
+// CK-32-NEXT: [[TMP22:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP22]], align 8
+// CK-32-NEXT: [[TMP23:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] [i32 -1, i32 0, i32 0], ptr [[TMP23]], align 4
+// CK-32-NEXT: [[TMP24:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP24]], align 4
+// CK-32-NEXT: [[TMP25:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP25]], align 4
+// CK-32-NEXT: [[TMP26:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 -1, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo7i_l98.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP27:%.*]] = icmp ne i32 [[TMP26]], 0
+// CK-32-NEXT: br i1 [[TMP27]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo7i_l98(ptr [[D]], ptr [[PVTARR]], ptr [[TMP1]]) #[[ATTR2]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo7i_l98(
+// CK-32-SAME: ptr nonnull align 4 dereferenceable(8) [[D:%.*]], ptr nonnull align 4 dereferenceable(40) [[PVTARR:%.*]], ptr [[PA:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[D_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PVTARR_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[PA_ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[D1:%.*]] = alloca double, align 8
+// CK-32-NEXT: [[PVTARR2:%.*]] = alloca [10 x i32], align 4
+// CK-32-NEXT: store ptr [[D]], ptr [[D_ADDR]], align 4
+// CK-32-NEXT: store ptr [[PVTARR]], ptr [[PVTARR_ADDR]], align 4
+// CK-32-NEXT: store ptr [[PA]], ptr [[PA_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[D_ADDR]], align 4, !nonnull [[META20]], !align [[META21]]
+// CK-32-NEXT: [[TMP1:%.*]] = load ptr, ptr [[PVTARR_ADDR]], align 4, !nonnull [[META20]], !align [[META21]]
+// CK-32-NEXT: [[TMP2:%.*]] = load double, ptr [[TMP0]], align 8
+// CK-32-NEXT: store double [[TMP2]], ptr [[D1]], align 8
+// CK-32-NEXT: call void @llvm.memcpy.p0.p0.i32(ptr align 4 [[PVTARR2]], ptr align 4 [[TMP1]], i32 40, i1 false)
+// CK-32-NEXT: [[TMP3:%.*]] = load double, ptr [[D1]], align 8
+// CK-32-NEXT: [[ADD:%.*]] = fadd double [[TMP3]], 1.000000e+00
+// CK-32-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// CK-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR2]], i32 0, i32 5
+// CK-32-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP4]], 1
+// CK-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// CK-32-NEXT: [[TMP5:%.*]] = load ptr, ptr [[PA_ADDR]], align 4
+// CK-32-NEXT: [[ARRAYIDX3:%.*]] = getelementptr inbounds i32, ptr [[TMP5]], i32 50
+// CK-32-NEXT: [[TMP6:%.*]] = load i32, ptr [[ARRAYIDX3]], align 4
+// CK-32-NEXT: [[INC4:%.*]] = add nsw i32 [[TMP6]], 1
+// CK-32-NEXT: store i32 [[INC4]], ptr [[ARRAYIDX3]], align 4
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define dso_local void @_Z4foo8v(
+// CK-32-SAME: ) #[[ATTR0]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[X:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[X_CASTED:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOFFLOAD_BASEPTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_PTRS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[DOTOFFLOAD_MAPPERS:%.*]] = alloca [1 x ptr], align 4
+// CK-32-NEXT: [[TMP:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[KERNEL_ARGS:%.*]] = alloca [[STRUCT___TGT_KERNEL_ARGUMENTS:%.*]], align 8
+// CK-32-NEXT: store i32 0, ptr [[X]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[X]], align 4
+// CK-32-NEXT: store i32 [[TMP0]], ptr [[X_CASTED]], align 4
+// CK-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[X_CASTED]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: store i32 [[TMP1]], ptr [[TMP2]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: store i32 [[TMP1]], ptr [[TMP3]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_MAPPERS]], i32 0, i32 0
+// CK-32-NEXT: store ptr null, ptr [[TMP4]], align 4
+// CK-32-NEXT: [[TMP5:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_BASEPTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP6:%.*]] = getelementptr inbounds [1 x ptr], ptr [[DOTOFFLOAD_PTRS]], i32 0, i32 0
+// CK-32-NEXT: [[TMP7:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 0
+// CK-32-NEXT: store i32 3, ptr [[TMP7]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 1
+// CK-32-NEXT: store i32 1, ptr [[TMP8]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 2
+// CK-32-NEXT: store ptr [[TMP5]], ptr [[TMP9]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 3
+// CK-32-NEXT: store ptr [[TMP6]], ptr [[TMP10]], align 4
+// CK-32-NEXT: [[TMP11:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 4
+// CK-32-NEXT: store ptr @.offload_sizes.13, ptr [[TMP11]], align 4
+// CK-32-NEXT: [[TMP12:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 5
+// CK-32-NEXT: store ptr @.offload_maptypes.14, ptr [[TMP12]], align 4
+// CK-32-NEXT: [[TMP13:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 6
+// CK-32-NEXT: store ptr null, ptr [[TMP13]], align 4
+// CK-32-NEXT: [[TMP14:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 7
+// CK-32-NEXT: store ptr null, ptr [[TMP14]], align 4
+// CK-32-NEXT: [[TMP15:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 8
+// CK-32-NEXT: store i64 10, ptr [[TMP15]], align 8
+// CK-32-NEXT: [[TMP16:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 9
+// CK-32-NEXT: store i64 0, ptr [[TMP16]], align 8
+// CK-32-NEXT: [[TMP17:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 10
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP17]], align 4
+// CK-32-NEXT: [[TMP18:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 11
+// CK-32-NEXT: store [3 x i32] zeroinitializer, ptr [[TMP18]], align 4
+// CK-32-NEXT: [[TMP19:%.*]] = getelementptr inbounds nuw [[STRUCT___TGT_KERNEL_ARGUMENTS]], ptr [[KERNEL_ARGS]], i32 0, i32 12
+// CK-32-NEXT: store i32 0, ptr [[TMP19]], align 4
+// CK-32-NEXT: [[TMP20:%.*]] = call i32 @__tgt_target_kernel(ptr @[[GLOB1]], i64 -1, i32 0, i32 0, ptr @.{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.region_id, ptr [[KERNEL_ARGS]])
+// CK-32-NEXT: [[TMP21:%.*]] = icmp ne i32 [[TMP20]], 0
+// CK-32-NEXT: br i1 [[TMP21]], label %[[OMP_OFFLOAD_FAILED:.*]], label %[[OMP_OFFLOAD_CONT:.*]]
+// CK-32: [[OMP_OFFLOAD_FAILED]]:
+// CK-32-NEXT: call void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112(i32 [[TMP1]]) #[[ATTR2]]
+// CK-32-NEXT: br label %[[OMP_OFFLOAD_CONT]]
+// CK-32: [[OMP_OFFLOAD_CONT]]:
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112(
+// CK-32-SAME: i32 [[X:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[X_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[X_CASTED:%.*]] = alloca i32, align 4
+// CK-32-NEXT: store i32 [[X]], ptr [[X_ADDR]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[X_ADDR]], align 4
+// CK-32-NEXT: store i32 [[TMP0]], ptr [[X_CASTED]], align 4
+// CK-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[X_CASTED]], align 4
+// CK-32-NEXT: call void (ptr, i32, ptr, ...) @__kmpc_fork_teams(ptr @[[GLOB1]], i32 1, ptr @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined, i32 [[TMP1]])
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined(
+// CK-32-SAME: ptr noalias [[DOTGLOBAL_TID_:%.*]], ptr noalias [[DOTBOUND_TID_:%.*]], i32 [[X:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[DOTGLOBAL_TID__ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[DOTBOUND_TID__ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[X_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_IV:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[TMP:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_COMB_LB:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_COMB_UB:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_STRIDE:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_IS_LAST:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[I:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[X_CASTED:%.*]] = alloca i32, align 4
+// CK-32-NEXT: store ptr [[DOTGLOBAL_TID_]], ptr [[DOTGLOBAL_TID__ADDR]], align 4
+// CK-32-NEXT: store ptr [[DOTBOUND_TID_]], ptr [[DOTBOUND_TID__ADDR]], align 4
+// CK-32-NEXT: store i32 [[X]], ptr [[X_ADDR]], align 4
+// CK-32-NEXT: store i32 0, ptr [[DOTOMP_COMB_LB]], align 4
+// CK-32-NEXT: store i32 9, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-32-NEXT: store i32 1, ptr [[DOTOMP_STRIDE]], align 4
+// CK-32-NEXT: store i32 0, ptr [[DOTOMP_IS_LAST]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[DOTGLOBAL_TID__ADDR]], align 4
+// CK-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[TMP0]], align 4
+// CK-32-NEXT: call void @__kmpc_for_static_init_4(ptr @[[GLOB2:[0-9]+]], i32 [[TMP1]], i32 92, ptr [[DOTOMP_IS_LAST]], ptr [[DOTOMP_COMB_LB]], ptr [[DOTOMP_COMB_UB]], ptr [[DOTOMP_STRIDE]], i32 1, i32 1)
+// CK-32-NEXT: [[TMP2:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-32-NEXT: [[CMP:%.*]] = icmp sgt i32 [[TMP2]], 9
+// CK-32-NEXT: br i1 [[CMP]], label %[[COND_TRUE:.*]], label %[[COND_FALSE:.*]]
+// CK-32: [[COND_TRUE]]:
+// CK-32-NEXT: br label %[[COND_END:.*]]
+// CK-32: [[COND_FALSE]]:
+// CK-32-NEXT: [[TMP3:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-32-NEXT: br label %[[COND_END]]
+// CK-32: [[COND_END]]:
+// CK-32-NEXT: [[COND:%.*]] = phi i32 [ 9, %[[COND_TRUE]] ], [ [[TMP3]], %[[COND_FALSE]] ]
+// CK-32-NEXT: store i32 [[COND]], ptr [[DOTOMP_COMB_UB]], align 4
+// CK-32-NEXT: [[TMP4:%.*]] = load i32, ptr [[DOTOMP_COMB_LB]], align 4
+// CK-32-NEXT: store i32 [[TMP4]], ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: br label %[[OMP_INNER_FOR_COND:.*]]
+// CK-32: [[OMP_INNER_FOR_COND]]:
+// CK-32-NEXT: [[TMP5:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: [[TMP6:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-32-NEXT: [[CMP1:%.*]] = icmp sle i32 [[TMP5]], [[TMP6]]
+// CK-32-NEXT: br i1 [[CMP1]], label %[[OMP_INNER_FOR_BODY:.*]], label %[[OMP_INNER_FOR_END:.*]]
+// CK-32: [[OMP_INNER_FOR_BODY]]:
+// CK-32-NEXT: [[TMP7:%.*]] = load i32, ptr [[DOTOMP_COMB_LB]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = load i32, ptr [[DOTOMP_COMB_UB]], align 4
+// CK-32-NEXT: [[TMP9:%.*]] = load i32, ptr [[X_ADDR]], align 4
+// CK-32-NEXT: store i32 [[TMP9]], ptr [[X_CASTED]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = load i32, ptr [[X_CASTED]], align 4
+// CK-32-NEXT: call void (ptr, i32, ptr, ...) @__kmpc_fork_call(ptr @[[GLOB1]], i32 3, ptr @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined.omp_outlined, i32 [[TMP7]], i32 [[TMP8]], i32 [[TMP10]])
+// CK-32-NEXT: br label %[[OMP_INNER_FOR_INC:.*]]
+// CK-32: [[OMP_INNER_FOR_INC]]:
+// CK-32-NEXT: [[TMP11:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: [[TMP12:%.*]] = load i32, ptr [[DOTOMP_STRIDE]], align 4
+// CK-32-NEXT: [[ADD:%.*]] = add nsw i32 [[TMP11]], [[TMP12]]
+// CK-32-NEXT: store i32 [[ADD]], ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: br label %[[OMP_INNER_FOR_COND]]
+// CK-32: [[OMP_INNER_FOR_END]]:
+// CK-32-NEXT: br label %[[OMP_LOOP_EXIT:.*]]
+// CK-32: [[OMP_LOOP_EXIT]]:
+// CK-32-NEXT: call void @__kmpc_for_static_fini(ptr @[[GLOB2]], i32 [[TMP1]])
+// CK-32-NEXT: ret void
+//
+//
+// CK-32-LABEL: define internal void @{{__omp_offloading_[0-9a-z]+_[0-9a-z]+}}__Z4foo8v_l112.omp_outlined.omp_outlined(
+// CK-32-SAME: ptr noalias [[DOTGLOBAL_TID_:%.*]], ptr noalias [[DOTBOUND_TID_:%.*]], i32 [[DOTPREVIOUS_LB_:%.*]], i32 [[DOTPREVIOUS_UB_:%.*]], i32 [[X:%.*]]) #[[ATTR1]] {
+// CK-32-NEXT: [[ENTRY:.*:]]
+// CK-32-NEXT: [[DOTGLOBAL_TID__ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[DOTBOUND_TID__ADDR:%.*]] = alloca ptr, align 4
+// CK-32-NEXT: [[DOTPREVIOUS_LB__ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTPREVIOUS_UB__ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[X_ADDR:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_IV:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[TMP:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_LB:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_UB:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_STRIDE:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[DOTOMP_IS_LAST:%.*]] = alloca i32, align 4
+// CK-32-NEXT: [[I:%.*]] = alloca i32, align 4
+// CK-32-NEXT: store ptr [[DOTGLOBAL_TID_]], ptr [[DOTGLOBAL_TID__ADDR]], align 4
+// CK-32-NEXT: store ptr [[DOTBOUND_TID_]], ptr [[DOTBOUND_TID__ADDR]], align 4
+// CK-32-NEXT: store i32 [[DOTPREVIOUS_LB_]], ptr [[DOTPREVIOUS_LB__ADDR]], align 4
+// CK-32-NEXT: store i32 [[DOTPREVIOUS_UB_]], ptr [[DOTPREVIOUS_UB__ADDR]], align 4
+// CK-32-NEXT: store i32 [[X]], ptr [[X_ADDR]], align 4
+// CK-32-NEXT: store i32 0, ptr [[DOTOMP_LB]], align 4
+// CK-32-NEXT: store i32 9, ptr [[DOTOMP_UB]], align 4
+// CK-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[DOTPREVIOUS_LB__ADDR]], align 4
+// CK-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[DOTPREVIOUS_UB__ADDR]], align 4
+// CK-32-NEXT: store i32 [[TMP0]], ptr [[DOTOMP_LB]], align 4
+// CK-32-NEXT: store i32 [[TMP1]], ptr [[DOTOMP_UB]], align 4
+// CK-32-NEXT: store i32 1, ptr [[DOTOMP_STRIDE]], align 4
+// CK-32-NEXT: store i32 0, ptr [[DOTOMP_IS_LAST]], align 4
+// CK-32-NEXT: [[TMP2:%.*]] = load ptr, ptr [[DOTGLOBAL_TID__ADDR]], align 4
+// CK-32-NEXT: [[TMP3:%.*]] = load i32, ptr [[TMP2]], align 4
+// CK-32-NEXT: call void @__kmpc_for_static_init_4(ptr @[[GLOB3:[0-9]+]], i32 [[TMP3]], i32 34, ptr [[DOTOMP_IS_LAST]], ptr [[DOTOMP_LB]], ptr [[DOTOMP_UB]], ptr [[DOTOMP_STRIDE]], i32 1, i32 1)
+// CK-32-NEXT: [[TMP4:%.*]] = load i32, ptr [[DOTOMP_UB]], align 4
+// CK-32-NEXT: [[CMP:%.*]] = icmp sgt i32 [[TMP4]], 9
+// CK-32-NEXT: br i1 [[CMP]], label %[[COND_TRUE:.*]], label %[[COND_FALSE:.*]]
+// CK-32: [[COND_TRUE]]:
+// CK-32-NEXT: br label %[[COND_END:.*]]
+// CK-32: [[COND_FALSE]]:
+// CK-32-NEXT: [[TMP5:%.*]] = load i32, ptr [[DOTOMP_UB]], align 4
+// CK-32-NEXT: br label %[[COND_END]]
+// CK-32: [[COND_END]]:
+// CK-32-NEXT: [[COND:%.*]] = phi i32 [ 9, %[[COND_TRUE]] ], [ [[TMP5]], %[[COND_FALSE]] ]
+// CK-32-NEXT: store i32 [[COND]], ptr [[DOTOMP_UB]], align 4
+// CK-32-NEXT: [[TMP6:%.*]] = load i32, ptr [[DOTOMP_LB]], align 4
+// CK-32-NEXT: store i32 [[TMP6]], ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: br label %[[OMP_INNER_FOR_COND:.*]]
+// CK-32: [[OMP_INNER_FOR_COND]]:
+// CK-32-NEXT: [[TMP7:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: [[TMP8:%.*]] = load i32, ptr [[DOTOMP_UB]], align 4
+// CK-32-NEXT: [[CMP1:%.*]] = icmp sle i32 [[TMP7]], [[TMP8]]
+// CK-32-NEXT: br i1 [[CMP1]], label %[[OMP_INNER_FOR_BODY:.*]], label %[[OMP_INNER_FOR_END:.*]]
+// CK-32: [[OMP_INNER_FOR_BODY]]:
+// CK-32-NEXT: [[TMP9:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: [[MUL:%.*]] = mul nsw i32 [[TMP9]], 1
+// CK-32-NEXT: [[ADD:%.*]] = add nsw i32 0, [[MUL]]
+// CK-32-NEXT: store i32 [[ADD]], ptr [[I]], align 4
+// CK-32-NEXT: [[TMP10:%.*]] = load i32, ptr [[X_ADDR]], align 4
+// CK-32-NEXT: [[ADD2:%.*]] = add nsw i32 [[TMP10]], 1
+// CK-32-NEXT: store i32 [[ADD2]], ptr [[X_ADDR]], align 4
+// CK-32-NEXT: br label %[[OMP_BODY_CONTINUE:.*]]
+// CK-32: [[OMP_BODY_CONTINUE]]:
+// CK-32-NEXT: br label %[[OMP_INNER_FOR_INC:.*]]
+// CK-32: [[OMP_INNER_FOR_INC]]:
+// CK-32-NEXT: [[TMP11:%.*]] = load i32, ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: [[ADD3:%.*]] = add nsw i32 [[TMP11]], 1
+// CK-32-NEXT: store i32 [[ADD3]], ptr [[DOTOMP_IV]], align 4
+// CK-32-NEXT: br label %[[OMP_INNER_FOR_COND]]
+// CK-32: [[OMP_INNER_FOR_END]]:
+// CK-32-NEXT: br label %[[OMP_LOOP_EXIT:.*]]
+// CK-32: [[OMP_LOOP_EXIT]]:
+// CK-32-NEXT: call void @__kmpc_for_static_fini(ptr @[[GLOB3]], i32 [[TMP3]])
+// CK-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo1i(
+// SIMD-ONLY-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0:[0-9]+]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-64-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-64-NEXT: [[D1:%.*]] = alloca double, align 8
+// SIMD-ONLY-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D1]], align 8
+// SIMD-ONLY-64-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-64-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo2v(
+// SIMD-ONLY-64-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-64-NEXT: [[PVTARR1:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR1]], i64 0, i64 5
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP0]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo3v(
+// SIMD-ONLY-64-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// SIMD-ONLY-64-NEXT: [[PA1:%.*]] = alloca ptr, align 8
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PA1]], align 8
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds i32, ptr [[TMP0]], i64 50
+// SIMD-ONLY-64-NEXT: [[TMP1:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP1]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo4v(
+// SIMD-ONLY-64-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[P:%.*]] = alloca i32, align 4
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[P]], align 4
+// SIMD-ONLY-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP0]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC]], ptr [[P]], align 4
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo5i(
+// SIMD-ONLY-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-64-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// SIMD-ONLY-64-NEXT: [[D1:%.*]] = alloca double, align 8
+// SIMD-ONLY-64-NEXT: [[PVTARR2:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-64-NEXT: [[PA3:%.*]] = alloca ptr, align 8
+// SIMD-ONLY-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D1]], align 8
+// SIMD-ONLY-64-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-64-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR2]], i64 0, i64 5
+// SIMD-ONLY-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA3]], align 8
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX4:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i64 50
+// SIMD-ONLY-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX4]], align 4
+// SIMD-ONLY-64-NEXT: [[INC5:%.*]] = add nsw i32 [[TMP4]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC5]], ptr [[ARRAYIDX4]], align 4
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo6i(
+// SIMD-ONLY-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-64-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// SIMD-ONLY-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-64-NEXT: store double [[ADD]], ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR]], i64 0, i64 5
+// SIMD-ONLY-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA]], align 8
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX1:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i64 50
+// SIMD-ONLY-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-64-NEXT: [[INC2:%.*]] = add nsw i32 [[TMP4]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC2]], ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo7i(
+// SIMD-ONLY-64-SAME: i32 signext [[A:%.*]]) #[[ATTR0]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-64-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-64-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-64-NEXT: [[PA:%.*]] = alloca ptr, align 8
+// SIMD-ONLY-64-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-64-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-64-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-64-NEXT: store double [[ADD]], ptr [[D]], align 8
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR]], i64 0, i64 5
+// SIMD-ONLY-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-64-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA]], align 8
+// SIMD-ONLY-64-NEXT: [[ARRAYIDX1:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i64 50
+// SIMD-ONLY-64-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-64-NEXT: [[INC2:%.*]] = add nsw i32 [[TMP4]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC2]], ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-64-LABEL: define dso_local void @_Z4foo8v(
+// SIMD-ONLY-64-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-64-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-64-NEXT: [[X:%.*]] = alloca i32, align 4
+// SIMD-ONLY-64-NEXT: [[I:%.*]] = alloca i32, align 4
+// SIMD-ONLY-64-NEXT: store i32 0, ptr [[X]], align 4
+// SIMD-ONLY-64-NEXT: store i32 0, ptr [[I]], align 4
+// SIMD-ONLY-64-NEXT: br label %[[FOR_COND:.*]]
+// SIMD-ONLY-64: [[FOR_COND]]:
+// SIMD-ONLY-64-NEXT: [[TMP0:%.*]] = load i32, ptr [[I]], align 4
+// SIMD-ONLY-64-NEXT: [[CMP:%.*]] = icmp slt i32 [[TMP0]], 10
+// SIMD-ONLY-64-NEXT: br i1 [[CMP]], label %[[FOR_BODY:.*]], label %[[FOR_END:.*]]
+// SIMD-ONLY-64: [[FOR_BODY]]:
+// SIMD-ONLY-64-NEXT: [[TMP1:%.*]] = load i32, ptr [[X]], align 4
+// SIMD-ONLY-64-NEXT: [[ADD:%.*]] = add nsw i32 [[TMP1]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[ADD]], ptr [[X]], align 4
+// SIMD-ONLY-64-NEXT: br label %[[FOR_INC:.*]]
+// SIMD-ONLY-64: [[FOR_INC]]:
+// SIMD-ONLY-64-NEXT: [[TMP2:%.*]] = load i32, ptr [[I]], align 4
+// SIMD-ONLY-64-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-64-NEXT: store i32 [[INC]], ptr [[I]], align 4
+// SIMD-ONLY-64-NEXT: br label %[[FOR_COND]], !llvm.loop [[LOOP2:![0-9]+]]
+// SIMD-ONLY-64: [[FOR_END]]:
+// SIMD-ONLY-64-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo1i(
+// SIMD-ONLY-32-SAME: i32 [[A:%.*]]) #[[ATTR0:[0-9]+]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-32-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-32-NEXT: [[D1:%.*]] = alloca double, align 8
+// SIMD-ONLY-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[TMP1:%.*]] = load double, ptr [[D1]], align 8
+// SIMD-ONLY-32-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-32-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// SIMD-ONLY-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo2v(
+// SIMD-ONLY-32-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-32-NEXT: [[PVTARR1:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR1]], i32 0, i32 5
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP0]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo3v(
+// SIMD-ONLY-32-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// SIMD-ONLY-32-NEXT: [[PA1:%.*]] = alloca ptr, align 4
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load ptr, ptr [[PA1]], align 4
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds i32, ptr [[TMP0]], i32 50
+// SIMD-ONLY-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP1]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo4v(
+// SIMD-ONLY-32-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[P:%.*]] = alloca i32, align 4
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[P]], align 4
+// SIMD-ONLY-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP0]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC]], ptr [[P]], align 4
+// SIMD-ONLY-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo5i(
+// SIMD-ONLY-32-SAME: i32 [[A:%.*]]) #[[ATTR0]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-32-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// SIMD-ONLY-32-NEXT: [[D1:%.*]] = alloca double, align 8
+// SIMD-ONLY-32-NEXT: [[PVTARR2:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-32-NEXT: [[PA3:%.*]] = alloca ptr, align 4
+// SIMD-ONLY-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[TMP1:%.*]] = load double, ptr [[D1]], align 8
+// SIMD-ONLY-32-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-32-NEXT: store double [[ADD]], ptr [[D1]], align 8
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR2]], i32 0, i32 5
+// SIMD-ONLY-32-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA3]], align 4
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX4:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i32 50
+// SIMD-ONLY-32-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX4]], align 4
+// SIMD-ONLY-32-NEXT: [[INC5:%.*]] = add nsw i32 [[TMP4]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC5]], ptr [[ARRAYIDX4]], align 4
+// SIMD-ONLY-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo6i(
+// SIMD-ONLY-32-SAME: i32 [[A:%.*]]) #[[ATTR0]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-32-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// SIMD-ONLY-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-32-NEXT: store double [[ADD]], ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR]], i32 0, i32 5
+// SIMD-ONLY-32-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA]], align 4
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX1:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i32 50
+// SIMD-ONLY-32-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-32-NEXT: [[INC2:%.*]] = add nsw i32 [[TMP4]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC2]], ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo7i(
+// SIMD-ONLY-32-SAME: i32 [[A:%.*]]) #[[ATTR0]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4
+// SIMD-ONLY-32-NEXT: [[D:%.*]] = alloca double, align 8
+// SIMD-ONLY-32-NEXT: [[PVTARR:%.*]] = alloca [10 x i32], align 4
+// SIMD-ONLY-32-NEXT: [[PA:%.*]] = alloca ptr, align 4
+// SIMD-ONLY-32-NEXT: store i32 [[A]], ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[A_ADDR]], align 4
+// SIMD-ONLY-32-NEXT: [[CONV:%.*]] = sitofp i32 [[TMP0]] to double
+// SIMD-ONLY-32-NEXT: store double [[CONV]], ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[TMP1:%.*]] = load double, ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[ADD:%.*]] = fadd double [[TMP1]], 1.000000e+00
+// SIMD-ONLY-32-NEXT: store double [[ADD]], ptr [[D]], align 8
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX:%.*]] = getelementptr inbounds [10 x i32], ptr [[PVTARR]], i32 0, i32 5
+// SIMD-ONLY-32-NEXT: [[TMP2:%.*]] = load i32, ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC]], ptr [[ARRAYIDX]], align 4
+// SIMD-ONLY-32-NEXT: [[TMP3:%.*]] = load ptr, ptr [[PA]], align 4
+// SIMD-ONLY-32-NEXT: [[ARRAYIDX1:%.*]] = getelementptr inbounds i32, ptr [[TMP3]], i32 50
+// SIMD-ONLY-32-NEXT: [[TMP4:%.*]] = load i32, ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-32-NEXT: [[INC2:%.*]] = add nsw i32 [[TMP4]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC2]], ptr [[ARRAYIDX1]], align 4
+// SIMD-ONLY-32-NEXT: ret void
+//
+//
+// SIMD-ONLY-32-LABEL: define dso_local void @_Z4foo8v(
+// SIMD-ONLY-32-SAME: ) #[[ATTR0]] {
+// SIMD-ONLY-32-NEXT: [[ENTRY:.*:]]
+// SIMD-ONLY-32-NEXT: [[X:%.*]] = alloca i32, align 4
+// SIMD-ONLY-32-NEXT: [[I:%.*]] = alloca i32, align 4
+// SIMD-ONLY-32-NEXT: store i32 0, ptr [[X]], align 4
+// SIMD-ONLY-32-NEXT: store i32 0, ptr [[I]], align 4
+// SIMD-ONLY-32-NEXT: br label %[[FOR_COND:.*]]
+// SIMD-ONLY-32: [[FOR_COND]]:
+// SIMD-ONLY-32-NEXT: [[TMP0:%.*]] = load i32, ptr [[I]], align 4
+// SIMD-ONLY-32-NEXT: [[CMP:%.*]] = icmp slt i32 [[TMP0]], 10
+// SIMD-ONLY-32-NEXT: br i1 [[CMP]], label %[[FOR_BODY:.*]], label %[[FOR_END:.*]]
+// SIMD-ONLY-32: [[FOR_BODY]]:
+// SIMD-ONLY-32-NEXT: [[TMP1:%.*]] = load i32, ptr [[X]], align 4
+// SIMD-ONLY-32-NEXT: [[ADD:%.*]] = add nsw i32 [[TMP1]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[ADD]], ptr [[X]], align 4
+// SIMD-ONLY-32-NEXT: br label %[[FOR_INC:.*]]
+// SIMD-ONLY-32: [[FOR_INC]]:
+// SIMD-ONLY-32-NEXT: [[TMP2:%.*]] = load i32, ptr [[I]], align 4
+// SIMD-ONLY-32-NEXT: [[INC:%.*]] = add nsw i32 [[TMP2]], 1
+// SIMD-ONLY-32-NEXT: store i32 [[INC]], ptr [[I]], align 4
+// SIMD-ONLY-32-NEXT: br label %[[FOR_COND]], !llvm.loop [[LOOP3:![0-9]+]]
+// SIMD-ONLY-32: [[FOR_END]]:
+// SIMD-ONLY-32-NEXT: ret void
+//
+//.
+// CK-64: [[META19]] = !{}
+// CK-64: [[META20]] = !{i64 4}
+//.
+// CK-32: [[META20]] = !{}
+// CK-32: [[META21]] = !{i64 4}
+//.
+// SIMD-ONLY-64: [[LOOP2]] = distinct !{[[LOOP2]], [[META3:![0-9]+]]}
+// SIMD-ONLY-64: [[META3]] = !{!"llvm.loop.mustprogress"}
+//.
+// SIMD-ONLY-32: [[LOOP3]] = distinct !{[[LOOP3]], [[META4:![0-9]+]]}
+// SIMD-ONLY-32: [[META4]] = !{!"llvm.loop.mustprogress"}
+//.
diff --git a/clang/test/OpenMP/target_default_messages.cpp b/clang/test/OpenMP/target_default_messages.cpp
index be677df..6a1a1f9 100644
--- a/clang/test/OpenMP/target_default_messages.cpp
+++ b/clang/test/OpenMP/target_default_messages.cpp
@@ -24,6 +24,8 @@ int main(int argc, char **argv) {
for (int i=0; i<200; i++) foo();
#pragma omp target default(x) // expected-error {{expected 'none', 'shared', 'private' or 'firstprivate' in OpenMP clause 'default'}}
for (int i=0; i<200; i++) foo();
+#pragma omp target default(none) // expected-note {{explicit data sharing attribute, data mapping attribute, or is_device_ptr clause requested here}}
+ x++; // expected-error {{variable 'x' must have explicitly specified data sharing attributes, data mapping attributes, or in an is_device_ptr clause}}
#endif
#ifdef OMP52