| ; NOTE: Assertions have been autogenerated by utils/update_test_checks.py |
| ; RUN: opt < %s -passes=vector-combine -S -mtriple=x86_64-- -mattr=sse2 | FileCheck %s --check-prefixes=CHECK,SSE2 |
| ; RUN: opt < %s -passes=vector-combine -S -mtriple=x86_64-- -mattr=avx2 | FileCheck %s --check-prefixes=CHECK,AVX2 |
| |
| target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128" |
| |
| define float @matching_fp_scalar(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @matching_fp_scalar( |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %r = load float, ptr %p, align 16 |
| ret float %r |
| } |
| |
| define float @matching_fp_scalar_volatile(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @matching_fp_scalar_volatile( |
| ; CHECK-NEXT: [[R:%.*]] = load volatile float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %r = load volatile float, ptr %p, align 16 |
| ret float %r |
| } |
| |
| define double @larger_fp_scalar(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @larger_fp_scalar( |
| ; CHECK-NEXT: [[R:%.*]] = load double, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret double [[R]] |
| ; |
| %r = load double, ptr %p, align 16 |
| ret double %r |
| } |
| |
| define float @smaller_fp_scalar(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @smaller_fp_scalar( |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %r = load float, ptr %p, align 16 |
| ret float %r |
| } |
| |
| define float @matching_fp_vector(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @matching_fp_vector( |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %r = load float, ptr %p, align 16 |
| ret float %r |
| } |
| |
| define float @matching_fp_vector_gep00(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @matching_fp_vector_gep00( |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %r = load float, ptr %p, align 16 |
| ret float %r |
| } |
| |
| define float @matching_fp_vector_gep01(ptr align 16 dereferenceable(20) %p) { |
| ; CHECK-LABEL: @matching_fp_vector_gep01( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <4 x float>, ptr [[P:%.*]], i64 0, i64 1 |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[GEP]], align 4 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %gep = getelementptr inbounds <4 x float>, ptr %p, i64 0, i64 1 |
| %r = load float, ptr %gep, align 4 |
| ret float %r |
| } |
| |
| define float @matching_fp_vector_gep01_deref(ptr align 16 dereferenceable(19) %p) { |
| ; CHECK-LABEL: @matching_fp_vector_gep01_deref( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <4 x float>, ptr [[P:%.*]], i64 0, i64 1 |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[GEP]], align 4 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %gep = getelementptr inbounds <4 x float>, ptr %p, i64 0, i64 1 |
| %r = load float, ptr %gep, align 4 |
| ret float %r |
| } |
| |
| define float @matching_fp_vector_gep10(ptr align 16 dereferenceable(32) %p) { |
| ; CHECK-LABEL: @matching_fp_vector_gep10( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <4 x float>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %gep = getelementptr inbounds <4 x float>, ptr %p, i64 1, i64 0 |
| %r = load float, ptr %gep, align 16 |
| ret float %r |
| } |
| |
| define float @matching_fp_vector_gep10_deref(ptr align 16 dereferenceable(31) %p) { |
| ; CHECK-LABEL: @matching_fp_vector_gep10_deref( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <4 x float>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %gep = getelementptr inbounds <4 x float>, ptr %p, i64 1, i64 0 |
| %r = load float, ptr %gep, align 16 |
| ret float %r |
| } |
| |
| define float @nonmatching_int_vector(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @nonmatching_int_vector( |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %r = load float, ptr %p, align 16 |
| ret float %r |
| } |
| |
| define double @less_aligned(ptr align 4 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @less_aligned( |
| ; CHECK-NEXT: [[R:%.*]] = load double, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: ret double [[R]] |
| ; |
| %r = load double, ptr %p, align 4 |
| ret double %r |
| } |
| |
| define float @matching_fp_scalar_small_deref(ptr align 16 dereferenceable(15) %p) { |
| ; CHECK-LABEL: @matching_fp_scalar_small_deref( |
| ; CHECK-NEXT: [[R:%.*]] = load float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret float [[R]] |
| ; |
| %r = load float, ptr %p, align 16 |
| ret float %r |
| } |
| |
| define i64 @larger_int_scalar(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @larger_int_scalar( |
| ; CHECK-NEXT: [[R:%.*]] = load i64, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret i64 [[R]] |
| ; |
| %r = load i64, ptr %p, align 16 |
| ret i64 %r |
| } |
| |
| define i8 @smaller_int_scalar(ptr align 16 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @smaller_int_scalar( |
| ; CHECK-NEXT: [[R:%.*]] = load i8, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: ret i8 [[R]] |
| ; |
| %r = load i8, ptr %p, align 16 |
| ret i8 %r |
| } |
| |
| define double @larger_fp_scalar_256bit_vec(ptr align 32 dereferenceable(32) %p) { |
| ; CHECK-LABEL: @larger_fp_scalar_256bit_vec( |
| ; CHECK-NEXT: [[R:%.*]] = load double, ptr [[P:%.*]], align 32 |
| ; CHECK-NEXT: ret double [[R]] |
| ; |
| %r = load double, ptr %p, align 32 |
| ret double %r |
| } |
| |
| define <4 x float> @load_f32_insert_v4f32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_f32_insert_v4f32( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 4 |
| %r = insertelement <4 x float> undef, float %s, i32 0 |
| ret <4 x float> %r |
| } |
| |
| define <4 x float> @casted_load_f32_insert_v4f32(ptr align 4 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @casted_load_f32_insert_v4f32( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 4 |
| %r = insertelement <4 x float> undef, float %s, i32 0 |
| ret <4 x float> %r |
| } |
| |
| ; Element type does not change cost. |
| |
| define <4 x i32> @load_i32_insert_v4i32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_i32_insert_v4i32( |
| ; CHECK-NEXT: [[S:%.*]] = load i32, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x i32> undef, i32 [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x i32> [[R]] |
| ; |
| %s = load i32, ptr %p, align 4 |
| %r = insertelement <4 x i32> undef, i32 %s, i32 0 |
| ret <4 x i32> %r |
| } |
| |
| ; Pointer type does not change cost. |
| |
| define <4 x i32> @casted_load_i32_insert_v4i32(ptr align 4 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @casted_load_i32_insert_v4i32( |
| ; CHECK-NEXT: [[S:%.*]] = load i32, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x i32> undef, i32 [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x i32> [[R]] |
| ; |
| %s = load i32, ptr %p, align 4 |
| %r = insertelement <4 x i32> undef, i32 %s, i32 0 |
| ret <4 x i32> %r |
| } |
| |
| ; This is canonical form for vector element access. |
| |
| define <4 x float> @gep00_load_f32_insert_v4f32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @gep00_load_f32_insert_v4f32( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i64 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 16 |
| %r = insertelement <4 x float> undef, float %s, i64 0 |
| ret <4 x float> %r |
| } |
| |
| ; Should work with addrspace as well. |
| |
| define <4 x float> @gep00_load_f32_insert_v4f32_addrspace(ptr addrspace(44) align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @gep00_load_f32_insert_v4f32_addrspace( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr addrspace(44) [[P:%.*]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i64 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %s = load float, ptr addrspace(44) %p, align 16 |
| %r = insertelement <4 x float> undef, float %s, i64 0 |
| ret <4 x float> %r |
| } |
| |
| ; If there are enough dereferenceable bytes, we can offset the vector load. |
| |
| define <8 x i16> @gep01_load_i16_insert_v8i16(ptr align 16 dereferenceable(18) %p) nofree nosync { |
| ; CHECK-LABEL: @gep01_load_i16_insert_v8i16( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 0, i64 1 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 2 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 0, i64 1 |
| %s = load i16, ptr %gep, align 2 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; Can't safely load the offset vector, but can load+shuffle if it is profitable. |
| |
| define <8 x i16> @gep01_load_i16_insert_v8i16_deref(ptr align 16 dereferenceable(17) %p) nofree nosync { |
| ; CHECK-LABEL: @gep01_load_i16_insert_v8i16_deref( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 0, i64 1 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 2 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 0, i64 1 |
| %s = load i16, ptr %gep, align 2 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; Verify that alignment of the new load is not over-specified. |
| |
| define <8 x i16> @gep01_load_i16_insert_v8i16_deref_minalign(ptr align 2 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @gep01_load_i16_insert_v8i16_deref_minalign( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 0, i64 1 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 8 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 0, i64 1 |
| %s = load i16, ptr %gep, align 8 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; Negative test - if we are shuffling a load from the base pointer, the address offset |
| ; must be a multiple of element size. |
| ; TODO: Could bitcast around this limitation. |
| |
| define <4 x i32> @gep01_bitcast_load_i32_insert_v4i32(ptr align 1 dereferenceable(16) %p) { |
| ; CHECK-LABEL: @gep01_bitcast_load_i32_insert_v4i32( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <16 x i8>, ptr [[P:%.*]], i64 0, i64 1 |
| ; CHECK-NEXT: [[S:%.*]] = load i32, ptr [[GEP]], align 1 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x i32> undef, i32 [[S]], i64 0 |
| ; CHECK-NEXT: ret <4 x i32> [[R]] |
| ; |
| %gep = getelementptr inbounds <16 x i8>, ptr %p, i64 0, i64 1 |
| %s = load i32, ptr %gep, align 1 |
| %r = insertelement <4 x i32> undef, i32 %s, i64 0 |
| ret <4 x i32> %r |
| } |
| |
| define <4 x i32> @gep012_bitcast_load_i32_insert_v4i32(ptr align 1 dereferenceable(20) %p) nofree nosync { |
| ; CHECK-LABEL: @gep012_bitcast_load_i32_insert_v4i32( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <16 x i8>, ptr [[P:%.*]], i64 0, i64 12 |
| ; CHECK-NEXT: [[S:%.*]] = load i32, ptr [[GEP]], align 1 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x i32> undef, i32 [[S]], i64 0 |
| ; CHECK-NEXT: ret <4 x i32> [[R]] |
| ; |
| %gep = getelementptr inbounds <16 x i8>, ptr %p, i64 0, i64 12 |
| %s = load i32, ptr %gep, align 1 |
| %r = insertelement <4 x i32> undef, i32 %s, i64 0 |
| ret <4 x i32> %r |
| } |
| |
| ; Negative test - if we are shuffling a load from the base pointer, the address offset |
| ; must be a multiple of element size and the offset must be low enough to fit in the vector |
| ; (bitcasting would not help this case). |
| |
| define <4 x i32> @gep013_bitcast_load_i32_insert_v4i32(ptr align 1 dereferenceable(20) %p) nofree nosync { |
| ; CHECK-LABEL: @gep013_bitcast_load_i32_insert_v4i32( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <16 x i8>, ptr [[P:%.*]], i64 0, i64 13 |
| ; CHECK-NEXT: [[S:%.*]] = load i32, ptr [[GEP]], align 1 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x i32> undef, i32 [[S]], i64 0 |
| ; CHECK-NEXT: ret <4 x i32> [[R]] |
| ; |
| %gep = getelementptr inbounds <16 x i8>, ptr %p, i64 0, i64 13 |
| %s = load i32, ptr %gep, align 1 |
| %r = insertelement <4 x i32> undef, i32 %s, i64 0 |
| ret <4 x i32> %r |
| } |
| |
| ; If there are enough dereferenceable bytes, we can offset the vector load. |
| |
| define <8 x i16> @gep10_load_i16_insert_v8i16(ptr align 16 dereferenceable(32) %p) nofree nosync { |
| ; CHECK-LABEL: @gep10_load_i16_insert_v8i16( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 1, i64 0 |
| %s = load i16, ptr %gep, align 16 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; Negative test - disable under asan because widened load can cause spurious |
| ; use-after-poison issues when __asan_poison_memory_region is used. |
| |
| define <8 x i16> @gep10_load_i16_insert_v8i16_asan(ptr align 16 dereferenceable(32) %p) sanitize_address nofree nosync { |
| ; CHECK-LABEL: @gep10_load_i16_insert_v8i16_asan( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 1, i64 0 |
| %s = load i16, ptr %gep, align 16 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; hwasan and memtag should be similarly suppressed. |
| |
| define <8 x i16> @gep10_load_i16_insert_v8i16_hwasan(ptr align 16 dereferenceable(32) %p) sanitize_hwaddress nofree nosync { |
| ; CHECK-LABEL: @gep10_load_i16_insert_v8i16_hwasan( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 1, i64 0 |
| %s = load i16, ptr %gep, align 16 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| define <8 x i16> @gep10_load_i16_insert_v8i16_memtag(ptr align 16 dereferenceable(32) %p) sanitize_memtag nofree nosync { |
| ; CHECK-LABEL: @gep10_load_i16_insert_v8i16_memtag( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 1, i64 0 |
| %s = load i16, ptr %gep, align 16 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; Negative test - disable under tsan because widened load may overlap bytes |
| ; being concurrently modified. tsan does not know that some bytes are undef. |
| |
| define <8 x i16> @gep10_load_i16_insert_v8i16_tsan(ptr align 16 dereferenceable(32) %p) sanitize_thread nofree nosync { |
| ; CHECK-LABEL: @gep10_load_i16_insert_v8i16_tsan( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 1, i64 0 |
| %s = load i16, ptr %gep, align 16 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; Negative test - can't safely load the offset vector, but could load+shuffle. |
| |
| define <8 x i16> @gep10_load_i16_insert_v8i16_deref(ptr align 16 dereferenceable(31) %p) nofree nosync { |
| ; CHECK-LABEL: @gep10_load_i16_insert_v8i16_deref( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <8 x i16>, ptr [[P:%.*]], i64 1, i64 0 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 16 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <8 x i16>, ptr %p, i64 1, i64 0 |
| %s = load i16, ptr %gep, align 16 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; Negative test - do not alter volatile. |
| |
| define <4 x float> @load_f32_insert_v4f32_volatile(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_f32_insert_v4f32_volatile( |
| ; CHECK-NEXT: [[S:%.*]] = load volatile float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %s = load volatile float, ptr %p, align 4 |
| %r = insertelement <4 x float> undef, float %s, i32 0 |
| ret <4 x float> %r |
| } |
| |
| ; Pointer is not as aligned as load, but that's ok. |
| ; The new load uses the larger alignment value. |
| |
| define <4 x float> @load_f32_insert_v4f32_align(ptr align 1 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_f32_insert_v4f32_align( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 4 |
| %r = insertelement <4 x float> undef, float %s, i32 0 |
| ret <4 x float> %r |
| } |
| |
| ; Negative test - not enough bytes. |
| |
| define <4 x float> @load_f32_insert_v4f32_deref(ptr align 4 dereferenceable(15) %p) nofree nosync { |
| ; CHECK-LABEL: @load_f32_insert_v4f32_deref( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 4 |
| %r = insertelement <4 x float> undef, float %s, i32 0 |
| ret <4 x float> %r |
| } |
| |
| define <8 x i32> @load_i32_insert_v8i32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_i32_insert_v8i32( |
| ; CHECK-NEXT: [[S:%.*]] = load i32, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i32> undef, i32 [[S]], i32 0 |
| ; CHECK-NEXT: ret <8 x i32> [[R]] |
| ; |
| %s = load i32, ptr %p, align 4 |
| %r = insertelement <8 x i32> undef, i32 %s, i32 0 |
| ret <8 x i32> %r |
| } |
| |
| define <8 x i32> @casted_load_i32_insert_v8i32(ptr align 4 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @casted_load_i32_insert_v8i32( |
| ; CHECK-NEXT: [[S:%.*]] = load i32, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i32> undef, i32 [[S]], i32 0 |
| ; CHECK-NEXT: ret <8 x i32> [[R]] |
| ; |
| %s = load i32, ptr %p, align 4 |
| %r = insertelement <8 x i32> undef, i32 %s, i32 0 |
| ret <8 x i32> %r |
| } |
| |
| define <16 x float> @load_f32_insert_v16f32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_f32_insert_v16f32( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <16 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <16 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 4 |
| %r = insertelement <16 x float> undef, float %s, i32 0 |
| ret <16 x float> %r |
| } |
| |
| define <2 x float> @load_f32_insert_v2f32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_f32_insert_v2f32( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <2 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <2 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 4 |
| %r = insertelement <2 x float> undef, float %s, i32 0 |
| ret <2 x float> %r |
| } |
| |
| ; Negative test - suppress load widening for asan/hwasan/memtag/tsan. |
| |
| define <2 x float> @load_f32_insert_v2f32_asan(ptr align 16 dereferenceable(16) %p) sanitize_address { |
| ; CHECK-LABEL: @load_f32_insert_v2f32_asan( |
| ; CHECK-NEXT: [[S:%.*]] = load float, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <2 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <2 x float> [[R]] |
| ; |
| %s = load float, ptr %p, align 4 |
| %r = insertelement <2 x float> undef, float %s, i32 0 |
| ret <2 x float> %r |
| } |
| |
| declare ptr @getscaleptr() |
| define void @PR47558_multiple_use_load(ptr nocapture nonnull %resultptr, ptr nocapture nonnull readonly %opptr) { |
| ; CHECK-LABEL: @PR47558_multiple_use_load( |
| ; CHECK-NEXT: [[SCALEPTR:%.*]] = tail call nonnull align 16 dereferenceable(64) ptr @getscaleptr() |
| ; CHECK-NEXT: [[OP:%.*]] = load <2 x float>, ptr [[OPPTR:%.*]], align 4 |
| ; CHECK-NEXT: [[SCALE:%.*]] = load float, ptr [[SCALEPTR]], align 16 |
| ; CHECK-NEXT: [[T1:%.*]] = insertelement <2 x float> undef, float [[SCALE]], i32 0 |
| ; CHECK-NEXT: [[T2:%.*]] = insertelement <2 x float> [[T1]], float [[SCALE]], i32 1 |
| ; CHECK-NEXT: [[T3:%.*]] = fmul <2 x float> [[OP]], [[T2]] |
| ; CHECK-NEXT: [[T4:%.*]] = extractelement <2 x float> [[T3]], i32 0 |
| ; CHECK-NEXT: [[RESULT0:%.*]] = insertelement <2 x float> undef, float [[T4]], i32 0 |
| ; CHECK-NEXT: [[RESULT1:%.*]] = shufflevector <2 x float> [[RESULT0]], <2 x float> [[T3]], <2 x i32> <i32 0, i32 3> |
| ; CHECK-NEXT: store <2 x float> [[RESULT1]], ptr [[RESULTPTR:%.*]], align 8 |
| ; CHECK-NEXT: ret void |
| ; |
| %scaleptr = tail call nonnull align 16 dereferenceable(64) ptr @getscaleptr() |
| %op = load <2 x float>, ptr %opptr, align 4 |
| %scale = load float, ptr %scaleptr, align 16 |
| %t1 = insertelement <2 x float> undef, float %scale, i32 0 |
| %t2 = insertelement <2 x float> %t1, float %scale, i32 1 |
| %t3 = fmul <2 x float> %op, %t2 |
| %t4 = extractelement <2 x float> %t3, i32 0 |
| %result0 = insertelement <2 x float> undef, float %t4, i32 0 |
| %t5 = extractelement <2 x float> %t3, i32 1 |
| %result1 = insertelement <2 x float> %result0, float %t5, i32 1 |
| store <2 x float> %result1, ptr %resultptr, align 8 |
| ret void |
| } |
| |
| define <4 x float> @load_v2f32_extract_insert_v4f32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @load_v2f32_extract_insert_v4f32( |
| ; CHECK-NEXT: [[L:%.*]] = load <2 x float>, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: [[S:%.*]] = extractelement <2 x float> [[L]], i32 0 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; CHECK-NEXT: ret <4 x float> [[R]] |
| ; |
| %l = load <2 x float>, ptr %p, align 4 |
| %s = extractelement <2 x float> %l, i32 0 |
| %r = insertelement <4 x float> undef, float %s, i32 0 |
| ret <4 x float> %r |
| } |
| |
| define <4 x float> @load_v8f32_extract_insert_v4f32(ptr align 16 dereferenceable(16) %p) nofree nosync { |
| ; SSE2-LABEL: @load_v8f32_extract_insert_v4f32( |
| ; SSE2-NEXT: [[S:%.*]] = load float, ptr [[TMP1:%.*]], align 4 |
| ; SSE2-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; SSE2-NEXT: ret <4 x float> [[R]] |
| ; |
| ; AVX2-LABEL: @load_v8f32_extract_insert_v4f32( |
| ; AVX2-NEXT: [[L:%.*]] = load <8 x float>, ptr [[P:%.*]], align 4 |
| ; AVX2-NEXT: [[S:%.*]] = extractelement <8 x float> [[L]], i32 0 |
| ; AVX2-NEXT: [[R:%.*]] = insertelement <4 x float> undef, float [[S]], i32 0 |
| ; AVX2-NEXT: ret <4 x float> [[R]] |
| ; |
| %l = load <8 x float>, ptr %p, align 4 |
| %s = extractelement <8 x float> %l, i32 0 |
| %r = insertelement <4 x float> undef, float %s, i32 0 |
| ret <4 x float> %r |
| } |
| |
| define <8 x i32> @load_v1i32_extract_insert_v8i32_extra_use(ptr align 16 dereferenceable(16) %p, ptr %store_ptr) nofree nosync { |
| ; CHECK-LABEL: @load_v1i32_extract_insert_v8i32_extra_use( |
| ; CHECK-NEXT: [[L:%.*]] = load <1 x i32>, ptr [[P:%.*]], align 4 |
| ; CHECK-NEXT: store <1 x i32> [[L]], ptr [[STORE_PTR:%.*]], align 4 |
| ; CHECK-NEXT: [[TMP1:%.*]] = shufflevector <1 x i32> [[L]], <1 x i32> poison, <8 x i32> <i32 0, i32 poison, i32 poison, i32 poison, i32 poison, i32 poison, i32 poison, i32 poison> |
| ; CHECK-NEXT: [[R:%.*]] = shufflevector <8 x i32> [[TMP1]], <8 x i32> undef, <8 x i32> <i32 0, i32 9, i32 10, i32 11, i32 12, i32 13, i32 14, i32 15> |
| ; CHECK-NEXT: ret <8 x i32> [[R]] |
| ; |
| %l = load <1 x i32>, ptr %p, align 4 |
| store <1 x i32> %l, ptr %store_ptr |
| %s = extractelement <1 x i32> %l, i32 0 |
| %r = insertelement <8 x i32> undef, i32 %s, i32 0 |
| ret <8 x i32> %r |
| } |
| |
| ; Can't safely load the offset vector, but can load+shuffle if it is profitable. |
| |
| define <8 x i16> @gep1_load_v2i16_extract_insert_v8i16(ptr align 1 dereferenceable(16) %p) nofree nosync { |
| ; CHECK-LABEL: @gep1_load_v2i16_extract_insert_v8i16( |
| ; CHECK-NEXT: [[GEP:%.*]] = getelementptr inbounds <2 x i16>, ptr [[P:%.*]], i64 1 |
| ; CHECK-NEXT: [[S:%.*]] = load i16, ptr [[GEP]], align 8 |
| ; CHECK-NEXT: [[R:%.*]] = insertelement <8 x i16> undef, i16 [[S]], i64 0 |
| ; CHECK-NEXT: ret <8 x i16> [[R]] |
| ; |
| %gep = getelementptr inbounds <2 x i16>, ptr %p, i64 1 |
| %l = load <2 x i16>, ptr %gep, align 8 |
| %s = extractelement <2 x i16> %l, i32 0 |
| %r = insertelement <8 x i16> undef, i16 %s, i64 0 |
| ret <8 x i16> %r |
| } |
| |
| ; PR30986 - split vector loads for scalarized operations |
| define <2 x i64> @PR30986(ptr %0) { |
| ; CHECK-LABEL: @PR30986( |
| ; CHECK-NEXT: [[TMP3:%.*]] = load i64, ptr [[TMP2:%.*]], align 16 |
| ; CHECK-NEXT: [[TMP4:%.*]] = tail call i64 @llvm.ctpop.i64(i64 [[TMP3]]) |
| ; CHECK-NEXT: [[TMP5:%.*]] = insertelement <2 x i64> undef, i64 [[TMP4]], i32 0 |
| ; CHECK-NEXT: [[TMP6:%.*]] = getelementptr inbounds <2 x i64>, ptr [[TMP2]], i32 0, i32 1 |
| ; CHECK-NEXT: [[TMP7:%.*]] = load i64, ptr [[TMP6]], align 8 |
| ; CHECK-NEXT: [[TMP8:%.*]] = tail call i64 @llvm.ctpop.i64(i64 [[TMP7]]) |
| ; CHECK-NEXT: [[TMP9:%.*]] = insertelement <2 x i64> [[TMP5]], i64 [[TMP8]], i32 1 |
| ; CHECK-NEXT: ret <2 x i64> [[TMP9]] |
| ; |
| %2 = load <2 x i64>, ptr %0, align 16 |
| %3 = extractelement <2 x i64> %2, i32 0 |
| %4 = tail call i64 @llvm.ctpop.i64(i64 %3) |
| %5 = insertelement <2 x i64> undef, i64 %4, i32 0 |
| %6 = extractelement <2 x i64> %2, i32 1 |
| %7 = tail call i64 @llvm.ctpop.i64(i64 %6) |
| %8 = insertelement <2 x i64> %5, i64 %7, i32 1 |
| ret <2 x i64> %8 |
| } |
| declare i64 @llvm.ctpop.i64(i64) |