| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| ; RUN: llc -mtriple=aarch64-linux-gnu < %s | FileCheck %s |
| ; RUN: not --crash llc -mtriple=aarch64_be < %s |
| |
| ; |
| ; bitcast to nxv16i8 |
| ; |
| |
| define <vscale x 16 x i8> @bitcast_nxv8i16_to_nxv16i8(<vscale x 8 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i16_to_nxv16i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i16> %v to <vscale x 16 x i8> |
| ret <vscale x 16 x i8> %bc |
| } |
| |
| define <vscale x 16 x i8> @bitcast_nxv4i32_to_nxv16i8(<vscale x 4 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i32_to_nxv16i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i32> %v to <vscale x 16 x i8> |
| ret <vscale x 16 x i8> %bc |
| } |
| |
| define <vscale x 16 x i8> @bitcast_nxv2i64_to_nxv16i8(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i64_to_nxv16i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i64> %v to <vscale x 16 x i8> |
| ret <vscale x 16 x i8> %bc |
| } |
| |
| define <vscale x 16 x i8> @bitcast_nxv8f16_to_nxv16i8(<vscale x 8 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8f16_to_nxv16i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x half> %v to <vscale x 16 x i8> |
| ret <vscale x 16 x i8> %bc |
| } |
| |
| define <vscale x 16 x i8> @bitcast_nxv4f32_to_nxv16i8(<vscale x 4 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f32_to_nxv16i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x float> %v to <vscale x 16 x i8> |
| ret <vscale x 16 x i8> %bc |
| } |
| |
| define <vscale x 16 x i8> @bitcast_nxv2f64_to_nxv16i8(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f64_to_nxv16i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x double> %v to <vscale x 16 x i8> |
| ret <vscale x 16 x i8> %bc |
| } |
| |
| define <vscale x 16 x i8> @bitcast_nxv8bf16_to_nxv16i8(<vscale x 8 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8bf16_to_nxv16i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x bfloat> %v to <vscale x 16 x i8> |
| ret <vscale x 16 x i8> %bc |
| } |
| |
| ; |
| ; bitcast to nxv8i16 |
| ; |
| |
| define <vscale x 8 x i16> @bitcast_nxv16i8_to_nxv8i16(<vscale x 16 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv16i8_to_nxv8i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 16 x i8> %v to <vscale x 8 x i16> |
| ret <vscale x 8 x i16> %bc |
| } |
| |
| define <vscale x 8 x i16> @bitcast_nxv4i32_to_nxv8i16(<vscale x 4 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i32_to_nxv8i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i32> %v to <vscale x 8 x i16> |
| ret <vscale x 8 x i16> %bc |
| } |
| |
| define <vscale x 8 x i16> @bitcast_nxv2i64_to_nxv8i16(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i64_to_nxv8i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i64> %v to <vscale x 8 x i16> |
| ret <vscale x 8 x i16> %bc |
| } |
| |
| define <vscale x 8 x i16> @bitcast_nxv8f16_to_nxv8i16(<vscale x 8 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8f16_to_nxv8i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x half> %v to <vscale x 8 x i16> |
| ret <vscale x 8 x i16> %bc |
| } |
| |
| define <vscale x 8 x i16> @bitcast_nxv4f32_to_nxv8i16(<vscale x 4 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f32_to_nxv8i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x float> %v to <vscale x 8 x i16> |
| ret <vscale x 8 x i16> %bc |
| } |
| |
| define <vscale x 8 x i16> @bitcast_nxv2f64_to_nxv8i16(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f64_to_nxv8i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x double> %v to <vscale x 8 x i16> |
| ret <vscale x 8 x i16> %bc |
| } |
| |
| define <vscale x 8 x i16> @bitcast_nxv8bf16_to_nxv8i16(<vscale x 8 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8bf16_to_nxv8i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x bfloat> %v to <vscale x 8 x i16> |
| ret <vscale x 8 x i16> %bc |
| } |
| |
| ; |
| ; bitcast to nxv4i32 |
| ; |
| |
| define <vscale x 4 x i32> @bitcast_nxv16i8_to_nxv4i32(<vscale x 16 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv16i8_to_nxv4i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 16 x i8> %v to <vscale x 4 x i32> |
| ret <vscale x 4 x i32> %bc |
| } |
| |
| define <vscale x 4 x i32> @bitcast_nxv8i16_to_nxv4i32(<vscale x 8 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i16_to_nxv4i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i16> %v to <vscale x 4 x i32> |
| ret <vscale x 4 x i32> %bc |
| } |
| |
| define <vscale x 4 x i32> @bitcast_nxv2i64_to_nxv4i32(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i64_to_nxv4i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i64> %v to <vscale x 4 x i32> |
| ret <vscale x 4 x i32> %bc |
| } |
| |
| define <vscale x 4 x i32> @bitcast_nxv8f16_to_nxv4i32(<vscale x 8 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8f16_to_nxv4i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x half> %v to <vscale x 4 x i32> |
| ret <vscale x 4 x i32> %bc |
| } |
| |
| define <vscale x 4 x i32> @bitcast_nxv4f32_to_nxv4i32(<vscale x 4 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f32_to_nxv4i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x float> %v to <vscale x 4 x i32> |
| ret <vscale x 4 x i32> %bc |
| } |
| |
| define <vscale x 4 x i32> @bitcast_nxv2f64_to_nxv4i32(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f64_to_nxv4i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x double> %v to <vscale x 4 x i32> |
| ret <vscale x 4 x i32> %bc |
| } |
| |
| define <vscale x 4 x i32> @bitcast_nxv8bf16_to_nxv4i32(<vscale x 8 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8bf16_to_nxv4i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x bfloat> %v to <vscale x 4 x i32> |
| ret <vscale x 4 x i32> %bc |
| } |
| |
| ; |
| ; bitcast to nxv2i64 |
| ; |
| |
| define <vscale x 2 x i64> @bitcast_nxv16i8_to_nxv2i64(<vscale x 16 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv16i8_to_nxv2i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 16 x i8> %v to <vscale x 2 x i64> |
| ret <vscale x 2 x i64> %bc |
| } |
| |
| define <vscale x 2 x i64> @bitcast_nxv8i16_to_nxv2i64(<vscale x 8 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i16_to_nxv2i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i16> %v to <vscale x 2 x i64> |
| ret <vscale x 2 x i64> %bc |
| } |
| |
| define <vscale x 2 x i64> @bitcast_nxv4i32_to_nxv2i64(<vscale x 4 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i32_to_nxv2i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i32> %v to <vscale x 2 x i64> |
| ret <vscale x 2 x i64> %bc |
| } |
| |
| define <vscale x 2 x i64> @bitcast_nxv8f16_to_nxv2i64(<vscale x 8 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8f16_to_nxv2i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x half> %v to <vscale x 2 x i64> |
| ret <vscale x 2 x i64> %bc |
| } |
| |
| define <vscale x 2 x i64> @bitcast_nxv4f32_to_nxv2i64(<vscale x 4 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f32_to_nxv2i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x float> %v to <vscale x 2 x i64> |
| ret <vscale x 2 x i64> %bc |
| } |
| |
| define <vscale x 2 x i64> @bitcast_nxv2f64_to_nxv2i64(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f64_to_nxv2i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x double> %v to <vscale x 2 x i64> |
| ret <vscale x 2 x i64> %bc |
| } |
| |
| define <vscale x 2 x i64> @bitcast_nxv8bf16_to_nxv2i64(<vscale x 8 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8bf16_to_nxv2i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x bfloat> %v to <vscale x 2 x i64> |
| ret <vscale x 2 x i64> %bc |
| } |
| |
| ; |
| ; bitcast to nxv8f16 |
| ; |
| |
| define <vscale x 8 x half> @bitcast_nxv16i8_to_nxv8f16(<vscale x 16 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv16i8_to_nxv8f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 16 x i8> %v to <vscale x 8 x half> |
| ret <vscale x 8 x half> %bc |
| } |
| |
| define <vscale x 8 x half> @bitcast_nxv8i16_to_nxv8f16(<vscale x 8 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i16_to_nxv8f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i16> %v to <vscale x 8 x half> |
| ret <vscale x 8 x half> %bc |
| } |
| |
| define <vscale x 8 x half> @bitcast_nxv4i32_to_nxv8f16(<vscale x 4 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i32_to_nxv8f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i32> %v to <vscale x 8 x half> |
| ret <vscale x 8 x half> %bc |
| } |
| |
| define <vscale x 8 x half> @bitcast_nxv2i64_to_nxv8f16(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i64_to_nxv8f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i64> %v to <vscale x 8 x half> |
| ret <vscale x 8 x half> %bc |
| } |
| |
| define <vscale x 8 x half> @bitcast_nxv4f32_to_nxv8f16(<vscale x 4 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f32_to_nxv8f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x float> %v to <vscale x 8 x half> |
| ret <vscale x 8 x half> %bc |
| } |
| |
| define <vscale x 8 x half> @bitcast_nxv2f64_to_nxv8f16(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f64_to_nxv8f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x double> %v to <vscale x 8 x half> |
| ret <vscale x 8 x half> %bc |
| } |
| |
| define <vscale x 8 x half> @bitcast_nxv8bf16_to_nxv8f16(<vscale x 8 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8bf16_to_nxv8f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x bfloat> %v to <vscale x 8 x half> |
| ret <vscale x 8 x half> %bc |
| } |
| |
| ; |
| ; bitcast to nxv4f32 |
| ; |
| |
| define <vscale x 4 x float> @bitcast_nxv16i8_to_nxv4f32(<vscale x 16 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv16i8_to_nxv4f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 16 x i8> %v to <vscale x 4 x float> |
| ret <vscale x 4 x float> %bc |
| } |
| |
| define <vscale x 4 x float> @bitcast_nxv8i16_to_nxv4f32(<vscale x 8 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i16_to_nxv4f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i16> %v to <vscale x 4 x float> |
| ret <vscale x 4 x float> %bc |
| } |
| |
| define <vscale x 4 x float> @bitcast_nxv4i32_to_nxv4f32(<vscale x 4 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i32_to_nxv4f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i32> %v to <vscale x 4 x float> |
| ret <vscale x 4 x float> %bc |
| } |
| |
| define <vscale x 4 x float> @bitcast_nxv2i64_to_nxv4f32(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i64_to_nxv4f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i64> %v to <vscale x 4 x float> |
| ret <vscale x 4 x float> %bc |
| } |
| |
| define <vscale x 4 x float> @bitcast_nxv8f16_to_nxv4f32(<vscale x 8 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8f16_to_nxv4f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x half> %v to <vscale x 4 x float> |
| ret <vscale x 4 x float> %bc |
| } |
| |
| define <vscale x 4 x float> @bitcast_nxv2f64_to_nxv4f32(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f64_to_nxv4f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x double> %v to <vscale x 4 x float> |
| ret <vscale x 4 x float> %bc |
| } |
| |
| define <vscale x 4 x float> @bitcast_nxv8bf16_to_nxv4f32(<vscale x 8 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8bf16_to_nxv4f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x bfloat> %v to <vscale x 4 x float> |
| ret <vscale x 4 x float> %bc |
| } |
| |
| ; |
| ; bitcast to nxv2f64 |
| ; |
| |
| define <vscale x 2 x double> @bitcast_nxv16i8_to_nxv2f64(<vscale x 16 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv16i8_to_nxv2f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 16 x i8> %v to <vscale x 2 x double> |
| ret <vscale x 2 x double> %bc |
| } |
| |
| define <vscale x 2 x double> @bitcast_nxv8i16_to_nxv2f64(<vscale x 8 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i16_to_nxv2f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i16> %v to <vscale x 2 x double> |
| ret <vscale x 2 x double> %bc |
| } |
| |
| define <vscale x 2 x double> @bitcast_nxv4i32_to_nxv2f64(<vscale x 4 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i32_to_nxv2f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i32> %v to <vscale x 2 x double> |
| ret <vscale x 2 x double> %bc |
| } |
| |
| define <vscale x 2 x double> @bitcast_nxv2i64_to_nxv2f64(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i64_to_nxv2f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i64> %v to <vscale x 2 x double> |
| ret <vscale x 2 x double> %bc |
| } |
| |
| define <vscale x 2 x double> @bitcast_nxv8f16_to_nxv2f64(<vscale x 8 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8f16_to_nxv2f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x half> %v to <vscale x 2 x double> |
| ret <vscale x 2 x double> %bc |
| } |
| |
| define <vscale x 2 x double> @bitcast_nxv4f32_to_nxv2f64(<vscale x 4 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f32_to_nxv2f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x float> %v to <vscale x 2 x double> |
| ret <vscale x 2 x double> %bc |
| } |
| |
| define <vscale x 2 x double> @bitcast_nxv8bf16_to_nxv2f64(<vscale x 8 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8bf16_to_nxv2f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x bfloat> %v to <vscale x 2 x double> |
| ret <vscale x 2 x double> %bc |
| } |
| |
| ; |
| ; bitcast to nxv8bf16 |
| ; |
| |
| define <vscale x 8 x bfloat> @bitcast_nxv16i8_to_nxv8bf16(<vscale x 16 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv16i8_to_nxv8bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 16 x i8> %v to <vscale x 8 x bfloat> |
| ret <vscale x 8 x bfloat> %bc |
| } |
| |
| define <vscale x 8 x bfloat> @bitcast_nxv8i16_to_nxv8bf16(<vscale x 8 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i16_to_nxv8bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i16> %v to <vscale x 8 x bfloat> |
| ret <vscale x 8 x bfloat> %bc |
| } |
| |
| define <vscale x 8 x bfloat> @bitcast_nxv4i32_to_nxv8bf16(<vscale x 4 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i32_to_nxv8bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i32> %v to <vscale x 8 x bfloat> |
| ret <vscale x 8 x bfloat> %bc |
| } |
| |
| define <vscale x 8 x bfloat> @bitcast_nxv2i64_to_nxv8bf16(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i64_to_nxv8bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i64> %v to <vscale x 8 x bfloat> |
| ret <vscale x 8 x bfloat> %bc |
| } |
| |
| define <vscale x 8 x bfloat> @bitcast_nxv8f16_to_nxv8bf16(<vscale x 8 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8f16_to_nxv8bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x half> %v to <vscale x 8 x bfloat> |
| ret <vscale x 8 x bfloat> %bc |
| } |
| |
| define <vscale x 8 x bfloat> @bitcast_nxv4f32_to_nxv8bf16(<vscale x 4 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f32_to_nxv8bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x float> %v to <vscale x 8 x bfloat> |
| ret <vscale x 8 x bfloat> %bc |
| } |
| |
| define <vscale x 8 x bfloat> @bitcast_nxv2f64_to_nxv8bf16(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f64_to_nxv8bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x double> %v to <vscale x 8 x bfloat> |
| ret <vscale x 8 x bfloat> %bc |
| } |
| |
| ; |
| ; bitcast to nxv8i8 |
| ; |
| |
| define <vscale x 8 x i8> @bitcast_nxv4i16_to_nxv8i8(<vscale x 4 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i16_to_nxv8i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: ld1b { z0.h }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i16> %v to <vscale x 8 x i8> |
| ret <vscale x 8 x i8> %bc |
| } |
| |
| define <vscale x 8 x i8> @bitcast_nxv2i32_to_nxv8i8(<vscale x 2 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i32_to_nxv8i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: ld1b { z0.h }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i32> %v to <vscale x 8 x i8> |
| ret <vscale x 8 x i8> %bc |
| } |
| |
| define <vscale x 8 x i8> @bitcast_nxv1i64_to_nxv8i8(<vscale x 1 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i64_to_nxv8i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.h, z0.b |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i64> %v to <vscale x 8 x i8> |
| ret <vscale x 8 x i8> %bc |
| } |
| |
| define <vscale x 8 x i8> @bitcast_nxv4f16_to_nxv8i8(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f16_to_nxv8i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: ld1b { z0.h }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x half> %v to <vscale x 8 x i8> |
| ret <vscale x 8 x i8> %bc |
| } |
| |
| define <vscale x 8 x i8> @bitcast_nxv2f32_to_nxv8i8(<vscale x 2 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f32_to_nxv8i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: ld1b { z0.h }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x float> %v to <vscale x 8 x i8> |
| ret <vscale x 8 x i8> %bc |
| } |
| |
| define <vscale x 8 x i8> @bitcast_nxv1f64_to_nxv8i8(<vscale x 1 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1f64_to_nxv8i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.h, z0.b |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x double> %v to <vscale x 8 x i8> |
| ret <vscale x 8 x i8> %bc |
| } |
| |
| define <vscale x 8 x i8> @bitcast_nxv4bf16_to_nxv8i8(<vscale x 4 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4bf16_to_nxv8i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: ld1b { z0.h }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x bfloat> %v to <vscale x 8 x i8> |
| ret <vscale x 8 x i8> %bc |
| } |
| |
| ; |
| ; bitcast to nxv4i16 |
| ; |
| |
| define <vscale x 4 x i16> @bitcast_nxv8i8_to_nxv4i16(<vscale x 8 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i8_to_nxv4i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: st1b { z0.h }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i8> %v to <vscale x 4 x i16> |
| ret <vscale x 4 x i16> %bc |
| } |
| |
| define <vscale x 4 x i16> @bitcast_nxv2i32_to_nxv4i16(<vscale x 2 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i32_to_nxv4i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i32> %v to <vscale x 4 x i16> |
| ret <vscale x 4 x i16> %bc |
| } |
| |
| define <vscale x 4 x i16> @bitcast_nxv1i64_to_nxv4i16(<vscale x 1 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i64_to_nxv4i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i64> %v to <vscale x 4 x i16> |
| ret <vscale x 4 x i16> %bc |
| } |
| |
| define <vscale x 4 x i16> @bitcast_nxv4f16_to_nxv4i16(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f16_to_nxv4i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x half> %v to <vscale x 4 x i16> |
| ret <vscale x 4 x i16> %bc |
| } |
| |
| define <vscale x 4 x i16> @bitcast_nxv2f32_to_nxv4i16(<vscale x 2 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f32_to_nxv4i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x float> %v to <vscale x 4 x i16> |
| ret <vscale x 4 x i16> %bc |
| } |
| |
| define <vscale x 4 x i16> @bitcast_nxv1f64_to_nxv4i16(<vscale x 1 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1f64_to_nxv4i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x double> %v to <vscale x 4 x i16> |
| ret <vscale x 4 x i16> %bc |
| } |
| |
| define <vscale x 4 x i16> @bitcast_nxv4bf16_to_nxv4i16(<vscale x 4 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4bf16_to_nxv4i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x bfloat> %v to <vscale x 4 x i16> |
| ret <vscale x 4 x i16> %bc |
| } |
| |
| ; |
| ; bitcast to nxv2i32 |
| ; |
| |
| define <vscale x 2 x i32> @bitcast_nxv8i8_to_nxv2i32(<vscale x 8 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i8_to_nxv2i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: st1b { z0.h }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i8> %v to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bc |
| } |
| |
| define <vscale x 2 x i32> @bitcast_nxv4i16_to_nxv2i32(<vscale x 4 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i16_to_nxv2i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i16> %v to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bc |
| } |
| |
| define <vscale x 2 x i32> @bitcast_nxv1i64_to_nxv2i32(<vscale x 1 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i64_to_nxv2i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.d, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i64> %v to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bc |
| } |
| |
| define <vscale x 2 x i32> @bitcast_nxv4f16_to_nxv2i32(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f16_to_nxv2i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x half> %v to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bc |
| } |
| |
| define <vscale x 2 x i32> @bitcast_nxv2f32_to_nxv2i32(<vscale x 2 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f32_to_nxv2i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x float> %v to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bc |
| } |
| |
| define <vscale x 2 x i32> @bitcast_nxv1f64_to_nxv2i32(<vscale x 1 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1f64_to_nxv2i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.d, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x double> %v to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bc |
| } |
| |
| define <vscale x 2 x i32> @bitcast_nxv4bf16_to_nxv2i32(<vscale x 4 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4bf16_to_nxv2i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x bfloat> %v to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bc |
| } |
| |
| ; |
| ; bitcast to nxv1i64 |
| ; |
| |
| define <vscale x 1 x i64> @bitcast_nxv8i8_to_nxv1i64(<vscale x 8 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i8_to_nxv1i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.b, z0.b, z0.b |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i8> %v to <vscale x 1 x i64> |
| ret <vscale x 1 x i64> %bc |
| } |
| |
| define <vscale x 1 x i64> @bitcast_nxv4i16_to_nxv1i64(<vscale x 4 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i16_to_nxv1i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i16> %v to <vscale x 1 x i64> |
| ret <vscale x 1 x i64> %bc |
| } |
| |
| define <vscale x 1 x i64> @bitcast_nxv2i32_to_nxv1i64(<vscale x 2 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i32_to_nxv1i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.s, z0.s, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i32> %v to <vscale x 1 x i64> |
| ret <vscale x 1 x i64> %bc |
| } |
| |
| define <vscale x 1 x i64> @bitcast_nxv4f16_to_nxv1i64(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f16_to_nxv1i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x half> %v to <vscale x 1 x i64> |
| ret <vscale x 1 x i64> %bc |
| } |
| |
| define <vscale x 1 x i64> @bitcast_nxv2f32_to_nxv1i64(<vscale x 2 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f32_to_nxv1i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.s, z0.s, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x float> %v to <vscale x 1 x i64> |
| ret <vscale x 1 x i64> %bc |
| } |
| |
| define <vscale x 1 x i64> @bitcast_nxv1f64_to_nxv1i64(<vscale x 1 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1f64_to_nxv1i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x double> %v to <vscale x 1 x i64> |
| ret <vscale x 1 x i64> %bc |
| } |
| |
| define <vscale x 1 x i64> @bitcast_nxv4bf16_to_nxv1i64(<vscale x 4 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4bf16_to_nxv1i64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x bfloat> %v to <vscale x 1 x i64> |
| ret <vscale x 1 x i64> %bc |
| } |
| |
| ; |
| ; bitcast to nxv4f16 |
| ; |
| |
| define <vscale x 4 x half> @bitcast_nxv8i8_to_nxv4f16(<vscale x 8 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i8_to_nxv4f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: st1b { z0.h }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i8> %v to <vscale x 4 x half> |
| ret <vscale x 4 x half> %bc |
| } |
| |
| define <vscale x 4 x half> @bitcast_nxv4i16_to_nxv4f16(<vscale x 4 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i16_to_nxv4f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i16> %v to <vscale x 4 x half> |
| ret <vscale x 4 x half> %bc |
| } |
| |
| define <vscale x 4 x half> @bitcast_nxv2i32_to_nxv4f16(<vscale x 2 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i32_to_nxv4f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i32> %v to <vscale x 4 x half> |
| ret <vscale x 4 x half> %bc |
| } |
| |
| define <vscale x 4 x half> @bitcast_nxv1i64_to_nxv4f16(<vscale x 1 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i64_to_nxv4f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i64> %v to <vscale x 4 x half> |
| ret <vscale x 4 x half> %bc |
| } |
| |
| define <vscale x 4 x half> @bitcast_nxv2f32_to_nxv4f16(<vscale x 2 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f32_to_nxv4f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x float> %v to <vscale x 4 x half> |
| ret <vscale x 4 x half> %bc |
| } |
| |
| define <vscale x 4 x half> @bitcast_nxv1f64_to_nxv4f16(<vscale x 1 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1f64_to_nxv4f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x double> %v to <vscale x 4 x half> |
| ret <vscale x 4 x half> %bc |
| } |
| |
| define <vscale x 4 x half> @bitcast_nxv4bf16_to_nxv4f16(<vscale x 4 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4bf16_to_nxv4f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x bfloat> %v to <vscale x 4 x half> |
| ret <vscale x 4 x half> %bc |
| } |
| |
| ; |
| ; bitcast to nxv2f32 |
| ; |
| |
| define <vscale x 2 x float> @bitcast_nxv8i8_to_nxv2f32(<vscale x 8 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i8_to_nxv2f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: st1b { z0.h }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i8> %v to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bc |
| } |
| |
| define <vscale x 2 x float> @bitcast_nxv4i16_to_nxv2f32(<vscale x 4 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i16_to_nxv2f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i16> %v to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bc |
| } |
| |
| define <vscale x 2 x float> @bitcast_nxv2i32_to_nxv2f32(<vscale x 2 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i32_to_nxv2f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i32> %v to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bc |
| } |
| |
| define <vscale x 2 x float> @bitcast_nxv1i64_to_nxv2f32(<vscale x 1 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i64_to_nxv2f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.d, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i64> %v to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bc |
| } |
| |
| define <vscale x 2 x float> @bitcast_nxv4f16_to_nxv2f32(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f16_to_nxv2f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x half> %v to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bc |
| } |
| |
| define <vscale x 2 x float> @bitcast_nxv1f64_to_nxv2f32(<vscale x 1 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1f64_to_nxv2f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.d, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x double> %v to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bc |
| } |
| |
| define <vscale x 2 x float> @bitcast_nxv4bf16_to_nxv2f32(<vscale x 4 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4bf16_to_nxv2f32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x bfloat> %v to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bc |
| } |
| |
| ; |
| ; bitcast to nxv1f64 |
| ; |
| |
| define <vscale x 1 x double> @bitcast_nxv8i8_to_nxv1f64(<vscale x 8 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i8_to_nxv1f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.b, z0.b, z0.b |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i8> %v to <vscale x 1 x double> |
| ret <vscale x 1 x double> %bc |
| } |
| |
| define <vscale x 1 x double> @bitcast_nxv4i16_to_nxv1f64(<vscale x 4 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i16_to_nxv1f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i16> %v to <vscale x 1 x double> |
| ret <vscale x 1 x double> %bc |
| } |
| |
| define <vscale x 1 x double> @bitcast_nxv2i32_to_nxv1f64(<vscale x 2 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i32_to_nxv1f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.s, z0.s, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i32> %v to <vscale x 1 x double> |
| ret <vscale x 1 x double> %bc |
| } |
| |
| define <vscale x 1 x double> @bitcast_nxv1i64_to_nxv1f64(<vscale x 1 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i64_to_nxv1f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i64> %v to <vscale x 1 x double> |
| ret <vscale x 1 x double> %bc |
| } |
| |
| define <vscale x 1 x double> @bitcast_nxv4f16_to_nxv1f64(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f16_to_nxv1f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x half> %v to <vscale x 1 x double> |
| ret <vscale x 1 x double> %bc |
| } |
| |
| define <vscale x 1 x double> @bitcast_nxv2f32_to_nxv1f64(<vscale x 2 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f32_to_nxv1f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.s, z0.s, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x float> %v to <vscale x 1 x double> |
| ret <vscale x 1 x double> %bc |
| } |
| |
| define <vscale x 1 x double> @bitcast_nxv4bf16_to_nxv1f64(<vscale x 4 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4bf16_to_nxv1f64: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x bfloat> %v to <vscale x 1 x double> |
| ret <vscale x 1 x double> %bc |
| } |
| |
| ; |
| ; bitcast to nxv4bf16 |
| ; |
| |
| define <vscale x 4 x bfloat> @bitcast_nxv8i8_to_nxv4bf16(<vscale x 8 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv8i8_to_nxv4bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.h |
| ; CHECK-NEXT: st1b { z0.h }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 8 x i8> %v to <vscale x 4 x bfloat> |
| ret <vscale x 4 x bfloat> %bc |
| } |
| |
| define <vscale x 4 x bfloat> @bitcast_nxv4i16_to_nxv4bf16(<vscale x 4 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i16_to_nxv4bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i16> %v to <vscale x 4 x bfloat> |
| ret <vscale x 4 x bfloat> %bc |
| } |
| |
| define <vscale x 4 x bfloat> @bitcast_nxv2i32_to_nxv4bf16(<vscale x 2 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i32_to_nxv4bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i32> %v to <vscale x 4 x bfloat> |
| ret <vscale x 4 x bfloat> %bc |
| } |
| |
| define <vscale x 4 x bfloat> @bitcast_nxv1i64_to_nxv4bf16(<vscale x 1 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i64_to_nxv4bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i64> %v to <vscale x 4 x bfloat> |
| ret <vscale x 4 x bfloat> %bc |
| } |
| |
| define <vscale x 4 x bfloat> @bitcast_nxv4f16_to_nxv4bf16(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4f16_to_nxv4bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x half> %v to <vscale x 4 x bfloat> |
| ret <vscale x 4 x bfloat> %bc |
| } |
| |
| define <vscale x 4 x bfloat> @bitcast_nxv2f32_to_nxv4bf16(<vscale x 2 x float> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f32_to_nxv4bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1w { z0.d }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1h { z0.s }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x float> %v to <vscale x 4 x bfloat> |
| ret <vscale x 4 x bfloat> %bc |
| } |
| |
| define <vscale x 4 x bfloat> @bitcast_nxv1f64_to_nxv4bf16(<vscale x 1 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1f64_to_nxv4bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x double> %v to <vscale x 4 x bfloat> |
| ret <vscale x 4 x bfloat> %bc |
| } |
| |
| ; |
| ; bitcast to nxv4i8 |
| ; |
| |
| define <vscale x 4 x i8> @bitcast_nxv2i16_to_nxv4i8(<vscale x 2 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i16_to_nxv4i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1h { z0.d }, p0, [sp, #3, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1b { z0.s }, p0/z, [sp, #3, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i16> %v to <vscale x 4 x i8> |
| ret <vscale x 4 x i8> %bc |
| } |
| |
| define <vscale x 4 x i8> @bitcast_nxv1i32_to_nxv4i8(<vscale x 1 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i32_to_nxv4i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.h, z0.b |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i32> %v to <vscale x 4 x i8> |
| ret <vscale x 4 x i8> %bc |
| } |
| |
| define <vscale x 4 x i8> @bitcast_nxv2f16_to_nxv4i8(<vscale x 2 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f16_to_nxv4i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1h { z0.d }, p0, [sp, #3, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1b { z0.s }, p0/z, [sp, #3, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x half> %v to <vscale x 4 x i8> |
| ret <vscale x 4 x i8> %bc |
| } |
| |
| ; @bitcast_nxv1f32_to_nxv4i8 is missing |
| |
| define <vscale x 4 x i8> @bitcast_nxv2bf16_to_nxv4i8(<vscale x 2 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2bf16_to_nxv4i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1h { z0.d }, p0, [sp, #3, mul vl] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1b { z0.s }, p0/z, [sp, #3, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x bfloat> %v to <vscale x 4 x i8> |
| ret <vscale x 4 x i8> %bc |
| } |
| |
| ; |
| ; bitcast to nxv2i16 |
| ; |
| |
| define <vscale x 2 x i16> @bitcast_nxv4i8_to_nxv2i16(<vscale x 4 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i8_to_nxv2i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1b { z0.s }, p0, [sp, #3, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1h { z0.d }, p0/z, [sp, #3, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i8> %v to <vscale x 2 x i16> |
| ret <vscale x 2 x i16> %bc |
| } |
| |
| define <vscale x 2 x i16> @bitcast_nxv1i32_to_nxv2i16(<vscale x 1 x i32> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i32_to_nxv2i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: uunpklo z0.d, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i32> %v to <vscale x 2 x i16> |
| ret <vscale x 2 x i16> %bc |
| } |
| |
| define <vscale x 2 x i16> @bitcast_nxv2f16_to_nxv2i16(<vscale x 2 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f16_to_nxv2i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x half> %v to <vscale x 2 x i16> |
| ret <vscale x 2 x i16> %bc |
| } |
| |
| ; @bitcast_nxv1f32_to_nxv2i16 is missing |
| |
| define <vscale x 2 x i16> @bitcast_nxv2bf16_to_nxv2i16(<vscale x 2 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2bf16_to_nxv2i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x bfloat> %v to <vscale x 2 x i16> |
| ret <vscale x 2 x i16> %bc |
| } |
| |
| ; |
| ; bitcast to nxv1i32 |
| ; |
| |
| define <vscale x 1 x i32> @bitcast_nxv4i8_to_nxv1i32(<vscale x 4 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i8_to_nxv1i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: uzp1 z0.b, z0.b, z0.b |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i8> %v to <vscale x 1 x i32> |
| ret <vscale x 1 x i32> %bc |
| } |
| |
| define <vscale x 1 x i32> @bitcast_nxv2i16_to_nxv1i32(<vscale x 2 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i16_to_nxv1i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.s, z0.s, z0.s |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i16> %v to <vscale x 1 x i32> |
| ret <vscale x 1 x i32> %bc |
| } |
| |
| define <vscale x 1 x i32> @bitcast_nxv2f16_to_nxv1i32(<vscale x 2 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f16_to_nxv1i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1h { z0.d }, p0, [sp] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1w { z0.s }, p0/z, [sp] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x half> %v to <vscale x 1 x i32> |
| ret <vscale x 1 x i32> %bc |
| } |
| |
| ; @bitcast_nxv1f32_to_nxv1i32 is missing |
| |
| define <vscale x 1 x i32> @bitcast_nxv2bf16_to_nxv1i32(<vscale x 2 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2bf16_to_nxv1i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: st1h { z0.d }, p0, [sp] |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: ld1w { z0.s }, p0/z, [sp] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x bfloat> %v to <vscale x 1 x i32> |
| ret <vscale x 1 x i32> %bc |
| } |
| |
| ; |
| ; bitcast to nxv2f16 |
| ; |
| |
| define <vscale x 2 x half> @bitcast_nxv4i8_to_nxv2f16(<vscale x 4 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i8_to_nxv2f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1b { z0.s }, p0, [sp, #3, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1h { z0.d }, p0/z, [sp, #3, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i8> %v to <vscale x 2 x half> |
| ret <vscale x 2 x half> %bc |
| } |
| |
| define <vscale x 2 x half> @bitcast_nxv2i16_to_nxv2f16(<vscale x 2 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i16_to_nxv2f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i16> %v to <vscale x 2 x half> |
| ret <vscale x 2 x half> %bc |
| } |
| |
| ; @bitcast_nxv1i32_to_nxv2f16 is missing |
| ; @bitcast_nxv1f32_to_nxv2f16 is missing |
| |
| define <vscale x 2 x half> @bitcast_nxv2bf16_to_nxv2f16(<vscale x 2 x bfloat> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2bf16_to_nxv2f16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x bfloat> %v to <vscale x 2 x half> |
| ret <vscale x 2 x half> %bc |
| } |
| |
| ; |
| ; bitcast to nxv1f32 |
| ; |
| |
| ; @bitcast_nxv4i8_to_nxv1f32 is missing |
| ; @bitcast_nxv2i16_to_nxv1f32 is missing |
| ; @bitcast_nxv1i32_to_nxv1f32 is missing |
| ; @bitcast_nxv2f16_to_nxv1f32 is missing |
| ; @bitcast_nxv2bf16_to_nxv1f32 is missing |
| |
| ; |
| ; bitcast to nxv2bf16 |
| ; |
| |
| define <vscale x 2 x bfloat> @bitcast_nxv4i8_to_nxv2bf16(<vscale x 4 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv4i8_to_nxv2bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: st1b { z0.s }, p0, [sp, #3, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1h { z0.d }, p0/z, [sp, #3, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 4 x i8> %v to <vscale x 2 x bfloat> |
| ret <vscale x 2 x bfloat> %bc |
| } |
| |
| define <vscale x 2 x bfloat> @bitcast_nxv2i16_to_nxv2bf16(<vscale x 2 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i16_to_nxv2bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i16> %v to <vscale x 2 x bfloat> |
| ret <vscale x 2 x bfloat> %bc |
| } |
| |
| ; @bitcast_nxv1i32_to_nxv2bf16 is missing |
| |
| define <vscale x 2 x bfloat> @bitcast_nxv2f16_to_nxv2bf16(<vscale x 2 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2f16_to_nxv2bf16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x half> %v to <vscale x 2 x bfloat> |
| ret <vscale x 2 x bfloat> %bc |
| } |
| |
| ; @bitcast_nxv1f32_to_nxv2bf16 is missing |
| |
| ; |
| ; bitcast to nxv2i8 |
| ; |
| |
| define <vscale x 2 x i8> @bitcast_nxv1i16_to_nxv2i8(<vscale x 1 x i16> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv1i16_to_nxv2i8: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uunpklo z0.h, z0.b |
| ; CHECK-NEXT: uunpklo z0.s, z0.h |
| ; CHECK-NEXT: uunpklo z0.d, z0.s |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 1 x i16> %v to <vscale x 2 x i8> |
| ret <vscale x 2 x i8> %bc |
| } |
| |
| ; @bitcast_nxv1f16_to_nxv2i8 is missing |
| ; @bitcast_nxv1bf16_to_nxv2i8 is missing |
| |
| ; |
| ; bitcast to nxv1i16 |
| ; |
| |
| define <vscale x 1 x i16> @bitcast_nxv2i8_to_nxv1i16(<vscale x 2 x i8> %v) #0 { |
| ; CHECK-LABEL: bitcast_nxv2i8_to_nxv1i16: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: uzp1 z0.s, z0.s, z0.s |
| ; CHECK-NEXT: uzp1 z0.h, z0.h, z0.h |
| ; CHECK-NEXT: uzp1 z0.b, z0.b, z0.b |
| ; CHECK-NEXT: ret |
| %bc = bitcast <vscale x 2 x i8> %v to <vscale x 1 x i16> |
| ret <vscale x 1 x i16> %bc |
| } |
| |
| ; @bitcast_nxv1f16_to_nxv1i16 is missing |
| ; @bitcast_nxv1bf16_to_nxv1i16 is missing |
| |
| ; |
| ; bitcast to nxv1f16 |
| ; |
| |
| ; @bitcast_nxv2i8_to_nxv1f16 is missing |
| ; @bitcast_nxv1i16_to_nxv1f16 is missing |
| ; @bitcast_nxv1bf16_to_nxv1f16 is missing |
| |
| ; |
| ; bitcast to nxv1bf16 |
| ; |
| |
| ; @bitcast_nxv2i8_to_nxv1bf16 is missing |
| ; @bitcast_nxv1i16_to_nxv1bf16 is missing |
| ; @bitcast_nxv1f16_to_nxv1bf16 is missing |
| |
| ; |
| ; Other |
| ; |
| |
| define <vscale x 2 x i32> @bitcast_short_float_to_i32(<vscale x 2 x double> %v) #0 { |
| ; CHECK-LABEL: bitcast_short_float_to_i32: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: fcvt z0.s, p0/m, z0.d |
| ; CHECK-NEXT: ret |
| %trunc = fptrunc <vscale x 2 x double> %v to <vscale x 2 x float> |
| %bitcast = bitcast <vscale x 2 x float> %trunc to <vscale x 2 x i32> |
| ret <vscale x 2 x i32> %bitcast |
| } |
| |
| define <vscale x 2 x double> @bitcast_short_i32_to_float(<vscale x 2 x i64> %v) #0 { |
| ; CHECK-LABEL: bitcast_short_i32_to_float: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: fcvt z0.d, p0/m, z0.s |
| ; CHECK-NEXT: ret |
| %trunc = trunc <vscale x 2 x i64> %v to <vscale x 2 x i32> |
| %bitcast = bitcast <vscale x 2 x i32> %trunc to <vscale x 2 x float> |
| %extended = fpext <vscale x 2 x float> %bitcast to <vscale x 2 x double> |
| ret <vscale x 2 x double> %extended |
| } |
| |
| define <vscale x 2 x float> @bitcast_short_half_to_float(<vscale x 4 x half> %v) #0 { |
| ; CHECK-LABEL: bitcast_short_half_to_float: |
| ; CHECK: // %bb.0: |
| ; CHECK-NEXT: str x29, [sp, #-16]! // 8-byte Folded Spill |
| ; CHECK-NEXT: addvl sp, sp, #-1 |
| ; CHECK-NEXT: ptrue p0.s |
| ; CHECK-NEXT: fadd z0.h, p0/m, z0.h, z0.h |
| ; CHECK-NEXT: st1h { z0.s }, p0, [sp, #1, mul vl] |
| ; CHECK-NEXT: ptrue p0.d |
| ; CHECK-NEXT: ld1w { z0.d }, p0/z, [sp, #1, mul vl] |
| ; CHECK-NEXT: addvl sp, sp, #1 |
| ; CHECK-NEXT: ldr x29, [sp], #16 // 8-byte Folded Reload |
| ; CHECK-NEXT: ret |
| %add = fadd <vscale x 4 x half> %v, %v |
| %bitcast = bitcast <vscale x 4 x half> %add to <vscale x 2 x float> |
| ret <vscale x 2 x float> %bitcast |
| } |
| |
| ; +bf16 is required for the bfloat version. |
| attributes #0 = { nounwind "target-features"="+sve,+bf16" } |