| ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py |
| ; RUN: llc < %s -mtriple=i686-- -mattr=-bmi | FileCheck %s --check-prefixes=X86,X86-NOBMI |
| ; RUN: llc < %s -mtriple=i686-- -mattr=+bmi | FileCheck %s --check-prefixes=X86,X86-BMI |
| ; RUN: llc < %s -mtriple=x86_64-- -mattr=-bmi | FileCheck %s --check-prefixes=X64,X64-NOBMI |
| ; RUN: llc < %s -mtriple=x86_64-- -mattr=+bmi | FileCheck %s --check-prefixes=X64,X64-BMI |
| |
| ; TODO - PR112425 - attempt to reconstruct andnot patterns through bitwise-agnostic operations |
| |
| declare void @use_i64(i64) |
| declare void @use_i32(i32) |
| |
| ; |
| ; Fold (and X, (rotl (not Y), Z))) -> (and X, (not (rotl Y, Z))) |
| ; |
| |
| define i64 @andnot_rotl_i64(i64 %a0, i64 %a1, i64 %a2) nounwind { |
| ; X86-NOBMI-LABEL: andnot_rotl_i64: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: pushl %esi |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %esi |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-NOBMI-NEXT: testb $32, %cl |
| ; X86-NOBMI-NEXT: jne .LBB0_1 |
| ; X86-NOBMI-NEXT: # %bb.2: |
| ; X86-NOBMI-NEXT: movl %eax, %edx |
| ; X86-NOBMI-NEXT: jmp .LBB0_3 |
| ; X86-NOBMI-NEXT: .LBB0_1: |
| ; X86-NOBMI-NEXT: movl %esi, %edx |
| ; X86-NOBMI-NEXT: movl %eax, %esi |
| ; X86-NOBMI-NEXT: .LBB0_3: |
| ; X86-NOBMI-NEXT: movl %esi, %eax |
| ; X86-NOBMI-NEXT: shldl %cl, %edx, %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X86-NOBMI-NEXT: shldl %cl, %esi, %edx |
| ; X86-NOBMI-NEXT: notl %edx |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx |
| ; X86-NOBMI-NEXT: popl %esi |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_rotl_i64: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: pushl %esi |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-BMI-NEXT: testb $32, %cl |
| ; X86-BMI-NEXT: jne .LBB0_1 |
| ; X86-BMI-NEXT: # %bb.2: |
| ; X86-BMI-NEXT: movl %eax, %esi |
| ; X86-BMI-NEXT: jmp .LBB0_3 |
| ; X86-BMI-NEXT: .LBB0_1: |
| ; X86-BMI-NEXT: movl %edx, %esi |
| ; X86-BMI-NEXT: movl %eax, %edx |
| ; X86-BMI-NEXT: .LBB0_3: |
| ; X86-BMI-NEXT: movl %edx, %eax |
| ; X86-BMI-NEXT: shldl %cl, %esi, %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X86-BMI-NEXT: shldl %cl, %edx, %esi |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %esi, %edx |
| ; X86-BMI-NEXT: popl %esi |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_rotl_i64: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movq %rdx, %rcx |
| ; X64-NOBMI-NEXT: movq %rsi, %rax |
| ; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $rcx |
| ; X64-NOBMI-NEXT: rolq %cl, %rax |
| ; X64-NOBMI-NEXT: notq %rax |
| ; X64-NOBMI-NEXT: andq %rdi, %rax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_rotl_i64: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: movq %rdx, %rcx |
| ; X64-BMI-NEXT: # kill: def $cl killed $cl killed $rcx |
| ; X64-BMI-NEXT: rolq %cl, %rsi |
| ; X64-BMI-NEXT: andnq %rdi, %rsi, %rax |
| ; X64-BMI-NEXT: retq |
| %not = xor i64 %a1, -1 |
| %rot = tail call i64 @llvm.fshl.i64(i64 %not, i64 %not, i64 %a2) |
| %and = and i64 %rot, %a0 |
| ret i64 %and |
| } |
| |
| define i32 @andnot_rotl_i32(i32 %a0, i32 %a1, i32 %a2) nounwind { |
| ; X86-NOBMI-LABEL: andnot_rotl_i32: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: roll %cl, %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_rotl_i32: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: roll %cl, %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_rotl_i32: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movl %edx, %ecx |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-NOBMI-NEXT: roll %cl, %eax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_rotl_i32: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: movl %edx, %ecx |
| ; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-BMI-NEXT: roll %cl, %esi |
| ; X64-BMI-NEXT: andnl %edi, %esi, %eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i32 %a1, -1 |
| %rot = tail call i32 @llvm.fshl.i32(i32 %not, i32 %not, i32 %a2) |
| %and = and i32 %rot, %a0 |
| ret i32 %and |
| } |
| |
| define i16 @andnot_rotl_i16(i16 %a0, i16 %a1, i16 %a2) nounwind { |
| ; X86-LABEL: andnot_rotl_i16: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: rolw %cl, %ax |
| ; X86-NEXT: notl %eax |
| ; X86-NEXT: andw {{[0-9]+}}(%esp), %ax |
| ; X86-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X86-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_rotl_i16: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movl %edx, %ecx |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-NOBMI-NEXT: rolw %cl, %ax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_rotl_i16: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: movl %edx, %ecx |
| ; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-BMI-NEXT: rolw %cl, %si |
| ; X64-BMI-NEXT: andnl %edi, %esi, %eax |
| ; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i16 %a1, -1 |
| %rot = tail call i16 @llvm.fshl.i16(i16 %not, i16 %not, i16 %a2) |
| %and = and i16 %rot, %a0 |
| ret i16 %and |
| } |
| |
| define i8 @andnot_rotl_i8(i8 %a0, i8 %a1, i8 %a2) nounwind { |
| ; X86-LABEL: andnot_rotl_i8: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: rolb %cl, %al |
| ; X86-NEXT: notb %al |
| ; X86-NEXT: andb {{[0-9]+}}(%esp), %al |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: andnot_rotl_i8: |
| ; X64: # %bb.0: |
| ; X64-NEXT: movl %edx, %ecx |
| ; X64-NEXT: movl %esi, %eax |
| ; X64-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-NEXT: rolb %cl, %al |
| ; X64-NEXT: notb %al |
| ; X64-NEXT: andb %dil, %al |
| ; X64-NEXT: # kill: def $al killed $al killed $eax |
| ; X64-NEXT: retq |
| %not = xor i8 %a1, -1 |
| %rot = tail call i8 @llvm.fshl.i8(i8 %not, i8 %not, i8 %a2) |
| %and = and i8 %rot, %a0 |
| ret i8 %and |
| } |
| |
| define i64 @andnot_rotl_i64_multiuse_rot(i64 %a0, i64 %a1, i64 %a2) nounwind { |
| ; X86-LABEL: andnot_rotl_i64_multiuse_rot: |
| ; X86: # %bb.0: |
| ; X86-NEXT: pushl %ebx |
| ; X86-NEXT: pushl %edi |
| ; X86-NEXT: pushl %esi |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %esi |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-NEXT: notl %edx |
| ; X86-NEXT: notl %esi |
| ; X86-NEXT: testb $32, %cl |
| ; X86-NEXT: jne .LBB4_1 |
| ; X86-NEXT: # %bb.2: |
| ; X86-NEXT: movl %esi, %eax |
| ; X86-NEXT: jmp .LBB4_3 |
| ; X86-NEXT: .LBB4_1: |
| ; X86-NEXT: movl %edx, %eax |
| ; X86-NEXT: movl %esi, %edx |
| ; X86-NEXT: .LBB4_3: |
| ; X86-NEXT: movl %edx, %ebx |
| ; X86-NEXT: shldl %cl, %eax, %ebx |
| ; X86-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X86-NEXT: shldl %cl, %edx, %eax |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %esi |
| ; X86-NEXT: andl %eax, %esi |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %edi |
| ; X86-NEXT: andl %ebx, %edi |
| ; X86-NEXT: pushl %ebx |
| ; X86-NEXT: pushl %eax |
| ; X86-NEXT: calll use_i64@PLT |
| ; X86-NEXT: addl $8, %esp |
| ; X86-NEXT: movl %esi, %eax |
| ; X86-NEXT: movl %edi, %edx |
| ; X86-NEXT: popl %esi |
| ; X86-NEXT: popl %edi |
| ; X86-NEXT: popl %ebx |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: andnot_rotl_i64_multiuse_rot: |
| ; X64: # %bb.0: |
| ; X64-NEXT: pushq %rbx |
| ; X64-NEXT: movq %rdx, %rcx |
| ; X64-NEXT: movq %rdi, %rbx |
| ; X64-NEXT: notq %rsi |
| ; X64-NEXT: # kill: def $cl killed $cl killed $rcx |
| ; X64-NEXT: rolq %cl, %rsi |
| ; X64-NEXT: andq %rsi, %rbx |
| ; X64-NEXT: movq %rsi, %rdi |
| ; X64-NEXT: callq use_i64@PLT |
| ; X64-NEXT: movq %rbx, %rax |
| ; X64-NEXT: popq %rbx |
| ; X64-NEXT: retq |
| %not = xor i64 %a1, -1 |
| %rot = tail call i64 @llvm.fshl.i64(i64 %not, i64 %not, i64 %a2) |
| %and = and i64 %rot, %a0 |
| call void @use_i64(i64 %rot) |
| ret i64 %and |
| } |
| |
| ; |
| ; Fold (and X, (rotr (not Y), Z))) -> (and X, (not (rotr Y, Z))) |
| ; |
| |
| define i64 @andnot_rotr_i64(i64 %a0, i64 %a1, i64 %a2) nounwind { |
| ; X86-NOBMI-LABEL: andnot_rotr_i64: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: pushl %esi |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %esi |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-NOBMI-NEXT: testb $32, %cl |
| ; X86-NOBMI-NEXT: je .LBB5_1 |
| ; X86-NOBMI-NEXT: # %bb.2: |
| ; X86-NOBMI-NEXT: movl %eax, %edx |
| ; X86-NOBMI-NEXT: jmp .LBB5_3 |
| ; X86-NOBMI-NEXT: .LBB5_1: |
| ; X86-NOBMI-NEXT: movl %esi, %edx |
| ; X86-NOBMI-NEXT: movl %eax, %esi |
| ; X86-NOBMI-NEXT: .LBB5_3: |
| ; X86-NOBMI-NEXT: movl %esi, %eax |
| ; X86-NOBMI-NEXT: shrdl %cl, %edx, %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X86-NOBMI-NEXT: shrdl %cl, %esi, %edx |
| ; X86-NOBMI-NEXT: notl %edx |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx |
| ; X86-NOBMI-NEXT: popl %esi |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_rotr_i64: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: pushl %esi |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-BMI-NEXT: testb $32, %cl |
| ; X86-BMI-NEXT: je .LBB5_1 |
| ; X86-BMI-NEXT: # %bb.2: |
| ; X86-BMI-NEXT: movl %eax, %esi |
| ; X86-BMI-NEXT: jmp .LBB5_3 |
| ; X86-BMI-NEXT: .LBB5_1: |
| ; X86-BMI-NEXT: movl %edx, %esi |
| ; X86-BMI-NEXT: movl %eax, %edx |
| ; X86-BMI-NEXT: .LBB5_3: |
| ; X86-BMI-NEXT: movl %edx, %eax |
| ; X86-BMI-NEXT: shrdl %cl, %esi, %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X86-BMI-NEXT: shrdl %cl, %edx, %esi |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %esi, %edx |
| ; X86-BMI-NEXT: popl %esi |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_rotr_i64: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movq %rdx, %rcx |
| ; X64-NOBMI-NEXT: movq %rsi, %rax |
| ; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $rcx |
| ; X64-NOBMI-NEXT: rorq %cl, %rax |
| ; X64-NOBMI-NEXT: notq %rax |
| ; X64-NOBMI-NEXT: andq %rdi, %rax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_rotr_i64: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: movq %rdx, %rcx |
| ; X64-BMI-NEXT: # kill: def $cl killed $cl killed $rcx |
| ; X64-BMI-NEXT: rorq %cl, %rsi |
| ; X64-BMI-NEXT: andnq %rdi, %rsi, %rax |
| ; X64-BMI-NEXT: retq |
| %not = xor i64 %a1, -1 |
| %rot = tail call i64 @llvm.fshr.i64(i64 %not, i64 %not, i64 %a2) |
| %and = and i64 %rot, %a0 |
| ret i64 %and |
| } |
| |
| define i32 @andnot_rotr_i32(i32 %a0, i32 %a1, i32 %a2) nounwind { |
| ; X86-NOBMI-LABEL: andnot_rotr_i32: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: rorl %cl, %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_rotr_i32: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: rorl %cl, %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_rotr_i32: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movl %edx, %ecx |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-NOBMI-NEXT: rorl %cl, %eax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_rotr_i32: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: movl %edx, %ecx |
| ; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-BMI-NEXT: rorl %cl, %esi |
| ; X64-BMI-NEXT: andnl %edi, %esi, %eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i32 %a1, -1 |
| %rot = tail call i32 @llvm.fshr.i32(i32 %not, i32 %not, i32 %a2) |
| %and = and i32 %rot, %a0 |
| ret i32 %and |
| } |
| |
| define i16 @andnot_rotr_i16(i16 %a0, i16 %a1, i16 %a2) nounwind { |
| ; X86-LABEL: andnot_rotr_i16: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: rorw %cl, %ax |
| ; X86-NEXT: notl %eax |
| ; X86-NEXT: andw {{[0-9]+}}(%esp), %ax |
| ; X86-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X86-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_rotr_i16: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movl %edx, %ecx |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-NOBMI-NEXT: rorw %cl, %ax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_rotr_i16: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: movl %edx, %ecx |
| ; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-BMI-NEXT: rorw %cl, %si |
| ; X64-BMI-NEXT: andnl %edi, %esi, %eax |
| ; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i16 %a1, -1 |
| %rot = tail call i16 @llvm.fshr.i16(i16 %not, i16 %not, i16 %a2) |
| %and = and i16 %rot, %a0 |
| ret i16 %and |
| } |
| |
| define i8 @andnot_rotr_i8(i8 %a0, i8 %a1, i8 %a2) nounwind { |
| ; X86-LABEL: andnot_rotr_i8: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: rorb %cl, %al |
| ; X86-NEXT: notb %al |
| ; X86-NEXT: andb {{[0-9]+}}(%esp), %al |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: andnot_rotr_i8: |
| ; X64: # %bb.0: |
| ; X64-NEXT: movl %edx, %ecx |
| ; X64-NEXT: movl %esi, %eax |
| ; X64-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-NEXT: rorb %cl, %al |
| ; X64-NEXT: notb %al |
| ; X64-NEXT: andb %dil, %al |
| ; X64-NEXT: # kill: def $al killed $al killed $eax |
| ; X64-NEXT: retq |
| %not = xor i8 %a1, -1 |
| %rot = tail call i8 @llvm.fshr.i8(i8 %not, i8 %not, i8 %a2) |
| %and = and i8 %rot, %a0 |
| ret i8 %and |
| } |
| |
| define i32 @andnot_rotr_i32_multiuse_not(i32 %a0, i32 %a1, i32 %a2) nounwind { |
| ; X86-NOBMI-LABEL: andnot_rotr_i32_multiuse_not: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: pushl %esi |
| ; X86-NOBMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: movl %eax, %esi |
| ; X86-NOBMI-NEXT: rorl %cl, %esi |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %esi |
| ; X86-NOBMI-NEXT: pushl %eax |
| ; X86-NOBMI-NEXT: calll use_i32@PLT |
| ; X86-NOBMI-NEXT: addl $4, %esp |
| ; X86-NOBMI-NEXT: movl %esi, %eax |
| ; X86-NOBMI-NEXT: popl %esi |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_rotr_i32_multiuse_not: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: pushl %esi |
| ; X86-BMI-NEXT: movzbl {{[0-9]+}}(%esp), %ecx |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: movl %eax, %edx |
| ; X86-BMI-NEXT: notl %edx |
| ; X86-BMI-NEXT: rorl %cl, %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %esi |
| ; X86-BMI-NEXT: pushl %edx |
| ; X86-BMI-NEXT: calll use_i32@PLT |
| ; X86-BMI-NEXT: addl $4, %esp |
| ; X86-BMI-NEXT: movl %esi, %eax |
| ; X86-BMI-NEXT: popl %esi |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_rotr_i32_multiuse_not: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: pushq %rbx |
| ; X64-NOBMI-NEXT: movl %edx, %ecx |
| ; X64-NOBMI-NEXT: notl %esi |
| ; X64-NOBMI-NEXT: movl %esi, %ebx |
| ; X64-NOBMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-NOBMI-NEXT: rorl %cl, %ebx |
| ; X64-NOBMI-NEXT: andl %edi, %ebx |
| ; X64-NOBMI-NEXT: movl %esi, %edi |
| ; X64-NOBMI-NEXT: callq use_i32@PLT |
| ; X64-NOBMI-NEXT: movl %ebx, %eax |
| ; X64-NOBMI-NEXT: popq %rbx |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_rotr_i32_multiuse_not: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: pushq %rbx |
| ; X64-BMI-NEXT: movl %edx, %ecx |
| ; X64-BMI-NEXT: movl %esi, %eax |
| ; X64-BMI-NEXT: notl %eax |
| ; X64-BMI-NEXT: # kill: def $cl killed $cl killed $ecx |
| ; X64-BMI-NEXT: rorl %cl, %esi |
| ; X64-BMI-NEXT: andnl %edi, %esi, %ebx |
| ; X64-BMI-NEXT: movl %eax, %edi |
| ; X64-BMI-NEXT: callq use_i32@PLT |
| ; X64-BMI-NEXT: movl %ebx, %eax |
| ; X64-BMI-NEXT: popq %rbx |
| ; X64-BMI-NEXT: retq |
| %not = xor i32 %a1, -1 |
| %rot = tail call i32 @llvm.fshr.i32(i32 %not, i32 %not, i32 %a2) |
| %and = and i32 %rot, %a0 |
| call void @use_i32(i32 %not) |
| ret i32 %and |
| } |
| |
| ; |
| ; Fold (and X, (bswap (not Y)))) -> (and X, (not (bswap Y))) |
| ; |
| |
| define i64 @andnot_bswap_i64(i64 %a0, i64 %a1) nounwind { |
| ; X86-NOBMI-LABEL: andnot_bswap_i64: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %edx |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: bswapl %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: bswapl %edx |
| ; X86-NOBMI-NEXT: notl %edx |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_bswap_i64: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: bswapl %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: bswapl %ecx |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %ecx, %edx |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_bswap_i64: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movq %rsi, %rax |
| ; X64-NOBMI-NEXT: bswapq %rax |
| ; X64-NOBMI-NEXT: notq %rax |
| ; X64-NOBMI-NEXT: andq %rdi, %rax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_bswap_i64: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: bswapq %rsi |
| ; X64-BMI-NEXT: andnq %rdi, %rsi, %rax |
| ; X64-BMI-NEXT: retq |
| %not = xor i64 %a1, -1 |
| %bswap = tail call i64 @llvm.bswap.i64(i64 %not) |
| %and = and i64 %bswap, %a0 |
| ret i64 %and |
| } |
| |
| define i32 @andnot_bswap_i32(i32 %a0, i32 %a1) nounwind { |
| ; X86-NOBMI-LABEL: andnot_bswap_i32: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: bswapl %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_bswap_i32: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: bswapl %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_bswap_i32: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: bswapl %eax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_bswap_i32: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: bswapl %esi |
| ; X64-BMI-NEXT: andnl %edi, %esi, %eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i32 %a1, -1 |
| %bswap = tail call i32 @llvm.bswap.i32(i32 %not) |
| %and = and i32 %bswap, %a0 |
| ret i32 %and |
| } |
| |
| define i16 @andnot_bswap_i16(i16 %a0, i16 %a1) nounwind { |
| ; X86-LABEL: andnot_bswap_i16: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: rolw $8, %ax |
| ; X86-NEXT: notl %eax |
| ; X86-NEXT: andw {{[0-9]+}}(%esp), %ax |
| ; X86-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X86-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_bswap_i16: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: rolw $8, %ax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_bswap_i16: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: rolw $8, %si |
| ; X64-BMI-NEXT: andnl %edi, %esi, %eax |
| ; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i16 %a1, -1 |
| %bswap = tail call i16 @llvm.bswap.i16(i16 %not) |
| %and = and i16 %bswap, %a0 |
| ret i16 %and |
| } |
| |
| define i32 @andnot_bswap_i32_multiuse_bswap(i32 %a0, i32 %a1) nounwind { |
| ; X86-LABEL: andnot_bswap_i32_multiuse_bswap: |
| ; X86: # %bb.0: |
| ; X86-NEXT: pushl %esi |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: notl %eax |
| ; X86-NEXT: bswapl %eax |
| ; X86-NEXT: movl {{[0-9]+}}(%esp), %esi |
| ; X86-NEXT: andl %eax, %esi |
| ; X86-NEXT: pushl %eax |
| ; X86-NEXT: calll use_i32@PLT |
| ; X86-NEXT: addl $4, %esp |
| ; X86-NEXT: movl %esi, %eax |
| ; X86-NEXT: popl %esi |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: andnot_bswap_i32_multiuse_bswap: |
| ; X64: # %bb.0: |
| ; X64-NEXT: pushq %rbx |
| ; X64-NEXT: movl %edi, %ebx |
| ; X64-NEXT: notl %esi |
| ; X64-NEXT: bswapl %esi |
| ; X64-NEXT: andl %esi, %ebx |
| ; X64-NEXT: movl %esi, %edi |
| ; X64-NEXT: callq use_i32@PLT |
| ; X64-NEXT: movl %ebx, %eax |
| ; X64-NEXT: popq %rbx |
| ; X64-NEXT: retq |
| %not = xor i32 %a1, -1 |
| %bswap = tail call i32 @llvm.bswap.i32(i32 %not) |
| %and = and i32 %bswap, %a0 |
| call void @use_i32(i32 %bswap) |
| ret i32 %and |
| } |
| |
| define i32 @andnot_bswap_i32_multiuse_not(i32 %a0, i32 %a1) nounwind { |
| ; X86-NOBMI-LABEL: andnot_bswap_i32_multiuse_not: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: pushl %esi |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: movl %eax, %esi |
| ; X86-NOBMI-NEXT: bswapl %esi |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %esi |
| ; X86-NOBMI-NEXT: pushl %eax |
| ; X86-NOBMI-NEXT: calll use_i32@PLT |
| ; X86-NOBMI-NEXT: addl $4, %esp |
| ; X86-NOBMI-NEXT: movl %esi, %eax |
| ; X86-NOBMI-NEXT: popl %esi |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_bswap_i32_multiuse_not: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: pushl %esi |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: movl %eax, %ecx |
| ; X86-BMI-NEXT: notl %ecx |
| ; X86-BMI-NEXT: bswapl %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %esi |
| ; X86-BMI-NEXT: pushl %ecx |
| ; X86-BMI-NEXT: calll use_i32@PLT |
| ; X86-BMI-NEXT: addl $4, %esp |
| ; X86-BMI-NEXT: movl %esi, %eax |
| ; X86-BMI-NEXT: popl %esi |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_bswap_i32_multiuse_not: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: pushq %rbx |
| ; X64-NOBMI-NEXT: notl %esi |
| ; X64-NOBMI-NEXT: movl %esi, %ebx |
| ; X64-NOBMI-NEXT: bswapl %ebx |
| ; X64-NOBMI-NEXT: andl %edi, %ebx |
| ; X64-NOBMI-NEXT: movl %esi, %edi |
| ; X64-NOBMI-NEXT: callq use_i32@PLT |
| ; X64-NOBMI-NEXT: movl %ebx, %eax |
| ; X64-NOBMI-NEXT: popq %rbx |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_bswap_i32_multiuse_not: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: pushq %rbx |
| ; X64-BMI-NEXT: movl %esi, %eax |
| ; X64-BMI-NEXT: notl %eax |
| ; X64-BMI-NEXT: bswapl %esi |
| ; X64-BMI-NEXT: andnl %edi, %esi, %ebx |
| ; X64-BMI-NEXT: movl %eax, %edi |
| ; X64-BMI-NEXT: callq use_i32@PLT |
| ; X64-BMI-NEXT: movl %ebx, %eax |
| ; X64-BMI-NEXT: popq %rbx |
| ; X64-BMI-NEXT: retq |
| %not = xor i32 %a1, -1 |
| %bswap = tail call i32 @llvm.bswap.i32(i32 %not) |
| %and = and i32 %bswap, %a0 |
| call void @use_i32(i32 %not) |
| ret i32 %and |
| } |
| |
| ; |
| ; Fold (and X, (bitreverse (not Y)))) -> (and X, (not (bitreverse Y))) |
| ; |
| |
| define i64 @andnot_bitreverse_i64(i64 %a0, i64 %a1) nounwind { |
| ; X86-NOBMI-LABEL: andnot_bitreverse_i64: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: bswapl %eax |
| ; X86-NOBMI-NEXT: movl %eax, %edx |
| ; X86-NOBMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F |
| ; X86-NOBMI-NEXT: shll $4, %edx |
| ; X86-NOBMI-NEXT: shrl $4, %eax |
| ; X86-NOBMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F |
| ; X86-NOBMI-NEXT: orl %edx, %eax |
| ; X86-NOBMI-NEXT: movl %eax, %edx |
| ; X86-NOBMI-NEXT: andl $858993459, %edx # imm = 0x33333333 |
| ; X86-NOBMI-NEXT: shrl $2, %eax |
| ; X86-NOBMI-NEXT: andl $858993459, %eax # imm = 0x33333333 |
| ; X86-NOBMI-NEXT: leal (%eax,%edx,4), %eax |
| ; X86-NOBMI-NEXT: movl %eax, %edx |
| ; X86-NOBMI-NEXT: andl $1431655765, %edx # imm = 0x55555555 |
| ; X86-NOBMI-NEXT: shrl %eax |
| ; X86-NOBMI-NEXT: andl $1431655765, %eax # imm = 0x55555555 |
| ; X86-NOBMI-NEXT: leal (%eax,%edx,2), %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: bswapl %ecx |
| ; X86-NOBMI-NEXT: movl %ecx, %edx |
| ; X86-NOBMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F |
| ; X86-NOBMI-NEXT: shll $4, %edx |
| ; X86-NOBMI-NEXT: shrl $4, %ecx |
| ; X86-NOBMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F |
| ; X86-NOBMI-NEXT: orl %edx, %ecx |
| ; X86-NOBMI-NEXT: movl %ecx, %edx |
| ; X86-NOBMI-NEXT: andl $858993459, %edx # imm = 0x33333333 |
| ; X86-NOBMI-NEXT: shrl $2, %ecx |
| ; X86-NOBMI-NEXT: andl $858993459, %ecx # imm = 0x33333333 |
| ; X86-NOBMI-NEXT: leal (%ecx,%edx,4), %ecx |
| ; X86-NOBMI-NEXT: movl %ecx, %edx |
| ; X86-NOBMI-NEXT: andl $1431655765, %edx # imm = 0x55555555 |
| ; X86-NOBMI-NEXT: shrl %ecx |
| ; X86-NOBMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555 |
| ; X86-NOBMI-NEXT: leal (%ecx,%edx,2), %edx |
| ; X86-NOBMI-NEXT: notl %edx |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %edx |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_bitreverse_i64: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %ecx |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: bswapl %eax |
| ; X86-BMI-NEXT: movl %eax, %edx |
| ; X86-BMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F |
| ; X86-BMI-NEXT: shll $4, %edx |
| ; X86-BMI-NEXT: shrl $4, %eax |
| ; X86-BMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F |
| ; X86-BMI-NEXT: orl %edx, %eax |
| ; X86-BMI-NEXT: movl %eax, %edx |
| ; X86-BMI-NEXT: andl $858993459, %edx # imm = 0x33333333 |
| ; X86-BMI-NEXT: shrl $2, %eax |
| ; X86-BMI-NEXT: andl $858993459, %eax # imm = 0x33333333 |
| ; X86-BMI-NEXT: leal (%eax,%edx,4), %eax |
| ; X86-BMI-NEXT: movl %eax, %edx |
| ; X86-BMI-NEXT: andl $1431655765, %edx # imm = 0x55555555 |
| ; X86-BMI-NEXT: shrl %eax |
| ; X86-BMI-NEXT: andl $1431655765, %eax # imm = 0x55555555 |
| ; X86-BMI-NEXT: leal (%eax,%edx,2), %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: bswapl %ecx |
| ; X86-BMI-NEXT: movl %ecx, %edx |
| ; X86-BMI-NEXT: andl $252645135, %edx # imm = 0xF0F0F0F |
| ; X86-BMI-NEXT: shll $4, %edx |
| ; X86-BMI-NEXT: shrl $4, %ecx |
| ; X86-BMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F |
| ; X86-BMI-NEXT: orl %edx, %ecx |
| ; X86-BMI-NEXT: movl %ecx, %edx |
| ; X86-BMI-NEXT: andl $858993459, %edx # imm = 0x33333333 |
| ; X86-BMI-NEXT: shrl $2, %ecx |
| ; X86-BMI-NEXT: andl $858993459, %ecx # imm = 0x33333333 |
| ; X86-BMI-NEXT: leal (%ecx,%edx,4), %ecx |
| ; X86-BMI-NEXT: movl %ecx, %edx |
| ; X86-BMI-NEXT: andl $1431655765, %edx # imm = 0x55555555 |
| ; X86-BMI-NEXT: shrl %ecx |
| ; X86-BMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555 |
| ; X86-BMI-NEXT: leal (%ecx,%edx,2), %ecx |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %ecx, %edx |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_bitreverse_i64: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: bswapq %rsi |
| ; X64-NOBMI-NEXT: movq %rsi, %rax |
| ; X64-NOBMI-NEXT: shrq $4, %rax |
| ; X64-NOBMI-NEXT: movabsq $1085102592571150095, %rcx # imm = 0xF0F0F0F0F0F0F0F |
| ; X64-NOBMI-NEXT: andq %rcx, %rax |
| ; X64-NOBMI-NEXT: andq %rcx, %rsi |
| ; X64-NOBMI-NEXT: shlq $4, %rsi |
| ; X64-NOBMI-NEXT: orq %rax, %rsi |
| ; X64-NOBMI-NEXT: movabsq $3689348814741910323, %rax # imm = 0x3333333333333333 |
| ; X64-NOBMI-NEXT: movq %rsi, %rcx |
| ; X64-NOBMI-NEXT: andq %rax, %rcx |
| ; X64-NOBMI-NEXT: shrq $2, %rsi |
| ; X64-NOBMI-NEXT: andq %rax, %rsi |
| ; X64-NOBMI-NEXT: leaq (%rsi,%rcx,4), %rax |
| ; X64-NOBMI-NEXT: movabsq $6148914691236517205, %rcx # imm = 0x5555555555555555 |
| ; X64-NOBMI-NEXT: movq %rax, %rdx |
| ; X64-NOBMI-NEXT: andq %rcx, %rdx |
| ; X64-NOBMI-NEXT: shrq %rax |
| ; X64-NOBMI-NEXT: andq %rcx, %rax |
| ; X64-NOBMI-NEXT: leaq (%rax,%rdx,2), %rax |
| ; X64-NOBMI-NEXT: notq %rax |
| ; X64-NOBMI-NEXT: andq %rdi, %rax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_bitreverse_i64: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: bswapq %rsi |
| ; X64-BMI-NEXT: movq %rsi, %rax |
| ; X64-BMI-NEXT: shrq $4, %rax |
| ; X64-BMI-NEXT: movabsq $1085102592571150095, %rcx # imm = 0xF0F0F0F0F0F0F0F |
| ; X64-BMI-NEXT: andq %rcx, %rax |
| ; X64-BMI-NEXT: andq %rcx, %rsi |
| ; X64-BMI-NEXT: shlq $4, %rsi |
| ; X64-BMI-NEXT: orq %rax, %rsi |
| ; X64-BMI-NEXT: movabsq $3689348814741910323, %rax # imm = 0x3333333333333333 |
| ; X64-BMI-NEXT: movq %rsi, %rcx |
| ; X64-BMI-NEXT: andq %rax, %rcx |
| ; X64-BMI-NEXT: shrq $2, %rsi |
| ; X64-BMI-NEXT: andq %rax, %rsi |
| ; X64-BMI-NEXT: leaq (%rsi,%rcx,4), %rax |
| ; X64-BMI-NEXT: movabsq $6148914691236517205, %rcx # imm = 0x5555555555555555 |
| ; X64-BMI-NEXT: movq %rax, %rdx |
| ; X64-BMI-NEXT: andq %rcx, %rdx |
| ; X64-BMI-NEXT: shrq %rax |
| ; X64-BMI-NEXT: andq %rcx, %rax |
| ; X64-BMI-NEXT: leaq (%rax,%rdx,2), %rax |
| ; X64-BMI-NEXT: andnq %rdi, %rax, %rax |
| ; X64-BMI-NEXT: retq |
| %not = xor i64 %a1, -1 |
| %bitrev = tail call i64 @llvm.bitreverse.i64(i64 %not) |
| %and = and i64 %bitrev, %a0 |
| ret i64 %and |
| } |
| |
| define i32 @andnot_bitreverse_i32(i32 %a0, i32 %a1) nounwind { |
| ; X86-NOBMI-LABEL: andnot_bitreverse_i32: |
| ; X86-NOBMI: # %bb.0: |
| ; X86-NOBMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: bswapl %eax |
| ; X86-NOBMI-NEXT: movl %eax, %ecx |
| ; X86-NOBMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F |
| ; X86-NOBMI-NEXT: shll $4, %ecx |
| ; X86-NOBMI-NEXT: shrl $4, %eax |
| ; X86-NOBMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F |
| ; X86-NOBMI-NEXT: orl %ecx, %eax |
| ; X86-NOBMI-NEXT: movl %eax, %ecx |
| ; X86-NOBMI-NEXT: andl $858993459, %ecx # imm = 0x33333333 |
| ; X86-NOBMI-NEXT: shrl $2, %eax |
| ; X86-NOBMI-NEXT: andl $858993459, %eax # imm = 0x33333333 |
| ; X86-NOBMI-NEXT: leal (%eax,%ecx,4), %eax |
| ; X86-NOBMI-NEXT: movl %eax, %ecx |
| ; X86-NOBMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555 |
| ; X86-NOBMI-NEXT: shrl %eax |
| ; X86-NOBMI-NEXT: andl $1431655765, %eax # imm = 0x55555555 |
| ; X86-NOBMI-NEXT: leal (%eax,%ecx,2), %eax |
| ; X86-NOBMI-NEXT: notl %eax |
| ; X86-NOBMI-NEXT: andl {{[0-9]+}}(%esp), %eax |
| ; X86-NOBMI-NEXT: retl |
| ; |
| ; X86-BMI-LABEL: andnot_bitreverse_i32: |
| ; X86-BMI: # %bb.0: |
| ; X86-BMI-NEXT: movl {{[0-9]+}}(%esp), %eax |
| ; X86-BMI-NEXT: bswapl %eax |
| ; X86-BMI-NEXT: movl %eax, %ecx |
| ; X86-BMI-NEXT: andl $252645135, %ecx # imm = 0xF0F0F0F |
| ; X86-BMI-NEXT: shll $4, %ecx |
| ; X86-BMI-NEXT: shrl $4, %eax |
| ; X86-BMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F |
| ; X86-BMI-NEXT: orl %ecx, %eax |
| ; X86-BMI-NEXT: movl %eax, %ecx |
| ; X86-BMI-NEXT: andl $858993459, %ecx # imm = 0x33333333 |
| ; X86-BMI-NEXT: shrl $2, %eax |
| ; X86-BMI-NEXT: andl $858993459, %eax # imm = 0x33333333 |
| ; X86-BMI-NEXT: leal (%eax,%ecx,4), %eax |
| ; X86-BMI-NEXT: movl %eax, %ecx |
| ; X86-BMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555 |
| ; X86-BMI-NEXT: shrl %eax |
| ; X86-BMI-NEXT: andl $1431655765, %eax # imm = 0x55555555 |
| ; X86-BMI-NEXT: leal (%eax,%ecx,2), %eax |
| ; X86-BMI-NEXT: andnl {{[0-9]+}}(%esp), %eax, %eax |
| ; X86-BMI-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_bitreverse_i32: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: # kill: def $esi killed $esi def $rsi |
| ; X64-NOBMI-NEXT: bswapl %esi |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F |
| ; X64-NOBMI-NEXT: shll $4, %eax |
| ; X64-NOBMI-NEXT: shrl $4, %esi |
| ; X64-NOBMI-NEXT: andl $252645135, %esi # imm = 0xF0F0F0F |
| ; X64-NOBMI-NEXT: orl %eax, %esi |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: andl $858993459, %eax # imm = 0x33333333 |
| ; X64-NOBMI-NEXT: shrl $2, %esi |
| ; X64-NOBMI-NEXT: andl $858993459, %esi # imm = 0x33333333 |
| ; X64-NOBMI-NEXT: leal (%rsi,%rax,4), %eax |
| ; X64-NOBMI-NEXT: movl %eax, %ecx |
| ; X64-NOBMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555 |
| ; X64-NOBMI-NEXT: shrl %eax |
| ; X64-NOBMI-NEXT: andl $1431655765, %eax # imm = 0x55555555 |
| ; X64-NOBMI-NEXT: leal (%rax,%rcx,2), %eax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_bitreverse_i32: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: # kill: def $esi killed $esi def $rsi |
| ; X64-BMI-NEXT: bswapl %esi |
| ; X64-BMI-NEXT: movl %esi, %eax |
| ; X64-BMI-NEXT: andl $252645135, %eax # imm = 0xF0F0F0F |
| ; X64-BMI-NEXT: shll $4, %eax |
| ; X64-BMI-NEXT: shrl $4, %esi |
| ; X64-BMI-NEXT: andl $252645135, %esi # imm = 0xF0F0F0F |
| ; X64-BMI-NEXT: orl %eax, %esi |
| ; X64-BMI-NEXT: movl %esi, %eax |
| ; X64-BMI-NEXT: andl $858993459, %eax # imm = 0x33333333 |
| ; X64-BMI-NEXT: shrl $2, %esi |
| ; X64-BMI-NEXT: andl $858993459, %esi # imm = 0x33333333 |
| ; X64-BMI-NEXT: leal (%rsi,%rax,4), %eax |
| ; X64-BMI-NEXT: movl %eax, %ecx |
| ; X64-BMI-NEXT: andl $1431655765, %ecx # imm = 0x55555555 |
| ; X64-BMI-NEXT: shrl %eax |
| ; X64-BMI-NEXT: andl $1431655765, %eax # imm = 0x55555555 |
| ; X64-BMI-NEXT: leal (%rax,%rcx,2), %eax |
| ; X64-BMI-NEXT: andnl %edi, %eax, %eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i32 %a1, -1 |
| %bitrev = tail call i32 @llvm.bitreverse.i32(i32 %not) |
| %and = and i32 %bitrev, %a0 |
| ret i32 %and |
| } |
| |
| define i16 @andnot_bitreverse_i16(i16 %a0, i16 %a1) nounwind { |
| ; X86-LABEL: andnot_bitreverse_i16: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movzwl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: rolw $8, %ax |
| ; X86-NEXT: movl %eax, %ecx |
| ; X86-NEXT: andl $3855, %ecx # imm = 0xF0F |
| ; X86-NEXT: shll $4, %ecx |
| ; X86-NEXT: shrl $4, %eax |
| ; X86-NEXT: andl $3855, %eax # imm = 0xF0F |
| ; X86-NEXT: orl %ecx, %eax |
| ; X86-NEXT: movl %eax, %ecx |
| ; X86-NEXT: andl $13107, %ecx # imm = 0x3333 |
| ; X86-NEXT: shrl $2, %eax |
| ; X86-NEXT: andl $13107, %eax # imm = 0x3333 |
| ; X86-NEXT: leal (%eax,%ecx,4), %eax |
| ; X86-NEXT: movl %eax, %ecx |
| ; X86-NEXT: andl $21845, %ecx # imm = 0x5555 |
| ; X86-NEXT: shrl %eax |
| ; X86-NEXT: andl $21845, %eax # imm = 0x5555 |
| ; X86-NEXT: leal (%eax,%ecx,2), %eax |
| ; X86-NEXT: notl %eax |
| ; X86-NEXT: andw {{[0-9]+}}(%esp), %ax |
| ; X86-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X86-NEXT: retl |
| ; |
| ; X64-NOBMI-LABEL: andnot_bitreverse_i16: |
| ; X64-NOBMI: # %bb.0: |
| ; X64-NOBMI-NEXT: # kill: def $esi killed $esi def $rsi |
| ; X64-NOBMI-NEXT: rolw $8, %si |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: andl $3855, %eax # imm = 0xF0F |
| ; X64-NOBMI-NEXT: shll $4, %eax |
| ; X64-NOBMI-NEXT: shrl $4, %esi |
| ; X64-NOBMI-NEXT: andl $3855, %esi # imm = 0xF0F |
| ; X64-NOBMI-NEXT: orl %eax, %esi |
| ; X64-NOBMI-NEXT: movl %esi, %eax |
| ; X64-NOBMI-NEXT: andl $13107, %eax # imm = 0x3333 |
| ; X64-NOBMI-NEXT: shrl $2, %esi |
| ; X64-NOBMI-NEXT: andl $13107, %esi # imm = 0x3333 |
| ; X64-NOBMI-NEXT: leal (%rsi,%rax,4), %eax |
| ; X64-NOBMI-NEXT: movl %eax, %ecx |
| ; X64-NOBMI-NEXT: andl $21845, %ecx # imm = 0x5555 |
| ; X64-NOBMI-NEXT: shrl %eax |
| ; X64-NOBMI-NEXT: andl $21845, %eax # imm = 0x5555 |
| ; X64-NOBMI-NEXT: leal (%rax,%rcx,2), %eax |
| ; X64-NOBMI-NEXT: notl %eax |
| ; X64-NOBMI-NEXT: andl %edi, %eax |
| ; X64-NOBMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-NOBMI-NEXT: retq |
| ; |
| ; X64-BMI-LABEL: andnot_bitreverse_i16: |
| ; X64-BMI: # %bb.0: |
| ; X64-BMI-NEXT: # kill: def $esi killed $esi def $rsi |
| ; X64-BMI-NEXT: rolw $8, %si |
| ; X64-BMI-NEXT: movl %esi, %eax |
| ; X64-BMI-NEXT: andl $3855, %eax # imm = 0xF0F |
| ; X64-BMI-NEXT: shll $4, %eax |
| ; X64-BMI-NEXT: shrl $4, %esi |
| ; X64-BMI-NEXT: andl $3855, %esi # imm = 0xF0F |
| ; X64-BMI-NEXT: orl %eax, %esi |
| ; X64-BMI-NEXT: movl %esi, %eax |
| ; X64-BMI-NEXT: andl $13107, %eax # imm = 0x3333 |
| ; X64-BMI-NEXT: shrl $2, %esi |
| ; X64-BMI-NEXT: andl $13107, %esi # imm = 0x3333 |
| ; X64-BMI-NEXT: leal (%rsi,%rax,4), %eax |
| ; X64-BMI-NEXT: movl %eax, %ecx |
| ; X64-BMI-NEXT: andl $21845, %ecx # imm = 0x5555 |
| ; X64-BMI-NEXT: shrl %eax |
| ; X64-BMI-NEXT: andl $21845, %eax # imm = 0x5555 |
| ; X64-BMI-NEXT: leal (%rax,%rcx,2), %eax |
| ; X64-BMI-NEXT: andnl %edi, %eax, %eax |
| ; X64-BMI-NEXT: # kill: def $ax killed $ax killed $eax |
| ; X64-BMI-NEXT: retq |
| %not = xor i16 %a1, -1 |
| %bitrev = tail call i16 @llvm.bitreverse.i16(i16 %not) |
| %and = and i16 %bitrev, %a0 |
| ret i16 %and |
| } |
| |
| define i8 @andnot_bitreverse_i8(i8 %a0, i8 %a1) nounwind { |
| ; X86-LABEL: andnot_bitreverse_i8: |
| ; X86: # %bb.0: |
| ; X86-NEXT: movzbl {{[0-9]+}}(%esp), %eax |
| ; X86-NEXT: rolb $4, %al |
| ; X86-NEXT: movl %eax, %ecx |
| ; X86-NEXT: andb $51, %cl |
| ; X86-NEXT: shlb $2, %cl |
| ; X86-NEXT: shrb $2, %al |
| ; X86-NEXT: andb $51, %al |
| ; X86-NEXT: orb %cl, %al |
| ; X86-NEXT: movl %eax, %ecx |
| ; X86-NEXT: andb $85, %cl |
| ; X86-NEXT: addb %cl, %cl |
| ; X86-NEXT: shrb %al |
| ; X86-NEXT: andb $85, %al |
| ; X86-NEXT: orb %cl, %al |
| ; X86-NEXT: notb %al |
| ; X86-NEXT: andb {{[0-9]+}}(%esp), %al |
| ; X86-NEXT: retl |
| ; |
| ; X64-LABEL: andnot_bitreverse_i8: |
| ; X64: # %bb.0: |
| ; X64-NEXT: rolb $4, %sil |
| ; X64-NEXT: movl %esi, %eax |
| ; X64-NEXT: andb $51, %al |
| ; X64-NEXT: shlb $2, %al |
| ; X64-NEXT: shrb $2, %sil |
| ; X64-NEXT: andb $51, %sil |
| ; X64-NEXT: orb %sil, %al |
| ; X64-NEXT: movl %eax, %ecx |
| ; X64-NEXT: andb $85, %cl |
| ; X64-NEXT: addb %cl, %cl |
| ; X64-NEXT: shrb %al |
| ; X64-NEXT: andb $85, %al |
| ; X64-NEXT: orb %cl, %al |
| ; X64-NEXT: notb %al |
| ; X64-NEXT: andb %dil, %al |
| ; X64-NEXT: retq |
| %not = xor i8 %a1, -1 |
| %bitrev = tail call i8 @llvm.bitreverse.i8(i8 %not) |
| %and = and i8 %bitrev, %a0 |
| ret i8 %and |
| } |