| // RUN: mlir-opt %s -split-input-file -quant-convert-simulated-quantization | FileCheck %s |
| |
| // ----- |
| // Verifies a quint8 single point. |
| // CHECK-LABEL: fakeQuantArgs_Quint8_0 |
| func @fakeQuantArgs_Quint8_0(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8:f32, 1.000000e+00>> |
| // CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<u8:f32, 1.000000e+00>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 0.0 : f32, num_bits = 8 |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a quint8 single point (with narrow_range = true). |
| // CHECK-LABEL: fakeQuantArgs_Quint8_0_NarrowRange |
| func @fakeQuantArgs_Quint8_0_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 1.000000e+00:1>> |
| // CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 1.000000e+00:1>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 0.0 : f32, num_bits = 8, narrow_range = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a quint8 asymmetric 0..1 range. |
| // CHECK-LABEL: fakeQuantArgs_Quint8_0_1 |
| func @fakeQuantArgs_Quint8_0_1(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8:f32, 0.0039215686274509803>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<u8:f32, 0.0039215686274509803>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 1.0 : f32, num_bits = 8 |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a quint8 asymmetric 0..1 range (with narrow_range = true). |
| // CHECK-LABEL: fakeQuantArgs_Quint8_NarrowRange |
| func @fakeQuantArgs_Quint8_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 0.003937007874015748:1>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<u8<1:255>:f32, 0.003937007874015748:1>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 1.0 : f32, num_bits = 8, narrow_range = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a quint8 symmetric range of -1..127/128. |
| // CHECK-LABEL: fakeQuantArgs_Quint8_SymmetricRange |
| func @fakeQuantArgs_Quint8_SymmetricRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<u8:f32, 7.812500e-03:128>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<u8:f32, 7.812500e-03:128>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = -1.0 : f32, max = 0.9921875 : f32, num_bits = 8, narrow_range = false |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a qint8 single point. |
| // CHECK-LABEL: fakeQuantArgs_Qint8_0 |
| func @fakeQuantArgs_Qint8_0(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 1.000000e+00:-128>> |
| // CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8:f32, 1.000000e+00:-128>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 0.0 : f32, num_bits = 8, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a qint8 single point (with narrow_range = true). |
| // CHECK-LABEL: fakeQuantArgs_Qint8_0_NarrowRange |
| func @fakeQuantArgs_Qint8_0_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 1.000000e+00:-127>> |
| // CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 1.000000e+00:-127>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 0.0 : f32, num_bits = 8, narrow_range = true, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a qint8 asymmetric 0..1 range. |
| // CHECK-LABEL: fakeQuantArgs_Qint8_0_1 |
| func @fakeQuantArgs_Qint8_0_1(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 1.0 : f32, num_bits = 8, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a qint8 asymmetric 0..1 range (with narrow_range = true). |
| // CHECK-LABEL: fakeQuantArgs_Qint8_NarrowRange |
| func @fakeQuantArgs_Qint8_NarrowRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 0.003937007874015748:-127>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i8<-127:127>:f32, 0.003937007874015748:-127>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 1.0 : f32, num_bits = 8, narrow_range = true, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a qint8 symmetric range of -1..127/128. |
| // CHECK-LABEL: fakeQuantArgs_Qint8_SymmetricRange |
| func @fakeQuantArgs_Qint8_SymmetricRange(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 7.812500e-03>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i8:f32, 7.812500e-03>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = -1.0 : f32, max = 0.9921875 : f32, num_bits = 8, narrow_range = false, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a commonly used -1..1 symmetric 16bit range with a zero point of |
| // 0 and range -1.0 .. 32767/32768. |
| // CHECK-LABEL: fakeQuantArgs_Qint16_Symmetric |
| func @fakeQuantArgs_Qint16_Symmetric(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i16:f32, 3.0517578125E-5>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<8x4x3x!quant.uniform<i16:f32, 3.0517578125E-5>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = -1.0 : f32, max = 0.999969482 : f32, num_bits = 16, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verify that lowering to barriers of unranked tensors functions. |
| // CHECK-LABEL: fakeQuantArgs_UnrankedTensor |
| func @fakeQuantArgs_UnrankedTensor(tensor<f32>) -> tensor<f32> { |
| ^bb0(%arg0: tensor<f32>): |
| // CHECK: %0 = "quant.qcast"(%arg0) : (tensor<f32>) |
| // CHECK-SAME: -> tensor<!quant.uniform<u8:f32, 0.0039215686274509803>> |
| // CHECK-NEXT: %1 = "quant.dcast"(%0) : (tensor<!quant.uniform<u8:f32, 0.0039215686274509803>>) |
| // CHECK-SAME: -> tensor<f32> |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.0 : f32, max = 1.0 : f32, num_bits = 8 |
| } : (tensor<f32>) -> tensor<f32> |
| return %0 : tensor<f32> |
| } |
| |
| // ----- |
| // CHECK-LABEL: fakeQuantArgs_all_positive |
| func @fakeQuantArgs_all_positive(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| |
| // CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>> |
| // CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:-128>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = 0.5 : f32, max = 1.5 : f32, num_bits = 8, narrow_range = false, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // CHECK-LABEL: fakeQuantArgs_all_negative |
| func @fakeQuantArgs_all_negative(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| |
| // CHECK: %[[qc:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:127>> |
| // CHECK-NEXT: "quant.dcast"(%[[qc]]) : (tensor<8x4x3x!quant.uniform<i8:f32, 0.0039215686274509803:127>>) |
| // CHECK-SAME: -> tensor<8x4x3xf32> |
| |
| %0 = "quant.const_fake_quant"(%arg0) { |
| min = -1.5 : f32, max = -0.5 : f32, num_bits = 8, narrow_range = false, is_signed = true |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |
| |
| // ----- |
| // Verifies a qint8 per axis |
| // CHECK-LABEL: fakeQuantPerAxis |
| func @fakeQuantPerAxis(tensor<8x4x3xf32>) -> tensor<8x4x3xf32> { |
| ^bb0(%arg0: tensor<8x4x3xf32>): |
| |
| // CHECK: %[[q:.*]] = "quant.qcast"(%arg0) : (tensor<8x4x3xf32>) |
| // CHECK-SAME: -> tensor<8x4x3x!quant.uniform<i8:f32:2, {7.812500e-03,1.000000e+00:-128,0.0039215686274509803:-128}>> |
| // CHECK: %[[d:.*]] = "quant.dcast"(%[[q]]) |
| // CHECK-SAME: (tensor<8x4x3x!quant.uniform<i8:f32:2, {7.812500e-03,1.000000e+00:-128,0.0039215686274509803:-128}>>) |
| |
| %0 = "quant.const_fake_quant_per_axis"(%arg0) { |
| min = [-1.0 : f32, 0.0 : f32, 0.0 : f32], |
| max = [0.9921875 : f32, 0.0: f32, 1.0 : f32], |
| num_bits = 8, narrow_range = false, is_signed = true, axis = 2 |
| } : (tensor<8x4x3xf32>) -> tensor<8x4x3xf32> |
| return %0 : tensor<8x4x3xf32> |
| } |