| // RUN: mlir-opt %s -split-input-file -verify-diagnostics |
| |
| func.func @broadcast_to_scalar(%arg0: f32) -> f32 { |
| // expected-error@+1 {{custom op 'vector.broadcast' invalid kind of type specified: expected builtin.vector, but found 'f32'}} |
| %0 = vector.broadcast %arg0 : f32 to f32 |
| } |
| |
| // ----- |
| |
| func.func @broadcast_rank_too_high(%arg0: vector<4x4xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op source rank higher than destination rank}} |
| %1 = vector.broadcast %arg0 : vector<4x4xf32> to vector<4xf32> |
| } |
| |
| // ----- |
| |
| func.func @broadcast_rank_too_high_0d(%arg0: vector<1xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op source rank higher than destination rank}} |
| %1 = vector.broadcast %arg0 : vector<1xf32> to vector<f32> |
| } |
| |
| // ----- |
| |
| func.func @broadcast_dim1_mismatch(%arg0: vector<7xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op dimension mismatch (7 vs. 3)}} |
| %1 = vector.broadcast %arg0 : vector<7xf32> to vector<3xf32> |
| } |
| |
| // ----- |
| |
| func.func @broadcast_dim2_mismatch(%arg0: vector<4x8xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op dimension mismatch (4 vs. 1)}} |
| %1 = vector.broadcast %arg0 : vector<4x8xf32> to vector<1x8xf32> |
| } |
| |
| // ----- |
| |
| func.func @broadcast_scalable_unit_dim(%arg0: vector<[1]xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op dimension mismatch ([1] vs. [4])}} |
| %0 = vector.broadcast %arg0 : vector<[1]xf32> to vector<[4]xf32> |
| } |
| |
| // ----- |
| |
| func.func @broadcast_fixed_to_scalable(%arg0: vector<2xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op dimension mismatch (2 vs. [2])}} |
| %0 = vector.broadcast %arg0 : vector<2xf32> to vector<[2]xf32> |
| } |
| |
| // ----- |
| |
| func.func @broadcast_scalable_to_fixed(%arg0: vector<[1]xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op dimension mismatch ([1] vs. 1)}} |
| %0 = vector.broadcast %arg0 : vector<[1]xf32> to vector<4x1xf32> |
| } |
| |
| // ----- |
| |
| func.func @broadcast_unknown(%arg0: memref<4x8xf32>) { |
| // expected-error@+1 {{'vector.broadcast' op source type is not a vector}} |
| %1 = vector.broadcast %arg0 : memref<4x8xf32> to vector<1x8xf32> |
| } |
| |
| // ----- |
| |
| func.func @fma_vector_4xi32(%arg0: vector<4xi32>) { |
| // expected-error@+1 {{'vector.fma' op operand #0 must be vector of floating-point value}} |
| %1 = vector.fma %arg0, %arg0, %arg0 : vector<4xi32> |
| } |
| |
| // ----- |
| |
| func.func @shuffle_elt_type_mismatch(%arg0: vector<2xf32>, %arg1: vector<2xi32>) { |
| // expected-error@+1 {{'vector.shuffle' op failed to verify that second operand v2 and result have same element type}} |
| %1 = vector.shuffle %arg0, %arg1 [0, 1] : vector<2xf32>, vector<2xi32> |
| } |
| |
| // ----- |
| |
| func.func @shuffle_rank_mismatch(%arg0: vector<2xf32>, %arg1: vector<4x2xf32>) { |
| // expected-error@+1 {{'vector.shuffle' op rank mismatch}} |
| %1 = vector.shuffle %arg0, %arg1 [0, 1] : vector<2xf32>, vector<4x2xf32> |
| } |
| |
| // ----- |
| |
| func.func @shuffle_rank_mismatch_0d(%arg0: vector<f32>, %arg1: vector<1xf32>) { |
| // expected-error@+1 {{'vector.shuffle' op rank mismatch}} |
| %1 = vector.shuffle %arg0, %arg1 [0, 1] : vector<f32>, vector<1xf32> |
| } |
| |
| // ----- |
| |
| func.func @shuffle_trailing_dim_size_mismatch(%arg0: vector<2x2xf32>, %arg1: vector<2x4xf32>) { |
| // expected-error@+1 {{'vector.shuffle' op dimension mismatch}} |
| %1 = vector.shuffle %arg0, %arg1 [0, 1] : vector<2x2xf32>, vector<2x4xf32> |
| } |
| |
| // ----- |
| |
| func.func @shuffle_index_out_of_range(%arg0: vector<2xf32>, %arg1: vector<2xf32>) { |
| // expected-error@+1 {{'vector.shuffle' op mask index #2 out of range}} |
| %1 = vector.shuffle %arg0, %arg1 [0, 4] : vector<2xf32>, vector<2xf32> |
| } |
| |
| // ----- |
| |
| func.func @shuffle_scalable_vec(%arg0: vector<[2]xf32>, %arg1: vector<[2]xf32>) { |
| // expected-error@+1 {{'vector.shuffle' op operand #0 must be fixed-length vector of any type values}} |
| %1 = vector.shuffle %arg0, %arg1 [0, 1, 2, 3] : vector<[2]xf32>, vector<[2]xf32> |
| } |
| |
| // ----- |
| |
| func.func @shuffle_empty_mask(%arg0: vector<2xf32>, %arg1: vector<2xf32>) { |
| // expected-error@+1 {{'vector.shuffle' op invalid mask length}} |
| %1 = vector.shuffle %arg0, %arg1 [] : vector<2xf32>, vector<2xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_vector_type(%arg0: index) { |
| // expected-error@+1 {{invalid kind of type specified: expected builtin.vector, but found 'index'}} |
| %1 = vector.extract %arg0[] : index from index |
| } |
| |
| // ----- |
| |
| func.func @extract_position_rank_overflow(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute of rank no greater than vector rank}} |
| %1 = vector.extract %arg0[0, 0, 0, 0] : f32 from vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_position_rank_overflow_generic(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute of rank no greater than vector rank}} |
| %1 = "vector.extract" (%arg0) <{static_position = array<i64: 0, 0, 0, 0>}> : (vector<4x8x16xf32>) -> (vector<16xf32>) |
| } |
| |
| // ----- |
| |
| func.func @extract_position_overflow(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute #2 to be a non-negative integer smaller than the corresponding vector dimension}} |
| %1 = vector.extract %arg0[0, 43, 0] : f32 from vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_precise_position_overflow(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute #3 to be a non-negative integer smaller than the corresponding vector dimension}} |
| %1 = vector.extract %arg0[3, 7, 16] : f32 from vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_0d_result(%arg0: vector<f32>) { |
| // expected-error@+1 {{expected a scalar instead of a 0-d vector as the result type}} |
| %1 = vector.extract %arg0[] : vector<f32> from vector<f32> |
| } |
| |
| // ----- |
| |
| func.func @extract_position_overflow(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute #3 to be a non-negative integer smaller than the corresponding vector dimension or poison (-1)}} |
| %1 = vector.extract %arg0[0, 0, -5] : f32 from vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_vector_type(%a: f32, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute of rank no greater than dest vector rank}} |
| %1 = vector.insert %a, %b[3, 3, 3, 3, 3, 3] : f32 into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_vector_type(%a: vector<4xf32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute rank + source rank to match dest vector rank}} |
| %1 = vector.insert %a, %b[3] : vector<4xf32> into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_vector_type(%a: f32, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute rank to match the dest vector rank}} |
| %1 = vector.insert %a, %b[3, 3] : f32 into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_position_overflow(%a: f32, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute #3 to be a non-negative integer smaller than the corresponding dest vector dimension}} |
| %1 = vector.insert %a, %b[0, 0, -5] : f32 into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_precise_position_overflow(%a: f32, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected position attribute #1 to be a non-negative integer smaller than the corresponding dest vector dimension}} |
| %1 = vector.insert %a, %b[4, 7, 15] : f32 into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_0d_value_to_store(%a: vector<f32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected a scalar instead of a 0-d vector as the source operand}} |
| %1 = vector.insert %a, %b[0, 0, 0] : vector<f32> into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_num_operands(%arg0: f32) { |
| // expected-error@+1 {{expected at least 2 operands}} |
| %1 = vector.outerproduct %arg0 : f32, f32 |
| } |
| // ----- |
| |
| func.func @outerproduct_non_vector_operand(%arg0: f32) { |
| // expected-error@+1 {{expected vector type for operand #1}} |
| %1 = vector.outerproduct %arg0, %arg0 : f32, f32 |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_operand_1(%arg0: vector<4xf32>, %arg1: vector<4x8xf32>) { |
| // expected-error@+1 {{expected 1-d vector for operand #1}} |
| %1 = vector.outerproduct %arg1, %arg1 : vector<4x8xf32>, vector<4x8xf32> |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_operand_2(%arg0: vector<4xf32>, %arg1: vector<4x8xf32>) { |
| // expected-error@+1 {{expected 1-d vector for operand #2}} |
| %1 = vector.outerproduct %arg0, %arg1 : vector<4xf32>, vector<4x8xf32> |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_result_generic(%arg0: vector<4xf32>, %arg1: vector<8xf32>) { |
| // expected-error@+1 {{expected 2-d vector result}} |
| %1 = "vector.outerproduct" (%arg0, %arg1) : (vector<4xf32>, vector<8xf32>) -> (vector<8xf32>) |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_operand_1_dim_generic(%arg0: vector<4xf32>, %arg1: vector<8xf32>) { |
| // expected-error@+1 {{expected #1 operand dim to match result dim #1}} |
| %1 = "vector.outerproduct" (%arg0, %arg1) : (vector<4xf32>, vector<8xf32>) -> (vector<8x16xf32>) |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_operand_2_dim_generic(%arg0: vector<4xf32>, %arg1: vector<8xf32>) { |
| // expected-error@+1 {{expected #2 operand dim to match result dim #2}} |
| %1 = "vector.outerproduct" (%arg0, %arg1) : (vector<4xf32>, vector<8xf32>) -> (vector<4x16xf32>) |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_axpy_operand(%arg0: vector<4x8xf32>, %arg1: f32) { |
| // expected-error@+1 {{expected 1-d vector for operand #1}} |
| %1 = vector.outerproduct %arg0, %arg1 : vector<4x8xf32>, f32 |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_axpy_result_generic(%arg0: vector<4xf32>, %arg1: f32) { |
| // expected-error@+1 {{expected 1-d vector result}} |
| %1 = "vector.outerproduct" (%arg0, %arg1) : (vector<4xf32>, f32) -> (vector<4x8xf32>) |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_axpy_operand_dim_generic(%arg0: vector<8xf32>, %arg1: f32) { |
| // expected-error@+1 {{expected #1 operand dim to match result dim #1}} |
| %1 = "vector.outerproduct" (%arg0, %arg1) : (vector<8xf32>, f32) -> (vector<16xf32>) |
| } |
| |
| // ----- |
| |
| func.func @outerproduct_operand_3_result_type_generic(%arg0: vector<4xf32>, %arg1: vector<8xf32>, %arg2: vector<4x16xf32>) { |
| // expected-error@+1 {{expected operand #3 of same type as result type}} |
| %1 = "vector.outerproduct" (%arg0, %arg1, %arg2) : (vector<4xf32>, vector<8xf32>, vector<4x16xf32>) -> (vector<4x8xf32>) |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires two types}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %cst { permutation_map = affine_map<()->(0)> } : memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| #map1 = affine_map<(d0, d1, d2) -> (d0, 0, 0)> |
| func.func @main(%m: memref<1xi32>, %2: vector<1x32xi1>) -> vector<1x32xi32> { |
| %0 = arith.constant 1 : index |
| %1 = arith.constant 1 : i32 |
| // expected-error@+1 {{expected the same rank for the vector and the results of the permutation map}} |
| %3 = vector.transfer_read %m[%0], %1, %2 { permutation_map = #map1 } : memref<1xi32>, vector<1x32xi32> |
| return %3 : vector<1x32xi32> |
| } |
| |
| // ----- |
| |
| #map1 = affine_map<(d0, d1, d2) -> (d0, 0, 0)> |
| func.func @test_vector.transfer_write(%m: memref<1xi32>, %2: vector<1x32xi32>) -> vector<1x32xi32> { |
| %0 = arith.constant 1 : index |
| %1 = arith.constant 1 : i32 |
| // expected-error@+1 {{expected the same rank for the vector and the results of the permutation map}} |
| %3 = vector.transfer_write %2, %m[%0], %1 { permutation_map = #map1 } : vector<1x32xi32>, memref<1xi32> |
| return %3 : vector<1x32xi32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: vector<4x3xf32>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<4x3xf32> |
| // expected-error@+1 {{ requires memref or ranked tensor type}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %vf0 : vector<4x3xf32>, vector<1x1x2x3xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<4x3xf32>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<4x3xf32> |
| // expected-error@+1 {{ requires vector type}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %vf0 : memref<4x3xf32>, f32 |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires 2 indices}} |
| %0 = vector.transfer_read %arg0[%c3, %c3, %c3], %cst { permutation_map = affine_map<()->(0)> } : memref<?x?xf32>, vector<128xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires 2 indices}} |
| %0 = vector.transfer_read %arg0[%c3], %cst { permutation_map = affine_map<()->(0)> } : memref<?x?xf32>, vector<128xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires a permutation_map with input dims of the same rank as the source type}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %cst {permutation_map = affine_map<(d0)->(d0)>} : memref<?x?xf32>, vector<128xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires a permutation_map with result dims of the same rank as the vector type}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %cst {permutation_map = affine_map<(d0, d1)->(d0, d1)>} : memref<?x?xf32>, vector<128xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires a projected permutation_map (at most one dim or the zero constant can appear in each result)}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %cst {permutation_map = affine_map<(d0, d1)->(d0 + d1)>} : memref<?x?xf32>, vector<128xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires a projected permutation_map (at most one dim or the zero constant can appear in each result)}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %cst {permutation_map = affine_map<(d0, d1)->(d0 + 1)>} : memref<?x?xf32>, vector<128xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires a projected permutation_map (at most one dim or the zero constant can appear in each result)}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %cst {permutation_map = affine_map<(d0, d1)->(1)>} : memref<?x?xf32>, vector<128xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires a permutation_map that is a permutation (found one dim used more than once)}} |
| %0 = vector.transfer_read %arg0[%c3, %c3, %c3], %cst {permutation_map = affine_map<(d0, d1, d2)->(d0, d0)>} : memref<?x?x?xf32>, vector<3x7xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?x?xf32>) { |
| %c1 = arith.constant 1 : i1 |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-note@+1 {{prior use here}} |
| %mask = vector.splat %c1 : vector<3x8x7xi1> |
| // expected-error@+1 {{expects different type than prior uses: 'vector<3x7xi1>' vs 'vector<3x8x7xi1>'}} |
| %0 = vector.transfer_read %arg0[%c3, %c3, %c3], %cst, %mask {permutation_map = affine_map<(d0, d1, d2)->(d0, 0, d2)>} : memref<?x?x?xf32>, vector<3x8x7xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xvector<4x3xf32>>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<4x3xf32> |
| // expected-error@+1 {{requires source vector element and vector result ranks to match}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %vf0 {permutation_map = affine_map<(d0, d1)->(d0, d1)>} : memref<?x?xvector<4x3xf32>>, vector<3xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xvector<6xf32>>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<6xf32> |
| // expected-error@+1 {{requires the bitwidth of the minor 1-D vector to be an integral multiple of the bitwidth of the minor 1-D vector of the source}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %vf0 : memref<?x?xvector<6xf32>>, vector<3xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xvector<2x3xf32>>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<2x3xf32> |
| // expected-error@+1 {{ expects the in_bounds attr of same rank as permutation_map results: affine_map<(d0, d1) -> (d0, d1)>}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %vf0 {in_bounds = [true], permutation_map = affine_map<(d0, d1)->(d0, d1)>} : memref<?x?xvector<2x3xf32>>, vector<1x1x2x3xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?x?xvector<2x3xf32>>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<2x3xf32> |
| %mask = vector.splat %c1 : vector<2x3xi1> |
| // expected-error@+1 {{does not support masks with vector element type}} |
| %0 = vector.transfer_read %arg0[%c3, %c3], %vf0, %mask {permutation_map = affine_map<(d0, d1)->(d0, d1)>} : memref<?x?xvector<2x3xf32>>, vector<1x1x2x3xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_read(%arg0: memref<?xindex>) -> vector<3x4xindex> { |
| %c3 = arith.constant 3 : index |
| // expected-error@+1 {{expected a custom permutation_map when rank(source) != rank(destination)}} |
| %0 = vector.transfer_read %arg0[%c3], %c3 : memref<?xindex>, vector<3x4xindex> |
| return %0 : vector<3x4xindex> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?xvector<2xindex>>) { |
| %c3 = arith.constant 3 : index |
| // expected-error@+1 {{expected a custom permutation_map when rank(source) != rank(destination)}} |
| %0 = vector.transfer_read %arg0[%c3], %c3 : memref<?xvector<2xindex>>, vector<2x3x4xindex> |
| return %0 : vector<2x3x4xindex> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{requires two types}} |
| vector.transfer_write %arg0, %arg0[%c3, %c3] : memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<vector<4x3xf32>>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<4x3xf32> |
| // expected-error@+1 {{ requires vector type}} |
| vector.transfer_write %arg0, %arg0[%c3, %c3] : memref<vector<4x3xf32>>, vector<4x3xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: vector<4x3xf32>) { |
| %c3 = arith.constant 3 : index |
| %f0 = arith.constant 0.0 : f32 |
| %vf0 = vector.splat %f0 : vector<4x3xf32> |
| // expected-error@+1 {{ requires memref or ranked tensor type}} |
| vector.transfer_write %arg0, %arg0[%c3, %c3] : vector<4x3xf32>, f32 |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{expected 5 operand types but had 4}} |
| %0 = "vector.transfer_write"(%cst, %arg0, %c3, %c3, %c3) {permutation_map = affine_map<()->(0)>} : (vector<128xf32>, memref<?x?xf32>, index, index) -> () |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{requires 2 indices}} |
| vector.transfer_write %cst, %arg0[%c3, %c3, %c3] {permutation_map = affine_map<()->(0)>} : vector<128xf32>, memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{requires 2 indices}} |
| vector.transfer_write %cst, %arg0[%c3] {permutation_map = affine_map<()->(0)>} : vector<128xf32>, memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{requires a permutation_map with input dims of the same rank as the source type}} |
| vector.transfer_write %cst, %arg0[%c3, %c3] {permutation_map = affine_map<(d0)->(d0)>} : vector<128xf32>, memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{requires a permutation_map with result dims of the same rank as the vector type}} |
| vector.transfer_write %cst, %arg0[%c3, %c3] {permutation_map = affine_map<(d0, d1)->(d0, d1)>} : vector<128xf32>, memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{requires a projected permutation_map (at most one dim or the zero constant can appear in each result)}} |
| vector.transfer_write %cst, %arg0[%c3, %c3] {permutation_map = affine_map<(d0, d1)->(d0 + d1)>} : vector<128xf32>, memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{requires a projected permutation_map (at most one dim or the zero constant can appear in each result)}} |
| vector.transfer_write %cst, %arg0[%c3, %c3] {permutation_map = affine_map<(d0, d1)->(d0 + 1)>} : vector<128xf32>, memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<128 x f32> |
| // expected-error@+1 {{requires a projected permutation_map (at most one dim or the zero constant can appear in each result)}} |
| vector.transfer_write %cst, %arg0[%c3, %c3] {permutation_map = affine_map<(d0, d1)->(1)>} : vector<128xf32>, memref<?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?x?x?xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant dense<3.0> : vector<3 x 7 x f32> |
| // expected-error@+1 {{requires a permutation_map that is a permutation (found one dim used more than once)}} |
| vector.transfer_write %cst, %arg0[%c3, %c3, %c3] {permutation_map = affine_map<(d0, d1, d2)->(d0, d0)>} : vector<3x7xf32>, memref<?x?x?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?xf32>, %arg1: vector<7xf32>) { |
| %c3 = arith.constant 3 : index |
| %cst = arith.constant 3.0 : f32 |
| // expected-error@+1 {{should not have broadcast dimensions}} |
| vector.transfer_write %arg1, %arg0[%c3] |
| {permutation_map = affine_map<(d0) -> (0)>} |
| : vector<7xf32>, memref<?xf32> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?xindex>, %arg1: vector<3x4xindex>) { |
| %c3 = arith.constant 3 : index |
| // expected-error@+1 {{expected a custom permutation_map when rank(source) != rank(destination)}} |
| vector.transfer_write %arg1, %arg0[%c3, %c3] : vector<3x4xindex>, memref<?xindex> |
| } |
| |
| // ----- |
| |
| func.func @test_vector.transfer_write(%arg0: memref<?xvector<2xindex>>, %arg1: vector<2x3x4xindex>) { |
| %c3 = arith.constant 3 : index |
| // expected-error@+1 {{expected a custom permutation_map when rank(source) != rank(destination)}} |
| vector.transfer_write %arg1, %arg0[%c3, %c3] : vector<2x3x4xindex>, memref<?xvector<2xindex>> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice(%a: vector<4x4xf32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected offsets of same size as destination vector rank}} |
| %1 = vector.insert_strided_slice %a, %b {offsets = [100], strides = [1, 1]} : vector<4x4xf32> into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice(%a: vector<4x4xf32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected strides of same size as source vector rank}} |
| %1 = vector.insert_strided_slice %a, %b {offsets = [2, 2, 2], strides = [1]} : vector<4x4xf32> into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice(%a: vector<4x4xf32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected source rank to be no greater than destination rank}} |
| %1 = vector.insert_strided_slice %b, %a {offsets = [2, 2], strides = [1, 1, 1]} : vector<4x8x16xf32> into vector<4x4xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice(%a: vector<4x4xf32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected offsets dimension 0 to be confined to [0, 4)}} |
| %1 = vector.insert_strided_slice %a, %b {offsets = [100,100,100], strides = [1, 1]} : vector<4x4xf32> into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice(%a: vector<4x4xf32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected strides to be confined to [1, 2)}} |
| %1 = vector.insert_strided_slice %a, %b {offsets = [2, 2, 2], strides = [100, 100]} : vector<4x4xf32> into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice(%a: vector<4x4xf32>, %b: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected sum(offsets, source vector shape) dimension 1 to be confined to [1, 9)}} |
| %1 = vector.insert_strided_slice %a, %b {offsets = [2, 7, 2], strides = [1, 1]} : vector<4x4xf32> into vector<4x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice_scalable(%a : vector<1x1x[2]xi32>, %b: vector<1x4x[4]xi32>) -> vector<1x4x[4]xi32> { |
| // expected-error@+1 {{op expected size at idx=2 to match the corresponding base size from the input vector (2 vs 4)}} |
| %0 = vector.insert_strided_slice %a, %b {offsets = [0, 3, 0], strides = [1, 1, 1]} : vector<1x1x[2]xi32> into vector<1x4x[4]xi32> |
| return %0 : vector<1x4x[4]xi32> |
| } |
| |
| // ----- |
| |
| func.func @insert_strided_slice_scalable(%a : vector<1x1x4xi32>, %b: vector<1x4x[4]xi32>) -> vector<1x4x[4]xi32> { |
| // expected-error@+1 {{op mismatching scalable flags (at source vector idx=2)}} |
| %0 = vector.insert_strided_slice %a, %b {offsets = [0, 3, 0], strides = [1, 1, 1]} : vector<1x1x4xi32> into vector<1x4x[4]xi32> |
| return %0 : vector<1x4x[4]xi32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected offsets, sizes and strides attributes of same size}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [100], sizes = [2, 2], strides = [1, 1]} : vector<4x8x16xf32> to vector<2x2x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{expected offsets attribute of rank no greater than vector rank}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [2, 2, 2, 2], sizes = [2, 2, 2, 2], strides = [1, 1, 1, 1]} : vector<4x8x16xf32> to vector<2x2x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected offsets dimension 0 to be confined to [0, 4)}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [100], sizes = [100], strides = [100]} : vector<4x8x16xf32> to vector<100x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected sizes dimension 0 to be confined to [1, 5)}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [2], sizes = [100], strides = [100]} : vector<4x8x16xf32> to vector<100x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected strides to be confined to [1, 2)}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [2], sizes = [1], strides = [100]} : vector<4x8x16xf32> to vector<1x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice_scalable(%arg0 : vector<1x4x[4]xi32>) -> vector<1x1x[2]xi32> { |
| // expected-error@+1 {{op expected size at idx=2 to match the corresponding base size from the input vector (2 vs 4)}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [0, 3, 0], sizes = [1, 1, 2], strides = [1, 1, 1]} : vector<1x4x[4]xi32> to vector<1x1x[2]xi32> |
| return %1 : vector<1x1x[2]xi32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected strides to be confined to [1, 2)}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [2], sizes = [1], strides = [100]} : vector<4x8x16xf32> to vector<1x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected sum(offsets, sizes) dimension 0 to be confined to [1, 5)}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [2], sizes = [3], strides = [1]} : vector<4x8x16xf32> to vector<3x8x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @extract_strided_slice(%arg0: vector<4x8x16xf32>) { |
| // expected-error@+1 {{op expected result type to be 'vector<2x8x16xf32>'}} |
| %1 = vector.extract_strided_slice %arg0 {offsets = [2], sizes = [2], strides = [1]} : vector<4x8x16xf32> to vector<3x1xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, b0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, c1, c0, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{expected an indexing map for each vector operand}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, c0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, c1, c0, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{expected indexing map 0 to be a projected permutation of its inputs}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, b0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1)[s0] -> (b0, s0, c0, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{op expected indexing map 1 to have no symbols}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, b0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, c1, c0, f1)>, |
| affine_map<(b0, f0, f1, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{expected indexing map 2 to have 5 number of inputs}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, b0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, c1, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{expected indexing map 1 to have 4 number of outputs}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, b1, b2) -> (b1, b0, b2, f0)>, |
| affine_map<(b0, f0, f1, b1, b2) -> (b0, b2, b1, f1)>, |
| affine_map<(b0, f0, f1, b1, b2) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "parallel", "parallel"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{op expected at least one contracting dimension pair}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c1, b0, c0, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, c1, c0, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{invalid contracting dimension map}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, b0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (f1, c1, c0, b0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{invalid batch dimension map}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, b0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, c1, c0, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<88x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| // expected-error@+1 {{invalid accumulator/result vector shape}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<88x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(b0, f0, f1, c0, c1) -> (c0, b0, c1, f0)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, c1, c0, f1)>, |
| affine_map<(b0, f0, f1, c0, c1) -> (b0, f0, f1)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "parallel", "reduction", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<7x8x16x15xf32>, %arg1: vector<8x16x7x5xf32>, |
| %arg2: vector<8x15x5xf32>, %arg3 : vector<8x15x8x5xf32>, |
| %arg4 : index) { |
| %lhs_mask = vector.constant_mask [7, 8, 16, 15] : vector<7x8x16x15xi1> |
| %rhs_mask = vector.constant_mask [8, 16, 7, 5] : vector<8x16x7x5xi1> |
| // expected-error@+1 {{expected zero or exactly 2 vector mask operands}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2, %lhs_mask |
| : vector<7x8x16x15xf32>, vector<8x16x7x5xf32> into vector<8x15x5xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(i, j, k) -> (i, k)>, |
| affine_map<(i, j, k) -> (k, j)>, |
| affine_map<(i, j, k) -> (i, j)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<4x3xi32>, |
| %arg1: vector<3x7xf32>, |
| %arg2: vector<4x7xf32>) -> vector<4x7xf32> { |
| // expected-error@+1 {{'vector.contract' op failed to verify that lhs and rhs have same element type}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<4x3xi32>, vector<3x7xf32> into vector<4x7xf32> |
| } |
| |
| // ----- |
| |
| #contraction_accesses = [ |
| affine_map<(m, n, k) -> (m, k)>, |
| affine_map<(m, n, k) -> (k, n)>, |
| affine_map<(m, n, k) -> (n, m)> |
| ] |
| #contraction_trait = { |
| indexing_maps = #contraction_accesses, |
| iterator_types = ["parallel", "parallel", "reduction"] |
| } |
| func.func @contraction(%arg0: vector<2x1xf32>, %arg1: vector<1x3xf32>, %arg2: vector<2x3xf32>) |
| -> vector<3x2xf32> |
| { |
| // expected-error@+1 {{invalid accumulator/result vector shape, expected: 'vector<3x2xf32>'}} |
| %0 = vector.contract #contraction_trait %arg0, %arg1, %arg2 |
| : vector<2x1xf32>, vector<1x3xf32> into vector<2x3xf32> |
| return %0 : vector<2x3xf32> |
| } |
| |
| // ----- |
| |
| func.func @contract_with_dim_unused_by_lhs_and_rhs(%arg0 : vector<1x2xi32>, %arg1 : vector<2xi32>, %arg2 : vector<1xi32>) -> vector<1xi32> { |
| // expected-error@+1 {{'vector.contract' op expected all dimensions to be either a LHS or a RHS dimension}} |
| %result = vector.contract { |
| indexing_maps = [ |
| affine_map<(d0, d1, d2) -> (d0, d2)>, |
| affine_map<(d0, d1, d2) -> (d2)>, |
| affine_map<(d0, d1, d2) -> (d1)> |
| ], |
| iterator_types = ["reduction", "parallel", "reduction"], |
| kind = #vector.kind<add>} %arg0, %arg1, %arg2 : vector<1x2xi32>, vector<2xi32> into vector<1xi32> |
| return %result : vector<1xi32> |
| } |
| |
| // ----- |
| |
| func.func @contract_missing_iterator_types(%arg0: vector<1x2xi32>, %arg1: vector<2xi32>, %arg2: vector<1xi32>) -> vector<1xi32> { |
| // expected-error@+1 {{'vector.contract' expected "iterator_types" array attribute}} |
| %0 = vector.contract {} %arg0, %arg1, %arg2 : vector<1x2xi32>, vector<2xi32> into vector<1xi32> |
| return %0 : vector<1xi32> |
| } |
| |
| // ----- |
| |
| func.func @create_mask_0d_no_operands() { |
| %c1 = arith.constant 1 : index |
| // expected-error@+1 {{must specify exactly one operand for 0-D create_mask}} |
| %0 = vector.create_mask : vector<i1> |
| } |
| |
| // ----- |
| |
| func.func @create_mask_0d_many_operands() { |
| %c1 = arith.constant 1 : index |
| %c2 = arith.constant 2 : index |
| %c3 = arith.constant 3 : index |
| // expected-error@+1 {{must specify exactly one operand for 0-D create_mask}} |
| %0 = vector.create_mask %c1, %c2, %c3 : vector<i1> |
| } |
| |
| // ----- |
| |
| func.func @create_mask() { |
| %c2 = arith.constant 2 : index |
| %c3 = arith.constant 3 : index |
| // expected-error@+1 {{must specify an operand for each result vector dimension}} |
| %0 = vector.create_mask %c3, %c2 : vector<4x3x7xi1> |
| } |
| |
| |
| // ----- |
| |
| func.func @constant_mask_0d_no_attr() { |
| // expected-error@+1 {{array attr must have length 1 for 0-D vectors}} |
| %0 = vector.constant_mask [] : vector<i1> |
| } |
| |
| // ----- |
| |
| func.func @constant_mask_0d_bad_attr() { |
| // expected-error@+1 {{mask dim size must be either 0 or 1 for 0-D vectors}} |
| %0 = vector.constant_mask [2] : vector<i1> |
| } |
| |
| // ----- |
| |
| func.func @constant_mask() { |
| // expected-error@+1 {{must specify array attr of size equal vector result rank}} |
| %0 = vector.constant_mask [3, 2, 7] : vector<4x3xi1> |
| } |
| |
| // ----- |
| |
| func.func @constant_mask_out_of_bounds() { |
| // expected-error@+1 {{array attr of size out of bounds of vector result dimension size}} |
| %0 = vector.constant_mask [-1, 2] : vector<4x3xi1> |
| } |
| |
| // ----- |
| |
| func.func @constant_mask_out_of_bounds() { |
| // expected-error@+1 {{array attr of size out of bounds of vector result dimension size}} |
| %0 = vector.constant_mask [3, 4] : vector<4x3xi1> |
| } |
| |
| // ----- |
| |
| func.func @constant_mask_with_zero_mask_dim_size() { |
| // expected-error@+1 {{expected all mask dim sizes to be zeros, as a result of conjunction with zero mask dim}} |
| %0 = vector.constant_mask [0, 2] : vector<4x3xi1> |
| } |
| |
| // ----- |
| |
| func.func @constant_mask_scalable_non_zero_dim_size() { |
| // expected-error@+1 {{only supports 'none set' or 'all set' scalable dimensions}} |
| %0 = vector.constant_mask [2] : vector<[8]xi1> |
| } |
| |
| // ----- |
| |
| func.func @print_no_result(%arg0 : f32) -> i32 { |
| // expected-error@+1 {{cannot name an operation with no results}} |
| %0 = vector.print %arg0 : f32 |
| } |
| |
| // ----- |
| |
| func.func private @print_needs_vector(%arg0: tensor<8xf32>) { |
| // expected-error@+1 {{op operand #0 must be , but got 'tensor<8xf32>'}} |
| vector.print %arg0 : tensor<8xf32> |
| return |
| } |
| |
| // ----- |
| |
| func.func @cannot_print_string_with_punctuation_set() { |
| // expected-error@+1 {{`source` or `punctuation` are not set when printing strings}} |
| vector.print str "Whoops!" punctuation <comma> |
| return |
| } |
| |
| // ----- |
| |
| func.func @cannot_print_string_with_source_set(%vec: vector<[4]xf32>) { |
| // expected-error@+1 {{`source` or `punctuation` are not set when printing strings}} |
| vector.print %vec: vector<[4]xf32> str "Yay!" |
| return |
| } |
| |
| // ----- |
| |
| |
| func.func @shape_cast_wrong_element_type(%arg0 : vector<5x1x3x2xf32>) { |
| // expected-error@+1 {{'vector.shape_cast' op has different source and result element types}} |
| %0 = vector.shape_cast %arg0 : vector<5x1x3x2xf32> to vector<15x2xi32> |
| } |
| |
| // ----- |
| |
| func.func @shape_cast_wrong_num_elements(%arg0 : vector<5x1x3x2xf32>) { |
| // expected-error@+1 {{'vector.shape_cast' op has different number of elements at source (30) and result (20)}} |
| %0 = vector.shape_cast %arg0 : vector<5x1x3x2xf32> to vector<10x2xf32> |
| } |
| |
| // ----- |
| |
| func.func @shape_cast_scalability_flag_is_dropped(%arg0 : vector<15x[2]xf32>) { |
| // expected-error@+1 {{different number of scalable dims at source (1) and result (0)}} |
| %0 = vector.shape_cast %arg0 : vector<15x[2]xf32> to vector<30xf32> |
| } |
| |
| // ----- |
| |
| func.func @shape_cast_scalability_flag_is_dropped(%arg0 : vector<2x[15]x[2]xf32>) { |
| // expected-error@+1 {{different number of scalable dims at source (2) and result (1)}} |
| %0 = vector.shape_cast %arg0 : vector<2x[15]x[2]xf32> to vector<30x[2]xf32> |
| } |
| |
| // ----- |
| |
| func.func @bitcast_not_vector(%arg0 : vector<5x1x3x2xf32>) { |
| // expected-error@+1 {{'vector.bitcast' invalid kind of type specified: expected builtin.vector, but found 'f32'}} |
| %0 = vector.bitcast %arg0 : vector<5x1x3x2xf32> to f32 |
| } |
| |
| // ----- |
| |
| func.func @bitcast_rank_mismatch_to_0d(%arg0 : vector<1xf32>) { |
| // expected-error@+1 {{op failed to verify that all of {source, result} have same rank}} |
| %0 = vector.bitcast %arg0 : vector<1xf32> to vector<f32> |
| } |
| |
| // ----- |
| |
| func.func @bitcast_rank_mismatch_from_0d(%arg0 : vector<f32>) { |
| // expected-error@+1 {{op failed to verify that all of {source, result} have same rank}} |
| %0 = vector.bitcast %arg0 : vector<f32> to vector<1xf32> |
| } |
| |
| // ----- |
| |
| func.func @bitcast_rank_mismatch(%arg0 : vector<5x1x3x2xf32>) { |
| // expected-error@+1 {{op failed to verify that all of {source, result} have same rank}} |
| %0 = vector.bitcast %arg0 : vector<5x1x3x2xf32> to vector<5x3x2xf32> |
| } |
| |
| // ----- |
| |
| func.func @bitcast_shape_mismatch(%arg0 : vector<5x1x3x2xf32>) { |
| // expected-error@+1 {{op dimension size mismatch}} |
| %0 = vector.bitcast %arg0 : vector<5x1x3x2xf32> to vector<5x2x3x2xf32> |
| } |
| |
| // ----- |
| |
| func.func @bitcast_sizemismatch(%arg0 : vector<5x1x3x2xf32>) { |
| // expected-error@+1 {{op source/result bitwidth of the minor 1-D vectors must be equal}} |
| %0 = vector.bitcast %arg0 : vector<5x1x3x2xf32> to vector<5x1x3x3xf16> |
| } |
| |
| // ----- |
| |
| func.func @reduce_unknown_kind(%arg0: vector<16xf32>) -> f32 { |
| // expected-error@+2 {{custom op 'vector.reduction' failed to parse Vector_CombiningKindAttr parameter 'value' which is to be a `::mlir::vector::CombiningKind`}} |
| // expected-error@+1 {{custom op 'vector.reduction' expected ::mlir::vector::CombiningKind to be one of: }} |
| %0 = vector.reduction <joho>, %arg0 : vector<16xf32> into f32 |
| } |
| |
| // ----- |
| |
| func.func @reduce_elt_type_mismatch(%arg0: vector<16xf32>) -> i32 { |
| // expected-error@+1 {{'vector.reduction' op failed to verify that source operand and result have same element type}} |
| %0 = vector.reduction <add>, %arg0 : vector<16xf32> into i32 |
| } |
| |
| // ----- |
| |
| func.func @reduce_unsupported_attr(%arg0: vector<16xf32>) -> i32 { |
| // expected-error@+1 {{expected '<'}} |
| %0 = vector.reduction 1234, %arg0 : vector<16xf32> into i32 |
| } |
| |
| // ----- |
| |
| func.func @reduce_unsupported_third_argument(%arg0: vector<16xf32>, %arg1: f32) -> f32 { |
| // expected-error@+1 {{expected ':'}} |
| %0 = vector.reduction <add>, %arg0, %arg1, %arg1 : vector<16xf32> into f32 |
| } |
| |
| // ----- |
| |
| func.func @reduce_unsupported_rank(%arg0: vector<4x16xf32>) -> f32 { |
| // expected-error@+1 {{'vector.reduction' op unsupported reduction rank: 2}} |
| %0 = vector.reduction <add>, %arg0 : vector<4x16xf32> into f32 |
| } |
| |
| // ----- |
| |
| func.func @multi_reduce_invalid_type(%arg0: vector<4x16xf32>, %acc: vector<16xf32>) -> f32 { |
| // expected-error@+1 {{'vector.multi_reduction' op destination type 'vector<16xf32>' is incompatible with source type 'vector<4x16xf32>'}} |
| %0 = vector.multi_reduction <mul>, %arg0, %acc [1] : vector<4x16xf32> to vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @transpose_rank_mismatch_0d(%arg0: vector<f32>) { |
| // expected-error@+1 {{'vector.transpose' op vector result rank mismatch: 1}} |
| %0 = vector.transpose %arg0, [] : vector<f32> to vector<100xf32> |
| } |
| |
| // ----- |
| |
| func.func @transpose_rank_mismatch(%arg0: vector<4x16x11xf32>) { |
| // expected-error@+1 {{'vector.transpose' op vector result rank mismatch: 1}} |
| %0 = vector.transpose %arg0, [2, 1, 0] : vector<4x16x11xf32> to vector<100xf32> |
| } |
| |
| // ----- |
| |
| func.func @transpose_length_mismatch_0d(%arg0: vector<f32>) { |
| // expected-error@+1 {{'vector.transpose' op transposition length mismatch: 1}} |
| %0 = vector.transpose %arg0, [1] : vector<f32> to vector<f32> |
| } |
| |
| // ----- |
| |
| func.func @transpose_length_mismatch(%arg0: vector<4x4xf32>) { |
| // expected-error@+1 {{'vector.transpose' op transposition length mismatch: 3}} |
| %0 = vector.transpose %arg0, [2, 0, 1] : vector<4x4xf32> to vector<4x4xf32> |
| } |
| |
| // ----- |
| |
| func.func @transpose_index_oob(%arg0: vector<4x4xf32>) { |
| // expected-error@+1 {{'vector.transpose' op transposition index out of range: 2}} |
| %0 = vector.transpose %arg0, [2, 0] : vector<4x4xf32> to vector<4x4xf32> |
| } |
| |
| // ----- |
| |
| func.func @transpose_index_dup(%arg0: vector<4x4xf32>) { |
| // expected-error@+1 {{'vector.transpose' op duplicate position index: 0}} |
| %0 = vector.transpose %arg0, [0, 0] : vector<4x4xf32> to vector<4x4xf32> |
| } |
| |
| // ----- |
| |
| func.func @transpose_dim_size_mismatch(%arg0: vector<11x7x3x2xi32>) { |
| // expected-error@+1 {{'vector.transpose' op dimension size mismatch at: 0}} |
| %0 = vector.transpose %arg0, [3, 0, 1, 2] : vector<11x7x3x2xi32> to vector<2x3x7x11xi32> |
| } |
| |
| // ----- |
| |
| func.func @type_cast_layout(%arg0: memref<4x3xf32, affine_map<(d0, d1)[s0, s1, s2] -> (d0 * s0 + d1 * s1 + s2)>>) { |
| // expected-error@+1 {{expects operand to be a memref with identity layout}} |
| %0 = vector.type_cast %arg0: memref<4x3xf32, affine_map<(d0, d1)[s0, s1, s2] -> (d0 * s0 + d1 * s1 + s2)>> to memref<vector<4x3xf32>> |
| } |
| |
| // ----- |
| |
| func.func @store_unsupported_layout(%memref : memref<200x100xf32, affine_map<(d0, d1) -> (200*d0 + 2*d1)>>, |
| %i : index, %j : index, %value : vector<8xf32>) { |
| // expected-error@+1 {{'vector.store' op most minor memref dim must have unit stride}} |
| vector.store %value, %memref[%i, %j] : memref<200x100xf32, affine_map<(d0, d1) -> (200*d0 + 2*d1)>>, |
| vector<8xf32> |
| return |
| } |
| |
| // ----- |
| |
| func.func @vector_memref_mismatch(%memref : memref<200x100xvector<4xf32>>, %i : index, |
| %j : index, %value : vector<8xf32>) { |
| // expected-error@+1 {{'vector.store' op base memref and valueToStore vector types should match}} |
| vector.store %value, %memref[%i, %j] : memref<200x100xvector<4xf32>>, vector<8xf32> |
| } |
| |
| // ----- |
| |
| func.func @store_base_type_mismatch(%base : memref<?xf64>, %value : vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.store' op base and valueToStore element type should match}} |
| vector.store %value, %base[%c0] : memref<?xf64>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @store_memref_index_mismatch(%base : memref<?xf32>, %value : vector<16xf32>) { |
| // expected-error@+1 {{'vector.store' op requires 1 indices}} |
| vector.store %value, %base[] : memref<?xf32>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @maskedload_base_type_mismatch(%base: memref<?xf64>, %mask: vector<16xi1>, %pass: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.maskedload' op base and result element type should match}} |
| %0 = vector.maskedload %base[%c0], %mask, %pass : memref<?xf64>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @maskedload_dim_mask_mismatch(%base: memref<?xf32>, %mask: vector<15xi1>, %pass: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.maskedload' op expected result shape to match mask shape}} |
| %0 = vector.maskedload %base[%c0], %mask, %pass : memref<?xf32>, vector<15xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @maskedload_pass_thru_type_mask_mismatch(%base: memref<?xf32>, %mask: vector<16xi1>, %pass: vector<16xi32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.maskedload' op expected pass_thru of same type as result type}} |
| %0 = vector.maskedload %base[%c0], %mask, %pass : memref<?xf32>, vector<16xi1>, vector<16xi32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @maskedload_memref_mismatch(%base: memref<?xf32>, %mask: vector<16xi1>, %pass: vector<16xf32>) { |
| // expected-error@+1 {{'vector.maskedload' op requires 1 indices}} |
| %0 = vector.maskedload %base[], %mask, %pass : memref<?xf32>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @maskedstore_base_type_mismatch(%base: memref<?xf64>, %mask: vector<16xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.maskedstore' op base and valueToStore element type should match}} |
| vector.maskedstore %base[%c0], %mask, %value : memref<?xf64>, vector<16xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @maskedstore_dim_mask_mismatch(%base: memref<?xf32>, %mask: vector<15xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.maskedstore' op expected valueToStore shape to match mask shape}} |
| vector.maskedstore %base[%c0], %mask, %value : memref<?xf32>, vector<15xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @maskedstore_memref_mismatch(%base: memref<?xf32>, %mask: vector<16xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.maskedstore' op requires 1 indices}} |
| vector.maskedstore %base[%c0, %c0], %mask, %value : memref<?xf32>, vector<16xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @gather_from_vector(%base: vector<16xf32>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.gather' op operand #0 must be Tensor or MemRef of any type values, but got 'vector<16xf32>'}} |
| %0 = vector.gather %base[%c0][%indices], %mask, %pass_thru |
| : vector<16xf32>, vector<16xi32>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @gather_base_type_mismatch(%base: memref<?xf64>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.gather' op base and result element type should match}} |
| %0 = vector.gather %base[%c0][%indices], %mask, %pass_thru |
| : memref<?xf64>, vector<16xi32>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @gather_memref_mismatch(%base: memref<?x?xf64>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %pass_thru: vector<16xf64>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.gather' op requires 2 indices}} |
| %0 = vector.gather %base[%c0][%indices], %mask, %pass_thru |
| : memref<?x?xf64>, vector<16xi32>, vector<16xi1>, vector<16xf64> into vector<16xf64> |
| } |
| |
| // ----- |
| |
| func.func @gather_rank_mismatch(%base: memref<?xf32>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.gather' op expected result dim to match indices dim}} |
| %0 = vector.gather %base[%c0][%indices], %mask, %pass_thru |
| : memref<?xf32>, vector<16xi32>, vector<16xi1>, vector<16xf32> into vector<2x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @gather_dim_indices_mismatch(%base: memref<?xf32>, %indices: vector<17xi32>, |
| %mask: vector<16xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.gather' op expected result dim to match indices dim}} |
| %0 = vector.gather %base[%c0][%indices], %mask, %pass_thru |
| : memref<?xf32>, vector<17xi32>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @gather_dim_mask_mismatch(%base: memref<?xf32>, %indices: vector<16xi32>, |
| %mask: vector<17xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.gather' op expected result dim to match mask dim}} |
| %0 = vector.gather %base[%c0][%indices], %mask, %pass_thru |
| : memref<?xf32>, vector<16xi32>, vector<17xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @gather_pass_thru_type_mismatch(%base: memref<?xf32>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %pass_thru: vector<16xf64>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.gather' op expected pass_thru of same type as result type}} |
| %0 = vector.gather %base[%c0][%indices], %mask, %pass_thru |
| : memref<?xf32>, vector<16xi32>, vector<16xi1>, vector<16xf64> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @scatter_to_vector(%base: vector<16xf32>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+2 {{custom op 'vector.scatter' invalid kind of type specified}} |
| vector.scatter %base[%c0][%indices], %mask, %pass_thru |
| : vector<16xf32>, vector<16xi32>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| |
| func.func @scatter_base_type_mismatch(%base: memref<?xf64>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.scatter' op base and valueToStore element type should match}} |
| vector.scatter %base[%c0][%indices], %mask, %value |
| : memref<?xf64>, vector<16xi32>, vector<16xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @scatter_memref_mismatch(%base: memref<?x?xf64>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %value: vector<16xf64>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.scatter' op requires 2 indices}} |
| vector.scatter %base[%c0][%indices], %mask, %value |
| : memref<?x?xf64>, vector<16xi32>, vector<16xi1>, vector<16xf64> |
| } |
| |
| // ----- |
| |
| func.func @scatter_rank_mismatch(%base: memref<?xf32>, %indices: vector<16xi32>, |
| %mask: vector<16xi1>, %value: vector<2x16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.scatter' op expected valueToStore dim to match indices dim}} |
| vector.scatter %base[%c0][%indices], %mask, %value |
| : memref<?xf32>, vector<16xi32>, vector<16xi1>, vector<2x16xf32> |
| } |
| |
| // ----- |
| |
| func.func @scatter_dim_indices_mismatch(%base: memref<?xf32>, %indices: vector<17xi32>, |
| %mask: vector<16xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.scatter' op expected valueToStore dim to match indices dim}} |
| vector.scatter %base[%c0][%indices], %mask, %value |
| : memref<?xf32>, vector<17xi32>, vector<16xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @scatter_dim_mask_mismatch(%base: memref<?xf32>, %indices: vector<16xi32>, |
| %mask: vector<17xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.scatter' op expected valueToStore dim to match mask dim}} |
| vector.scatter %base[%c0][%indices], %mask, %value |
| : memref<?xf32>, vector<16xi32>, vector<17xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @expand_base_type_mismatch(%base: memref<?xf64>, %mask: vector<16xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.expandload' op base and result element type should match}} |
| %0 = vector.expandload %base[%c0], %mask, %pass_thru : memref<?xf64>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @expand_base_scalable(%base: memref<?xf32>, %mask: vector<[16]xi1>, %pass_thru: vector<[16]xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.expandload' op operand #2 must be fixed-length vector of 1-bit signless integer values, but got 'vector<[16]xi1>}} |
| %0 = vector.expandload %base[%c0], %mask, %pass_thru : memref<?xf32>, vector<[16]xi1>, vector<[16]xf32> into vector<[16]xf32> |
| } |
| |
| // ----- |
| |
| func.func @expand_dim_mask_mismatch(%base: memref<?xf32>, %mask: vector<17xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.expandload' op expected result dim to match mask dim}} |
| %0 = vector.expandload %base[%c0], %mask, %pass_thru : memref<?xf32>, vector<17xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @expand_pass_thru_mismatch(%base: memref<?xf32>, %mask: vector<16xi1>, %pass_thru: vector<17xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.expandload' op expected pass_thru of same type as result type}} |
| %0 = vector.expandload %base[%c0], %mask, %pass_thru : memref<?xf32>, vector<16xi1>, vector<17xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @expand_memref_mismatch(%base: memref<?x?xf32>, %mask: vector<16xi1>, %pass_thru: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.expandload' op requires 2 indices}} |
| %0 = vector.expandload %base[%c0], %mask, %pass_thru : memref<?x?xf32>, vector<16xi1>, vector<16xf32> into vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @compress_base_type_mismatch(%base: memref<?xf64>, %mask: vector<16xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.compressstore' op base and valueToStore element type should match}} |
| vector.compressstore %base[%c0], %mask, %value : memref<?xf64>, vector<16xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @compress_scalable(%base: memref<?xf32>, %mask: vector<[16]xi1>, %value: vector<[16]xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.compressstore' op operand #2 must be fixed-length vector of 1-bit signless integer values, but got 'vector<[16]xi1>}} |
| vector.compressstore %base[%c0], %mask, %value : memref<?xf32>, vector<[16]xi1>, vector<[16]xf32> |
| } |
| |
| // ----- |
| |
| func.func @compress_dim_mask_mismatch(%base: memref<?xf32>, %mask: vector<17xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.compressstore' op expected valueToStore dim to match mask dim}} |
| vector.compressstore %base[%c0], %mask, %value : memref<?xf32>, vector<17xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @compress_memref_mismatch(%base: memref<?x?xf32>, %mask: vector<16xi1>, %value: vector<16xf32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error@+1 {{'vector.compressstore' op requires 2 indices}} |
| vector.compressstore %base[%c0, %c0, %c0], %mask, %value : memref<?x?xf32>, vector<16xi1>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @scan_reduction_dim_constraint(%arg0: vector<2x3xi32>, %arg1: vector<3xi32>) -> vector<3xi32> { |
| // expected-error@+1 {{'vector.scan' op reduction dimension 5 has to be less than 2}} |
| %0:2 = vector.scan <add>, %arg0, %arg1 {inclusive = true, reduction_dim = 5} : |
| vector<2x3xi32>, vector<3xi32> |
| return %0#1 : vector<3xi32> |
| } |
| |
| // ----- |
| |
| func.func @scan_ival_rank_constraint(%arg0: vector<2x3xi32>, %arg1: vector<1x3xi32>) -> vector<1x3xi32> { |
| // expected-error@+1 {{initial value rank 2 has to be equal to 1}} |
| %0:2 = vector.scan <add>, %arg0, %arg1 {inclusive = true, reduction_dim = 0} : |
| vector<2x3xi32>, vector<1x3xi32> |
| return %0#1 : vector<1x3xi32> |
| } |
| |
| // ----- |
| |
| func.func @scan_incompatible_shapes(%arg0: vector<2x3xi32>, %arg1: vector<5xi32>) -> vector<2x3xi32> { |
| // expected-error@+1 {{incompatible input/initial value shapes}} |
| %0:2 = vector.scan <add>, %arg0, %arg1 {inclusive = true, reduction_dim = 0} : |
| vector<2x3xi32>, vector<5xi32> |
| return %0#0 : vector<2x3xi32> |
| } |
| |
| // ----- |
| |
| func.func @scan_unsupported_kind(%arg0: vector<2x3xf32>, %arg1: vector<3xf32>) -> vector<2x3xf32> { |
| // expected-error@+1 {{'vector.scan' op unsupported reduction type 'f32' for kind 'xor'}} |
| %0:2 = vector.scan <xor>, %arg0, %arg1 {inclusive = true, reduction_dim = 0} : |
| vector<2x3xf32>, vector<3xf32> |
| return %0#0 : vector<2x3xf32> |
| } |
| |
| |
| // ----- |
| |
| func.func @vector_mask_multiple_ops(%t0: tensor<?xf32>, %t1: tensor<?xf32>, %idx: index, %val: vector<16xf32>, %m0: vector<16xi1>) { |
| %ft0 = arith.constant 0.0 : f32 |
| // expected-error@+1 {{'vector.mask' op expects only one operation to mask}} |
| vector.mask %m0 { |
| vector.transfer_write %val, %t0[%idx] : vector<16xf32>, tensor<?xf32> |
| vector.transfer_write %val, %t1[%idx] : vector<16xf32>, tensor<?xf32> |
| } : vector<16xi1> |
| return |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_shape_mismatch(%a: vector<8xi32>, %m0: vector<16xi1>) -> i32 { |
| // expected-error@+1 {{'vector.mask' op expects a 'vector<8xi1>' mask for the maskable operation}} |
| %0 = vector.mask %m0 { vector.reduction <add>, %a : vector<8xi32> into i32 } : vector<16xi1> -> i32 |
| return %0 : i32 |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_passthru_type_mismatch(%t0: tensor<f32>, %m0: vector<i1>) -> vector<f32> { |
| %ft0 = arith.constant 0.0 : f32 |
| // expected-error@+1 {{'vector.mask' op operand #0 must be vector of 1-bit signless integer values, but got 'vector<i1>'}} |
| %0 = vector.mask %m0 { vector.transfer_read %t0[], %ft0 : tensor<f32>, vector<f32> } : vector<i1> -> vector<f32> |
| return %0 : vector<f32> |
| } |
| |
| // ----- |
| |
| // expected-note@+1 {{prior use here}} |
| func.func @vector_mask_passthru_type_mismatch(%t0: tensor<?xf32>, %idx: index, %m0: vector<16xi1>, %pt0: vector<16xi32>) -> vector<16xf32> { |
| %ft0 = arith.constant 0.0 : f32 |
| // expected-error@+1 {{use of value '%pt0' expects different type than prior uses: 'vector<16xf32>' vs 'vector<16xi32>'}} |
| %0 = vector.mask %m0, %pt0 { vector.transfer_read %t0[%idx], %ft0 : tensor<?xf32>, vector<16xf32> } : vector<16xi1> -> vector<16xf32> |
| return %0 : vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_passthru_no_return(%val: vector<16xf32>, %t0: tensor<?xf32>, %idx: index, %m0: vector<16xi1>, %pt0: vector<16xf32>) { |
| // expected-error@+1 {{'vector.mask' op expects result type to match maskable operation result type}} |
| vector.mask %m0, %pt0 { vector.transfer_write %val, %t0[%idx] : vector<16xf32>, tensor<?xf32> } : vector<16xi1> -> vector<16xf32> |
| return |
| } |
| // ----- |
| |
| func.func @vector_mask_non_maskable_op(%a : vector<3x4xf32>) -> vector<3x4xf32> { |
| %m0 = vector.constant_mask [2, 2] : vector<3x4xi1> |
| // expected-error@+1 {{'vector.mask' op expects a MaskableOpInterface within the mask region}} |
| %0 = vector.mask %m0 { arith.addf %a, %a : vector<3x4xf32> } : vector<3x4xi1> -> vector<3x4xf32> |
| return %0 : vector<3x4xf32> |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_0d_mask(%arg0: tensor<2x4xi32>, |
| %idx0: index, %idx1: index, |
| %m0: vector<i1>) -> vector<1x1x4xi32> { |
| %cst = arith.constant 0 : i32 |
| // expected-error@+1 {{'vector.mask' op operand #0 must be vector of 1-bit signless integer values, but got 'vector<i1>'}} |
| %res = vector.mask %m0 { |
| %0 = vector.transfer_read %arg0[%idx0, %idx1], %cst {permutation_map = affine_map<(d0, d1) -> (0, 0, 0)>} |
| : tensor<2x4xi32>, vector<1x1x4xi32> |
| vector.yield %0 : vector<1x1x4xi32> |
| } : vector<i1> -> vector<1x1x4xi32> |
| return %res : vector<1x1x4xi32> |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_empty_passthru_no_return_type(%mask : vector<8xi1>, |
| %passthru : vector<8xi32>) { |
| // expected-error@+1 {{'vector.mask' expects a result if passthru operand is provided}} |
| vector.mask %mask, %passthru { } : vector<8xi1> |
| return |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_non_empty_external_return(%t: tensor<?xf32>, %idx: index, |
| %m: vector<16xi1>, %ext: vector<16xf32>) -> vector<16xf32> { |
| %ft0 = arith.constant 0.0 : f32 |
| // expected-error@+1 {{'vector.mask' op expects all the results from the MaskableOpInterface to match all the values returned by the terminator}} |
| %0 = vector.mask %m { |
| %1 =vector.transfer_read %t[%idx], %ft0 : tensor<?xf32>, vector<16xf32> |
| vector.yield %ext : vector<16xf32> |
| } : vector<16xi1> -> vector<16xf32> |
| |
| return %0 : vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_empty_passthru_empty_return_type(%mask : vector<8xi1>, |
| %passthru : vector<8xi32>) { |
| // expected-error@+1 {{'vector.mask' expects a result if passthru operand is provided}} |
| vector.mask %mask, %passthru { } : vector<8xi1> -> () |
| return |
| } |
| |
| // ----- |
| |
| func.func @vector_mask_non_empty_mixed_return(%t: tensor<?xf32>, %idx: index, |
| %m: vector<16xi1>, %ext: vector<16xf32>) -> (vector<16xf32>, vector<16xf32>) { |
| %ft0 = arith.constant 0.0 : f32 |
| // expected-error@+1 {{'vector.mask' op expects number of results to match maskable operation number of results}} |
| %0:2 = vector.mask %m { |
| %1 =vector.transfer_read %t[%idx], %ft0 : tensor<?xf32>, vector<16xf32> |
| vector.yield %1, %ext : vector<16xf32>, vector<16xf32> |
| } : vector<16xi1> -> (vector<16xf32>, vector<16xf32>) |
| |
| return %0#0, %0#1 : vector<16xf32>, vector<16xf32> |
| } |
| |
| // ----- |
| |
| func.func @vector_scalable_insert_unaligned(%subv: vector<4xi32>, %vec: vector<[16]xi32>) { |
| // expected-error@+1 {{op failed to verify that position is a multiple of the source length.}} |
| %0 = vector.scalable.insert %subv, %vec[2] : vector<4xi32> into vector<[16]xi32> |
| } |
| |
| // ----- |
| |
| func.func @vector_scalable_extract_unaligned(%vec: vector<[16]xf32>) { |
| // expected-error@+1 {{op failed to verify that position is a multiple of the result length.}} |
| %0 = vector.scalable.extract %vec[5] : vector<4xf32> from vector<[16]xf32> |
| } |
| |
| // ----- |
| |
| func.func @integer_vector_contract(%arg0: vector<16x32xsi8>, %arg1: vector<32x16xsi8>, %arg2: vector<16x16xsi32>) -> vector<16x16xsi32> { |
| // expected-error@+1 {{op only supports signless integer types}} |
| %0 = vector.contract { |
| indexing_maps = [affine_map<(d0, d1, d2) -> (d0, d2)>, affine_map<(d0, d1, d2) -> (d2, d1)>, affine_map<(d0, d1, d2) -> (d0, d1)>], |
| iterator_types = ["parallel", "parallel", "reduction"], kind = #vector.kind<add> |
| } %arg0, %arg1, %arg2 : vector<16x32xsi8>, vector<32x16xsi8> into vector<16x16xsi32> |
| return %0: vector<16x16xsi32> |
| } |
| |
| // ----- |
| |
| func.func @invalid_outerproduct(%src : memref<?xf32>) { |
| %idx = arith.constant 0 : index |
| %0 = vector.load %src[%idx] : memref<?xf32>, vector<[4]xf32> |
| %1 = vector.load %src[%idx] : memref<?xf32>, vector<4xf32> |
| |
| // expected-error @+1 {{expected either both or only #2 operand dim to be scalable}} |
| %op = vector.outerproduct %0, %1 : vector<[4]xf32>, vector<4xf32> |
| |
| return |
| } |
| |
| // ----- |
| |
| func.func @invalid_outerproduct1(%src : memref<?xf32>, %lhs : vector<[4]x[4]xf32>, %rhs : vector<[4]xf32>) { |
| %idx = arith.constant 0 : index |
| |
| // expected-error @+1 {{'vector.outerproduct' op expected 1-d vector for operand #1}} |
| %op = vector.outerproduct %lhs, %rhs : vector<[4]x[4]xf32>, vector<[4]xf32> |
| } |
| |
| // ----- |
| |
| func.func @deinterleave_zero_dim_fail(%vec : vector<f32>) { |
| // expected-error @+1 {{'vector.deinterleave' op operand #0 must be vector of any type values, but got 'vector<f32>}} |
| %0, %1 = vector.deinterleave %vec : vector<f32> -> vector<f32> |
| return |
| } |
| |
| // ----- |
| |
| func.func @deinterleave_one_dim_fail(%vec : vector<1xf32>) { |
| // expected-error @+1 {{'vector.deinterleave' op failed to verify that the trailing dimension of the source vector has an even number of elements}} |
| %0, %1 = vector.deinterleave %vec : vector<1xf32> -> vector<1xf32> |
| return |
| } |
| |
| // ----- |
| |
| func.func @deinterleave_oversized_output_fail(%vec : vector<4xf32>) { |
| // expected-error @+1 {{'vector.deinterleave' op failed to verify that the trailing dimension of the results is half the width of source trailing dimension}} |
| %0, %1 = "vector.deinterleave" (%vec) : (vector<4xf32>) -> (vector<8xf32>, vector<8xf32>) |
| return |
| } |
| |
| // ----- |
| |
| func.func @deinterleave_output_dim_size_mismatch(%vec : vector<4xf32>) { |
| // expected-error @+1 {{'vector.deinterleave' op failed to verify that the trailing dimension of the results is half the width of source trailing dimension}} |
| %0, %1 = "vector.deinterleave" (%vec) : (vector<4xf32>) -> (vector<4xf32>, vector<2xf32>) |
| return |
| } |
| |
| // ----- |
| |
| func.func @deinterleave_n_dim_rank_fail(%vec : vector<2x3x4xf32>) { |
| // expected-error @+1 {{'vector.deinterleave' op failed to verify that the trailing dimension of the results is half the width of source trailing dimension}} |
| %0, %1 = "vector.deinterleave" (%vec) : (vector<2x3x4xf32>) -> (vector<2x3x4xf32>, vector<2x3x2xf32>) |
| return |
| } |
| |
| // ----- |
| |
| func.func @deinterleave_scalable_dim_size_fail(%vec : vector<2x[4]xf32>) { |
| // expected-error @+1 {{'vector.deinterleave' op failed to verify that all of {res1, res2} have same type}} |
| %0, %1 = "vector.deinterleave" (%vec) : (vector<2x[4]xf32>) -> (vector<2x[2]xf32>, vector<2x[1]xf32>) |
| return |
| } |
| |
| // ----- |
| |
| func.func @deinterleave_scalable_rank_fail(%vec : vector<2x[4]xf32>) { |
| // expected-error @+1 {{'vector.deinterleave' op failed to verify that all of {res1, res2} have same type}} |
| %0, %1 = "vector.deinterleave" (%vec) : (vector<2x[4]xf32>) -> (vector<2x[2]xf32>, vector<[2]xf32>) |
| return |
| } |
| |
| // ----- |
| |
| func.func @to_elements_wrong_num_results(%a: vector<1x1x2xf32>) { |
| // expected-error @+1 {{operation defines 2 results but was provided 4 to bind}} |
| %0:4 = vector.to_elements %a : vector<1x1x2xf32> |
| return |
| } |
| |
| // ----- |
| |
| func.func @to_elements_wrong_result_type(%a: vector<2xf32>) -> i32 { |
| // expected-error @+3 {{use of value '%0' expects different type than prior uses: 'i32'}} |
| // expected-note @+1 {{prior use here}} |
| %0:2 = vector.to_elements %a : vector<2xf32> |
| return %0#0 : i32 |
| } |
| |
| // ----- |
| |
| func.func @from_elements_wrong_num_operands(%a: f32) { |
| // expected-error @+1 {{'vector.from_elements' number of operands and types do not match: got 1 operands and 2 types}} |
| vector.from_elements %a : vector<2xf32> |
| return |
| } |
| |
| // ----- |
| |
| // expected-note @+1 {{prior use here}} |
| func.func @from_elements_wrong_operand_type(%a: f32, %b: i32) { |
| // expected-error @+1 {{use of value '%b' expects different type than prior uses: 'f32' vs 'i32'}} |
| vector.from_elements %a, %b : vector<2xf32> |
| return |
| } |
| // ----- |
| |
| func.func @invalid_from_elements_scalable(%a: f32, %b: i32) { |
| // expected-error @+1 {{'dest' must be fixed-length vector of any type values, but got 'vector<[2]xf32>'}} |
| vector.from_elements %a, %b : vector<[2]xf32> |
| return |
| } |
| |
| // ----- |
| |
| func.func @invalid_step_0d() { |
| // expected-error @+1 {{vector.step' op result #0 must be vector of index values of ranks 1, but got 'vector<f32>'}} |
| vector.step : vector<f32> |
| return |
| } |
| |
| // ----- |
| |
| func.func @invalid_step_2d() { |
| // expected-error @+1 {{vector.step' op result #0 must be vector of index values of ranks 1, but got 'vector<2x4xf32>'}} |
| vector.step : vector<2x4xf32> |
| return |
| } |
| |
| // ----- |
| |
| //===----------------------------------------------------------------------===// |
| // vector.splat |
| //===----------------------------------------------------------------------===// |
| |
| // ----- |
| |
| func.func @vector_splat_invalid_result(%v : f32) { |
| // expected-error@+1 {{invalid kind of type specified: expected builtin.vector, but found 'memref<8xf32>'}} |
| vector.splat %v : memref<8xf32> |
| return |
| } |
| |
| // ----- |
| |
| // expected-note @+1 {{prior use here}} |
| func.func @vector_splat_type_mismatch(%a: f32) { |
| // expected-error @+1 {{expects different type than prior uses: 'i32' vs 'f32'}} |
| %0 = vector.splat %a : vector<1xi32> |
| return |
| } |
| |
| // ----- |
| |
| //===----------------------------------------------------------------------===// |
| // vector.load |
| //===----------------------------------------------------------------------===// |
| |
| func.func @vector_load(%src : memref<?xi8>) { |
| %c0 = arith.constant 0 : index |
| // expected-error @+1 {{'vector.load' op destination memref has lower rank than the result vector}} |
| %0 = vector.load %src[%c0] : memref<?xi8>, vector<16x16xi8> |
| return |
| } |
| |
| // ----- |
| |
| func.func @invalid_load_alignment(%memref: memref<4xi32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error @below {{'vector.load' op attribute 'alignment' failed to satisfy constraint: 64-bit signless integer attribute whose value is positive and whose value is a power of two > 0}} |
| %val = vector.load %memref[%c0] { alignment = -1 } : memref<4xi32>, vector<4xi32> |
| return |
| } |
| |
| // ----- |
| |
| //===----------------------------------------------------------------------===// |
| // vector.store |
| //===----------------------------------------------------------------------===// |
| |
| func.func @vector_store(%dest : memref<?xi8>, %vec : vector<16x16xi8>) { |
| %c0 = arith.constant 0 : index |
| // expected-error @+1 {{'vector.store' op source memref has lower rank than the vector to store}} |
| vector.store %vec, %dest[%c0] : memref<?xi8>, vector<16x16xi8> |
| return |
| } |
| |
| // ----- |
| |
| func.func @invalid_store_alignment(%memref: memref<4xi32>, %val: vector<4xi32>) { |
| %c0 = arith.constant 0 : index |
| // expected-error @below {{'vector.store' op attribute 'alignment' failed to satisfy constraint: 64-bit signless integer attribute whose value is positive and whose value is a power of two > 0}} |
| vector.store %val, %memref[%c0] { alignment = 3 } : memref<4xi32>, vector<4xi32> |
| return |
| } |