Index: lib/Target/WebAssembly/MCTargetDesc/WebAssemblyMCCodeEmitter.cpp =================================================================== --- lib/Target/WebAssembly/MCTargetDesc/WebAssemblyMCCodeEmitter.cpp +++ lib/Target/WebAssembly/MCTargetDesc/WebAssemblyMCCodeEmitter.cpp @@ -67,7 +67,8 @@ OS << uint8_t(Binary); } else { assert(Binary <= UINT16_MAX && "Several-byte opcodes not supported yet"); - OS << uint8_t(Binary >> 8) << uint8_t(Binary); + OS << uint8_t(Binary >> 8); + encodeULEB128(uint8_t(Binary), OS); } // For br_table instructions, encode the size of the table. In the MCInst, Index: lib/Target/WebAssembly/WebAssemblyInstrSIMD.td =================================================================== --- lib/Target/WebAssembly/WebAssemblyInstrSIMD.td +++ lib/Target/WebAssembly/WebAssemblyInstrSIMD.td @@ -36,6 +36,58 @@ foreach SIZE = [2, 4, 8, 16, 32] in def LaneIdx#SIZE : ImmLeaf; +//===----------------------------------------------------------------------===// +// Load and store +//===----------------------------------------------------------------------===// + +// Load: v128.load +multiclass SIMDLoad { + let mayLoad = 1 in + defm LOAD_#vec_t : + SIMD_I<(outs V128:$dst), (ins P2Align:$align, offset32_op:$off, I32:$addr), + (outs), (ins P2Align:$align, offset32_op:$off), [], + "v128.load\t$dst, ${off}(${addr})$align", + "v128.load\t$off$align", 0>; +} + +foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in { +defm "" : SIMDLoad; + +// Def load and store patterns from WebAssemblyInstrMemory.td for vector types +def : LoadPatNoOffset("LOAD_"#vec_t)>; +def : LoadPatImmOff("LOAD_"#vec_t)>; +def : LoadPatImmOff("LOAD_"#vec_t)>; +def : LoadPatGlobalAddr("LOAD_"#vec_t)>; +def : LoadPatExternalSym("LOAD_"#vec_t)>; +def : LoadPatOffsetOnly("LOAD_"#vec_t)>; +def : LoadPatGlobalAddrOffOnly("LOAD_"#vec_t)>; +def : LoadPatExternSymOffOnly("LOAD_"#vec_t)>; +} + +// Store: v128.store +multiclass SIMDStore { + let mayStore = 1 in + defm STORE_#vec_t : + SIMD_I<(outs), (ins P2Align:$align, offset32_op:$off, I32:$addr, V128:$vec), + (outs), (ins P2Align:$align, offset32_op:$off), [], + "v128.store\t${off}(${addr})$align, $vec", + "v128.store\t$off$align", 1>; +} + +foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in { +defm "" : SIMDStore; + +// Def load and store patterns from WebAssemblyInstrMemory.td for vector types +def : StorePatNoOffset("STORE_"#vec_t)>; +def : StorePatImmOff("STORE_"#vec_t)>; +def : StorePatImmOff("STORE_"#vec_t)>; +def : StorePatGlobalAddr("STORE_"#vec_t)>; +def : StorePatExternalSym("STORE_"#vec_t)>; +def : StorePatOffsetOnly("STORE_"#vec_t)>; +def : StorePatGlobalAddrOffOnly("STORE_"#vec_t)>; +def : StorePatExternSymOffOnly("STORE_"#vec_t)>; +} + //===----------------------------------------------------------------------===// // Constructing SIMD values //===----------------------------------------------------------------------===// @@ -46,7 +98,7 @@ defm CONST_V128_#vec_t : SIMD_I<(outs V128:$dst), ops, (outs), ops, [(set V128:$dst, (vec_t pat))], "v128.const\t$dst, "#args, - "v128.const\t"#args, 0>; + "v128.const\t"#args, 2>; } defm "" : ConstVec; } -defm "" : Splat; -defm "" : Splat; -defm "" : Splat; -defm "" : Splat; -defm "" : Splat; -defm "" : Splat; +defm "" : Splat; +defm "" : Splat; +defm "" : Splat; +defm "" : Splat; +defm "" : Splat; +defm "" : Splat; + +// Shuffle lanes: shuffle +defm SHUFFLE : + SIMD_I<(outs V128:$dst), + (ins V128:$x, V128:$y, + vec_i8imm_op:$m0, vec_i8imm_op:$m1, + vec_i8imm_op:$m2, vec_i8imm_op:$m3, + vec_i8imm_op:$m4, vec_i8imm_op:$m5, + vec_i8imm_op:$m6, vec_i8imm_op:$m7, + vec_i8imm_op:$m8, vec_i8imm_op:$m9, + vec_i8imm_op:$mA, vec_i8imm_op:$mB, + vec_i8imm_op:$mC, vec_i8imm_op:$mD, + vec_i8imm_op:$mE, vec_i8imm_op:$mF), + (outs), + (ins + vec_i8imm_op:$m0, vec_i8imm_op:$m1, + vec_i8imm_op:$m2, vec_i8imm_op:$m3, + vec_i8imm_op:$m4, vec_i8imm_op:$m5, + vec_i8imm_op:$m6, vec_i8imm_op:$m7, + vec_i8imm_op:$m8, vec_i8imm_op:$m9, + vec_i8imm_op:$mA, vec_i8imm_op:$mB, + vec_i8imm_op:$mC, vec_i8imm_op:$mD, + vec_i8imm_op:$mE, vec_i8imm_op:$mF), + [], + "v8x16.shuffle\t$dst, $x, $y, "# + "$m0, $m1, $m2, $m3, $m4, $m5, $m6, $m7, "# + "$m8, $m9, $mA, $mB, $mC, $mD, $mE, $mF", + "v8x16.shuffle\t"# + "$m0, $m1, $m2, $m3, $m4, $m5, $m6, $m7, "# + "$m8, $m9, $mA, $mB, $mC, $mD, $mE, $mF", + 3>; + +// Shuffles after custom lowering +def wasm_shuffle_t : SDTypeProfile<1, 18, []>; +def wasm_shuffle : SDNode<"WebAssemblyISD::SHUFFLE", wasm_shuffle_t>; +foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in { +def : Pat<(vec_t (wasm_shuffle (vec_t V128:$x), (vec_t V128:$y), + (i32 LaneIdx32:$m0), (i32 LaneIdx32:$m1), + (i32 LaneIdx32:$m2), (i32 LaneIdx32:$m3), + (i32 LaneIdx32:$m4), (i32 LaneIdx32:$m5), + (i32 LaneIdx32:$m6), (i32 LaneIdx32:$m7), + (i32 LaneIdx32:$m8), (i32 LaneIdx32:$m9), + (i32 LaneIdx32:$mA), (i32 LaneIdx32:$mB), + (i32 LaneIdx32:$mC), (i32 LaneIdx32:$mD), + (i32 LaneIdx32:$mE), (i32 LaneIdx32:$mF))), + (vec_t (SHUFFLE (vec_t V128:$x), (vec_t V128:$y), + (i32 LaneIdx32:$m0), (i32 LaneIdx32:$m1), + (i32 LaneIdx32:$m2), (i32 LaneIdx32:$m3), + (i32 LaneIdx32:$m4), (i32 LaneIdx32:$m5), + (i32 LaneIdx32:$m6), (i32 LaneIdx32:$m7), + (i32 LaneIdx32:$m8), (i32 LaneIdx32:$m9), + (i32 LaneIdx32:$mA), (i32 LaneIdx32:$mB), + (i32 LaneIdx32:$mC), (i32 LaneIdx32:$mD), + (i32 LaneIdx32:$mE), (i32 LaneIdx32:$mF)))>; +} //===----------------------------------------------------------------------===// // Accessing lanes @@ -164,16 +271,16 @@ multiclass ExtractLaneExtended baseInst> { defm "" : ExtractLane("extract_i8x16"#sign)>; - defm "" : ExtractLane("extract_i16x8"#sign)>; } -defm "" : ExtractLaneExtended<"_s", 9>; -defm "" : ExtractLaneExtended<"_u", 10>; +defm "" : ExtractLaneExtended<"_s", 5>; +defm "" : ExtractLaneExtended<"_u", 6>; defm "" : ExtractLane; -defm "" : ExtractLane; -defm "" : ExtractLane; -defm "" : ExtractLane; +defm "" : ExtractLane; +defm "" : ExtractLane; +defm "" : ExtractLane; // Follow convention of making implicit expansions unsigned def : Pat<(i32 (vector_extract (v16i8 V128:$vec), (i32 LaneIdx16:$idx))), @@ -216,12 +323,12 @@ vec#".replace_lane\t$idx", simdop>; } -defm "" : ReplaceLane; -defm "" : ReplaceLane; -defm "" : ReplaceLane; -defm "" : ReplaceLane; -defm "" : ReplaceLane; -defm "" : ReplaceLane; +defm "" : ReplaceLane; +defm "" : ReplaceLane; +defm "" : ReplaceLane; +defm "" : ReplaceLane; +defm "" : ReplaceLane; +defm "" : ReplaceLane; // Lower undef lane indices to zero def : Pat<(vector_insert (v16i8 V128:$vec), I32:$x, undef), @@ -349,63 +456,80 @@ (v2f64 (REPLACE_LANE_v2f64 (v2f64 (SPLAT_v2f64 (f64 F64:$x0))), 1, F64:$x1))>; -// Shuffle lanes: shuffle -defm SHUFFLE : - SIMD_I<(outs V128:$dst), - (ins V128:$x, V128:$y, - vec_i8imm_op:$m0, vec_i8imm_op:$m1, - vec_i8imm_op:$m2, vec_i8imm_op:$m3, - vec_i8imm_op:$m4, vec_i8imm_op:$m5, - vec_i8imm_op:$m6, vec_i8imm_op:$m7, - vec_i8imm_op:$m8, vec_i8imm_op:$m9, - vec_i8imm_op:$mA, vec_i8imm_op:$mB, - vec_i8imm_op:$mC, vec_i8imm_op:$mD, - vec_i8imm_op:$mE, vec_i8imm_op:$mF), - (outs), - (ins - vec_i8imm_op:$m0, vec_i8imm_op:$m1, - vec_i8imm_op:$m2, vec_i8imm_op:$m3, - vec_i8imm_op:$m4, vec_i8imm_op:$m5, - vec_i8imm_op:$m6, vec_i8imm_op:$m7, - vec_i8imm_op:$m8, vec_i8imm_op:$m9, - vec_i8imm_op:$mA, vec_i8imm_op:$mB, - vec_i8imm_op:$mC, vec_i8imm_op:$mD, - vec_i8imm_op:$mE, vec_i8imm_op:$mF), - [], - "v8x16.shuffle\t$dst, $x, $y, "# - "$m0, $m1, $m2, $m3, $m4, $m5, $m6, $m7, "# - "$m8, $m9, $mA, $mB, $mC, $mD, $mE, $mF", - "v8x16.shuffle\t"# - "$m0, $m1, $m2, $m3, $m4, $m5, $m6, $m7, "# - "$m8, $m9, $mA, $mB, $mC, $mD, $mE, $mF", - 23>; +//===----------------------------------------------------------------------===// +// Comparisons +//===----------------------------------------------------------------------===// -// Shuffles after custom lowering -def wasm_shuffle_t : SDTypeProfile<1, 18, []>; -def wasm_shuffle : SDNode<"WebAssemblyISD::SHUFFLE", wasm_shuffle_t>; -foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in { -def : Pat<(vec_t (wasm_shuffle (vec_t V128:$x), (vec_t V128:$y), - (i32 LaneIdx32:$m0), (i32 LaneIdx32:$m1), - (i32 LaneIdx32:$m2), (i32 LaneIdx32:$m3), - (i32 LaneIdx32:$m4), (i32 LaneIdx32:$m5), - (i32 LaneIdx32:$m6), (i32 LaneIdx32:$m7), - (i32 LaneIdx32:$m8), (i32 LaneIdx32:$m9), - (i32 LaneIdx32:$mA), (i32 LaneIdx32:$mB), - (i32 LaneIdx32:$mC), (i32 LaneIdx32:$mD), - (i32 LaneIdx32:$mE), (i32 LaneIdx32:$mF))), - (vec_t (SHUFFLE (vec_t V128:$x), (vec_t V128:$y), - (i32 LaneIdx32:$m0), (i32 LaneIdx32:$m1), - (i32 LaneIdx32:$m2), (i32 LaneIdx32:$m3), - (i32 LaneIdx32:$m4), (i32 LaneIdx32:$m5), - (i32 LaneIdx32:$m6), (i32 LaneIdx32:$m7), - (i32 LaneIdx32:$m8), (i32 LaneIdx32:$m9), - (i32 LaneIdx32:$mA), (i32 LaneIdx32:$mB), - (i32 LaneIdx32:$mC), (i32 LaneIdx32:$mD), - (i32 LaneIdx32:$mE), (i32 LaneIdx32:$mF)))>; +multiclass SIMDCondition simdop> { + defm _#vec_t : + SIMD_I<(outs V128:$dst), (ins V128:$lhs, V128:$rhs), (outs), (ins), + [(set (out_t V128:$dst), + (setcc (vec_t V128:$lhs), (vec_t V128:$rhs), cond) + )], + vec#"."#name#"\t$dst, $lhs, $rhs", vec#"."#name, simdop>; +} + +multiclass SIMDConditionInt baseInst> { + defm "" : SIMDCondition; + defm "" : SIMDCondition; + defm "" : SIMDCondition; } +multiclass SIMDConditionFP baseInst> { + defm "" : SIMDCondition; + defm "" : SIMDCondition; +} + +// Equality: eq +let isCommutable = 1 in { +defm EQ : SIMDConditionInt<"eq", SETEQ, 24>; +defm EQ : SIMDConditionFP<"eq", SETOEQ, 64>; +} // isCommutable = 1 + +// Non-equality: ne +let isCommutable = 1 in { +defm NE : SIMDConditionInt<"ne", SETNE, 25>; +defm NE : SIMDConditionFP<"ne", SETUNE, 65>; +} // isCommutable = 1 + +// Less than: lt_s / lt_u / lt +defm LT_S : SIMDConditionInt<"lt_s", SETLT, 26>; +defm LT_U : SIMDConditionInt<"lt_u", SETULT, 27>; +defm LT : SIMDConditionFP<"lt", SETOLT, 66>; + +// Greater than: gt_s / gt_u / gt +defm GT_S : SIMDConditionInt<"gt_s", SETGT, 28>; +defm GT_U : SIMDConditionInt<"gt_u", SETUGT, 29>; +defm GT : SIMDConditionFP<"gt", SETOGT, 67>; + +// Less than or equal: le_s / le_u / le +defm LE_S : SIMDConditionInt<"le_s", SETLE, 30>; +defm LE_U : SIMDConditionInt<"le_u", SETULE, 31>; +defm LE : SIMDConditionFP<"le", SETOLE, 68>; + +// Greater than or equal: ge_s / ge_u / ge +defm GE_S : SIMDConditionInt<"ge_s", SETGE, 32>; +defm GE_U : SIMDConditionInt<"ge_u", SETUGE, 33>; +defm GE : SIMDConditionFP<"ge", SETOGE, 69>; + +// Lower float comparisons that don't care about NaN to standard WebAssembly +// float comparisons. These instructions are generated in the target-independent +// expansion of unordered comparisons and ordered ne. +def : Pat<(v4i32 (seteq (v4f32 V128:$lhs), (v4f32 V128:$rhs))), + (v4i32 (EQ_v4f32 (v4f32 V128:$lhs), (v4f32 V128:$rhs)))>; +def : Pat<(v4i32 (setne (v4f32 V128:$lhs), (v4f32 V128:$rhs))), + (v4i32 (NE_v4f32 (v4f32 V128:$lhs), (v4f32 V128:$rhs)))>; +def : Pat<(v2i64 (seteq (v2f64 V128:$lhs), (v2f64 V128:$rhs))), + (v2i64 (EQ_v2f64 (v2f64 V128:$lhs), (v2f64 V128:$rhs)))>; +def : Pat<(v2i64 (setne (v2f64 V128:$lhs), (v2f64 V128:$rhs))), + (v2i64 (NE_v2f64 (v2f64 V128:$lhs), (v2f64 V128:$rhs)))>; + //===----------------------------------------------------------------------===// -// Integer arithmetic +// Bitwise operations //===----------------------------------------------------------------------===// multiclass SIMDBinary; } -multiclass SIMDBinaryIntNoI64x2 baseInst> { - defm "" : SIMDBinary; - defm "" : SIMDBinary; - defm "" : SIMDBinary; -} - -multiclass SIMDBinaryInt baseInst> { - defm "" : SIMDBinaryIntNoI64x2; - defm "" : SIMDBinary; +multiclass SIMDBitwise simdop> { + defm "" : SIMDBinary; + defm "" : SIMDBinary; + defm "" : SIMDBinary; + defm "" : SIMDBinary; } -// Integer vector negation -def ivneg : PatFrag<(ops node:$in), (sub immAllZerosV, node:$in)>; - -// Integer addition: add -let isCommutable = 1 in -defm ADD : SIMDBinaryInt; - -// Integer subtraction: sub -defm SUB : SIMDBinaryInt; +// Bitwise logic: v128.and / v128.or / v128.xor +let isCommutable = 1 in { +defm AND : SIMDBitwise; +defm OR : SIMDBitwise; +defm XOR : SIMDBitwise; +} // isCommutable = 1 -// Integer multiplication: mul -defm MUL : SIMDBinaryIntNoI64x2; +// Bitwise logic: v128.not +foreach vec_t = [v16i8, v8i16, v4i32, v2i64] in + defm NOT_#vec_t : SIMD_I<(outs V128:$dst), (ins V128:$vec), (outs), (ins), + [(set (vec_t V128:$dst), (vec_t (vnot V128:$vec)))], + "v128.not\t$dst, $vec", "v128.not", 79>; -// Integer negation: neg -multiclass SIMDNeg simdop> { - defm NEG_#vec_t : SIMD_I<(outs V128:$dst), (ins V128:$vec), (outs), (ins), - [(set (vec_t V128:$dst), - (vec_t (neg (vec_t V128:$vec))) - )], - vec#".neg\t$dst, $vec", vec#".neg", simdop>; -} +// Bitwise select: v128.bitselect +foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in + defm BITSELECT_#vec_t : + SIMD_I<(outs V128:$dst), (ins V128:$v1, V128:$v2, V128:$c), (outs), (ins), + [(set (vec_t V128:$dst), + (vec_t (int_wasm_bitselect + (vec_t V128:$c), (vec_t V128:$v1), (vec_t V128:$v2) + )) + )], + "v128.bitselect\t$dst, $v1, $v2, $c", "v128.bitselect", 80>; -defm "" : SIMDNeg; -defm "" : SIMDNeg; -defm "" : SIMDNeg; -defm "" : SIMDNeg; +// Bitselect is equivalent to (c & v1) | (~c & v2) +foreach vec_t = [v16i8, v8i16, v4i32, v2i64] in + def : Pat<(vec_t (or (and (vec_t V128:$c), (vec_t V128:$v1)), + (and (vnot V128:$c), (vec_t V128:$v2)))), + (!cast("BITSELECT_"#vec_t) + V128:$v1, V128:$v2, V128:$c)>; //===----------------------------------------------------------------------===// -// Saturating integer arithmetic +// Integer arithmetic //===----------------------------------------------------------------------===// -multiclass SIMDBinarySat baseInst> { +multiclass SIMDBinaryIntSmall baseInst> { defm "" : SIMDBinary; - defm "" : SIMDBinary; + defm "" : SIMDBinary; +} + +multiclass SIMDBinaryIntNoI64x2 baseInst> { + defm "" : SIMDBinaryIntSmall; + defm "" : SIMDBinary; } -// Saturating integer addition: add_saturate_s / add_saturate_u +multiclass SIMDBinaryInt baseInst> { + defm "" : SIMDBinaryIntNoI64x2; + defm "" : SIMDBinary; +} + +// Integer vector negation +def ivneg : PatFrag<(ops node:$in), (sub immAllZerosV, node:$in)>; + +// Integer addition: add / add_saturate_s / add_saturate_u let isCommutable = 1 in { -defm ADD_SAT_S : - SIMDBinarySat; -defm ADD_SAT_U : - SIMDBinarySat; +defm ADD : SIMDBinaryInt; +defm ADD_SAT_S : SIMDBinaryIntSmall; +defm ADD_SAT_U : SIMDBinaryIntSmall; } // isCommutable = 1 -// Saturating integer subtraction: sub_saturate_s / sub_saturate_u +// Integer subtraction: sub / sub_saturate_s / sub_saturate_u +defm SUB : SIMDBinaryInt; defm SUB_SAT_S : - SIMDBinarySat; + SIMDBinaryIntSmall; defm SUB_SAT_U : - SIMDBinarySat; + SIMDBinaryIntSmall; + +// Integer multiplication: mul +defm MUL : SIMDBinaryIntNoI64x2; //===----------------------------------------------------------------------===// // Bit shifts @@ -493,22 +633,22 @@ vec#"."#name#"\t$dst, $vec, $x", vec#"."#name, simdop>; } -multiclass SIMDShiftInt baseInst, int skip> { +multiclass SIMDShiftInt baseInst> { defm "" : SIMDShift; defm "" : SIMDShift; + !add(baseInst, 15)>; defm "" : SIMDShift; + !add(baseInst, 30)>; defm "" : SIMDShift; + name, !add(baseInst, 45)>; } // Left shift by scalar: shl -defm SHL : SIMDShiftInt; +defm SHL : SIMDShiftInt; // Right shift by scalar: shr_s / shr_u -defm SHR_S : SIMDShiftInt; -defm SHR_U : SIMDShiftInt; +defm SHR_S : SIMDShiftInt; +defm SHR_U : SIMDShiftInt; // Truncate i64 shift operands to i32s foreach shifts = [[shl, SHL_v2i64], [sra, SHR_S_v2i64], [srl, SHR_U_v2i64]] in @@ -529,56 +669,27 @@ (v2i64 (shifts[1] (v2i64 V128:$vec), I32:$x))>; //===----------------------------------------------------------------------===// -// Bitwise operations +// Integer negations //===----------------------------------------------------------------------===// -multiclass SIMDBitwise simdop> { - defm "" : SIMDBinary; - defm "" : SIMDBinary; - defm "" : SIMDBinary; - defm "" : SIMDBinary; -} - -// Bitwise logic: v128.and / v128.or / v128.xor -let isCommutable = 1 in { -defm AND : SIMDBitwise; -defm OR : SIMDBitwise; -defm XOR : SIMDBitwise; -} // isCommutable = 1 - -// Bitwise logic: v128.not -multiclass SIMDNot { - defm NOT_#vec_t : SIMD_I<(outs V128:$dst), (ins V128:$vec), (outs), (ins), - [(set (vec_t V128:$dst), (vec_t (vnot V128:$vec)))], - "v128.not\t$dst, $vec", "v128.not", 63>; +multiclass SIMDUnary simdop> { + defm _#vec_t : SIMD_I<(outs V128:$dst), (ins V128:$vec), (outs), (ins), + [(set (vec_t V128:$dst), + (vec_t (node (vec_t V128:$vec))) + )], + vec#"."#name#"\t$dst, $vec", vec#"."#name, simdop>; } -defm "" : SIMDNot; -defm "" : SIMDNot; -defm "" : SIMDNot; -defm "" : SIMDNot; - -// Bitwise select: v128.bitselect -multiclass Bitselect { - defm BITSELECT_#vec_t : - SIMD_I<(outs V128:$dst), (ins V128:$v1, V128:$v2, V128:$c), (outs), (ins), - [(set (vec_t V128:$dst), - (vec_t (int_wasm_bitselect - (vec_t V128:$c), (vec_t V128:$v1), (vec_t V128:$v2) - )) - )], - "v128.bitselect\t$dst, $v1, $v2, $c", "v128.bitselect", 64>; +multiclass SIMDUnaryInt baseInst> { + defm "" : SIMDUnary; + defm "" : SIMDUnary; + defm "" : SIMDUnary; + defm "" : SIMDUnary; } -foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in -defm "" : Bitselect; - -// Bitselect is equivalent to (c & v1) | (~c & v2) -foreach vec_t = [v16i8, v8i16, v4i32, v2i64] in - def : Pat<(vec_t (or (and (vec_t V128:$c), (vec_t V128:$v1)), - (and (vnot V128:$c), (vec_t V128:$v2)))), - (!cast("BITSELECT_"#vec_t) - V128:$v1, V128:$v2, V128:$c)>; +// Integer negation: neg +defm NEG : SIMDUnaryInt; //===----------------------------------------------------------------------===// // Boolean horizontal reductions @@ -593,203 +704,59 @@ multiclass SIMDReduce baseInst> { defm "" : SIMDReduceVec; - defm "" : SIMDReduceVec; - defm "" : SIMDReduceVec; - defm "" : SIMDReduceVec; + defm "" : SIMDReduceVec; + defm "" : SIMDReduceVec; + defm "" : SIMDReduceVec; } // Any lane true: any_true -defm ANYTRUE : SIMDReduce<"any_true", int_wasm_anytrue, 65>; +defm ANYTRUE : SIMDReduce<"any_true", int_wasm_anytrue, 94>; // All lanes true: all_true -defm ALLTRUE : SIMDReduce<"all_true", int_wasm_alltrue, 69>; - -//===----------------------------------------------------------------------===// -// Comparisons -//===----------------------------------------------------------------------===// - -multiclass SIMDCondition simdop> { - defm _#vec_t : - SIMD_I<(outs V128:$dst), (ins V128:$lhs, V128:$rhs), (outs), (ins), - [(set (out_t V128:$dst), - (setcc (vec_t V128:$lhs), (vec_t V128:$rhs), cond) - )], - vec#"."#name#"\t$dst, $lhs, $rhs", vec#"."#name, simdop>; -} - -multiclass SIMDConditionInt baseInst, - int step = 1> { - defm "" : SIMDCondition; - defm "" : SIMDCondition; - defm "" : SIMDCondition; -} - -multiclass SIMDConditionFP baseInst> { - defm "" : SIMDCondition; - defm "" : SIMDCondition; -} - -// Equality: eq -let isCommutable = 1 in { -defm EQ : SIMDConditionInt<"eq", SETEQ, 73>; -defm EQ : SIMDConditionFP<"eq", SETOEQ, 77>; -} // isCommutable = 1 - -// Non-equality: ne -let isCommutable = 1 in { -defm NE : SIMDConditionInt<"ne", SETNE, 79>; -defm NE : SIMDConditionFP<"ne", SETUNE, 83>; -} // isCommutable = 1 - -// Less than: lt_s / lt_u / lt -defm LT_S : SIMDConditionInt<"lt_s", SETLT, 85, 2>; -defm LT_U : SIMDConditionInt<"lt_u", SETULT, 86, 2>; -defm LT : SIMDConditionFP<"lt", SETOLT, 93>; - -// Less than or equal: le_s / le_u / le -defm LE_S : SIMDConditionInt<"le_s", SETLE, 95, 2>; -defm LE_U : SIMDConditionInt<"le_u", SETULE, 96, 2>; -defm LE : SIMDConditionFP<"le", SETOLE, 103>; - -// Greater than: gt_s / gt_u / gt -defm GT_S : SIMDConditionInt<"gt_s", SETGT, 105, 2>; -defm GT_U : SIMDConditionInt<"gt_u", SETUGT, 106, 2>; -defm GT : SIMDConditionFP<"gt", SETOGT, 113>; - -// Greater than or equal: ge_s / ge_u / ge -defm GE_S : SIMDConditionInt<"ge_s", SETGE, 115, 2>; -defm GE_U : SIMDConditionInt<"ge_u", SETUGE, 116, 2>; -defm GE : SIMDConditionFP<"ge", SETOGE, 123>; - -// Lower float comparisons that don't care about NaN to standard WebAssembly -// float comparisons. These instructions are generated in the target-independent -// expansion of unordered comparisons and ordered ne. -def : Pat<(v4i32 (seteq (v4f32 V128:$lhs), (v4f32 V128:$rhs))), - (v4i32 (EQ_v4f32 (v4f32 V128:$lhs), (v4f32 V128:$rhs)))>; -def : Pat<(v4i32 (setne (v4f32 V128:$lhs), (v4f32 V128:$rhs))), - (v4i32 (NE_v4f32 (v4f32 V128:$lhs), (v4f32 V128:$rhs)))>; -def : Pat<(v2i64 (seteq (v2f64 V128:$lhs), (v2f64 V128:$rhs))), - (v2i64 (EQ_v2f64 (v2f64 V128:$lhs), (v2f64 V128:$rhs)))>; -def : Pat<(v2i64 (setne (v2f64 V128:$lhs), (v2f64 V128:$rhs))), - (v2i64 (NE_v2f64 (v2f64 V128:$lhs), (v2f64 V128:$rhs)))>; +defm ALLTRUE : SIMDReduce<"all_true", int_wasm_alltrue, 95>; //===----------------------------------------------------------------------===// -// Load and store +// Floating-point arithmetic //===----------------------------------------------------------------------===// -// Load: v128.load -multiclass SIMDLoad { - let mayLoad = 1 in - defm LOAD_#vec_t : - SIMD_I<(outs V128:$dst), (ins P2Align:$align, offset32_op:$off, I32:$addr), - (outs), (ins P2Align:$align, offset32_op:$off), [], - "v128.load\t$dst, ${off}(${addr})$align", - "v128.load\t$off$align", 1>; -} - -foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in { -defm "" : SIMDLoad; - -// Def load and store patterns from WebAssemblyInstrMemory.td for vector types -def : LoadPatNoOffset("LOAD_"#vec_t)>; -def : LoadPatImmOff("LOAD_"#vec_t)>; -def : LoadPatImmOff("LOAD_"#vec_t)>; -def : LoadPatGlobalAddr("LOAD_"#vec_t)>; -def : LoadPatExternalSym("LOAD_"#vec_t)>; -def : LoadPatOffsetOnly("LOAD_"#vec_t)>; -def : LoadPatGlobalAddrOffOnly("LOAD_"#vec_t)>; -def : LoadPatExternSymOffOnly("LOAD_"#vec_t)>; -} - -// Store: v128.store -multiclass SIMDStore { - let mayStore = 1 in - defm STORE_#vec_t : - SIMD_I<(outs), (ins P2Align:$align, offset32_op:$off, I32:$addr, V128:$vec), - (outs), (ins P2Align:$align, offset32_op:$off), [], - "v128.store\t${off}(${addr})$align, $vec", - "v128.store\t$off$align", 2>; +multiclass SIMDBinaryFP baseInst> { + defm "" : SIMDBinary; + defm "" : SIMDBinary; } -foreach vec_t = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in { -defm "" : SIMDStore; - -// Def load and store patterns from WebAssemblyInstrMemory.td for vector types -def : StorePatNoOffset("STORE_"#vec_t)>; -def : StorePatImmOff("STORE_"#vec_t)>; -def : StorePatImmOff("STORE_"#vec_t)>; -def : StorePatGlobalAddr("STORE_"#vec_t)>; -def : StorePatExternalSym("STORE_"#vec_t)>; -def : StorePatOffsetOnly("STORE_"#vec_t)>; -def : StorePatGlobalAddrOffOnly("STORE_"#vec_t)>; -def : StorePatExternSymOffOnly("STORE_"#vec_t)>; +multiclass SIMDUnaryFP baseInst> { + defm "" : SIMDUnary; + defm "" : SIMDUnary; } -//===----------------------------------------------------------------------===// -// Floating-point sign bit operations -//===----------------------------------------------------------------------===// - -// Negation: neg -defm "" : SIMDNeg; -defm "" : SIMDNeg; - -// Absolute value: abs -multiclass SIMDAbs simdop> { - defm ABS_#vec_t : SIMD_I<(outs V128:$dst), (ins V128:$vec), (outs), (ins), - [(set (vec_t V128:$dst), (vec_t (fabs V128:$vec)))], - vec#".abs\t$dst, $vec", vec#".abs", simdop>; -} +// Addition: add +let isCommutable = 1 in +defm ADD : SIMDBinaryFP; -defm "" : SIMDAbs; -defm "" : SIMDAbs; +// Subtraction: sub +defm SUB : SIMDBinaryFP; -//===----------------------------------------------------------------------===// -// Floating-point min and max -//===----------------------------------------------------------------------===// +// Multiplication: mul +let isCommutable = 1 in +defm MUL : SIMDBinaryFP; -multiclass SIMDBinaryFP baseInst> { - defm "" : SIMDBinary; - defm "" : SIMDBinary; -} +// Division: div +defm DIV : SIMDBinaryFP; // NaN-propagating minimum: min -defm MIN : SIMDBinaryFP; +defm MIN : SIMDBinaryFP; // NaN-propagating maximum: max -defm MAX : SIMDBinaryFP; - -//===----------------------------------------------------------------------===// -// Floating-point arithmetic -//===----------------------------------------------------------------------===// +defm MAX : SIMDBinaryFP; -// Addition: add -let isCommutable = 1 in -defm ADD : SIMDBinaryFP; - -// Subtraction: sub -defm SUB : SIMDBinaryFP; - -// Division: div -defm DIV : SIMDBinaryFP; +// Negation: neg +defm NEG : SIMDUnaryFP; -// Multiplication: mul -let isCommutable = 1 in -defm MUL : SIMDBinaryFP; +// Absolute value: abs +defm ABS : SIMDUnaryFP; // Square root: sqrt -multiclass SIMDSqrt simdop> { - defm SQRT_#vec_t : - SIMD_I<(outs V128:$dst), (ins V128:$vec), (outs), (ins), - [(set (vec_t V128:$dst), (vec_t (fsqrt V128:$vec)))], - vec#".sqrt\t$dst, $vec", vec#".sqrt", simdop>; -} - -defm "" : SIMDSqrt; -defm "" : SIMDSqrt; +defm SQRT : SIMDUnaryFP; //===----------------------------------------------------------------------===// // Conversions @@ -803,17 +770,11 @@ name#"\t$dst, $vec", name, simdop>; } -// Integer to floating point: convert_s / convert_u -defm "" : SIMDConvert; -defm "" : SIMDConvert; -defm "" : SIMDConvert; -defm "" : SIMDConvert; - // Floating point to integer with saturation: trunc_sat_s / trunc_sat_u -defm "" : SIMDConvert; -defm "" : SIMDConvert; -defm "" : SIMDConvert; -defm "" : SIMDConvert; +defm "" : SIMDConvert; +defm "" : SIMDConvert; +defm "" : SIMDConvert; +defm "" : SIMDConvert; // Lower llvm.wasm.trunc.saturate.* to saturating instructions def : Pat<(v4i32 (int_wasm_trunc_saturate_signed (v4f32 V128:$src))), @@ -825,6 +786,12 @@ def : Pat<(v2i64 (int_wasm_trunc_saturate_unsigned (v2f64 V128:$src))), (fp_to_uint_v2i64_v2f64 (v2f64 V128:$src))>; +// Integer to floating point: convert_s / convert_u +defm "" : SIMDConvert; +defm "" : SIMDConvert; +defm "" : SIMDConvert; +defm "" : SIMDConvert; + // Bitcasts are nops // Matching bitcast t1 to t1 causes strange errors, so avoid repeating types foreach t1 = [v16i8, v8i16, v4i32, v2i64, v4f32, v2f64] in Index: test/MC/Disassembler/WebAssembly/wasm.txt =================================================================== --- test/MC/Disassembler/WebAssembly/wasm.txt +++ test/MC/Disassembler/WebAssembly/wasm.txt @@ -33,7 +33,7 @@ # v128.const is arbitrarily disassembled as v16i8 # CHECK: v128.const 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 -0xFD 0x00 0x00 0x01 0x02 0x03 0x04 0x05 0x06 0x07 0x08 0x09 0x0A 0x0B 0x0C 0x0D 0x0E 0x0F +0xFD 0x02 0x00 0x01 0x02 0x03 0x04 0x05 0x06 0x07 0x08 0x09 0x0A 0x0B 0x0C 0x0D 0x0E 0x0F # CHECK: v8x16.shuffle 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 -0xFD 0x17 0x00 0x01 0x02 0x03 0x04 0x05 0x06 0x07 0x08 0x09 0x0A 0x0B 0x0C 0x0D 0x0E 0x0F +0xFD 0x03 0x00 0x01 0x02 0x03 0x04 0x05 0x06 0x07 0x08 0x09 0x0A 0x0B 0x0C 0x0D 0x0E 0x0F Index: test/MC/WebAssembly/simd-encodings.s =================================================================== --- test/MC/WebAssembly/simd-encodings.s +++ test/MC/WebAssembly/simd-encodings.s @@ -1,13 +1,19 @@ # RUN: llvm-mc -show-encoding -triple=wasm32-unkown-unknown -mattr=+sign-ext,+simd128 < %s | FileCheck %s + # CHECK: v128.load 48:p2align=0 # encoding: [0xfd,0x00,0x00,0x30] + v128.load 48 + + # CHECK: v128.store 48:p2align=0 # encoding: [0xfd,0x01,0x00,0x30] + v128.store 48 + # CHECK: v128.const 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 - # CHECK-SAME: # encoding: [0xfd,0x00, + # CHECK-SAME: # encoding: [0xfd,0x02, # CHECK-SAME: 0x00,0x01,0x02,0x03,0x04,0x05,0x06,0x07, # CHECK-SAME: 0x08,0x09,0x0a,0x0b,0x0c,0x0d,0x0e,0x0f] v128.const 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15 # CHECK: v128.const 256, 770, 1284, 1798, 2312, 2826, 3340, 3854 - # CHECK-SAME: # encoding: [0xfd,0x00, + # CHECK-SAME: # encoding: [0xfd,0x02, # CHECK-SAME: 0x00,0x01,0x02,0x03,0x04,0x05,0x06,0x07, # CHECK-SAME: 0x08,0x09,0x0a,0x0b,0x0c,0x0d,0x0e,0x0f] v128.const 256, 770, 1284, 1798, 2312, 2826, 3340, 3854 @@ -16,436 +22,430 @@ # CHECK: v128.const 0x1.0402p-121, 0x1.0c0a08p-113, # CHECK-SAME: 0x1.14121p-105, 0x1.1c1a18p-97 - # CHECK-SAME: # encoding: [0xfd,0x00, + # CHECK-SAME: # encoding: [0xfd,0x02, # CHECK-SAME: 0x00,0x01,0x02,0x03,0x04,0x05,0x06,0x07, # CHECK-SAME: 0x08,0x09,0x0a,0x0b,0x0c,0x0d,0x0e,0x0f] v128.const 0x1.0402p-121, 0x1.0c0a08p-113, 0x1.14121p-105, 0x1.1c1a18p-97 # CHECK: v128.const 0x1.60504030201p-911, 0x1.e0d0c0b0a0908p-783 - # CHECK-SAME: # encoding: [0xfd,0x00, + # CHECK-SAME: # encoding: [0xfd,0x02, # CHECK-SAME: 0x00,0x01,0x02,0x03,0x04,0x05,0x06,0x07, # CHECK-SAME: 0x08,0x09,0x0a,0x0b,0x0c,0x0d,0x0e,0x0f] v128.const 0x1.60504030201p-911, 0x1.e0d0c0b0a0908p-783 - # CHECK: v128.load 48:p2align=0 # encoding: [0xfd,0x01,0x00,0x30] - v128.load 48 - - # CHECK: v128.store 48:p2align=0 # encoding: [0xfd,0x02,0x00,0x30] - v128.store 48 + # CHECK: v8x16.shuffle 0, 17, 2, 19, 4, 21, 6, 23, + # CHECK-SAME: 8, 25, 10, 27, 12, 29, 14, 31 + # CHECK-SAME: # encoding: [0xfd,0x03, + # CHECK-SAME: 0x00,0x11,0x02,0x13,0x04,0x15,0x06,0x17, + # CHECK-SAME: 0x08,0x19,0x0a,0x1b,0x0c,0x1d,0x0e,0x1f] + v8x16.shuffle 0, 17, 2, 19, 4, 21, 6, 23, 8, 25, 10, 27, 12, 29, 14, 31 - # CHECK: i8x16.splat # encoding: [0xfd,0x03] + # CHECK: i8x16.splat # encoding: [0xfd,0x04] i8x16.splat - # CHECK: i16x8.splat # encoding: [0xfd,0x04] - i16x8.splat - - # CHECK: i32x4.splat # encoding: [0xfd,0x05] - i32x4.splat - - # CHECK: i64x2.splat # encoding: [0xfd,0x06] - i64x2.splat - - # CHECK: f32x4.splat # encoding: [0xfd,0x07] - f32x4.splat - - # CHECK: f64x2.splat # encoding: [0xfd,0x08] - f64x2.splat - - # CHECK: i8x16.extract_lane_s 15 # encoding: [0xfd,0x09,0x0f] + # CHECK: i8x16.extract_lane_s 15 # encoding: [0xfd,0x05,0x0f] i8x16.extract_lane_s 15 - # CHECK: i8x16.extract_lane_u 15 # encoding: [0xfd,0x0a,0x0f] + # CHECK: i8x16.extract_lane_u 15 # encoding: [0xfd,0x06,0x0f] i8x16.extract_lane_u 15 - # CHECK: i16x8.extract_lane_s 7 # encoding: [0xfd,0x0b,0x07] + # CHECK: i8x16.replace_lane 15 # encoding: [0xfd,0x07,0x0f] + i8x16.replace_lane 15 + + # CHECK: i16x8.splat # encoding: [0xfd,0x08] + i16x8.splat + + # CHECK: i16x8.extract_lane_s 7 # encoding: [0xfd,0x09,0x07] i16x8.extract_lane_s 7 - # CHECK: i16x8.extract_lane_u 7 # encoding: [0xfd,0x0c,0x07] + # CHECK: i16x8.extract_lane_u 7 # encoding: [0xfd,0x0a,0x07] i16x8.extract_lane_u 7 + # CHECK: i16x8.replace_lane 7 # encoding: [0xfd,0x0b,0x07] + i16x8.replace_lane 7 + + # CHECK: i32x4.splat # encoding: [0xfd,0x0c] + i32x4.splat + # CHECK: i32x4.extract_lane 3 # encoding: [0xfd,0x0d,0x03] i32x4.extract_lane 3 - # CHECK: i64x2.extract_lane 1 # encoding: [0xfd,0x0e,0x01] + # CHECK: i32x4.replace_lane 3 # encoding: [0xfd,0x0e,0x03] + i32x4.replace_lane 3 + + # CHECK: i64x2.splat # encoding: [0xfd,0x0f] + i64x2.splat + + # CHECK: i64x2.extract_lane 1 # encoding: [0xfd,0x10,0x01] i64x2.extract_lane 1 - # CHECK: f32x4.extract_lane 3 # encoding: [0xfd,0x0f,0x03] + # CHECK: i64x2.replace_lane 1 # encoding: [0xfd,0x11,0x01] + i64x2.replace_lane 1 + + # CHECK: f32x4.splat # encoding: [0xfd,0x12] + f32x4.splat + + # CHECK: f32x4.extract_lane 3 # encoding: [0xfd,0x13,0x03] f32x4.extract_lane 3 - # CHECK: f64x2.extract_lane 1 # encoding: [0xfd,0x10,0x01] + # CHECK: f32x4.replace_lane 3 # encoding: [0xfd,0x14,0x03] + f32x4.replace_lane 3 + + # CHECK: f64x2.splat # encoding: [0xfd,0x15] + f64x2.splat + + # CHECK: f64x2.extract_lane 1 # encoding: [0xfd,0x16,0x01] f64x2.extract_lane 1 - # CHECK: i8x16.replace_lane 15 # encoding: [0xfd,0x11,0x0f] - i8x16.replace_lane 15 + # CHECK: f64x2.replace_lane 1 # encoding: [0xfd,0x17,0x01] + f64x2.replace_lane 1 - # CHECK: i16x8.replace_lane 7 # encoding: [0xfd,0x12,0x07] - i16x8.replace_lane 7 + # CHECK: i8x16.eq # encoding: [0xfd,0x18] + i8x16.eq - # CHECK: i32x4.replace_lane 3 # encoding: [0xfd,0x13,0x03] - i32x4.replace_lane 3 + # CHECK: i8x16.ne # encoding: [0xfd,0x19] + i8x16.ne - # CHECK: i64x2.replace_lane 1 # encoding: [0xfd,0x14,0x01] - i64x2.replace_lane 1 + # CHECK: i8x16.lt_s # encoding: [0xfd,0x1a] + i8x16.lt_s - # CHECK: f32x4.replace_lane 3 # encoding: [0xfd,0x15,0x03] - f32x4.replace_lane 3 + # CHECK: i8x16.lt_u # encoding: [0xfd,0x1b] + i8x16.lt_u - # CHECK: f64x2.replace_lane 1 # encoding: [0xfd,0x16,0x01] - f64x2.replace_lane 1 + # CHECK: i8x16.gt_s # encoding: [0xfd,0x1c] + i8x16.gt_s - # CHECK: v8x16.shuffle 0, 17, 2, 19, 4, 21, 6, 23, - # CHECK-SAME: 8, 25, 10, 27, 12, 29, 14, 31 - # CHECK-SAME: # encoding: [0xfd,0x17, - # CHECK-SAME: 0x00,0x11,0x02,0x13,0x04,0x15,0x06,0x17, - # CHECK-SAME: 0x08,0x19,0x0a,0x1b,0x0c,0x1d,0x0e,0x1f] - v8x16.shuffle 0, 17, 2, 19, 4, 21, 6, 23, 8, 25, 10, 27, 12, 29, 14, 31 + # CHECK: i8x16.gt_u # encoding: [0xfd,0x1d] + i8x16.gt_u - # CHECK: i8x16.add # encoding: [0xfd,0x18] - i8x16.add + # CHECK: i8x16.le_s # encoding: [0xfd,0x1e] + i8x16.le_s - # CHECK: i16x8.add # encoding: [0xfd,0x19] - i16x8.add + # CHECK: i8x16.le_u # encoding: [0xfd,0x1f] + i8x16.le_u - # CHECK: i32x4.add # encoding: [0xfd,0x1a] - i32x4.add + # CHECK: i8x16.ge_s # encoding: [0xfd,0x20] + i8x16.ge_s - # CHECK: i64x2.add # encoding: [0xfd,0x1b] - i64x2.add + # CHECK: i8x16.ge_u # encoding: [0xfd,0x21] + i8x16.ge_u - # CHECK: i8x16.sub # encoding: [0xfd,0x1c] - i8x16.sub + # CHECK: i16x8.eq # encoding: [0xfd,0x22] + i16x8.eq - # CHECK: i16x8.sub # encoding: [0xfd,0x1d] - i16x8.sub + # CHECK: i16x8.ne # encoding: [0xfd,0x23] + i16x8.ne - # CHECK: i32x4.sub # encoding: [0xfd,0x1e] - i32x4.sub + # CHECK: i16x8.lt_s # encoding: [0xfd,0x24] + i16x8.lt_s - # CHECK: i64x2.sub # encoding: [0xfd,0x1f] - i64x2.sub + # CHECK: i16x8.lt_u # encoding: [0xfd,0x25] + i16x8.lt_u - # CHECK: i8x16.mul # encoding: [0xfd,0x20] - i8x16.mul + # CHECK: i16x8.gt_s # encoding: [0xfd,0x26] + i16x8.gt_s - # CHECK: i16x8.mul # encoding: [0xfd,0x21] - i16x8.mul + # CHECK: i16x8.gt_u # encoding: [0xfd,0x27] + i16x8.gt_u - # CHECK: i32x4.mul # encoding: [0xfd,0x22] - i32x4.mul + # CHECK: i16x8.le_s # encoding: [0xfd,0x28] + i16x8.le_s - # CHECK: i8x16.neg # encoding: [0xfd,0x24] - i8x16.neg + # CHECK: i16x8.le_u # encoding: [0xfd,0x29] + i16x8.le_u - # CHECK: i16x8.neg # encoding: [0xfd,0x25] - i16x8.neg + # CHECK: i16x8.ge_s # encoding: [0xfd,0x2a] + i16x8.ge_s - # CHECK: i32x4.neg # encoding: [0xfd,0x26] - i32x4.neg + # CHECK: i16x8.ge_u # encoding: [0xfd,0x2b] + i16x8.ge_u - # CHECK: i64x2.neg # encoding: [0xfd,0x27] - i64x2.neg + # CHECK: i32x4.eq # encoding: [0xfd,0x2c] + i32x4.eq - # CHECK: i8x16.add_saturate_s # encoding: [0xfd,0x28] - i8x16.add_saturate_s + # CHECK: i32x4.ne # encoding: [0xfd,0x2d] + i32x4.ne - # CHECK: i8x16.add_saturate_u # encoding: [0xfd,0x29] - i8x16.add_saturate_u + # CHECK: i32x4.lt_s # encoding: [0xfd,0x2e] + i32x4.lt_s - # CHECK: i16x8.add_saturate_s # encoding: [0xfd,0x2a] - i16x8.add_saturate_s + # CHECK: i32x4.lt_u # encoding: [0xfd,0x2f] + i32x4.lt_u - # CHECK: i16x8.add_saturate_u # encoding: [0xfd,0x2b] - i16x8.add_saturate_u + # CHECK: i32x4.gt_s # encoding: [0xfd,0x30] + i32x4.gt_s - # CHECK: i8x16.sub_saturate_s # encoding: [0xfd,0x2c] - i8x16.sub_saturate_s + # CHECK: i32x4.gt_u # encoding: [0xfd,0x31] + i32x4.gt_u - # CHECK: i8x16.sub_saturate_u # encoding: [0xfd,0x2d] - i8x16.sub_saturate_u + # CHECK: i32x4.le_s # encoding: [0xfd,0x32] + i32x4.le_s - # CHECK: i16x8.sub_saturate_s # encoding: [0xfd,0x2e] - i16x8.sub_saturate_s + # CHECK: i32x4.le_u # encoding: [0xfd,0x33] + i32x4.le_u - # CHECK: i16x8.sub_saturate_u # encoding: [0xfd,0x2f] - i16x8.sub_saturate_u + # CHECK: i32x4.ge_s # encoding: [0xfd,0x34] + i32x4.ge_s - # CHECK: i8x16.shl # encoding: [0xfd,0x30] - i8x16.shl + # CHECK: i32x4.ge_u # encoding: [0xfd,0x35] + i32x4.ge_u - # CHECK: i16x8.shl # encoding: [0xfd,0x31] - i16x8.shl + # CHECK: f32x4.eq # encoding: [0xfd,0x40] + f32x4.eq - # CHECK: i32x4.shl # encoding: [0xfd,0x32] - i32x4.shl + # CHECK: f32x4.ne # encoding: [0xfd,0x41] + f32x4.ne - # CHECK: i64x2.shl # encoding: [0xfd,0x33] - i64x2.shl + # CHECK: f32x4.lt # encoding: [0xfd,0x42] + f32x4.lt - # CHECK: i8x16.shr_s # encoding: [0xfd,0x34] - i8x16.shr_s + # CHECK: f32x4.gt # encoding: [0xfd,0x43] + f32x4.gt - # CHECK: i8x16.shr_u # encoding: [0xfd,0x35] - i8x16.shr_u + # CHECK: f32x4.le # encoding: [0xfd,0x44] + f32x4.le - # CHECK: i16x8.shr_s # encoding: [0xfd,0x36] - i16x8.shr_s + # CHECK: f32x4.ge # encoding: [0xfd,0x45] + f32x4.ge - # CHECK: i16x8.shr_u # encoding: [0xfd,0x37] - i16x8.shr_u + # CHECK: f64x2.eq # encoding: [0xfd,0x46] + f64x2.eq - # CHECK: i32x4.shr_s # encoding: [0xfd,0x38] - i32x4.shr_s + # CHECK: f64x2.ne # encoding: [0xfd,0x47] + f64x2.ne - # CHECK: i32x4.shr_u # encoding: [0xfd,0x39] - i32x4.shr_u + # CHECK: f64x2.lt # encoding: [0xfd,0x48] + f64x2.lt - # CHECK: i64x2.shr_s # encoding: [0xfd,0x3a] - i64x2.shr_s + # CHECK: f64x2.gt # encoding: [0xfd,0x49] + f64x2.gt - # CHECK: i64x2.shr_u # encoding: [0xfd,0x3b] - i64x2.shr_u + # CHECK: f64x2.le # encoding: [0xfd,0x4a] + f64x2.le + + # CHECK: f64x2.ge # encoding: [0xfd,0x4b] + f64x2.ge - # CHECK: v128.and # encoding: [0xfd,0x3c] + # CHECK: v128.and # encoding: [0xfd,0x4c] v128.and - # CHECK: v128.or # encoding: [0xfd,0x3d] + # CHECK: v128.or # encoding: [0xfd,0x4d] v128.or - # CHECK: v128.xor # encoding: [0xfd,0x3e] + # CHECK: v128.xor # encoding: [0xfd,0x4e] v128.xor - # CHECK: v128.not # encoding: [0xfd,0x3f] + # CHECK: v128.not # encoding: [0xfd,0x4f] v128.not - # CHECK: v128.bitselect # encoding: [0xfd,0x40] + # CHECK: v128.bitselect # encoding: [0xfd,0x50] v128.bitselect - # CHECK: i8x16.any_true # encoding: [0xfd,0x41] - i8x16.any_true - - # CHECK: i16x8.any_true # encoding: [0xfd,0x42] - i16x8.any_true + # CHECK: i8x16.add # encoding: [0xfd,0x51] + i8x16.add - # CHECK: i32x4.any_true # encoding: [0xfd,0x43] - i32x4.any_true + # CHECK: i8x16.add_saturate_s # encoding: [0xfd,0x52] + i8x16.add_saturate_s - # CHECK: i64x2.any_true # encoding: [0xfd,0x44] - i64x2.any_true + # CHECK: i8x16.add_saturate_u # encoding: [0xfd,0x53] + i8x16.add_saturate_u - # CHECK: i8x16.all_true # encoding: [0xfd,0x45] - i8x16.all_true + # CHECK: i8x16.sub # encoding: [0xfd,0x54] + i8x16.sub - # CHECK: i16x8.all_true # encoding: [0xfd,0x46] - i16x8.all_true + # CHECK: i8x16.sub_saturate_s # encoding: [0xfd,0x55] + i8x16.sub_saturate_s - # CHECK: i32x4.all_true # encoding: [0xfd,0x47] - i32x4.all_true + # CHECK: i8x16.sub_saturate_u # encoding: [0xfd,0x56] + i8x16.sub_saturate_u - # CHECK: i64x2.all_true # encoding: [0xfd,0x48] - i64x2.all_true + # CHECK: i8x16.mul # encoding: [0xfd,0x57] + i8x16.mul - # CHECK: i8x16.eq # encoding: [0xfd,0x49] - i8x16.eq + # CHECK: i8x16.shl # encoding: [0xfd,0x5a] + i8x16.shl - # CHECK: i16x8.eq # encoding: [0xfd,0x4a] - i16x8.eq + # CHECK: i8x16.shr_s # encoding: [0xfd,0x5b] + i8x16.shr_s - # CHECK: i32x4.eq # encoding: [0xfd,0x4b] - i32x4.eq + # CHECK: i8x16.shr_u # encoding: [0xfd,0x5c] + i8x16.shr_u - # CHECK: f32x4.eq # encoding: [0xfd,0x4d] - f32x4.eq + # CHECK: i8x16.neg # encoding: [0xfd,0x5d] + i8x16.neg - # CHECK: f64x2.eq # encoding: [0xfd,0x4e] - f64x2.eq + # CHECK: i8x16.any_true # encoding: [0xfd,0x5e] + i8x16.any_true - # CHECK: i8x16.ne # encoding: [0xfd,0x4f] - i8x16.ne + # CHECK: i8x16.all_true # encoding: [0xfd,0x5f] + i8x16.all_true - # CHECK: i16x8.ne # encoding: [0xfd,0x50] - i16x8.ne + # CHECK: i16x8.add # encoding: [0xfd,0x60] + i16x8.add - # CHECK: i32x4.ne # encoding: [0xfd,0x51] - i32x4.ne + # CHECK: i16x8.add_saturate_s # encoding: [0xfd,0x61] + i16x8.add_saturate_s - # CHECK: f32x4.ne # encoding: [0xfd,0x53] - f32x4.ne + # CHECK: i16x8.add_saturate_u # encoding: [0xfd,0x62] + i16x8.add_saturate_u - # CHECK: f64x2.ne # encoding: [0xfd,0x54] - f64x2.ne + # CHECK: i16x8.sub # encoding: [0xfd,0x63] + i16x8.sub - # CHECK: i8x16.lt_s # encoding: [0xfd,0x55] - i8x16.lt_s + # CHECK: i16x8.sub_saturate_s # encoding: [0xfd,0x64] + i16x8.sub_saturate_s - # CHECK: i8x16.lt_u # encoding: [0xfd,0x56] - i8x16.lt_u + # CHECK: i16x8.sub_saturate_u # encoding: [0xfd,0x65] + i16x8.sub_saturate_u - # CHECK: i16x8.lt_s # encoding: [0xfd,0x57] - i16x8.lt_s + # CHECK: i16x8.mul # encoding: [0xfd,0x66] + i16x8.mul - # CHECK: i16x8.lt_u # encoding: [0xfd,0x58] - i16x8.lt_u + # CHECK: i16x8.shl # encoding: [0xfd,0x69] + i16x8.shl - # CHECK: i32x4.lt_s # encoding: [0xfd,0x59] - i32x4.lt_s + # CHECK: i16x8.shr_s # encoding: [0xfd,0x6a] + i16x8.shr_s - # CHECK: i32x4.lt_u # encoding: [0xfd,0x5a] - i32x4.lt_u + # CHECK: i16x8.shr_u # encoding: [0xfd,0x6b] + i16x8.shr_u - # CHECK: f32x4.lt # encoding: [0xfd,0x5d] - f32x4.lt + # CHECK: i16x8.neg # encoding: [0xfd,0x6c] + i16x8.neg - # CHECK: f64x2.lt # encoding: [0xfd,0x5e] - f64x2.lt + # CHECK: i16x8.any_true # encoding: [0xfd,0x6d] + i16x8.any_true - # CHECK: i8x16.le_s # encoding: [0xfd,0x5f] - i8x16.le_s + # CHECK: i16x8.all_true # encoding: [0xfd,0x6e] + i16x8.all_true - # CHECK: i8x16.le_u # encoding: [0xfd,0x60] - i8x16.le_u + # CHECK: i32x4.add # encoding: [0xfd,0x6f] + i32x4.add - # CHECK: i16x8.le_s # encoding: [0xfd,0x61] - i16x8.le_s + # CHECK: i32x4.sub # encoding: [0xfd,0x72] + i32x4.sub - # CHECK: i16x8.le_u # encoding: [0xfd,0x62] - i16x8.le_u + # CHECK: i32x4.mul # encoding: [0xfd,0x75] + i32x4.mul - # CHECK: i32x4.le_s # encoding: [0xfd,0x63] - i32x4.le_s + # CHECK: i32x4.shl # encoding: [0xfd,0x78] + i32x4.shl - # CHECK: i32x4.le_u # encoding: [0xfd,0x64] - i32x4.le_u + # CHECK: i32x4.shr_s # encoding: [0xfd,0x79] + i32x4.shr_s - # CHECK: f32x4.le # encoding: [0xfd,0x67] - f32x4.le + # CHECK: i32x4.shr_u # encoding: [0xfd,0x7a] + i32x4.shr_u - # CHECK: f64x2.le # encoding: [0xfd,0x68] - f64x2.le + # CHECK: i32x4.neg # encoding: [0xfd,0x7b] + i32x4.neg - # CHECK: i8x16.gt_s # encoding: [0xfd,0x69] - i8x16.gt_s + # CHECK: i32x4.any_true # encoding: [0xfd,0x7c] + i32x4.any_true - # CHECK: i8x16.gt_u # encoding: [0xfd,0x6a] - i8x16.gt_u + # CHECK: i32x4.all_true # encoding: [0xfd,0x7d] + i32x4.all_true - # CHECK: i16x8.gt_s # encoding: [0xfd,0x6b] - i16x8.gt_s + # CHECK: i64x2.add # encoding: [0xfd,0x7e] + i64x2.add - # CHECK: i16x8.gt_u # encoding: [0xfd,0x6c] - i16x8.gt_u + # CHECK: i64x2.sub # encoding: [0xfd,0x81,0x01] + i64x2.sub - # CHECK: i32x4.gt_s # encoding: [0xfd,0x6d] - i32x4.gt_s + # CHECK: i64x2.shl # encoding: [0xfd,0x87,0x01] + i64x2.shl - # CHECK: i32x4.gt_u # encoding: [0xfd,0x6e] - i32x4.gt_u + # CHECK: i64x2.shr_s # encoding: [0xfd,0x88,0x01] + i64x2.shr_s - # CHECK: f32x4.gt # encoding: [0xfd,0x71] - f32x4.gt + # CHECK: i64x2.shr_u # encoding: [0xfd,0x89,0x01] + i64x2.shr_u - # CHECK: f64x2.gt # encoding: [0xfd,0x72] - f64x2.gt + # CHECK: i64x2.neg # encoding: [0xfd,0x8a,0x01] + i64x2.neg - # CHECK: i8x16.ge_s # encoding: [0xfd,0x73] - i8x16.ge_s + # CHECK: i64x2.any_true # encoding: [0xfd,0x8b,0x01] + i64x2.any_true - # CHECK: i8x16.ge_u # encoding: [0xfd,0x74] - i8x16.ge_u + # CHECK: i64x2.all_true # encoding: [0xfd,0x8c,0x01] + i64x2.all_true - # CHECK: i16x8.ge_s # encoding: [0xfd,0x75] - i16x8.ge_s + # CHECK: f32x4.add # encoding: [0xfd,0x8d,0x01] + f32x4.add - # CHECK: i16x8.ge_u # encoding: [0xfd,0x76] - i16x8.ge_u + # CHECK: f32x4.sub # encoding: [0xfd,0x8e,0x01] + f32x4.sub - # CHECK: i32x4.ge_s # encoding: [0xfd,0x77] - i32x4.ge_s + # CHECK: f32x4.mul # encoding: [0xfd,0x8f,0x01] + f32x4.mul - # CHECK: i32x4.ge_u # encoding: [0xfd,0x78] - i32x4.ge_u + # CHECK: f32x4.div # encoding: [0xfd,0x90,0x01] + f32x4.div - # CHECK: f32x4.ge # encoding: [0xfd,0x7b] - f32x4.ge + # CHECK: f32x4.min # encoding: [0xfd,0x91,0x01] + f32x4.min - # CHECK: f64x2.ge # encoding: [0xfd,0x7c] - f64x2.ge + # CHECK: f32x4.max # encoding: [0xfd,0x92,0x01] + f32x4.max - # CHECK: f32x4.neg # encoding: [0xfd,0x7d] + # CHECK: f32x4.neg # encoding: [0xfd,0x93,0x01] f32x4.neg - # CHECK: f64x2.neg # encoding: [0xfd,0x7e] - f64x2.neg - - # CHECK: f32x4.abs # encoding: [0xfd,0x7f] + # CHECK: f32x4.abs # encoding: [0xfd,0x94,0x01] f32x4.abs - # CHECK: f64x2.abs # encoding: [0xfd,0x80] - f64x2.abs - - # CHECK: f32x4.min # encoding: [0xfd,0x81] - f32x4.min + # CHECK: f32x4.sqrt # encoding: [0xfd,0x95,0x01] + f32x4.sqrt - # CHECK: f64x2.min # encoding: [0xfd,0x82] - f64x2.min + # CHECK: f64x2.add # encoding: [0xfd,0x96,0x01] + f64x2.add - # CHECK: f32x4.max # encoding: [0xfd,0x83] - f32x4.max + # CHECK: f64x2.sub # encoding: [0xfd,0x97,0x01] + f64x2.sub - # CHECK: f64x2.max # encoding: [0xfd,0x84] - f64x2.max + # CHECK: f64x2.mul # encoding: [0xfd,0x98,0x01] + f64x2.mul - # CHECK: f32x4.add # encoding: [0xfd,0x85] - f32x4.add + # CHECK: f64x2.div # encoding: [0xfd,0x99,0x01] + f64x2.div - # CHECK: f64x2.add # encoding: [0xfd,0x86] - f64x2.add + # CHECK: f64x2.min # encoding: [0xfd,0x9a,0x01] + f64x2.min - # CHECK: f32x4.sub # encoding: [0xfd,0x87] - f32x4.sub + # CHECK: f64x2.max # encoding: [0xfd,0x9b,0x01] + f64x2.max - # CHECK: f64x2.sub # encoding: [0xfd,0x88] - f64x2.sub + # CHECK: f64x2.neg # encoding: [0xfd,0x9c,0x01] + f64x2.neg - # CHECK: f32x4.div # encoding: [0xfd,0x89] - f32x4.div + # CHECK: f64x2.abs # encoding: [0xfd,0x9d,0x01] + f64x2.abs - # CHECK: f64x2.div # encoding: [0xfd,0x8a] - f64x2.div + # CHECK: f64x2.sqrt # encoding: [0xfd,0x9e,0x01] + f64x2.sqrt - # CHECK: f32x4.mul # encoding: [0xfd,0x8b] - f32x4.mul + # CHECK: i32x4.trunc_sat_s/f32x4 # encoding: [0xfd,0x9f,0x01] + i32x4.trunc_sat_s/f32x4 - # CHECK: f64x2.mul # encoding: [0xfd,0x8c] - f64x2.mul + # CHECK: i32x4.trunc_sat_u/f32x4 # encoding: [0xfd,0xa0,0x01] + i32x4.trunc_sat_u/f32x4 - # CHECK: f32x4.sqrt # encoding: [0xfd,0x8d] - f32x4.sqrt + # CHECK: i64x2.trunc_sat_s/f64x2 # encoding: [0xfd,0xa1,0x01] + i64x2.trunc_sat_s/f64x2 - # CHECK: f64x2.sqrt # encoding: [0xfd,0x8e] - f64x2.sqrt + # CHECK: i64x2.trunc_sat_u/f64x2 # encoding: [0xfd,0xa2,0x01] + i64x2.trunc_sat_u/f64x2 - # CHECK: f32x4.convert_s/i32x4 # encoding: [0xfd,0x8f] + # CHECK: f32x4.convert_s/i32x4 # encoding: [0xfd,0xa3,0x01] f32x4.convert_s/i32x4 - # CHECK: f32x4.convert_u/i32x4 # encoding: [0xfd,0x90] + # CHECK: f32x4.convert_u/i32x4 # encoding: [0xfd,0xa4,0x01] f32x4.convert_u/i32x4 - # CHECK: f64x2.convert_s/i64x2 # encoding: [0xfd,0x91] + # CHECK: f64x2.convert_s/i64x2 # encoding: [0xfd,0xa5,0x01] f64x2.convert_s/i64x2 - # CHECK: f64x2.convert_u/i64x2 # encoding: [0xfd,0x92] + # CHECK: f64x2.convert_u/i64x2 # encoding: [0xfd,0xa6,0x01] f64x2.convert_u/i64x2 - # CHECK: i32x4.trunc_sat_s/f32x4 # encoding: [0xfd,0x93] - i32x4.trunc_sat_s/f32x4 - - # CHECK: i32x4.trunc_sat_u/f32x4 # encoding: [0xfd,0x94] - i32x4.trunc_sat_u/f32x4 - - # CHECK: i64x2.trunc_sat_s/f64x2 # encoding: [0xfd,0x95] - i64x2.trunc_sat_s/f64x2 - - # CHECK: i64x2.trunc_sat_u/f64x2 # encoding: [0xfd,0x96] - i64x2.trunc_sat_u/f64x2 - end_function