Lines Matching defs:__b
284 #define wasm_i8x16_replace_lane(__a, __i, __b) \ argument
297 #define wasm_i16x8_replace_lane(__a, __i, __b) \ argument
307 #define wasm_i32x4_replace_lane(__a, __i, __b) \ argument
317 #define wasm_i64x2_replace_lane(__a, __i, __b) \ argument
327 #define wasm_f32x4_replace_lane(__a, __i, __b) \ argument
337 #define wasm_f64x2_replace_lane(__a, __i, __b) \ argument
341 v128_t __b) { in wasm_i8x16_eq()
346 v128_t __b) { in wasm_i8x16_ne()
351 v128_t __b) { in wasm_i8x16_lt()
356 v128_t __b) { in wasm_u8x16_lt()
361 v128_t __b) { in wasm_i8x16_gt()
366 v128_t __b) { in wasm_u8x16_gt()
371 v128_t __b) { in wasm_i8x16_le()
376 v128_t __b) { in wasm_u8x16_le()
381 v128_t __b) { in wasm_i8x16_ge()
386 v128_t __b) { in wasm_u8x16_ge()
391 v128_t __b) { in wasm_i16x8_eq()
396 v128_t __b) { in wasm_i16x8_ne()
401 v128_t __b) { in wasm_i16x8_lt()
406 v128_t __b) { in wasm_u16x8_lt()
411 v128_t __b) { in wasm_i16x8_gt()
416 v128_t __b) { in wasm_u16x8_gt()
421 v128_t __b) { in wasm_i16x8_le()
426 v128_t __b) { in wasm_u16x8_le()
431 v128_t __b) { in wasm_i16x8_ge()
436 v128_t __b) { in wasm_u16x8_ge()
441 v128_t __b) { in wasm_i32x4_eq()
446 v128_t __b) { in wasm_i32x4_ne()
451 v128_t __b) { in wasm_i32x4_lt()
456 v128_t __b) { in wasm_u32x4_lt()
461 v128_t __b) { in wasm_i32x4_gt()
466 v128_t __b) { in wasm_u32x4_gt()
471 v128_t __b) { in wasm_i32x4_le()
476 v128_t __b) { in wasm_u32x4_le()
481 v128_t __b) { in wasm_i32x4_ge()
486 v128_t __b) { in wasm_u32x4_ge()
491 v128_t __b) { in wasm_f32x4_eq()
496 v128_t __b) { in wasm_f32x4_ne()
501 v128_t __b) { in wasm_f32x4_lt()
506 v128_t __b) { in wasm_f32x4_gt()
511 v128_t __b) { in wasm_f32x4_le()
516 v128_t __b) { in wasm_f32x4_ge()
521 v128_t __b) { in wasm_f64x2_eq()
526 v128_t __b) { in wasm_f64x2_ne()
531 v128_t __b) { in wasm_f64x2_lt()
536 v128_t __b) { in wasm_f64x2_gt()
541 v128_t __b) { in wasm_f64x2_le()
546 v128_t __b) { in wasm_f64x2_ge()
555 v128_t __b) { in wasm_v128_and()
560 v128_t __b) { in wasm_v128_or()
565 v128_t __b) { in wasm_v128_xor()
570 v128_t __b) { in wasm_v128_andnot()
575 v128_t __b, in wasm_v128_bitselect()
598 int32_t __b) { in wasm_i8x16_shl()
603 int32_t __b) { in wasm_i8x16_shr()
608 int32_t __b) { in wasm_u8x16_shr()
613 v128_t __b) { in wasm_i8x16_add()
618 wasm_i8x16_add_saturate(v128_t __a, v128_t __b) { in wasm_i8x16_add_saturate()
624 wasm_u8x16_add_saturate(v128_t __a, v128_t __b) { in wasm_u8x16_add_saturate()
630 v128_t __b) { in wasm_i8x16_sub()
635 wasm_i8x16_sub_saturate(v128_t __a, v128_t __b) { in wasm_i8x16_sub_saturate()
641 wasm_u8x16_sub_saturate(v128_t __a, v128_t __b) { in wasm_u8x16_sub_saturate()
647 v128_t __b) { in wasm_i8x16_min()
652 v128_t __b) { in wasm_u8x16_min()
657 v128_t __b) { in wasm_i8x16_max()
662 v128_t __b) { in wasm_u8x16_max()
667 v128_t __b) { in wasm_u8x16_avgr()
688 int32_t __b) { in wasm_i16x8_shl()
693 int32_t __b) { in wasm_i16x8_shr()
698 int32_t __b) { in wasm_u16x8_shr()
703 v128_t __b) { in wasm_i16x8_add()
708 wasm_i16x8_add_saturate(v128_t __a, v128_t __b) { in wasm_i16x8_add_saturate()
714 wasm_u16x8_add_saturate(v128_t __a, v128_t __b) { in wasm_u16x8_add_saturate()
720 v128_t __b) { in wasm_i16x8_sub()
725 wasm_i16x8_sub_saturate(v128_t __a, v128_t __b) { in wasm_i16x8_sub_saturate()
731 wasm_u16x8_sub_saturate(v128_t __a, v128_t __b) { in wasm_u16x8_sub_saturate()
737 v128_t __b) { in wasm_i16x8_mul()
742 v128_t __b) { in wasm_i16x8_min()
747 v128_t __b) { in wasm_u16x8_min()
752 v128_t __b) { in wasm_i16x8_max()
757 v128_t __b) { in wasm_u16x8_max()
762 v128_t __b) { in wasm_u16x8_avgr()
783 int32_t __b) { in wasm_i32x4_shl()
788 int32_t __b) { in wasm_i32x4_shr()
793 int32_t __b) { in wasm_u32x4_shr()
798 v128_t __b) { in wasm_i32x4_add()
803 v128_t __b) { in wasm_i32x4_sub()
808 v128_t __b) { in wasm_i32x4_mul()
813 v128_t __b) { in wasm_i32x4_min()
818 v128_t __b) { in wasm_u32x4_min()
823 v128_t __b) { in wasm_i32x4_max()
828 v128_t __b) { in wasm_u32x4_max()
849 int32_t __b) { in wasm_i64x2_shl()
854 int32_t __b) { in wasm_i64x2_shr()
859 int32_t __b) { in wasm_u64x2_shr()
864 v128_t __b) { in wasm_i64x2_add()
869 v128_t __b) { in wasm_i64x2_sub()
874 v128_t __b) { in wasm_i64x2_mul()
893 v128_t __b, in wasm_f32x4_qfma()
900 v128_t __b, in wasm_f32x4_qfms()
909 v128_t __b) { in wasm_f32x4_add()
914 v128_t __b) { in wasm_f32x4_sub()
919 v128_t __b) { in wasm_f32x4_mul()
924 v128_t __b) { in wasm_f32x4_div()
929 v128_t __b) { in wasm_f32x4_min()
934 v128_t __b) { in wasm_f32x4_max()
939 v128_t __b) { in wasm_f32x4_pmin()
944 v128_t __b) { in wasm_f32x4_pmax()
963 v128_t __b, in wasm_f64x2_qfma()
970 v128_t __b, in wasm_f64x2_qfms()
979 v128_t __b) { in wasm_f64x2_add()
984 v128_t __b) { in wasm_f64x2_sub()
989 v128_t __b) { in wasm_f64x2_mul()
994 v128_t __b) { in wasm_f64x2_div()
999 v128_t __b) { in wasm_f64x2_min()
1004 v128_t __b) { in wasm_f64x2_max()
1009 v128_t __b) { in wasm_f64x2_pmin()
1014 v128_t __b) { in wasm_f64x2_pmax()
1038 #define wasm_v8x16_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ argument
1045 #define wasm_v16x8_shuffle(__a, __b, __c0, __c1, __c2, __c3, __c4, __c5, __c6, \ argument
1053 #define wasm_v32x4_shuffle(__a, __b, __c0, __c1, __c2, __c3) \ argument
1060 #define wasm_v64x2_shuffle(__a, __b, __c0, __c1) \ argument
1068 v128_t __b) { in wasm_v8x16_swizzle()
1073 wasm_i8x16_narrow_i16x8(v128_t __a, v128_t __b) { in wasm_i8x16_narrow_i16x8()
1079 wasm_u8x16_narrow_i16x8(v128_t __a, v128_t __b) { in wasm_u8x16_narrow_i16x8()
1085 wasm_i16x8_narrow_i32x4(v128_t __a, v128_t __b) { in wasm_i16x8_narrow_i32x4()
1091 wasm_u16x8_narrow_i32x4(v128_t __a, v128_t __b) { in wasm_u16x8_narrow_i32x4()