Home | History | Annotate | Download | only in priv

Lines Matching refs:xx

142 static inline Short qadd16S ( Short xx, Short yy ) 
144 Int t = ((Int)xx) + ((Int)yy);
150 static inline Char qadd8S ( Char xx, Char yy )
152 Int t = ((Int)xx) + ((Int)yy);
158 static inline UShort qadd16U ( UShort xx, UShort yy )
160 UInt t = ((UInt)xx) + ((UInt)yy);
165 static inline UChar qadd8U ( UChar xx, UChar yy )
167 UInt t = ((UInt)xx) + ((UInt)yy);
172 static inline Short qsub16S ( Short xx, Short yy )
174 Int t = ((Int)xx) - ((Int)yy);
180 static inline Char qsub8S ( Char xx, Char yy )
182 Int t = ((Int)xx) - ((Int)yy);
188 static inline UShort qsub16U ( UShort xx, UShort yy )
190 Int t = ((Int)xx) - ((Int)yy);
196 static inline UChar qsub8U ( UChar xx, UChar yy )
198 Int t = ((Int)xx) - ((Int)yy);
204 static inline Short mul16 ( Short xx, Short yy )
206 Int t = ((Int)xx) * ((Int)yy);
210 static inline Int mul32 ( Int xx, Int yy )
212 Int t = ((Int)xx) * ((Int)yy);
216 static inline Short mulhi16S ( Short xx, Short yy )
218 Int t = ((Int)xx) * ((Int)yy);
223 static inline UShort mulhi16U ( UShort xx, UShort yy )
225 UInt t = ((UInt)xx) * ((UInt)yy);
230 static inline UInt cmpeq32 ( UInt xx, UInt yy )
232 return xx==yy ? 0xFFFFFFFF : 0;
235 static inline UShort cmpeq16 ( UShort xx, UShort yy )
237 return toUShort(xx==yy ? 0xFFFF : 0);
240 static inline UChar cmpeq8 ( UChar xx, UChar yy )
242 return toUChar(xx==yy ? 0xFF : 0);
245 static inline UInt cmpgt32S ( Int xx, Int yy )
247 return xx>yy ? 0xFFFFFFFF : 0;
250 static inline UShort cmpgt16S ( Short xx, Short yy )
252 return toUShort(xx>yy ? 0xFFFF : 0);
255 static inline UChar cmpgt8S ( Char xx, Char yy )
257 return toUChar(xx>yy ? 0xFF : 0);
260 static inline UInt cmpnez32 ( UInt xx )
262 return xx==0 ? 0 : 0xFFFFFFFF;
265 static inline UShort cmpnez16 ( UShort xx )
267 return toUShort(xx==0 ? 0 : 0xFFFF);
270 static inline UChar cmpnez8 ( UChar xx )
272 return toUChar(xx==0 ? 0 : 0xFF);
277 Int xx = (Int)xx0;
278 if (xx < -32768) xx = -32768;
279 if (xx > 32767) xx = 32767;
280 return (Short)xx;
285 Short xx = (Short)xx0;
286 if (xx < -128) xx = -128;
287 if (xx > 127) xx = 127;
288 return (Char)xx;
293 Short xx = (Short)xx0;
294 if (xx < 0) xx = 0;
295 if (xx > 255) xx = 255;
296 return (UChar)xx;
342 static inline UChar avg8U ( UChar xx, UChar yy )
344 UInt xxi = (UInt)xx;
350 static inline UShort avg16U ( UShort xx, UShort yy )
352 UInt xxi = (UInt)xx;
358 static inline Short max16S ( Short xx, Short yy )
360 return toUShort((xx > yy) ? xx : yy);
363 static inline UChar max8U ( UChar xx, UChar yy )
365 return toUChar((xx > yy) ? xx : yy);
368 static inline Short min16S ( Short xx, Short yy )
370 return toUShort((xx < yy) ? xx : yy);
373 static inline UChar min8U ( UChar xx, UChar yy )
375 return toUChar((xx < yy) ? xx : yy);
378 static inline UShort hadd16U ( UShort xx, UShort yy )
380 UInt xxi = (UInt)xx;
386 static inline Short hadd16S ( Short xx, Short yy )
388 Int xxi = (Int)xx;
394 static inline UShort hsub16U ( UShort xx, UShort yy )
396 UInt xxi = (UInt)xx;
402 static inline Short hsub16S ( Short xx, Short yy )
404 Int xxi = (Int)xx;
410 static inline UChar hadd8U ( UChar xx, UChar yy )
412 UInt xxi = (UInt)xx;
418 static inline Char hadd8S ( Char xx, Char yy )
420 Int xxi = (Int)xx;
426 static inline UChar hsub8U ( UChar xx, UChar yy )
428 UInt xxi = (UInt)xx;
434 static inline Char hsub8S ( Char xx, Char yy )
436 Int xxi = (Int)xx;
442 static inline UInt absdiff8U ( UChar xx, UChar yy )
444 UInt xxu = (UChar)xx;
456 ULong h_generic_calc_Add32x2 ( ULong xx, ULong yy )
459 sel32x2_1(xx) + sel32x2_1(yy),
460 sel32x2_0(xx) + sel32x2_0(yy)
464 ULong h_generic_calc_Add16x4 ( ULong xx, ULong yy )
467 toUShort( sel16x4_3(xx) + sel16x4_3(yy) ),
468 toUShort( sel16x4_2(xx) + sel16x4_2(yy) ),
469 toUShort( sel16x4_1(xx) + sel16x4_1(yy) ),
470 toUShort( sel16x4_0(xx) + sel16x4_0(yy) )
474 ULong h_generic_calc_Add8x8 ( ULong xx, ULong yy )
477 toUChar( sel8x8_7(xx) + sel8x8_7(yy) ),
478 toUChar( sel8x8_6(xx) + sel8x8_6(yy) ),
479 toUChar( sel8x8_5(xx) + sel8x8_5(yy) ),
480 toUChar( sel8x8_4(xx) + sel8x8_4(yy) ),
481 toUChar( sel8x8_3(xx) + sel8x8_3(yy) ),
482 toUChar( sel8x8_2(xx) + sel8x8_2(yy) ),
483 toUChar( sel8x8_1(xx) + sel8x8_1(yy) ),
484 toUChar( sel8x8_0(xx) + sel8x8_0(yy) )
490 ULong h_generic_calc_QAdd16Sx4 ( ULong xx, ULong yy )
493 qadd16S( sel16x4_3(xx), sel16x4_3(yy) ),
494 qadd16S( sel16x4_2(xx), sel16x4_2(yy) ),
495 qadd16S( sel16x4_1(xx), sel16x4_1(yy) ),
496 qadd16S( sel16x4_0(xx), sel16x4_0(yy) )
500 ULong h_generic_calc_QAdd8Sx8 ( ULong xx, ULong yy )
503 qadd8S( sel8x8_7(xx), sel8x8_7(yy) ),
504 qadd8S( sel8x8_6(xx), sel8x8_6(yy) ),
505 qadd8S( sel8x8_5(xx), sel8x8_5(yy) ),
506 qadd8S( sel8x8_4(xx), sel8x8_4(yy) ),
507 qadd8S( sel8x8_3(xx), sel8x8_3(yy) ),
508 qadd8S( sel8x8_2(xx), sel8x8_2(yy) ),
509 qadd8S( sel8x8_1(xx), sel8x8_1(yy) ),
510 qadd8S( sel8x8_0(xx), sel8x8_0(yy) )
514 ULong h_generic_calc_QAdd16Ux4 ( ULong xx, ULong yy )
517 qadd16U( sel16x4_3(xx), sel16x4_3(yy) ),
518 qadd16U( sel16x4_2(xx), sel16x4_2(yy) ),
519 qadd16U( sel16x4_1(xx), sel16x4_1(yy) ),
520 qadd16U( sel16x4_0(xx), sel16x4_0(yy) )
524 ULong h_generic_calc_QAdd8Ux8 ( ULong xx, ULong yy )
527 qadd8U( sel8x8_7(xx), sel8x8_7(yy) ),
528 qadd8U( sel8x8_6(xx), sel8x8_6(yy) ),
529 qadd8U( sel8x8_5(xx), sel8x8_5(yy) ),
530 qadd8U( sel8x8_4(xx), sel8x8_4(yy) ),
531 qadd8U( sel8x8_3(xx), sel8x8_3(yy) ),
532 qadd8U( sel8x8_2(xx), sel8x8_2(yy) ),
533 qadd8U( sel8x8_1(xx), sel8x8_1(yy) ),
534 qadd8U( sel8x8_0(xx), sel8x8_0(yy) )
540 ULong h_generic_calc_Sub32x2 ( ULong xx, ULong yy )
543 sel32x2_1(xx) - sel32x2_1(yy),
544 sel32x2_0(xx) - sel32x2_0(yy)
548 ULong h_generic_calc_Sub16x4 ( ULong xx, ULong yy )
551 toUShort( sel16x4_3(xx) - sel16x4_3(yy) ),
552 toUShort( sel16x4_2(xx) - sel16x4_2(yy) ),
553 toUShort( sel16x4_1(xx) - sel16x4_1(yy) ),
554 toUShort( sel16x4_0(xx) - sel16x4_0(yy) )
558 ULong h_generic_calc_Sub8x8 ( ULong xx, ULong yy )
561 toUChar( sel8x8_7(xx) - sel8x8_7(yy) ),
562 toUChar( sel8x8_6(xx) - sel8x8_6(yy) ),
563 toUChar( sel8x8_5(xx) - sel8x8_5(yy) ),
564 toUChar( sel8x8_4(xx) - sel8x8_4(yy) ),
565 toUChar( sel8x8_3(xx) - sel8x8_3(yy) ),
566 toUChar( sel8x8_2(xx) - sel8x8_2(yy) ),
567 toUChar( sel8x8_1(xx) - sel8x8_1(yy) ),
568 toUChar( sel8x8_0(xx) - sel8x8_0(yy) )
574 ULong h_generic_calc_QSub16Sx4 ( ULong xx, ULong yy )
577 qsub16S( sel16x4_3(xx), sel16x4_3(yy) ),
578 qsub16S( sel16x4_2(xx), sel16x4_2(yy) ),
579 qsub16S( sel16x4_1(xx), sel16x4_1(yy) ),
580 qsub16S( sel16x4_0(xx), sel16x4_0(yy) )
584 ULong h_generic_calc_QSub8Sx8 ( ULong xx, ULong yy )
587 qsub8S( sel8x8_7(xx), sel8x8_7(yy) ),
588 qsub8S( sel8x8_6(xx), sel8x8_6(yy) ),
589 qsub8S( sel8x8_5(xx), sel8x8_5(yy) ),
590 qsub8S( sel8x8_4(xx), sel8x8_4(yy) ),
591 qsub8S( sel8x8_3(xx), sel8x8_3(yy) ),
592 qsub8S( sel8x8_2(xx), sel8x8_2(yy) ),
593 qsub8S( sel8x8_1(xx), sel8x8_1(yy) ),
594 qsub8S( sel8x8_0(xx), sel8x8_0(yy) )
598 ULong h_generic_calc_QSub16Ux4 ( ULong xx, ULong yy )
601 qsub16U( sel16x4_3(xx), sel16x4_3(yy) ),
602 qsub16U( sel16x4_2(xx), sel16x4_2(yy) ),
603 qsub16U( sel16x4_1(xx), sel16x4_1(yy) ),
604 qsub16U( sel16x4_0(xx), sel16x4_0(yy) )
608 ULong h_generic_calc_QSub8Ux8 ( ULong xx, ULong yy )
611 qsub8U( sel8x8_7(xx), sel8x8_7(yy) ),
612 qsub8U( sel8x8_6(xx), sel8x8_6(yy) ),
613 qsub8U( sel8x8_5(xx), sel8x8_5(yy) ),
614 qsub8U( sel8x8_4(xx), sel8x8_4(yy) ),
615 qsub8U( sel8x8_3(xx), sel8x8_3(yy) ),
616 qsub8U( sel8x8_2(xx), sel8x8_2(yy) ),
617 qsub8U( sel8x8_1(xx), sel8x8_1(yy) ),
618 qsub8U( sel8x8_0(xx), sel8x8_0(yy) )
624 ULong h_generic_calc_Mul16x4 ( ULong xx, ULong yy )
627 mul16( sel16x4_3(xx), sel16x4_3(yy) ),
628 mul16( sel16x4_2(xx), sel16x4_2(yy) ),
629 mul16( sel16x4_1(xx), sel16x4_1(yy) ),
630 mul16( sel16x4_0(xx), sel16x4_0(yy) )
634 ULong h_generic_calc_Mul32x2 ( ULong xx, ULong yy )
637 mul32( sel32x2_1(xx), sel32x2_1(yy) ),
638 mul32( sel32x2_0(xx
642 ULong h_generic_calc_MulHi16Sx4 ( ULong xx, ULong yy )
645 mulhi16S( sel16x4_3(xx), sel16x4_3(yy) ),
646 mulhi16S( sel16x4_2(xx), sel16x4_2(yy) ),
647 mulhi16S( sel16x4_1(xx), sel16x4_1(yy) ),
648 mulhi16S( sel16x4_0(xx), sel16x4_0(yy) )
652 ULong h_generic_calc_MulHi16Ux4 ( ULong xx, ULong yy )
655 mulhi16U( sel16x4_3(xx), sel16x4_3(yy) ),
656 mulhi16U( sel16x4_2(xx), sel16x4_2(yy) ),
657 mulhi16U( sel16x4_1(xx), sel16x4_1(yy) ),
658 mulhi16U( sel16x4_0(xx), sel16x4_0(yy) )
664 ULong h_generic_calc_CmpEQ32x2 ( ULong xx, ULong yy )
667 cmpeq32( sel32x2_1(xx), sel32x2_1(yy) ),
668 cmpeq32( sel32x2_0(xx), sel32x2_0(yy) )
672 ULong h_generic_calc_CmpEQ16x4 ( ULong xx, ULong yy )
675 cmpeq16( sel16x4_3(xx), sel16x4_3(yy) ),
676 cmpeq16( sel16x4_2(xx), sel16x4_2(yy) ),
677 cmpeq16( sel16x4_1(xx), sel16x4_1(yy) ),
678 cmpeq16( sel16x4_0(xx), sel16x4_0(yy) )
682 ULong h_generic_calc_CmpEQ8x8 ( ULong xx, ULong yy )
685 cmpeq8( sel8x8_7(xx), sel8x8_7(yy) ),
686 cmpeq8( sel8x8_6(xx), sel8x8_6(yy) ),
687 cmpeq8( sel8x8_5(xx), sel8x8_5(yy) ),
688 cmpeq8( sel8x8_4(xx), sel8x8_4(yy) ),
689 cmpeq8( sel8x8_3(xx), sel8x8_3(yy) ),
690 cmpeq8( sel8x8_2(xx), sel8x8_2(yy) ),
691 cmpeq8( sel8x8_1(xx), sel8x8_1(yy) ),
692 cmpeq8( sel8x8_0(xx), sel8x8_0(yy) )
696 ULong h_generic_calc_CmpGT32Sx2 ( ULong xx, ULong yy )
699 cmpgt32S( sel32x2_1(xx), sel32x2_1(yy) ),
700 cmpgt32S( sel32x2_0(xx), sel32x2_0(yy) )
704 ULong h_generic_calc_CmpGT16Sx4 ( ULong xx, ULong yy )
707 cmpgt16S( sel16x4_3(xx), sel16x4_3(yy) ),
708 cmpgt16S( sel16x4_2(xx), sel16x4_2(yy) ),
709 cmpgt16S( sel16x4_1(xx), sel16x4_1(yy) ),
710 cmpgt16S( sel16x4_0(xx), sel16x4_0(yy) )
714 ULong h_generic_calc_CmpGT8Sx8 ( ULong xx, ULong yy )
717 cmpgt8S( sel8x8_7(xx), sel8x8_7(yy) ),
718 cmpgt8S( sel8x8_6(xx), sel8x8_6(yy) ),
719 cmpgt8S( sel8x8_5(xx), sel8x8_5(yy) ),
720 cmpgt8S( sel8x8_4(xx), sel8x8_4(yy) ),
721 cmpgt8S( sel8x8_3(xx), sel8x8_3(yy) ),
722 cmpgt8S( sel8x8_2(xx), sel8x8_2(yy) ),
723 cmpgt8S( sel8x8_1(xx), sel8x8_1(yy) ),
724 cmpgt8S( sel8x8_0(xx), sel8x8_0(yy) )
728 ULong h_generic_calc_CmpNEZ32x2 ( ULong xx )
731 cmpnez32( sel32x2_1(xx) ),
732 cmpnez32( sel32x2_0(xx) )
736 ULong h_generic_calc_CmpNEZ16x4 ( ULong xx )
739 cmpnez16( sel16x4_3(xx) ),
740 cmpnez16( sel16x4_2(xx) ),
741 cmpnez16( sel16x4_1(xx) ),
742 cmpnez16( sel16x4_0(xx) )
746 ULong h_generic_calc_CmpNEZ8x8 ( ULong xx )
749 cmpnez8( sel8x8_7(xx) ),
750 cmpnez8( sel8x8_6(xx) ),
751 cmpnez8( sel8x8_5(xx) ),
752 cmpnez8( sel8x8_4(xx) ),
753 cmpnez8( sel8x8_3(xx) ),
754 cmpnez8( sel8x8_2(xx) ),
755 cmpnez8( sel8x8_1(xx) ),
756 cmpnez8( sel8x8_0(xx) )
930 ULong h_generic_calc_ShlN32x2 ( ULong xx, UInt nn )
935 shl32( sel32x2_1(xx), nn ),
936 shl32( sel32x2_0(xx), nn )
940 ULong h_generic_calc_ShlN16x4 ( ULong xx, UInt nn )
945 shl16( sel16x4_3(xx), nn ),
946 shl16( sel16x4_2(xx), nn ),
947 shl16( sel16x4_1(xx), nn ),
948 shl16( sel16x4_0(xx), nn )
952 ULong h_generic_calc_ShlN8x8 ( ULong xx, UInt nn )
957 shl8( sel8x8_7(xx), nn ),
958 shl8( sel8x8_6(xx), nn ),
959 shl8( sel8x8_5(xx), nn ),
960 shl8( sel8x8_4(xx), nn ),
961 shl8( sel8x8_3(xx), nn ),
962 shl8( sel8x8_2(xx), nn ),
963 shl8( sel8x8_1(xx), nn ),
964 shl8( sel8x8_0(xx), nn )
968 ULong h_generic_calc_ShrN32x2 ( ULong xx, UInt nn )
973 shr32( sel32x2_1(xx), nn ),
974 shr32( sel32x2_0(xx), nn )
978 ULong h_generic_calc_ShrN16x4 ( ULong xx, UInt nn )
983 shr16( sel16x4_3(xx), nn ),
984 shr16( sel16x4_2(xx), nn ),
985 shr16( sel16x4_1(xx), nn ),
986 shr16( sel16x4_0(xx), nn )
990 ULong h_generic_calc_SarN32x2 ( ULong xx, UInt nn )
995 sar32( sel32x2_1(xx), nn ),
996 sar32( sel32x2_0(xx), nn )
1000 ULong h_generic_calc_SarN16x4 ( ULong xx, UInt nn )
1005 sar16( sel16x4_3(xx), nn ),
1006 sar16( sel16x4_2(xx), nn ),
1007 sar16( sel16x4_1(xx), nn ),
1008 sar16( sel16x4_0(xx), nn )
1012 ULong h_generic_calc_SarN8x8 ( ULong xx, UInt nn )
1017 sar8( sel8x8_7(xx), nn ),
1018 sar8( sel8x8_6(xx), nn ),
1019 sar8( sel8x8_5(xx), nn ),
1020 sar8( sel8x8_4(xx), nn ),
1021 sar8( sel8x8_3(xx), nn ),
1022 sar8( sel8x8_2(xx), nn ),
1023 sar8( sel8x8_1(xx), nn ),
1024 sar8( sel8x8_0(xx), nn )
1030 ULong h_generic_calc_Avg8Ux8 ( ULong xx, ULong yy )
1033 avg8U( sel8x8_7(xx), sel8x8_7(yy) ),
1034 avg8U( sel8x8_6(xx), sel8x8_6(yy) ),
1035 avg8U( sel8x8_5(xx), sel8x8_5(yy) ),
1036 avg8U( sel8x8_4(xx), sel8x8_4(yy) ),
1037 avg8U( sel8x8_3(xx), sel8x8_3(yy) ),
1038 avg8U( sel8x8_2(xx), sel8x8_2(yy) ),
1039 avg8U( sel8x8_1(xx), sel8x8_1(yy) ),
1040 avg8U( sel8x8_0(xx), sel8x8_0(yy) )
1044 ULong h_generic_calc_Avg16Ux4 ( ULong xx, ULong yy )
1047 avg16U( sel16x4_3(xx), sel16x4_3(yy) ),
1048 avg16U( sel16x4_2(xx), sel16x4_2(yy) ),
1049 avg16U( sel16x4_1(xx), sel16x4_1(yy) ),
1050 avg16U( sel16x4_0(xx), sel16x4_0(yy) )
1056 ULong h_generic_calc_Max16Sx4 ( ULong xx, ULong yy )
1059 max16S( sel16x4_3(xx), sel16x4_3(yy) ),
1060 max16S( sel16x4_2(xx), sel16x4_2(yy) ),
1061 max16S( sel16x4_1(xx), sel16x4_1(yy) ),
1062 max16S( sel16x4_0(xx), sel16x4_0(yy) )
1066 ULong h_generic_calc_Max8Ux8 ( ULong xx, ULong yy )
1069 max8U( sel8x8_7(xx), sel8x8_7(yy) ),
1070 max8U( sel8x8_6(xx), sel8x8_6(yy) ),
1071 max8U( sel8x8_5(xx), sel8x8_5(yy) ),
1072 max8U( sel8x8_4(xx), sel8x8_4(yy) ),
1073 max8U( sel8x8_3(xx), sel8x8_3(yy) ),
1074 max8U( sel8x8_2(xx), sel8x8_2(yy) ),
1075 max8U( sel8x8_1(xx), sel8x8_1(yy) ),
1076 max8U( sel8x8_0(xx), sel8x8_0(yy) )
1080 ULong h_generic_calc_Min16Sx4 ( ULong xx, ULong yy )
1083 min16S( sel16x4_3(xx), sel16x4_3(yy) ),
1084 min16S( sel16x4_2(xx), sel16x4_2(yy) ),
1085 min16S( sel16x4_1(xx), sel16x4_1(yy) ),
1086 min16S( sel16x4_0(xx), sel16x4_0(yy) )
1090 ULong h_generic_calc_Min8Ux8 ( ULong xx, ULong yy )
1093 min8U( sel8x8_7(xx), sel8x8_7(yy) ),
1094 min8U( sel8x8_6(xx), sel8x8_6(yy) ),
1095 min8U( sel8x8_5(xx), sel8x8_5(yy) ),
1096 min8U( sel8x8_4(xx), sel8x8_4(yy) ),
1097 min8U( sel8x8_3(xx), sel8x8_3(yy) ),
1098 min8U( sel8x8_2(xx), sel8x8_2(yy) ),
1099 min8U( sel8x8_1(xx), sel8x8_1(yy) ),
1100 min8U( sel8x8_0(xx), sel8x8_0(yy) )
1146 UInt h_generic_calc_Add16x2 ( UInt xx, UInt yy )
1148 return mk16x2( sel16x2_1(xx) + sel16x2_1(yy),
1149 sel16x2_0(xx) + sel16x2_0(yy) );
1152 UInt h_generic_calc_Sub16x2 ( UInt xx, UInt yy )
1154 return mk16x2( sel16x2_1(xx) - sel16x2_1(yy),
1155 sel16x2_0(xx) - sel16x2_0(yy) );
1158 UInt h_generic_calc_HAdd16Ux2 ( UInt xx, UInt yy )
1160 return mk16x2( hadd16U( sel16x2_1(xx), sel16x2_1(yy) ),
1161 hadd16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1164 UInt h_generic_calc_HAdd16Sx2 ( UInt xx, UInt yy )
1166 return mk16x2( hadd16S( sel16x2_1(xx), sel16x2_1(yy) ),
1167 hadd16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1170 UInt h_generic_calc_HSub16Ux2 ( UInt xx, UInt yy )
1172 return mk16x2( hsub16U( sel16x2_1(xx), sel16x2_1(yy) ),
1173 hsub16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1176 UInt h_generic_calc_HSub16Sx2 ( UInt xx, UInt yy )
1178 return mk16x2( hsub16S( sel16x2_1(xx), sel16x2_1(yy) ),
1179 hsub16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1182 UInt h_generic_calc_QAdd16Ux2 ( UInt xx, UInt yy )
1184 return mk16x2( qadd16U( sel16x2_1(xx), sel16x2_1(yy) ),
1185 qadd16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1188 UInt h_generic_calc_QAdd16Sx2 ( UInt xx, UInt yy )
1190 return mk16x2( qadd16S( sel16x2_1(xx), sel16x2_1(yy) ),
1191 qadd16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1194 UInt h_generic_calc_QSub16Ux2 ( UInt xx, UInt yy )
1196 return mk16x2( qsub16U( sel16x2_1(xx), sel16x2_1(yy) ),
1197 qsub16U( sel16x2_0(xx), sel16x2_0(yy) ) );
1200 UInt h_generic_calc_QSub16Sx2 ( UInt xx, UInt yy )
1202 return mk16x2( qsub16S( sel16x2_1(xx), sel16x2_1(yy) ),
1203 qsub16S( sel16x2_0(xx), sel16x2_0(yy) ) );
1208 UInt h_generic_calc_Add8x4 ( UInt xx, UInt yy )
1211 sel8x4_3(xx) + sel8x4_3(yy),
1212 sel8x4_2(xx) + sel8x4_2(yy),
1213 sel8x4_1(xx) + sel8x4_1(yy),
1214 sel8x4_0(xx) + sel8x4_0(yy)
1218 UInt h_generic_calc_Sub8x4 ( UInt xx, UInt yy )
1221 sel8x4_3(xx) - sel8x4_3(yy),
1222 sel8x4_2(xx) - sel8x4_2(yy),
1223 sel8x4_1(xx) - sel8x4_1(yy),
1224 sel8x4_0(xx) - sel8x4_0(yy)
1228 UInt h_generic_calc_HAdd8Ux4 ( UInt xx, UInt yy )
1231 hadd8U( sel8x4_3(xx), sel8x4_3(yy) ),
1232 hadd8U( sel8x4_2(xx), sel8x4_2(yy) ),
1233 hadd8U( sel8x4_1(xx), sel8x4_1(yy) ),
1234 hadd8U( sel8x4_0(xx), sel8x4_0(yy) )
1238 UInt h_generic_calc_HAdd8Sx4 ( UInt xx, UInt yy )
1241 hadd8S( sel8x4_3(xx), sel8x4_3(yy) ),
1242 hadd8S( sel8x4_2(xx), sel8x4_2(yy) ),
1243 xx), sel8x4_1(yy) ),
1244 hadd8S( sel8x4_0(xx), sel8x4_0(yy) )
1248 UInt h_generic_calc_HSub8Ux4 ( UInt xx, UInt yy )
1251 hsub8U( sel8x4_3(xx), sel8x4_3(yy) ),
1252 hsub8U( sel8x4_2(xx), sel8x4_2(yy) ),
1253 hsub8U( sel8x4_1(xx), sel8x4_1(yy) ),
1254 hsub8U( sel8x4_0(xx), sel8x4_0(yy) )
1258 UInt h_generic_calc_HSub8Sx4 ( UInt xx, UInt yy )
1261 hsub8S( sel8x4_3(xx), sel8x4_3(yy) ),
1262 hsub8S( sel8x4_2(xx), sel8x4_2(yy) ),
1263 hsub8S( sel8x4_1(xx), sel8x4_1(yy) ),
1264 hsub8S( sel8x4_0(xx), sel8x4_0(yy) )
1268 UInt h_generic_calc_QAdd8Ux4 ( UInt xx, UInt yy )
1271 qadd8U( sel8x4_3(xx), sel8x4_3(yy) ),
1272 qadd8U( sel8x4_2(xx), sel8x4_2(yy) ),
1273 qadd8U( sel8x4_1(xx), sel8x4_1(yy) ),
1274 qadd8U( sel8x4_0(xx), sel8x4_0(yy) )
1278 UInt h_generic_calc_QAdd8Sx4 ( UInt xx, UInt yy )
1281 qadd8S( sel8x4_3(xx), sel8x4_3(yy) ),
1282 qadd8S( sel8x4_2(xx), sel8x4_2(yy) ),
1283 qadd8S( sel8x4_1(xx), sel8x4_1(yy) ),
1284 qadd8S( sel8x4_0(xx), sel8x4_0(yy) )
1288 UInt h_generic_calc_QSub8Ux4 ( UInt xx, UInt yy )
1291 qsub8U( sel8x4_3(xx), sel8x4_3(yy) ),
1292 qsub8U( sel8x4_2(xx), sel8x4_2(yy) ),
1293 qsub8U( sel8x4_1(xx), sel8x4_1(yy) ),
1294 qsub8U( sel8x4_0(xx), sel8x4_0(yy) )
1298 UInt h_generic_calc_QSub8Sx4 ( UInt xx, UInt yy )
1301 qsub8S( sel8x4_3(xx), sel8x4_3(yy) ),
1302 qsub8S( sel8x4_2(xx), sel8x4_2(yy) ),
1303 qsub8S( sel8x4_1(xx), sel8x4_1(yy) ),
1304 qsub8S( sel8x4_0(xx), sel8x4_0(yy) )
1308 UInt h_generic_calc_CmpNEZ16x2 ( UInt xx )
1311 cmpnez16( sel16x2_1(xx) ),
1312 cmpnez16( sel16x2_0(xx) )
1316 UInt h_generic_calc_CmpNEZ8x4 ( UInt xx )
1319 cmpnez8( sel8x4_3(xx) ),
1320 cmpnez8( sel8x4_2(xx) ),
1321 cmpnez8( sel8x4_1(xx) ),
1322 cmpnez8( sel8x4_0(xx) )
1326 UInt h_generic_calc_Sad8Ux4 ( UInt xx, UInt yy )
1328 return absdiff8U( sel8x4_3(xx), sel8x4_3(yy) )
1329 + absdiff8U( sel8x4_2(xx), sel8x4_2(yy) )
1330 + absdiff8U( sel8x4_1(xx), sel8x4_1(yy) )
1331 + absdiff8U( sel8x4_0(xx), sel8x4_0(yy) );