/external/valgrind/auxprogs/ |
ppcfround.c | 117 /* Give an insn string such as "fmadd %%f4, %%f1,%%f2,%%f3". Args are 119 #define INSN(name,insn) \ 141 insn "\n\t" \ 161 INSN(fabs, "fabs %%f4, %%f1"); 162 INSN(fabs_, "fabs. %%f4, %%f1"); 164 INSN(fnabs, "fnabs %%f4, %%f1"); 165 INSN(fnabs_, "fnabs. %%f4, %%f1"); 167 INSN(fadd, "fadd %%f4, %%f1,%%f2"); 168 INSN(fadd_, "fadd. %%f4, %%f1,%%f2") [all...] |
/external/arm-neon-tests/ |
ref_vqdmull.c | 34 #define INSN vqdmull 39 FNNAME (INSN) 42 #define TEST_VQDMULL2(INSN, T1, T2, W, W2, N) \ 45 INSN##_##T2##W(VECT_VAR(vector, T1, W, N), \ 49 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_##T2##W), \ 52 /* Two auxliary macros are necessary to expand INSN */ 53 #define TEST_VQDMULL1(INSN, T1, T2, W, W2, N) \ 54 TEST_VQDMULL2(INSN, T1, T2, W, W2, N) 57 TEST_VQDMULL1(INSN, T1, T2, W, W2, N)
|
ref_vuzp.c | 55 #define TEST_VUZP(INSN, Q, T1, T2, W, N) \ 57 INSN##Q##_##T2##W(VECT_VAR(vector1, T1, W, N), \ 123 #define TEST_ALL_VUZP(INSN) \ 124 TEST_VUZP(INSN, , int, s, 8, 8); \ 125 TEST_VUZP(INSN, , int, s, 16, 4); \ 126 TEST_VUZP(INSN, , int, s, 32, 2); \ 127 TEST_VUZP(INSN, , uint, u, 8, 8); \ 128 TEST_VUZP(INSN, , uint, u, 16, 4); \ 129 TEST_VUZP(INSN, , uint, u, 32, 2); \ 130 TEST_VUZP(INSN, , poly, p, 8, 8); [all...] |
ref_vqdmulh_lane.c | 34 #define INSN vqdmulh 39 FNNAME (INSN) 42 #define TEST_VQDMULH_LANE2(INSN, Q, T1, T2, W, N, N2, L) \ 45 INSN##Q##_lane_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_lane_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQDMULH_LANE1(INSN, Q, T1, T2, W, N, N2, L) \ 55 TEST_VQDMULH_LANE2(INSN, Q, T1, T2, W, N, N2, L) 58 TEST_VQDMULH_LANE1(INSN, Q, T1, T2, W, N, N2, L)
|
ref_vqdmulh_n.c | 34 #define INSN vqdmulh 39 FNNAME (INSN) 44 #define TEST_VQDMULH_N2(INSN, Q, T1, T2, W, N, L) \ 47 INSN##Q##_n_##T2##W(VECT_VAR(vector, T1, W, N), \ 51 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_n_##T2##W), \ 54 /* Two auxliary macros are necessary to expand INSN */ 55 #define TEST_VQDMULH_N1(INSN, Q, T1, T2, W, N, L) \ 56 TEST_VQDMULH_N2(INSN, Q, T1, T2, W, N, L) 59 TEST_VQDMULH_N1(INSN, Q, T1, T2, W, N, L)
|
ref_vqdmull_lane.c | 34 #define INSN vqdmull 39 FNNAME (INSN) 44 #define TEST_VQDMULL_LANE2(INSN, T1, T2, W, W2, N, L) \ 47 INSN##_lane_##T2##W(VECT_VAR(vector, T1, W, N), \ 52 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_lane_##T2##W), \ 55 /* Two auxliary macros are necessary to expand INSN */ 56 #define TEST_VQDMULL_LANE1(INSN, T1, T2, W, W2, N, L) \ 57 TEST_VQDMULL_LANE2(INSN, T1, T2, W, W2, N, L) 60 TEST_VQDMULL_LANE1(INSN, T1, T2, W, W2, N, L)
|
ref_vqdmull_n.c | 34 #define INSN vqdmull 39 FNNAME (INSN) 44 #define TEST_VQDMULL_N2(INSN, T1, T2, W, W2, N, L) \ 47 INSN##_n_##T2##W(VECT_VAR(vector, T1, W, N), \ 51 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_n_##T2##W), \ 54 /* Two auxliary macros are necessary to expand INSN */ 55 #define TEST_VQDMULL_N1(INSN, T1, T2, W, W2, N, L) \ 56 TEST_VQDMULL_N2(INSN, T1, T2, W, W2, N, L) 59 TEST_VQDMULL_N1(INSN, T1, T2, W, W2, N, L)
|
ref_vqshrun_n.c | 34 #define INSN vqshrun_n 40 FNNAME (INSN) 43 #define TEST_VQSHRUN_N2(INSN, T1, T2, W, W2, N, V) \ 46 INSN##_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQSHRUN_N1(INSN, T1, T2, W, W2, N, V) \ 55 TEST_VQSHRUN_N2(INSN, T1, T2, W, W2, N, V) 58 TEST_VQSHRUN_N1(INSN, T1, T2, W, W2, N, V)
|
ref_v_unary_sat_op.c | 42 #define TEST_UNARY_SAT_OP1(INSN, Q, T1, T2, W, N) \ 45 INSN##Q##_##T2##W(VECT_VAR(vector, T1, W, N)); \ 48 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_##T2##W), \ 51 #define TEST_UNARY_SAT_OP(INSN, Q, T1, T2, W, N) \ 52 TEST_UNARY_SAT_OP1(INSN, Q, T1, T2, W, N)
|
ref_vqdmlal_n.c | 46 #define TEST_VQDMLXL_N1(INSN, T1, T2, W, W2, N, V) \ 49 INSN##_##T2##W2(VECT_VAR(vector, T1, W, N), \ 54 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_##T2##W2), \ 57 #define TEST_VQDMLXL_N(INSN, T1, T2, W, W2, N, V) \ 58 TEST_VQDMLXL_N1(INSN, T1, T2, W, W2, N, V)
|
ref_vqmovun.c | 43 #define TEST_UNARY_OP1(INSN, T1, T2, W, W2, N) \ 46 INSN##_s##W2(VECT_VAR(vector, int, W2, N)); \ 49 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_s##W2), \ 52 #define TEST_UNARY_OP(INSN, T1, T2, W, W2, N) \ 53 TEST_UNARY_OP1(INSN, T1, T2, W, W2, N) \
|
ref_vqdmulh.c | 34 #define INSN vqdmulh 40 FNNAME (INSN) 43 #define TEST_VQDMULH2(INSN, Q, T1, T2, W, N) \ 46 INSN##Q##_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQDMULH1(INSN, Q, T1, T2, W, N) \ 55 TEST_VQDMULH2(INSN, Q, T1, T2, W, N) 58 TEST_VQDMULH1(INSN, Q, T1, T2, W, N)
|
ref_vqrdmulh.c | 34 #define INSN vqrdmulh 40 FNNAME (INSN) 43 #define TEST_VQRDMULH2(INSN, Q, T1, T2, W, N) \ 46 INSN##Q##_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQRDMULH1(INSN, Q, T1, T2, W, N) \ 55 TEST_VQRDMULH2(INSN, Q, T1, T2, W, N) 58 TEST_VQRDMULH1(INSN, Q, T1, T2, W, N)
|
ref_vqrdmulh_lane.c | 34 #define INSN vqrdmulh 40 FNNAME (INSN) 43 #define TEST_VQRDMULH_LANE2(INSN, Q, T1, T2, W, N, N2, L) \ 46 INSN##Q##_lane_##T2##W(VECT_VAR(vector, T1, W, N), \ 51 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_lane_##T2##W), \ 54 /* Two auxliary macros are necessary to expand INSN */ 55 #define TEST_VQRDMULH_LANE1(INSN, Q, T1, T2, W, N, N2, L) \ 56 TEST_VQRDMULH_LANE2(INSN, Q, T1, T2, W, N, N2, L) 59 TEST_VQRDMULH_LANE1(INSN, Q, T1, T2, W, N, N2, L)
|
ref_vqrdmulh_n.c | 34 #define INSN vqrdmulh 40 FNNAME (INSN) 45 #define TEST_VQRDMULH_N2(INSN, Q, T1, T2, W, N, L) \ 48 INSN##Q##_n_##T2##W(VECT_VAR(vector, T1, W, N), \ 52 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_n_##T2##W), \ 55 /* Two auxliary macros are necessary to expand INSN */ 56 #define TEST_VQRDMULH_N1(INSN, Q, T1, T2, W, N, L) \ 57 TEST_VQRDMULH_N2(INSN, Q, T1, T2, W, N, L) 60 TEST_VQRDMULH_N1(INSN, Q, T1, T2, W, N, L)
|
ref_vqrshrn_n.c | 34 #define INSN vqrshrn_n 40 FNNAME (INSN) 43 #define TEST_VQRSHRN_N2(INSN, T1, T2, W, W2, N, V) \ 46 INSN##_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQRSHRN_N1(INSN, T1, T2, W, W2, N, V) \ 55 TEST_VQRSHRN_N2(INSN, T1, T2, W, W2, N, V) 58 TEST_VQRSHRN_N1(INSN, T1, T2, W, W2, N, V)
|
ref_vqrshrun_n.c | 34 #define INSN vqrshrun_n 40 FNNAME (INSN) 43 #define TEST_VQRSHRUN_N2(INSN, T1, T2, W, W2, N, V) \ 46 INSN##_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQRSHRUN_N1(INSN, T1, T2, W, W2, N, V) \ 55 TEST_VQRSHRUN_N2(INSN, T1, T2, W, W2, N, V) 58 TEST_VQRSHRUN_N1(INSN, T1, T2, W, W2, N, V)
|
ref_vqshl_n.c | 34 #define INSN vqshl 40 FNNAME (INSN) 43 #define TEST_VQSHL_N2(INSN, Q, T1, T2, W, N, V) \ 46 INSN##Q##_n_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##Q##_n_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQSHL_N1(INSN, T3, Q, T1, T2, W, N) \ 55 TEST_VQSHL_N2(INSN, T3, Q, T1, T2, W, N) 58 TEST_VQSHL_N1(INSN, T3, Q, T1, T2, W, N)
|
ref_vqshrn_n.c | 34 #define INSN vqshrn_n 40 FNNAME (INSN) 43 #define TEST_VQSHRN_N2(INSN, T1, T2, W, W2, N, V) \ 46 INSN##_##T2##W(VECT_VAR(vector, T1, W, N), \ 50 dump_neon_cumulative_sat(TEST_MSG, xSTR(INSN##_##T2##W), \ 53 /* Two auxliary macros are necessary to expand INSN */ 54 #define TEST_VQSHRN_N1(INSN, T1, T2, W, W2, N, V) \ 55 TEST_VQSHRN_N2(INSN, T1, T2, W, W2, N, V) 58 TEST_VQSHRN_N1(INSN, T1, T2, W, W2, N, V)
|
ref_v_binary_op.c | 42 #define TEST_BINARY_OP1(INSN, Q, T1, T2, W, N) \ 44 INSN##Q##_##T2##W(VECT_VAR(vector, T1, W, N), \ 48 #define TEST_BINARY_OP(INSN, Q, T1, T2, W, N) \ 49 TEST_BINARY_OP1(INSN, Q, T1, T2, W, N) \
|
ref_v_comp_f_op.c | 43 #define TEST_VCOMP1(INSN, Q, T1, T2, T3, W, N) \ 45 INSN##Q##_##T2##W(VECT_VAR(vector, T1, W, N), \ 49 #define TEST_VCOMP(INSN, Q, T1, T2, T3, W, N) \ 50 TEST_VCOMP1(INSN, Q, T1, T2, T3, W, N)
|
ref_v_unary_op.c | 42 #define TEST_UNARY_OP1(INSN, Q, T1, T2, W, N) \ 44 INSN##Q##_##T2##W(VECT_VAR(vector, T1, W, N)); \ 47 #define TEST_UNARY_OP(INSN, Q, T1, T2, W, N) \ 48 TEST_UNARY_OP1(INSN, Q, T1, T2, W, N) \
|
ref_vcnt.c | 43 #define TEST_UNARY_OP1(INSN, Q, T1, T2, W, N) \ 45 INSN##Q##_##T2##W(VECT_VAR(vector, T1, W, N)); \ 48 #define TEST_UNARY_OP(INSN, Q, T1, T2, W, N) \ 49 TEST_UNARY_OP1(INSN, Q, T1, T2, W, N) \
|
ref_vmlal_n.c | 46 #define TEST_VMLXL_N1(INSN, T1, T2, W, W2, N, V) \ 47 VECT_VAR(vector_res, T1, W, N) = INSN##_##T2##W2(VECT_VAR(vector, T1, W, N), \ 52 #define TEST_VMLXL_N(INSN, T1, T2, W, W2, N, V) \ 53 TEST_VMLXL_N1(INSN, T1, T2, W, W2, N, V)
|
/prebuilts/gcc/linux-x86/host/x86_64-w64-mingw32-4.8/lib/gcc/x86_64-w64-mingw32/4.8.3/plugin/include/ |
basic-block.h | 135 /* The first insn of the block is embedded into bb->il.x. */ 136 /* The last insn of the block. */ 139 /* In CFGlayout mode points to insn notes/jumptables to be placed just before 166 Basic blocks need not start with a label nor end with a jump insn. 169 end with a jump insn. Block 0 is a descendant of the entry block. 369 #define FOR_BB_INSNS(BB, INSN) \ 370 for ((INSN) = BB_HEAD (BB); \ 371 (INSN) && (INSN) != NEXT_INSN (BB_END (BB)); \ 372 (INSN) = NEXT_INSN (INSN) [all...] |