/external/libvpx/libvpx/third_party/libyuv/source/ |
convert_argb.cc | 83 if (IS_ALIGNED(width, 8)) { 91 if (IS_ALIGNED(width, 16)) { 99 if (IS_ALIGNED(width, 8)) { 151 if (IS_ALIGNED(width, 8)) { 159 if (IS_ALIGNED(width, 16)) { 167 if (IS_ALIGNED(width, 8)) { 173 if (TestCpuFlag(kCpuHasMIPS_DSPR2) && IS_ALIGNED(width, 4) && 174 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) && 175 IS_ALIGNED(src_u, 2) && IS_ALIGNED(src_stride_u, 2) & [all...] |
convert_from_argb.cc | 55 if (IS_ALIGNED(width, 16)) { 63 if (IS_ALIGNED(width, 8)) { 71 if (IS_ALIGNED(width, 16)) { 79 if (IS_ALIGNED(width, 32)) { 87 if (IS_ALIGNED(width, 8)) { 136 if (IS_ALIGNED(width, 16)) { 144 if (IS_ALIGNED(width, 16)) { 152 if (IS_ALIGNED(width, 16)) { 160 if (IS_ALIGNED(width, 32)) { 168 if (IS_ALIGNED(width, 8)) [all...] |
rotate.cc | 39 if (IS_ALIGNED(width, 8)) { 47 if (IS_ALIGNED(width, 16)) { 54 if (IS_ALIGNED(width, 4) && 55 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4)) { 115 if (IS_ALIGNED(width, 16)) { 123 if (IS_ALIGNED(width, 16)) { 131 if (IS_ALIGNED(width, 16)) { 139 if (IS_ALIGNED(width, 32)) { 147 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4) & [all...] |
convert_from.cc | 178 if (IS_ALIGNED(width, 16)) { 186 if (IS_ALIGNED(width, 16)) { 225 if (IS_ALIGNED(width, 16)) { 233 if (IS_ALIGNED(width, 16)) { 286 if (IS_ALIGNED(width, 16)) { 294 if (IS_ALIGNED(width, 16)) { 333 if (IS_ALIGNED(width, 16)) { 341 if (IS_ALIGNED(width, 16)) { 405 if (IS_ALIGNED(halfwidth, 16)) { 413 if (IS_ALIGNED(halfwidth, 32)) [all...] |
rotate_argb.cc | 49 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(height, 4)) { // Width of dest. 54 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(height, 4)) { // Width of dest. 100 if (IS_ALIGNED(width, 4)) { 108 if (IS_ALIGNED(width, 4)) { 116 if (IS_ALIGNED(width, 8)) { 123 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2; 128 CopyRow = IS_ALIGNED(width * 4, 64) ? CopyRow_AVX : CopyRow_Any_AVX; 138 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
|
scale_argb.cc | 60 if (IS_ALIGNED(dst_width, 4)) { 72 if (IS_ALIGNED(dst_width, 8)) { 112 if (IS_ALIGNED(dst_width, 4)) { 120 if (IS_ALIGNED(dst_width, 8)) { 152 assert(IS_ALIGNED(src_width, 2)); 153 assert(IS_ALIGNED(src_height, 2)); 159 if (IS_ALIGNED(dst_width, 4)) { 169 if (IS_ALIGNED(dst_width, 4)) { 216 if (IS_ALIGNED(clip_src_width, 16)) { 224 if (IS_ALIGNED(clip_src_width, 16)) [all...] |
planar_functions.cc | 46 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2; 51 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX; 61 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON; 92 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(width, 32)) { 102 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(width, 32)) { 235 if (IS_ALIGNED(width, 16)) { 243 if (IS_ALIGNED(width, 16)) { 251 if (IS_ALIGNED(width, 16)) { 259 if (IS_ALIGNED(width, 32)) { 267 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) & [all...] |
convert.cc | 193 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2; 198 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX; 208 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON; 285 if (IS_ALIGNED(halfwidth, 16)) { 293 if (IS_ALIGNED(halfwidth, 32)) { 301 if (IS_ALIGNED(halfwidth, 16)) { 308 IS_ALIGNED(src_uv, 4) && IS_ALIGNED(src_stride_uv, 4) && 309 IS_ALIGNED(dst_u, 4) && IS_ALIGNED(dst_stride_u, 4) & [all...] |
scale.cc | 57 if (IS_ALIGNED(dst_width, 16)) { 69 if (IS_ALIGNED(dst_width, 16)) { 81 if (IS_ALIGNED(dst_width, 32)) { 89 if (TestCpuFlag(kCpuHasMIPS_DSPR2) && IS_ALIGNED(src_ptr, 4) && 90 IS_ALIGNED(src_stride, 4) && IS_ALIGNED(row_stride, 4) && 91 IS_ALIGNED(dst_ptr, 4) && IS_ALIGNED(dst_stride, 4)) { 126 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(dst_width, 16)) { 132 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 16)) [all...] |
scale_common.cc | 882 if (IS_ALIGNED(dst_width_bytes, 16)) { 890 if (IS_ALIGNED(dst_width_bytes, 16)) { 898 if (IS_ALIGNED(dst_width_bytes, 32)) { 906 if (IS_ALIGNED(dst_width_bytes, 16)) { 913 IS_ALIGNED(src_argb, 4) && IS_ALIGNED(src_stride, 4) && 914 IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride, 4)) { 916 if (IS_ALIGNED(dst_width_bytes, 4)) { 956 if (IS_ALIGNED(dst_width_bytes, 16)) [all...] |
/external/libyuv/files/source/ |
convert_from_argb.cc | 55 if (IS_ALIGNED(width, 16)) { 63 if (IS_ALIGNED(width, 8)) { 71 if (IS_ALIGNED(width, 16)) { 79 if (IS_ALIGNED(width, 32)) { 87 if (IS_ALIGNED(width, 8)) { 140 if (IS_ALIGNED(width, 16)) { 150 if (IS_ALIGNED(width, 32)) { 159 if (IS_ALIGNED(width, 8)) { 167 if (IS_ALIGNED(width, 16)) { 216 if (IS_ALIGNED(width, 16)) [all...] |
convert_argb.cc | 75 if (IS_ALIGNED(width, 8)) { 83 if (IS_ALIGNED(width, 16)) { 91 if (IS_ALIGNED(width, 8)) { 97 if (TestCpuFlag(kCpuHasDSPR2) && IS_ALIGNED(width, 4) && 98 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) && 99 IS_ALIGNED(src_u, 2) && IS_ALIGNED(src_stride_u, 2) && 100 IS_ALIGNED(src_v, 2) && IS_ALIGNED(src_stride_v, 2) & [all...] |
rotate.cc | 39 if (IS_ALIGNED(width, 8)) { 47 if (IS_ALIGNED(width, 16)) { 54 if (IS_ALIGNED(width, 4) && 55 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4)) { 115 if (IS_ALIGNED(width, 16)) { 123 if (IS_ALIGNED(width, 16)) { 131 if (IS_ALIGNED(width, 32)) { 139 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4) & [all...] |
rotate_argb.cc | 49 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(height, 4)) { // Width of dest. 54 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(height, 4)) { // Width of dest. 100 if (IS_ALIGNED(width, 4)) { 108 if (IS_ALIGNED(width, 4)) { 116 if (IS_ALIGNED(width, 8)) { 123 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2; 128 CopyRow = IS_ALIGNED(width * 4, 64) ? CopyRow_AVX : CopyRow_Any_AVX; 138 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
|
convert_from.cc | 178 if (IS_ALIGNED(width, 16)) { 186 if (IS_ALIGNED(width, 16)) { 225 if (IS_ALIGNED(width, 16)) { 233 if (IS_ALIGNED(width, 16)) { 286 if (IS_ALIGNED(width, 16)) { 294 if (IS_ALIGNED(width, 16)) { 333 if (IS_ALIGNED(width, 16)) { 341 if (IS_ALIGNED(width, 16)) { 405 if (IS_ALIGNED(halfwidth, 16)) { 413 if (IS_ALIGNED(halfwidth, 32)) [all...] |
planar_functions.cc | 47 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2; 52 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX; 62 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON; 93 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(width, 32)) { 103 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(width, 32)) { 236 if (IS_ALIGNED(width, 16)) { 244 if (IS_ALIGNED(width, 16)) { 252 if (IS_ALIGNED(width, 32)) { 260 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) & [all...] |
convert.cc | 193 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2; 198 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX; 208 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON; 285 if (IS_ALIGNED(halfwidth, 16)) { 293 if (IS_ALIGNED(halfwidth, 32)) { 301 if (IS_ALIGNED(halfwidth, 16)) { 308 IS_ALIGNED(src_uv, 4) && IS_ALIGNED(src_stride_uv, 4) && 309 IS_ALIGNED(dst_u, 4) && IS_ALIGNED(dst_stride_u, 4) & [all...] |
scale_argb.cc | 60 if (IS_ALIGNED(dst_width, 4)) { 72 if (IS_ALIGNED(dst_width, 8)) { 112 if (IS_ALIGNED(dst_width, 4)) { 120 if (IS_ALIGNED(dst_width, 8)) { 152 assert(IS_ALIGNED(src_width, 2)); 153 assert(IS_ALIGNED(src_height, 2)); 159 if (IS_ALIGNED(dst_width, 4)) { 169 if (IS_ALIGNED(dst_width, 4)) { 216 if (IS_ALIGNED(clip_src_width, 16)) { 224 if (IS_ALIGNED(clip_src_width, 32)) [all...] |
scale.cc | 57 if (IS_ALIGNED(dst_width, 16)) { 69 if (IS_ALIGNED(dst_width, 16)) { 81 if (IS_ALIGNED(dst_width, 32)) { 89 if (TestCpuFlag(kCpuHasDSPR2) && IS_ALIGNED(src_ptr, 4) && 90 IS_ALIGNED(src_stride, 4) && IS_ALIGNED(row_stride, 4) && 91 IS_ALIGNED(dst_ptr, 4) && IS_ALIGNED(dst_stride, 4)) { 126 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(dst_width, 16)) { 132 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 16)) [all...] |
scale_common.cc | 912 if (IS_ALIGNED(dst_width_bytes, 16)) { 920 if (IS_ALIGNED(dst_width_bytes, 32)) { 928 if (IS_ALIGNED(dst_width_bytes, 16)) { 935 IS_ALIGNED(src_argb, 4) && IS_ALIGNED(src_stride, 4) && 936 IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride, 4)) { 938 if (IS_ALIGNED(dst_width_bytes, 4)) { 978 if (IS_ALIGNED(dst_width_bytes, 16)) { 986 if (IS_ALIGNED(dst_width_bytes, 16)) [all...] |
/prebuilts/gcc/linux-x86/host/x86_64-w64-mingw32-4.8/x86_64-w64-mingw32/include/ |
rtinfo.h | 17 #define IS_ALIGNED(ptr) (((UINT_PTR)(ptr) & ALIGN_SHIFT)==0x00000000)
|
uastrfnc.h | 20 #define IS_ALIGNED(p) (((ULONG_PTR)(p) & (sizeof(*(p))-1))==0)
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/libyuv/source/ |
scale.c | [all...] |
/external/e2fsprogs/lib/ext2fs/ |
unix_io.c | 96 #define IS_ALIGNED(n, align) ((((unsigned long) n) & \ 179 (IS_ALIGNED(buf, channel->align) && 180 IS_ALIGNED(size, channel->align))) { 250 (IS_ALIGNED(buf, channel->align) && 251 IS_ALIGNED(size, channel->align))) {
|
/external/libjpeg-turbo/simd/ |
jsimd_x86_64.c | 28 #define IS_ALIGNED(ptr, order) (((size_t)ptr & ((1 << order) - 1)) == 0) 30 #define IS_ALIGNED_SSE(ptr) (IS_ALIGNED(ptr, 4)) /* 16 byte alignment */
|