HomeSort by relevance Sort by last modified time
    Searched refs:IS_ALIGNED (Results 1 - 25 of 30) sorted by null

1 2

  /external/libvpx/libvpx/third_party/libyuv/source/
convert_argb.cc 83 if (IS_ALIGNED(width, 8)) {
91 if (IS_ALIGNED(width, 16)) {
99 if (IS_ALIGNED(width, 8)) {
151 if (IS_ALIGNED(width, 8)) {
159 if (IS_ALIGNED(width, 16)) {
167 if (IS_ALIGNED(width, 8)) {
173 if (TestCpuFlag(kCpuHasMIPS_DSPR2) && IS_ALIGNED(width, 4) &&
174 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) &&
175 IS_ALIGNED(src_u, 2) && IS_ALIGNED(src_stride_u, 2) &
    [all...]
convert_from_argb.cc 55 if (IS_ALIGNED(width, 16)) {
63 if (IS_ALIGNED(width, 8)) {
71 if (IS_ALIGNED(width, 16)) {
79 if (IS_ALIGNED(width, 32)) {
87 if (IS_ALIGNED(width, 8)) {
136 if (IS_ALIGNED(width, 16)) {
144 if (IS_ALIGNED(width, 16)) {
152 if (IS_ALIGNED(width, 16)) {
160 if (IS_ALIGNED(width, 32)) {
168 if (IS_ALIGNED(width, 8))
    [all...]
rotate.cc 39 if (IS_ALIGNED(width, 8)) {
47 if (IS_ALIGNED(width, 16)) {
54 if (IS_ALIGNED(width, 4) &&
55 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4)) {
115 if (IS_ALIGNED(width, 16)) {
123 if (IS_ALIGNED(width, 16)) {
131 if (IS_ALIGNED(width, 16)) {
139 if (IS_ALIGNED(width, 32)) {
147 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4) &
    [all...]
convert_from.cc 178 if (IS_ALIGNED(width, 16)) {
186 if (IS_ALIGNED(width, 16)) {
225 if (IS_ALIGNED(width, 16)) {
233 if (IS_ALIGNED(width, 16)) {
286 if (IS_ALIGNED(width, 16)) {
294 if (IS_ALIGNED(width, 16)) {
333 if (IS_ALIGNED(width, 16)) {
341 if (IS_ALIGNED(width, 16)) {
405 if (IS_ALIGNED(halfwidth, 16)) {
413 if (IS_ALIGNED(halfwidth, 32))
    [all...]
rotate_argb.cc 49 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(height, 4)) { // Width of dest.
54 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(height, 4)) { // Width of dest.
100 if (IS_ALIGNED(width, 4)) {
108 if (IS_ALIGNED(width, 4)) {
116 if (IS_ALIGNED(width, 8)) {
123 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2;
128 CopyRow = IS_ALIGNED(width * 4, 64) ? CopyRow_AVX : CopyRow_Any_AVX;
138 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
scale_argb.cc 60 if (IS_ALIGNED(dst_width, 4)) {
72 if (IS_ALIGNED(dst_width, 8)) {
112 if (IS_ALIGNED(dst_width, 4)) {
120 if (IS_ALIGNED(dst_width, 8)) {
152 assert(IS_ALIGNED(src_width, 2));
153 assert(IS_ALIGNED(src_height, 2));
159 if (IS_ALIGNED(dst_width, 4)) {
169 if (IS_ALIGNED(dst_width, 4)) {
216 if (IS_ALIGNED(clip_src_width, 16)) {
224 if (IS_ALIGNED(clip_src_width, 16))
    [all...]
planar_functions.cc 46 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2;
51 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX;
61 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
92 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(width, 32)) {
102 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(width, 32)) {
235 if (IS_ALIGNED(width, 16)) {
243 if (IS_ALIGNED(width, 16)) {
251 if (IS_ALIGNED(width, 16)) {
259 if (IS_ALIGNED(width, 32)) {
267 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) &
    [all...]
convert.cc 193 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2;
198 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX;
208 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
285 if (IS_ALIGNED(halfwidth, 16)) {
293 if (IS_ALIGNED(halfwidth, 32)) {
301 if (IS_ALIGNED(halfwidth, 16)) {
308 IS_ALIGNED(src_uv, 4) && IS_ALIGNED(src_stride_uv, 4) &&
309 IS_ALIGNED(dst_u, 4) && IS_ALIGNED(dst_stride_u, 4) &
    [all...]
scale.cc 57 if (IS_ALIGNED(dst_width, 16)) {
69 if (IS_ALIGNED(dst_width, 16)) {
81 if (IS_ALIGNED(dst_width, 32)) {
89 if (TestCpuFlag(kCpuHasMIPS_DSPR2) && IS_ALIGNED(src_ptr, 4) &&
90 IS_ALIGNED(src_stride, 4) && IS_ALIGNED(row_stride, 4) &&
91 IS_ALIGNED(dst_ptr, 4) && IS_ALIGNED(dst_stride, 4)) {
126 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(dst_width, 16)) {
132 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 16))
    [all...]
scale_common.cc 882 if (IS_ALIGNED(dst_width_bytes, 16)) {
890 if (IS_ALIGNED(dst_width_bytes, 16)) {
898 if (IS_ALIGNED(dst_width_bytes, 32)) {
906 if (IS_ALIGNED(dst_width_bytes, 16)) {
913 IS_ALIGNED(src_argb, 4) && IS_ALIGNED(src_stride, 4) &&
914 IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride, 4)) {
916 if (IS_ALIGNED(dst_width_bytes, 4)) {
956 if (IS_ALIGNED(dst_width_bytes, 16))
    [all...]
  /external/libyuv/files/source/
convert_from_argb.cc 55 if (IS_ALIGNED(width, 16)) {
63 if (IS_ALIGNED(width, 8)) {
71 if (IS_ALIGNED(width, 16)) {
79 if (IS_ALIGNED(width, 32)) {
87 if (IS_ALIGNED(width, 8)) {
140 if (IS_ALIGNED(width, 16)) {
150 if (IS_ALIGNED(width, 32)) {
159 if (IS_ALIGNED(width, 8)) {
167 if (IS_ALIGNED(width, 16)) {
216 if (IS_ALIGNED(width, 16))
    [all...]
convert_argb.cc 75 if (IS_ALIGNED(width, 8)) {
83 if (IS_ALIGNED(width, 16)) {
91 if (IS_ALIGNED(width, 8)) {
97 if (TestCpuFlag(kCpuHasDSPR2) && IS_ALIGNED(width, 4) &&
98 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) &&
99 IS_ALIGNED(src_u, 2) && IS_ALIGNED(src_stride_u, 2) &&
100 IS_ALIGNED(src_v, 2) && IS_ALIGNED(src_stride_v, 2) &
    [all...]
rotate.cc 39 if (IS_ALIGNED(width, 8)) {
47 if (IS_ALIGNED(width, 16)) {
54 if (IS_ALIGNED(width, 4) &&
55 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4)) {
115 if (IS_ALIGNED(width, 16)) {
123 if (IS_ALIGNED(width, 16)) {
131 if (IS_ALIGNED(width, 32)) {
139 IS_ALIGNED(src, 4) && IS_ALIGNED(src_stride, 4) &
    [all...]
rotate_argb.cc 49 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(height, 4)) { // Width of dest.
54 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(height, 4)) { // Width of dest.
100 if (IS_ALIGNED(width, 4)) {
108 if (IS_ALIGNED(width, 4)) {
116 if (IS_ALIGNED(width, 8)) {
123 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2;
128 CopyRow = IS_ALIGNED(width * 4, 64) ? CopyRow_AVX : CopyRow_Any_AVX;
138 CopyRow = IS_ALIGNED(width * 4, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
convert_from.cc 178 if (IS_ALIGNED(width, 16)) {
186 if (IS_ALIGNED(width, 16)) {
225 if (IS_ALIGNED(width, 16)) {
233 if (IS_ALIGNED(width, 16)) {
286 if (IS_ALIGNED(width, 16)) {
294 if (IS_ALIGNED(width, 16)) {
333 if (IS_ALIGNED(width, 16)) {
341 if (IS_ALIGNED(width, 16)) {
405 if (IS_ALIGNED(halfwidth, 16)) {
413 if (IS_ALIGNED(halfwidth, 32))
    [all...]
planar_functions.cc 47 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2;
52 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX;
62 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
93 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(width, 32)) {
103 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(width, 32)) {
236 if (IS_ALIGNED(width, 16)) {
244 if (IS_ALIGNED(width, 16)) {
252 if (IS_ALIGNED(width, 32)) {
260 IS_ALIGNED(src_y, 4) && IS_ALIGNED(src_stride_y, 4) &
    [all...]
convert.cc 193 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_SSE2 : CopyRow_Any_SSE2;
198 CopyRow = IS_ALIGNED(width, 64) ? CopyRow_AVX : CopyRow_Any_AVX;
208 CopyRow = IS_ALIGNED(width, 32) ? CopyRow_NEON : CopyRow_Any_NEON;
285 if (IS_ALIGNED(halfwidth, 16)) {
293 if (IS_ALIGNED(halfwidth, 32)) {
301 if (IS_ALIGNED(halfwidth, 16)) {
308 IS_ALIGNED(src_uv, 4) && IS_ALIGNED(src_stride_uv, 4) &&
309 IS_ALIGNED(dst_u, 4) && IS_ALIGNED(dst_stride_u, 4) &
    [all...]
scale_argb.cc 60 if (IS_ALIGNED(dst_width, 4)) {
72 if (IS_ALIGNED(dst_width, 8)) {
112 if (IS_ALIGNED(dst_width, 4)) {
120 if (IS_ALIGNED(dst_width, 8)) {
152 assert(IS_ALIGNED(src_width, 2));
153 assert(IS_ALIGNED(src_height, 2));
159 if (IS_ALIGNED(dst_width, 4)) {
169 if (IS_ALIGNED(dst_width, 4)) {
216 if (IS_ALIGNED(clip_src_width, 16)) {
224 if (IS_ALIGNED(clip_src_width, 32))
    [all...]
scale.cc 57 if (IS_ALIGNED(dst_width, 16)) {
69 if (IS_ALIGNED(dst_width, 16)) {
81 if (IS_ALIGNED(dst_width, 32)) {
89 if (TestCpuFlag(kCpuHasDSPR2) && IS_ALIGNED(src_ptr, 4) &&
90 IS_ALIGNED(src_stride, 4) && IS_ALIGNED(row_stride, 4) &&
91 IS_ALIGNED(dst_ptr, 4) && IS_ALIGNED(dst_stride, 4)) {
126 if (TestCpuFlag(kCpuHasNEON) && IS_ALIGNED(dst_width, 16)) {
132 if (TestCpuFlag(kCpuHasSSE2) && IS_ALIGNED(dst_width, 16))
    [all...]
scale_common.cc 912 if (IS_ALIGNED(dst_width_bytes, 16)) {
920 if (IS_ALIGNED(dst_width_bytes, 32)) {
928 if (IS_ALIGNED(dst_width_bytes, 16)) {
935 IS_ALIGNED(src_argb, 4) && IS_ALIGNED(src_stride, 4) &&
936 IS_ALIGNED(dst_argb, 4) && IS_ALIGNED(dst_stride, 4)) {
938 if (IS_ALIGNED(dst_width_bytes, 4)) {
978 if (IS_ALIGNED(dst_width_bytes, 16)) {
986 if (IS_ALIGNED(dst_width_bytes, 16))
    [all...]
  /prebuilts/gcc/linux-x86/host/x86_64-w64-mingw32-4.8/x86_64-w64-mingw32/include/
rtinfo.h 17 #define IS_ALIGNED(ptr) (((UINT_PTR)(ptr) & ALIGN_SHIFT)==0x00000000)
uastrfnc.h 20 #define IS_ALIGNED(p) (((ULONG_PTR)(p) & (sizeof(*(p))-1))==0)
  /hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/third_party/libyuv/source/
scale.c     [all...]
  /external/e2fsprogs/lib/ext2fs/
unix_io.c 96 #define IS_ALIGNED(n, align) ((((unsigned long) n) & \
179 (IS_ALIGNED(buf, channel->align) &&
180 IS_ALIGNED(size, channel->align))) {
250 (IS_ALIGNED(buf, channel->align) &&
251 IS_ALIGNED(size, channel->align))) {
  /external/libjpeg-turbo/simd/
jsimd_x86_64.c 28 #define IS_ALIGNED(ptr, order) (((size_t)ptr & ((1 << order) - 1)) == 0)
30 #define IS_ALIGNED_SSE(ptr) (IS_ALIGNED(ptr, 4)) /* 16 byte alignment */

Completed in 5081 milliseconds

1 2