/device/moto/shamu/camera/QCamera/stack/mm-camera-test/src/ |
mm_qcamera_display.c | 69 } yuv; variable in typeref:union:__anon4099 178 /* Initialize yuv structure */ 179 yuv.list.count = 1; 181 e = &yuv.list.req[0]; 240 result = ioctl(fb_fd, MSMFB_BLIT, &yuv.list); 461 /* Initialize yuv structure */ 462 yuv.list.count = 1; 463 e = &yuv.list.req[0]; 521 /* Initialize yuv structure */ 522 yuv.list.count = 1 [all...] |
/frameworks/rs/java/tests/LivePreview/src/com/android/rs/livepreview/ |
RsYuv.java | 112 void execute(byte[] yuv) { 113 mAllocationIn.copyFrom(yuv);
|
/frameworks/rs/ |
rsType.h | 108 bool dimLOD, bool dimFaces, uint32_t yuv) { 109 ObjectBaseRef<Type> type = getTypeRef(rsc, e, dimX, dimY, dimZ, dimLOD, dimFaces, yuv);
|
rsAllocation.h | 63 uint32_t yuv; member in struct:android::renderscript::Allocation::Hal::State 90 } yuv; member in struct:android::renderscript::Allocation::Hal::DrvState
|
/hardware/intel/img/hwcomposer/ips/anniedale/ |
PlaneCapabilities.cpp | 127 if (stride.yuv.yStride > maxStride) { 128 VLOGTRACE("stride %d is too large", stride.yuv.yStride);
|
/prebuilts/ndk/9/platforms/android-19/arch-arm/usr/include/rs/ |
rsType.h | 106 bool dimLOD, bool dimFaces, uint32_t yuv) { 107 ObjectBaseRef<Type> type = getTypeRef(rsc, e, dimX, dimY, dimZ, dimLOD, dimFaces, yuv);
|
rsAllocation.h | 63 uint32_t yuv; member in struct:android::renderscript::Allocation::Hal::State 90 } yuv; member in struct:android::renderscript::Allocation::Hal::DrvState
|
/prebuilts/ndk/9/platforms/android-19/arch-mips/usr/include/rs/ |
rsType.h | 106 bool dimLOD, bool dimFaces, uint32_t yuv) { 107 ObjectBaseRef<Type> type = getTypeRef(rsc, e, dimX, dimY, dimZ, dimLOD, dimFaces, yuv);
|
rsAllocation.h | 63 uint32_t yuv; member in struct:android::renderscript::Allocation::Hal::State 90 } yuv; member in struct:android::renderscript::Allocation::Hal::DrvState
|
/prebuilts/ndk/9/platforms/android-19/arch-x86/usr/include/rs/ |
rsType.h | 106 bool dimLOD, bool dimFaces, uint32_t yuv) { 107 ObjectBaseRef<Type> type = getTypeRef(rsc, e, dimX, dimY, dimZ, dimLOD, dimFaces, yuv);
|
rsAllocation.h | 63 uint32_t yuv; member in struct:android::renderscript::Allocation::Hal::State 90 } yuv; member in struct:android::renderscript::Allocation::Hal::DrvState
|
/external/webp/src/ |
Android.mk | 52 dsp/yuv.c \ 106 dsp/yuv.c \
|
/frameworks/av/media/libstagefright/codecs/avc/common/include/ |
avcapi_common.h | 199 @param yuv The address of the yuv pointer returned to the AVC lib. 202 typedef int (*FunctionType_FrameBind)(void *userData, int indx, uint8 **yuv);
|
/hardware/intel/img/hwcomposer/include/ |
DataBuffer.h | 40 } yuv; member in union:android::intel::stride::__anon40829
|
/external/chromium_org/content/browser/media/capture/ |
web_contents_video_capture_device_unittest.cc | 67 uint8 yuv[3]; local 69 yuv, yuv + 1, yuv + 2, 1, 1, 1, 1, 1); local 70 return SkColorSetRGB(yuv[0], yuv[1], yuv[2]); 358 uint8 yuv[3]; variable 360 yuv[plane] = frame->data(plane)[0]; 364 color_callback_.Run((SkColorSetRGB(yuv[0], yuv[1], yuv[2]))) [all...] |
/hardware/intel/img/hwcomposer/ips/common/ |
OverlayPlaneBase.cpp | 395 stride.yuv.yStride = yStride; 396 stride.yuv.uvStride = uvStride; 401 stride.yuv.yStride = yStride; 402 stride.yuv.uvStride = uvStride; 408 stride.yuv.yStride = yStride; 409 stride.yuv.uvStride = uvStride; 415 stride.yuv.yStride = yStride; 416 stride.yuv.uvStride = uvStride; 668 uint32_t yStride = mapper.getStride().yuv.yStride; 669 uint32_t uvStride = mapper.getStride().yuv.uvStride [all...] |
/device/moto/shamu/camera/QCamera/HAL/usbcamcore/src/ |
QCameraMjpegDecode.cpp | 382 jpeg_buffer_init(&p_output_buffers.data.yuv.luma_buf); 383 jpeg_buffer_init(&p_output_buffers.data.yuv.chroma_buf); 408 p_output_buffers.data.yuv.luma_buf, 413 p_output_buffers.data.yuv.chroma_buf, 527 jpeg_buffer_destroy(&p_output_buffers.data.yuv.luma_buf); 528 jpeg_buffer_destroy(&p_output_buffers.data.yuv.chroma_buf);
|
/external/chromium_org/third_party/libjpeg_turbo/ |
tjbench.c | 49 int flags=TJFLAG_NOREALLOC, decomponly=0, yuv=0, quiet=0, dotile=0, variable 99 int scaledw=(yuv==YUVDECODE)? w : TJSCALED(w, sf); 100 int scaledh=(yuv==YUVDECODE)? h : TJSCALED(h, sf); 114 bufsize=(yuv==YUVDECODE? yuvsize:pitch*scaledh); 126 if(yuv==YUVDECODE) 139 if(yuv==YUVDECODE) 171 if(yuv==YUVDECODE) 173 snprintf(tempstr, 1024, "%s_%s%s.yuv", filename, subName[subsamp], 176 _throwunix("opening YUV image for output"); 178 _throwunix("writing YUV image") [all...] |
/external/qemu/android/camera/ |
camera-format-converters.c | 149 * Basics of RGB -> YUV conversion 153 * RGB -> YUV conversion macros 159 /* Converts R8 G8 B8 color to YUV. */ 168 /* Converts RGB565 color to YUV. */ 175 /* Converts RGB32 color to YUV. */ 185 * Basics of YUV -> RGB conversion. 189 * YUV -> RGB conversion macros 209 /* Converts YUV color to RGB565. */ 221 /* Converts YUV color to RGB32. */ 234 /* Converts YUV color to separated RGB32 colors. * [all...] |
/cts/apps/CameraITS/pymodules/its/ |
device.py | 65 CAP_YUV = {"format":"yuv"} 67 CAP_RAW_YUV = [{"format":"raw"}, {"format":"yuv"}] 68 CAP_DNG_YUV = [{"format":"dng"}, {"format":"yuv"}] 71 CAP_YUV_JPEG = [{"format":"yuv"}, {"format":"jpeg"}] 72 CAP_RAW_YUV_JPEG = [{"format":"raw"}, {"format":"yuv"}, {"format":"jpeg"}] 73 CAP_DNG_YUV_JPEG = [{"format":"dng"}, {"format":"yuv"}, {"format":"jpeg"}] 354 format(s) of the captured image. The formats may be "yuv", "jpeg", 355 "dng", "raw", or "raw10". The default is a YUV420 frame ("yuv") 359 request images back in multiple formats (e.g.) raw+yuv, raw+jpeg, 360 yuv+jpeg, raw+yuv+jpeg. If the size is omitted for a surface, th [all...] |
/cts/tests/tests/media/src/android/media/cts/ |
Vp8CodecTestBase.java | 82 // were calculated and were written to yuv file. 259 outputIvfBaseName + resolutionScales[i]+ ".yuv"; 316 private static byte[] YUV420ToNV(int width, int height, byte[] yuv) { 317 byte[] nv = new byte[yuv.length]; 319 System.arraycopy(yuv, 0, nv, 0, width * height); 326 nv[nv_offset++] = yuv[u_offset++]; 327 nv[nv_offset++] = yuv[v_offset++]; 338 byte[] yuv = new byte[width * height * 3 / 2]; 342 System.arraycopy(nv12, i * stride, yuv, i * width, width); 352 yuv[u_offset++] = nv12[nv_offset++] 541 FileOutputStream yuv = null; local [all...] |
/external/chromium_org/third_party/mesa/src/src/gallium/auxiliary/vl/ |
vl_compositor.h | 110 void *yuv; member in struct:vl_compositor::__anon18416 127 * set yuv -> rgba conversion matrix
|
/external/mesa3d/src/gallium/auxiliary/vl/ |
vl_compositor.h | 110 void *yuv; member in struct:vl_compositor::__anon31813 127 * set yuv -> rgba conversion matrix
|
/external/chromium_org/third_party/mesa/src/src/gallium/auxiliary/util/ |
u_format.csv | 140 # YUV formats 141 # http://www.fourcc.org/yuv.php#UYVY 142 PIPE_FORMAT_UYVY , subsampled, 2, 1, x32 , , , , xyz1, yuv 143 # http://www.fourcc.org/yuv.php#YUYV (a.k.a http://www.fourcc.org/yuv.php#YUY2) 144 PIPE_FORMAT_YUYV , subsampled, 2, 1, x32 , , , , xyz1, yuv 268 PIPE_FORMAT_YV12 , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv 269 PIPE_FORMAT_YV16 , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv 270 PIPE_FORMAT_IYUV , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv 271 PIPE_FORMAT_NV12 , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv [all...] |
/external/mesa3d/src/gallium/auxiliary/util/ |
u_format.csv | 140 # YUV formats 141 # http://www.fourcc.org/yuv.php#UYVY 142 PIPE_FORMAT_UYVY , subsampled, 2, 1, x32 , , , , xyz1, yuv 143 # http://www.fourcc.org/yuv.php#YUYV (a.k.a http://www.fourcc.org/yuv.php#YUY2) 144 PIPE_FORMAT_YUYV , subsampled, 2, 1, x32 , , , , xyz1, yuv 268 PIPE_FORMAT_YV12 , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv 269 PIPE_FORMAT_YV16 , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv 270 PIPE_FORMAT_IYUV , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv 271 PIPE_FORMAT_NV12 , other, 1, 1, x8 , x8 , x8 , x8 , xyzw, yuv [all...] |