/frameworks/compile/libbcc/lib/Renderscript/runtime/ |
rs_allocation.c | 261 const uint32_t yuvID = alloc->mHal.state.yuv; 285 const uint32_t yuvID = alloc->mHal.state.yuv;
|
/frameworks/rs/driver/ |
rsdAllocation.cpp | 254 static size_t DeriveYUVLayout(int yuv, Allocation::Hal::DrvState *state) { 255 // YUV only supports basic 2d 259 switch(yuv) { 327 } else if (alloc->mHal.state.yuv) { 328 o += DeriveYUVLayout(alloc->mHal.state.yuv, &alloc->mHal.drvState); 809 if (alloc->mHal.state.yuv) { 810 DeriveYUVLayout(alloc->mHal.state.yuv, &alloc->mHal.drvState); 878 if (alloc->mHal.state.yuv) { [all...] |
/hardware/samsung_slsi/exynos5/gralloc/ |
gralloc.cpp | 207 ALOGE("invalid yuv format %d\n", format); 274 ALOGE("invalid yuv format %d\n", format);
|
/hardware/ti/omap3/omx/image/src/openmax_il/jpeg_dec/tests/ |
JPEGTest.c | 357 PRINT("THIS IS A YUV 411 ENCODED IMAGE \n" ); 668 printf("usage: %S <input.jpg> <output.yuv> <nOutformat-1:420 4:422 9:RGB16 10: RGB24 11:RGB32 12:BGR32> <nResizeMode> <nRepeat> <nSections> <nXOrigin> <nYOrigin> <nXLenght> <nYLenght>\n", 671 printf("usage: %s <input.jpg> <output.yuv> <nOutformat-1:Original(default) 4:422 9:RGB16 10: RGB24 11:RGB32 12:BGR32> <nResizeMode> <nRepeat> <nSections> <nXOrigin> <nYOrigin> <nXLenght> <nYLenght>\n", 676 printf("Example ./JpegTestCommon patterns/shrek.jpg shrek.yuv 1 50 \n"); 679 printf("Example ./JpegTestCommon patterns/shrek.jpg shrek.yuv 1 50 0\n"); 680 printf("Example ./JpegTestCommon patterns/shrek.jpg shrek.yuv 1 50 9\n"); 683 printf("Example ./JpegTestCommon patterns/shrek.jpg shrek.yuv 4 50 0 1\n"); 687 printf("Example ./JpegTestCommon patterns/shrek.jpg shrek.yuv 1 100 0 0 192 48 288 256\n"); [all...] |
/hardware/ti/omap3/omx/video/src/openmax_il/prepost_processor/inc/ |
OMX_VPP_Utils.h | 268 OMX_U32 ulOutWidth; /* RGB/YUV picture buffer width */
269 OMX_U32 ulOutHeight; /* RGB/YUV picture buffer height */
|
/device/generic/goldfish/camera/ |
EmulatedFakeCameraDevice.h | 126 * Pixel colors in YUV format used when drawing the checker board.
|
/device/generic/goldfish/camera/fake-pipeline2/ |
Scene.h | 145 // RGB->YUV conversion
|
/device/lge/mako/camera/QCamera/HAL/core/inc/ |
QCameraHWI.h | 303 * Request the camera HAL to store meta data or real YUV data in the video 305 * it is not called, the default camera HAL behavior is to store real YUV 317 * but all camera HALs should support storing real YUV data in the video 327 * the camera HAL to store real YUV data in the video
|
/device/lge/mako/camera/QCamera/stack/mm-jpeg-interface/inc/ |
mm_jpeg_interface.h | 88 /* src img format: YUV, Bitstream */
|
/device/lge/mako/camera/ |
QCameraHWI.h | 306 * Request the camera HAL to store meta data or real YUV data in the video 308 * it is not called, the default camera HAL behavior is to store real YUV 320 * but all camera HALs should support storing real YUV data in the video 330 * the camera HAL to store real YUV data in the video
|
/external/chromium/third_party/libjingle/source/talk/session/phone/ |
videocommon.h | 60 // http://www.fourcc.org/yuv.php
|
/external/kernel-headers/original/video/ |
dsscomp.h | 45 OMAP_DSS_COLOR_NV12 = 1 << 14, /* NV12 format: YUV 4:2:0 */ 92 /* YUV to RGB color conversion info */ 198 /* YUV VC1 range mapping info */ 311 * 4) cconv and vc1 only needs to be filled for YUV color modes.
|
/external/libvpx/libvpx/test/ |
keyframe_test.cc | 122 ::libvpx_test::I420VideoSource video("hantro_collage_w352h288.yuv", 352, 288,
|
/external/libyuv/files/include/libyuv/ |
video_common.h | 35 // http://www.fourcc.org/yuv.php
|
/external/webrtc/test/testsupport/ |
fileutils.h | 123 // extension - File extension, without the dot, i.e. "bmp" or "yuv".
|
/frameworks/av/libvideoeditor/vss/inc/ |
M4PTO3GPP_InternalTypes.h | 189 M4OSA_UInt32 m_NbCurrentFrame; /**< Index of the current YUV frame encoded */
|
/frameworks/av/libvideoeditor/vss/video_filters/src/ |
M4VIFI_RGB888toYUV420.c | 92 /* two YUV rows are computed at each pass */
|
/frameworks/av/media/libstagefright/codecs/avc/enc/ |
SoftAVCEncoder.cpp | 158 void *userData, int32_t index, uint8_t **yuv) { 161 return encoder->bindOutputBuffer(index, yuv); [all...] |
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/include/ |
mp4dec_api.h | 64 uint8 *data; /* pointer to output YUV buffer */
|
/frameworks/av/services/camera/libcameraservice/ |
CameraHardwareInterface.h | 223 * Request the camera hal to store meta data or real YUV data in 226 * hal behavior is to store real YUV data in the video buffers. 238 * buffers, but all camera hal should support storing real YUV data 248 * the camera hal to store real YUV data in the video
|
/frameworks/base/graphics/java/android/graphics/ |
ImageFormat.java | 34 * <p>Android YUV format.</p>
|
/frameworks/native/libs/ui/ |
PixelFormat.cpp | 100 // YUV format from the HAL are handled here
|
/frameworks/rs/ |
rs.spec | 42 param uint32_t yuv
|
rsAllocation.h | 54 uint32_t yuv; member in struct:android::renderscript::Allocation::Hal::State
|
/hardware/libhardware/tests/camera2/ |
CameraBurstTests.cpp | 89 /* this assumes the format is YUV420sp or flexible YUV */
|