/hardware/intel/img/hwcomposer/merrifield/ips/common/ |
GrallocBufferBase.cpp | 61 mStride.yuv.yStride = yStride; 62 mStride.yuv.uvStride = uvStride; 67 mStride.yuv.yStride = yStride; 68 mStride.yuv.uvStride = uvStride; 74 mStride.yuv.yStride = yStride; 75 mStride.yuv.uvStride = uvStride; 81 mStride.yuv.yStride = yStride; 82 mStride.yuv.uvStride = uvStride;
|
PlaneCapabilities.cpp | 128 if (stride.yuv.yStride > maxStride) { 129 VTRACE("stride %d is too large", stride.yuv.yStride);
|
/hardware/intel/img/hwcomposer/moorefield_hdmi/ips/common/ |
GrallocBufferBase.cpp | 63 mStride.yuv.yStride = yStride; 64 mStride.yuv.uvStride = uvStride; 69 mStride.yuv.yStride = yStride; 70 mStride.yuv.uvStride = uvStride; 76 mStride.yuv.yStride = yStride; 77 mStride.yuv.uvStride = uvStride; 83 mStride.yuv.yStride = yStride; 84 mStride.yuv.uvStride = uvStride;
|
OverlayPlaneBase.cpp | 394 stride.yuv.yStride = yStride; 395 stride.yuv.uvStride = uvStride; 400 stride.yuv.yStride = yStride; 401 stride.yuv.uvStride = uvStride; 407 stride.yuv.yStride = yStride; 408 stride.yuv.uvStride = uvStride; 414 stride.yuv.yStride = yStride; 415 stride.yuv.uvStride = uvStride; 667 uint32_t yStride = mapper.getStride().yuv.yStride; 668 uint32_t uvStride = mapper.getStride().yuv.uvStride [all...] |
/prebuilts/misc/windows/sdl2/test/ |
testoverlay2.c | 154 RGBtoYUV(Uint8 * rgb, int *yuv, int monochrome, int luminance) 158 yuv[0] = (int)(0.299 * rgb[0] + 0.587 * rgb[1] + 0.114 * rgb[2]); 159 yuv[1] = 128; 160 yuv[2] = 128; 162 yuv[0] = (int)(0.257 * rgb[0]) + (0.504 * rgb[1]) + (0.098 * rgb[2]) + 16; 163 yuv[1] = 128; 164 yuv[2] = 128; 168 yuv[0] = (int)(0.299 * rgb[0] + 0.587 * rgb[1] + 0.114 * rgb[2]); 169 yuv[1] = (int)((rgb[2] - yuv[0]) * 0.565 + 128) 190 int yuv[3]; local [all...] |
/frameworks/base/core/jni/android/graphics/ |
YuvToJpegEncoder.h | 13 /** Create an encoder based on the YUV format. 15 * @param pixelFormat The yuv pixel format as defined in ui/PixelFormat.h. 23 /** Encode YUV data to jpeg, which is output to a stream. 26 * @param inYuv The input yuv data. 27 * @param width Width of the the Yuv data in terms of pixels. 28 * @param height Height of the Yuv data in terms of pixels. 45 uint8_t* yuv, int* offsets) = 0; 55 void deinterleaveYuv(uint8_t* yuv, int width, int height, 59 void compress(jpeg_compress_struct* cinfo, uint8_t* yuv, int* offsets); 69 void compress(jpeg_compress_struct* cinfo, uint8_t* yuv, int* offsets) [all...] |
YuvToJpegEncoder.cpp | 89 uint8_t* yuv, int* offsets) { 101 uint8_t* yPlanar = yuv + offsets[0]; 102 uint8_t* vuPlanar = yuv + offsets[1]; //width * height; 165 uint8_t* yuv, int* offsets) { 181 uint8_t* yuvOffset = yuv + offsets[0]; 207 void Yuv422IToJpegEncoder::deinterleave(uint8_t* yuv, uint8_t* yRows, uint8_t* uRows, 212 uint8_t* yuvSeg = yuv + (rowIndex + row) * fStrides[0]; 240 jbyte* yuv = env->GetByteArrayElements(inYuv, NULL); local 248 encoder->encode(strm, yuv, width, height, imgOffsets, jpegQuality); 253 env->ReleaseByteArrayElements(inYuv, yuv, 0) [all...] |
/cts/tests/camera/src/android/hardware/camera2/cts/ |
crop_yuvf_420_to_yuvx_444.rs | 19 // Must be YUV 420 888 (flexible YUV) 33 yuvx_444 yuv = { py, pu, pv }; 35 return yuv;
|
/cts/tests/tests/graphics/src/android/graphics/cts/ |
YuvImageTest.java | 96 byte[] yuv = new byte[width * height * 2]; 102 new YuvImage(yuv, FORMATS[i], width, height, null); 120 image = new YuvImage(yuv, FORMATS[i], width, height, null); 197 byte[] yuv = convertArgbsToYuvs(argb, stride, height, ImageFormat.NV21); 201 YuvImage image = new YuvImage(yuv, ImageFormat.NV21, width, height, strides); 202 assertEquals(yuv, image.getYuvData()); 234 byte[] yuv = convertArgbsToYuvs(argb, stride, height, format); 242 image = new YuvImage(yuv, format, width, height, strides); 299 byte[] yuv = new byte[width * height * 308 yuv[idx] = yuvColor[0] [all...] |
/external/libdrm/tests/util/ |
format.h | 59 const struct util_yuv_info yuv; member in struct:util_format_info
|
pattern.c | 76 static void fill_smpte_yuv_planar(const struct util_yuv_info *yuv, 109 unsigned int cs = yuv->chroma_stride; 110 unsigned int xsub = yuv->xsub; 111 unsigned int ysub = yuv->ysub; 180 static void fill_smpte_yuv_packed(const struct util_yuv_info *yuv, void *mem, 212 unsigned char *y_mem = (yuv->order & YUV_YC) ? mem : mem + 1; 213 unsigned char *c_mem = (yuv->order & YUV_CY) ? mem : mem + 1; 214 unsigned int u = (yuv->order & YUV_YCrCb) ? 2 : 0; 215 unsigned int v = (yuv->order & YUV_YCbCr) ? 2 : 0; 475 return fill_smpte_yuv_packed(&info->yuv, planes[0], width 614 const struct util_yuv_info *yuv = &info->yuv; local 647 const struct util_yuv_info *yuv = &info->yuv; local [all...] |
/external/mesa3d/src/gallium/state_trackers/xa/ |
xa_yuv.c | 70 xa_yuv_bind_samplers(struct xa_context *r, struct xa_surface *yuv[]) 88 u_sampler_view_default_template(&view_templ, yuv[i]->tex, 89 yuv[i]->tex->format); 92 r->pipe->create_sampler_view(r->pipe, yuv[i]->tex, &view_templ); 121 struct xa_surface *dst, struct xa_surface *yuv[]) 137 xa_yuv_bind_samplers(r, yuv); 153 scale_x * w, scale_y * h, x, y, w, h, yuv); local
|
xa_context.h | 51 * @yuv is an array of three xa_yuv_component surfaces. 65 struct xa_surface *dst, struct xa_surface *yuv[]);
|
/packages/apps/LegacyCamera/jni/feature_mos/src/mosaic/ |
ImageUtils.cpp | 366 // VC 8 does not like calling free on yuv->Y.ptr since it is in 368 // calling mapYUVInforToImage yuv->Y.ptr points to the begginning 370 YUVinfo *yuv = (YUVinfo *) calloc(sizeof(YUVinfo), 1); local 371 if (yuv) { 372 yuv->Y.width = yuv->Y.pitch = width; 373 yuv->Y.height = height; 374 yuv->Y.border = yuv->U.border = yuv->V.border = (unsigned short) 0 [all...] |
/device/linaro/bootloader/edk2/AppPkg/Applications/Python/Python-2.7.2/Modules/ |
yuvconvert.c | 2 #include "yuv.h"
5 yuv_sv411_to_cl422dc(int invert, void *data, void *yuv, int width, int height)
8 struct yuv422 *out_even = yuv;
42 yuv_sv411_to_cl422dc_quartersize(int invert, void *data, void *yuv,
48 struct yuv422 *out_even = yuv;
81 yuv_sv411_to_cl422dc_sixteenthsize(int invert, void *data, void *yuv,
88 struct yuv422 *out_even = yuv;
|
/external/python/cpython2/Modules/ |
yuvconvert.c | 2 #include "yuv.h" 5 yuv_sv411_to_cl422dc(int invert, void *data, void *yuv, int width, int height) 8 struct yuv422 *out_even = yuv; 42 yuv_sv411_to_cl422dc_quartersize(int invert, void *data, void *yuv, 48 struct yuv422 *out_even = yuv; 81 yuv_sv411_to_cl422dc_sixteenthsize(int invert, void *data, void *yuv, 88 struct yuv422 *out_even = yuv;
|
/external/zxing/qr_scanner/src/com/google/zxing/client/android/ |
PlanarYUVLuminanceSource.java | 24 * This object extends LuminanceSource around an array of YUV data returned from the camera driver, 95 byte[] yuv = yuvData; 98 System.arraycopy(yuv, inputOffset, matrix, outputOffset, width); 113 byte[] yuv = yuvData; 119 int grey = yuv[inputOffset + x] & 0xff;
|
/frameworks/base/graphics/java/android/graphics/ |
YuvImage.java | 22 * YuvImage contains YUV data and provides a method that compresses a region of 23 * the YUV data to a Jpeg. The YUV data should be provided as a single byte 27 * To compress a rectangle region in the YUV data, users have to specify the 39 * The YUV format as defined in {@link ImageFormat}. 44 * The raw YUV data. 68 * @param yuv The YUV data. In the case of more than one image plane, all the planes must be 70 * @param format The YUV data format as defined in {@link ImageFormat}. 73 * @param strides (Optional) Row bytes of each image plane. If yuv contains padding, the strid [all...] |
/hardware/libhardware/modules/camera/3_4/arc/ |
jpeg_compressor.h | 60 bool Compress(jpeg_compress_struct* cinfo, const uint8_t* yuv);
|
jpeg_compressor.cpp | 146 bool JpegCompressor::Compress(jpeg_compress_struct* cinfo, const uint8_t* yuv) { 154 uint8_t* y_plane = const_cast<uint8_t*>(yuv); 155 uint8_t* u_plane = const_cast<uint8_t*>(yuv + y_plane_size); 156 uint8_t* v_plane = const_cast<uint8_t*>(yuv + y_plane_size + uv_plane_size);
|
/frameworks/rs/driver/runtime/ |
rs_structs.h | 42 uint32_t yuv; member in struct:Allocation::__anon46692::__anon46693 77 } yuv; member in struct:Allocation::__anon46692::DrvState
|
/frameworks/rs/tests/java_api/LivePreview/src/com/android/rs/livepreview/ |
RsYuv.java | 112 void execute(byte[] yuv) { 113 mAllocationIn.copyFrom(yuv);
|
/hardware/intel/img/hwcomposer/moorefield_hdmi/ips/anniedale/ |
PlaneCapabilities.cpp | 128 if (stride.yuv.yStride > maxStride) { 129 VLOGTRACE("stride %d is too large", stride.yuv.yStride);
|
/hardware/intel/img/hwcomposer/merrifield/include/ |
DataBuffer.h | 40 } yuv; member in union:android::intel::stride::__anon47681
|
/hardware/intel/img/hwcomposer/moorefield_hdmi/include/ |
DataBuffer.h | 40 } yuv; member in union:android::intel::stride::__anon47726
|