/hardware/intel/img/hwcomposer/merrifield/ips/common/ |
GrallocBufferBase.cpp | 61 mStride.yuv.yStride = yStride; 62 mStride.yuv.uvStride = uvStride; 67 mStride.yuv.yStride = yStride; 68 mStride.yuv.uvStride = uvStride; 74 mStride.yuv.yStride = yStride; 75 mStride.yuv.uvStride = uvStride; 81 mStride.yuv.yStride = yStride; 82 mStride.yuv.uvStride = uvStride;
|
PlaneCapabilities.cpp | 128 if (stride.yuv.yStride > maxStride) { 129 VTRACE("stride %d is too large", stride.yuv.yStride);
|
/hardware/intel/img/hwcomposer/moorefield_hdmi/ips/common/ |
GrallocBufferBase.cpp | 63 mStride.yuv.yStride = yStride; 64 mStride.yuv.uvStride = uvStride; 69 mStride.yuv.yStride = yStride; 70 mStride.yuv.uvStride = uvStride; 76 mStride.yuv.yStride = yStride; 77 mStride.yuv.uvStride = uvStride; 83 mStride.yuv.yStride = yStride; 84 mStride.yuv.uvStride = uvStride;
|
/prebuilts/misc/windows/sdl2/test/ |
testoverlay2.c | 154 RGBtoYUV(Uint8 * rgb, int *yuv, int monochrome, int luminance) 158 yuv[0] = (int)(0.299 * rgb[0] + 0.587 * rgb[1] + 0.114 * rgb[2]); 159 yuv[1] = 128; 160 yuv[2] = 128; 162 yuv[0] = (int)(0.257 * rgb[0]) + (0.504 * rgb[1]) + (0.098 * rgb[2]) + 16; 163 yuv[1] = 128; 164 yuv[2] = 128; 168 yuv[0] = (int)(0.299 * rgb[0] + 0.587 * rgb[1] + 0.114 * rgb[2]); 169 yuv[1] = (int)((rgb[2] - yuv[0]) * 0.565 + 128) 190 int yuv[3]; local [all...] |
/frameworks/base/core/jni/android/graphics/ |
YuvToJpegEncoder.h | 13 /** Create an encoder based on the YUV format. 15 * @param pixelFormat The yuv pixel format as defined in ui/PixelFormat.h. 23 /** Encode YUV data to jpeg, which is output to a stream. 26 * @param inYuv The input yuv data. 27 * @param width Width of the the Yuv data in terms of pixels. 28 * @param height Height of the Yuv data in terms of pixels. 45 uint8_t* yuv, int* offsets) = 0; 55 void deinterleaveYuv(uint8_t* yuv, int width, int height, 59 void compress(jpeg_compress_struct* cinfo, uint8_t* yuv, int* offsets); 69 void compress(jpeg_compress_struct* cinfo, uint8_t* yuv, int* offsets) [all...] |
YuvToJpegEncoder.cpp | 74 uint8_t* yuv, int* offsets) { 86 uint8_t* yPlanar = yuv + offsets[0]; 87 uint8_t* vuPlanar = yuv + offsets[1]; //width * height; 150 uint8_t* yuv, int* offsets) { 166 uint8_t* yuvOffset = yuv + offsets[0]; 192 void Yuv422IToJpegEncoder::deinterleave(uint8_t* yuv, uint8_t* yRows, uint8_t* uRows, 197 uint8_t* yuvSeg = yuv + (rowIndex + row) * fStrides[0]; 225 jbyte* yuv = env->GetByteArrayElements(inYuv, NULL); local 233 encoder->encode(strm, yuv, width, height, imgOffsets, jpegQuality); 238 env->ReleaseByteArrayElements(inYuv, yuv, 0) [all...] |
/external/opencv3/modules/imgproc/test/ |
test_cvtyuv.cpp | 7 #undef YUV 9 typedef Vec3b YUV; 23 virtual YUV read(const Mat& yuv, int row, int col) = 0; 74 virtual void write(Mat& yuv, int row, int col, const YUV& val) = 0; 135 void write(Mat& yuv, int row, int col, const YUV& val) 137 int h = yuv.rows * 2 / 3; 139 yuv.ptr<uchar>(row)[col] = val[0] [all...] |
/cts/tests/camera/src/android/hardware/camera2/cts/ |
crop_yuvf_420_to_yuvx_444.rs | 19 // Must be YUV 420 888 (flexible YUV) 33 yuvx_444 yuv = { py, pu, pv }; 35 return yuv;
|
/cts/tests/tests/graphics/src/android/graphics/cts/ |
YuvImageTest.java | 95 byte[] yuv = new byte[width * height * 2]; 101 image = new YuvImage(yuv, mFormats[i], width, height, null); 119 image = new YuvImage(yuv, mFormats[i], width, height, null); 133 image = new YuvImage(yuv, format, -1, height, null); 141 image = new YuvImage(yuv, format, width, -1, null); 147 // abnormal case: yuv array is null 150 fail("not catching null yuv data"); 205 byte[] yuv = convertArgbsToYuvs(argb, stride, height, ImageFormat.NV21); 209 YuvImage image = new YuvImage(yuv, ImageFormat.NV21, width, height, strides); 210 assertEquals(yuv, image.getYuvData()) [all...] |
/external/libdrm/tests/util/ |
format.h | 59 const struct util_yuv_info yuv; member in struct:util_format_info
|
pattern.c | 76 static void fill_smpte_yuv_planar(const struct util_yuv_info *yuv, 109 unsigned int cs = yuv->chroma_stride; 110 unsigned int xsub = yuv->xsub; 111 unsigned int ysub = yuv->ysub; 180 static void fill_smpte_yuv_packed(const struct util_yuv_info *yuv, void *mem, 212 unsigned char *y_mem = (yuv->order & YUV_YC) ? mem : mem + 1; 213 unsigned char *c_mem = (yuv->order & YUV_CY) ? mem : mem + 1; 214 unsigned int u = (yuv->order & YUV_YCrCb) ? 2 : 0; 215 unsigned int v = (yuv->order & YUV_YCbCr) ? 2 : 0; 475 return fill_smpte_yuv_packed(&info->yuv, planes[0], width 614 const struct util_yuv_info *yuv = &info->yuv; local 647 const struct util_yuv_info *yuv = &info->yuv; local [all...] |
/external/mesa3d/src/gallium/state_trackers/xa/ |
xa_yuv.c | 70 xa_yuv_bind_samplers(struct xa_context *r, struct xa_surface *yuv[]) 88 u_sampler_view_default_template(&view_templ, yuv[i]->tex, 89 yuv[i]->tex->format); 92 r->pipe->create_sampler_view(r->pipe, yuv[i]->tex, &view_templ); 121 struct xa_surface *dst, struct xa_surface *yuv[]) 137 xa_yuv_bind_samplers(r, yuv); 152 scale_x * w, scale_y * h, x, y, w, h, yuv); local
|
xa_context.h | 49 * @yuv is an array of three xa_yuv_component surfaces. 63 struct xa_surface *dst, struct xa_surface *yuv[]);
|
/development/perftests/panorama/feature_mos/src/mosaic/ |
ImageUtils.cpp | 366 // VC 8 does not like calling free on yuv->Y.ptr since it is in 368 // calling mapYUVInforToImage yuv->Y.ptr points to the begginning 370 YUVinfo *yuv = (YUVinfo *) calloc(sizeof(YUVinfo), 1); local 371 if (yuv) { 372 yuv->Y.width = yuv->Y.pitch = width; 373 yuv->Y.height = height; 374 yuv->Y.border = yuv->U.border = yuv->V.border = (unsigned short) 0 [all...] |
/packages/apps/LegacyCamera/jni/feature_mos/src/mosaic/ |
ImageUtils.cpp | 366 // VC 8 does not like calling free on yuv->Y.ptr since it is in 368 // calling mapYUVInforToImage yuv->Y.ptr points to the begginning 370 YUVinfo *yuv = (YUVinfo *) calloc(sizeof(YUVinfo), 1); local 371 if (yuv) { 372 yuv->Y.width = yuv->Y.pitch = width; 373 yuv->Y.height = height; 374 yuv->Y.border = yuv->U.border = yuv->V.border = (unsigned short) 0 [all...] |
/external/zxing/qr_scanner/src/com/google/zxing/client/android/ |
PlanarYUVLuminanceSource.java | 24 * This object extends LuminanceSource around an array of YUV data returned from the camera driver, 95 byte[] yuv = yuvData; 98 System.arraycopy(yuv, inputOffset, matrix, outputOffset, width); 113 byte[] yuv = yuvData; 119 int grey = yuv[inputOffset + x] & 0xff;
|
/frameworks/base/graphics/java/android/graphics/ |
YuvImage.java | 22 * YuvImage contains YUV data and provides a method that compresses a region of 23 * the YUV data to a Jpeg. The YUV data should be provided as a single byte 27 * To compress a rectangle region in the YUV data, users have to specify the 39 * The YUV format as defined in {@link ImageFormat}. 44 * The raw YUV data. 68 * @param yuv The YUV data. In the case of more than one image plane, all the planes must be 70 * @param format The YUV data format as defined in {@link ImageFormat}. 73 * @param strides (Optional) Row bytes of each image plane. If yuv contains padding, the strid [all...] |
/frameworks/rs/ |
rsGrallocConsumer.cpp | 55 if (a->mHal.state.yuv) { 56 bq->setDefaultBufferFormat(a->mHal.state.yuv); 165 // For YUV Allocations, we need to populate the drvState with details of how 170 // mHal.state.yuv contains the requested format, 172 if (mAlloc[idx]->mHal.state.yuv == HAL_PIXEL_FORMAT_YCbCr_420_888) { 194 mAlloc[idx]->mHal.drvState.yuv.shift = 1; 195 mAlloc[idx]->mHal.drvState.yuv.step = ycbcr.chroma_step; 225 mAlloc[idx]->mHal.drvState.yuv.shift = 1; 226 mAlloc[idx]->mHal.drvState.yuv.step = 1;
|
/frameworks/rs/driver/runtime/ |
rs_structs.h | 42 uint32_t yuv; member in struct:Allocation::__anon29361::__anon29362 77 } yuv; member in struct:Allocation::__anon29361::DrvState
|
/frameworks/av/media/libstagefright/codecs/avc/enc/ |
SoftAVCEncoder.h | 53 int32_t bindOutputBuffer(int32_t index, uint8_t **yuv);
|
/device/moto/shamu/camera/QCamera/stack/mm-camera-test/src/ |
mm_qcamera_display.c | 69 } yuv; variable in typeref:union:__anon5304 178 /* Initialize yuv structure */ 179 yuv.list.count = 1; 181 e = &yuv.list.req[0]; 240 result = ioctl(fb_fd, MSMFB_BLIT, &yuv.list); 461 /* Initialize yuv structure */ 462 yuv.list.count = 1; 463 e = &yuv.list.req[0]; 521 /* Initialize yuv structure */ 522 yuv.list.count = 1 [all...] |
/frameworks/rs/java/tests/LivePreview/src/com/android/rs/livepreview/ |
RsYuv.java | 112 void execute(byte[] yuv) { 113 mAllocationIn.copyFrom(yuv);
|
/hardware/intel/img/hwcomposer/moorefield_hdmi/ips/anniedale/ |
PlaneCapabilities.cpp | 128 if (stride.yuv.yStride > maxStride) { 129 VLOGTRACE("stride %d is too large", stride.yuv.yStride);
|
/frameworks/av/media/libstagefright/codecs/avc/common/include/ |
avcapi_common.h | 199 @param yuv The address of the yuv pointer returned to the AVC lib. 202 typedef int (*FunctionType_FrameBind)(void *userData, int indx, uint8 **yuv);
|
/hardware/intel/img/hwcomposer/merrifield/include/ |
DataBuffer.h | 40 } yuv; member in union:android::intel::stride::__anon30794
|