/frameworks/base/core/jni/android/graphics/ |
YuvToJpegEncoder.h | 13 /** Create an encoder based on the YUV format. 15 * @param pixelFormat The yuv pixel format as defined in ui/PixelFormat.h. 23 /** Encode YUV data to jpeg, which is output to a stream. 26 * @param inYuv The input yuv data. 27 * @param width Width of the the Yuv data in terms of pixels. 28 * @param height Height of the Yuv data in terms of pixels. 45 uint8_t* yuv, int* offsets) = 0; 55 void deinterleaveYuv(uint8_t* yuv, int width, int height, 59 void compress(jpeg_compress_struct* cinfo, uint8_t* yuv, int* offsets); 69 void compress(jpeg_compress_struct* cinfo, uint8_t* yuv, int* offsets) [all...] |
YuvToJpegEncoder.cpp | 72 uint8_t* yuv, int* offsets) { 84 uint8_t* yPlanar = yuv + offsets[0]; 85 uint8_t* vuPlanar = yuv + offsets[1]; //width * height; 145 uint8_t* yuv, int* offsets) { 161 uint8_t* yuvOffset = yuv + offsets[0]; 186 void Yuv422IToJpegEncoder::deinterleave(uint8_t* yuv, uint8_t* yRows, uint8_t* uRows, 189 uint8_t* yuvSeg = yuv + (rowIndex + row) * fStrides[0]; 217 jbyte* yuv = env->GetByteArrayElements(inYuv, NULL); local 226 encoder->encode(strm, yuv, width, height, imgOffsets, jpegQuality); 229 env->ReleaseByteArrayElements(inYuv, yuv, 0) [all...] |
/cts/tests/tests/graphics/src/android/graphics/cts/ |
YuvImageTest.java | 105 byte[] yuv = new byte[width * height * 2]; 111 image = new YuvImage(yuv, mFormats[i], width, height, null); 129 image = new YuvImage(yuv, mFormats[i], width, height, null); 143 image = new YuvImage(yuv, format, -1, height, null); 151 image = new YuvImage(yuv, format, width, -1, null); 157 // abnormal case: yuv array is null 160 fail("not catching null yuv data"); 228 byte[] yuv = convertArgbsToYuvs(argb, stride, height, format); 236 image = new YuvImage(yuv, format, width, height, strides); 289 byte[] yuv = new byte[width * height [all...] |
/external/opencore/codecs_v2/video/m4v_h263/dec/include/ |
pvvideodecoderinterface.h | 39 virtual bool DecodeVideoFrame(uint8 *bitstream[], uint32 *timestamp, int32 *buffer_size, uint *use_ext_ts, uint8 *yuv) = 0; 42 virtual bool DecodeVideoFrame(uint8 *bitstream[], uint32 *timestamp, int32 *buffer_size, uint *use_ext_ts, uint8 *yuv, OsclAOStatus *asynch) 48 OSCL_UNUSED_ARG(yuv); 71 virtual void SetReferenceYUV(uint8 *YUV) = 0; 77 virtual void DecPostProcess(uint8 *YUV) = 0;
|
pvm4vdecoder.h | 57 This function takes the compressed bitstreams of a multiple layer video and decodes the next YUV 4:2:0 frame to 59 should be passed into the decoder through yuv parameter. The allocated memory should be WORD aligned. The input 60 bitstream is decoded into this passed YUV buffer. The unpadded (non-multiple of 16) size of the frame can be 75 buffers. The frames are always decoded into the YUV-buffer that is passed in using the yuv frame pointer 76 parameter. This YUV frame buffer is kept as reference frame for decoding of the next frame. After decoding of 80 virtual bool DecodeVideoFrame(uint8 *bitstream[], uint32 *timestamp, int32 *buffer_size, uint *use_ext_timestamp, uint8 *yuv); 87 virtual void SetReferenceYUV(uint8 *YUV); 127 *yuv frame. If a separate yuv frame is not used for postprocessed frames NULL pointer can be passed in. In thi [all...] |
/external/opencore/codecs_v2/utilities/colorconvert/src/ |
ccrgb24toyuv420.cpp | 163 int32 ccrgb24toyuv(uint8 *rgb24, uint8 *yuv[], uint32 *param, uint8 *table[]); 171 uint8 *yuv[3]; local 185 yuv[0] = yuv420; 186 yuv[1] = yuv420 + size16; 187 yuv[2] = yuv[1] + (size16 >> 2); 189 return ccrgb24toyuv(rgb24, yuv, param, table); 218 int32 ccrgb24toyuv(uint8 *rgb24, uint8 *yuv[], uint32 *param, uint8 *table[]) 239 /* do padding if input RGB size(height) is different from the output YUV size(height_16) */ 250 tempY = yuv[0] + offset [all...] |
ccrgb16toyuv420.cpp | 173 int32 ccrgb16toyuv(uint8 *rgb16, uint8 *yuv[], uint32 *param, uint8 *table[]); 174 int32 ccrgb16toyuv_wo_colorkey(uint8 *rgb16, uint8 *yuv[], uint32 *param, uint8 *table[]); 182 uint8 *yuv[3]; local 200 yuv[0] = yuv420; 201 yuv[1] = yuv420 + size16; 202 yuv[2] = yuv[1] + (size16 >> 2); 206 return ccrgb16toyuv(rgb16, yuv, param, table); 210 return ccrgb16toyuv_wo_colorkey(rgb16, yuv, param, table); 217 int32 ccrgb16toyuv(uint8 *rgb16, uint8 *yuv[], uint32 *param, uint8 *table[] [all...] |
ccrgb12toyuv420.cpp | 162 int32 ccrgb12toyuv(uint8 *rgb12, uint8 *yuv[], uint32 *param, uint8 *table[]); 170 uint8 *yuv[3]; local 184 yuv[0] = yuv420; 185 yuv[1] = yuv420 + size16; 186 yuv[2] = yuv[1] + (size16 >> 2); 188 return ccrgb12toyuv(rgb12, yuv, param, table); 216 int32 ccrgb12toyuv(uint8 *rgb12, uint8 *yuv[], uint32 *param, uint8 *table[]) 246 /* do padding if input RGB size(height) is different from the output YUV size(height_16) */ 258 tempY = (uint32 *)yuv[0] + (offset >> 2) [all...] |
/frameworks/base/media/libstagefright/include/ |
AVCDecoder.h | 68 void *userData, int32_t index, uint8_t **yuv); 75 int32_t bindFrame(int32_t index, uint8_t **yuv);
|
/frameworks/base/graphics/java/android/graphics/ |
YuvImage.java | 22 * YuvImage contains YUV data and provides a method that compresses a region of 23 * the YUV data to a Jpeg. The YUV data should be provided as a single byte 27 * To compress a rectangle region in the YUV data, users have to specify the 39 * The YUV format as defined in {@link PixelFormat}. 44 * The raw YUV data. 68 * @param yuv The YUV data. In the case of more than one image plane, all the planes must be 70 * @param format The YUV data format as defined in {@link PixelFormat}. 73 * @param strides (Optional) Row bytes of each image plane. If yuv contains padding, the strid [all...] |
/external/opencore/codecs_v2/omx/omx_h264enc/include/ |
avc_enc.h | 92 int AVC_FrameBind(int indx, uint8** yuv); 99 void CopyToYUVIn(uint8* YUV, int width, int height, int width_16, int height_16); 101 /* RGB->YUV conversion */
|
/external/opencore/codecs_v2/video/avc_h264/common/include/ |
avcapi_common.h | 197 @param yuv The address of the yuv pointer returned to the AVC lib. 200 typedef int (*FunctionType_FrameBind)(void *userData, int indx, uint8 **yuv);
|
/frameworks/base/media/libstagefright/codecs/avc/common/include/ |
avcapi_common.h | 199 @param yuv The address of the yuv pointer returned to the AVC lib. 202 typedef int (*FunctionType_FrameBind)(void *userData, int indx, uint8 **yuv);
|
/external/opencore/codecs_v2/video/avc_h264/enc/include/ |
pvavcencoder.h | 61 int AVC_FrameBind(int indx, uint8** yuv); 71 void CopyToYUVIn(uint8* YUV, int width, int height, int width_16, int height_16); 74 /** Color conversion instance RGB24/RGB12/YUV420SEMI to YUV 420 */
|
/frameworks/base/media/libstagefright/codecs/avc/dec/ |
AVCDecoder.cpp | 445 void *userData, int32_t index, uint8_t **yuv) { 446 return static_cast<AVCDecoder *>(userData)->bindFrame(index, yuv); 468 int32_t AVCDecoder::bindFrame(int32_t index, uint8_t **yuv) { 477 *yuv = (uint8_t *)mFrames[index]->data();
|
/frameworks/base/camera/libcameraservice/ |
FakeCamera.cpp | 30 // TODO: All this rgb to yuv should probably be in a util class. 48 /* the equation used by the video code to translate YUV to RGB looks like this 280 static void convert_rgb16_to_yuv422(uint8_t *rgb, uint8_t *yuv, int width, int height) 302 ccrgb16toyuv_wo_colorkey(rgb, yuv, param, table);
|
/external/opencore/codecs_v2/omx/omx_h264enc/src/ |
avc_enc.cpp | 76 int CbAvcEncFrameBind_OMX(void *userData, int indx, uint8 **yuv) 80 return pAvcEnc->AVC_FrameBind(indx, yuv); 920 int AvcEncoder_OMX::AVC_FrameBind(int indx, uint8** yuv) 928 *yuv = iFramePtr[indx]; 935 /* Copy from YUV input to YUV frame inside AvcEnc lib */ 936 /* When input is not YUV, the color conv will write it directly to iVideoInOut. */ 939 void AvcEncoder_OMX::CopyToYUVIn(uint8 *YUV, int width, int height, int width_16, int height_16) 947 /* do padding if input RGB size(height) is different from the output YUV size(height_16) */ 974 u = (uint8*)(&(YUV[size])) [all...] |
/external/opencore/codecs_v2/video/avc_h264/enc/src/ |
pvavcencoder.cpp | 71 int CbAvcEncFrameBind(void *userData, int indx, uint8 **yuv) 75 return pAvcEnc->AVC_FrameBind(indx, yuv); 726 /* Copy from YUV input to YUV frame inside M4VEnc lib */ 727 /* When input is not YUV, the color conv will write it directly to iVideoInOut. */ 730 void PVAVCEncoder::CopyToYUVIn(uint8 *YUV, int width, int height, int width_16, int height_16) 738 /* do padding if input RGB size(height) is different from the output YUV size(height_16) */ 765 u = (uint8*)(&(YUV[size])); 766 v = (uint8*)(&(YUV[size*5/4])); 782 YUV += ((width - width_16) >> 1) [all...] |