/cts/apps/CameraITS/tests/scene3/ |
test_lens_position.py | 82 chart.img = its.image.normalize_img(its.image.get_image_patch( 84 its.image.write_image(chart.img, '%s_stat_i=%d_chart.jpg' % (NAME, i)) 86 chart.img) 109 chart.img = its.image.normalize_img(its.image.get_image_patch( 111 its.image.write_image(chart.img, '%s_move_i=%d_chart.jpg' % (NAME, i)) 113 chart.img)
|
/external/tensorflow/tensorflow/examples/android/jni/object_tracking/ |
image_utils.h | 52 Image<U>* const img) { 53 SCHECK(img->ValidPixel(x, y), "Marking invalid pixel in image! %d, %d", x, y); 63 const int max_y = MIN(y + d_y, img->height_less_one_); 74 const int max_x = MIN(x + d_x, img->width_less_one_); 77 bool* const top_row_start = (*img)[min_y] + min_x; 78 bool* const bottom_row_start = (*img)[max_y] + min_x;
|
/prebuilts/go/darwin-x86/src/image/jpeg/ |
writer_test.go | 246 img := image.NewRGBA(image.Rect(0, 0, 640, 480)) 247 bo := img.Bounds() 251 img.SetRGBA(x, y, color.RGBA{ 264 Encode(ioutil.Discard, img, options) 269 img := image.NewYCbCr(image.Rect(0, 0, 640, 480), image.YCbCrSubsampleRatio420) 270 bo := img.Bounds() 274 cy := img.YOffset(x, y) 275 ci := img.COffset(x, y) 276 img.Y[cy] = uint8(rnd.Intn(256)) 277 img.Cb[ci] = uint8(rnd.Intn(256) [all...] |
/prebuilts/go/linux-x86/src/image/jpeg/ |
writer_test.go | 246 img := image.NewRGBA(image.Rect(0, 0, 640, 480)) 247 bo := img.Bounds() 251 img.SetRGBA(x, y, color.RGBA{ 264 Encode(ioutil.Discard, img, options) 269 img := image.NewYCbCr(image.Rect(0, 0, 640, 480), image.YCbCrSubsampleRatio420) 270 bo := img.Bounds() 274 cy := img.YOffset(x, y) 275 ci := img.COffset(x, y) 276 img.Y[cy] = uint8(rnd.Intn(256)) 277 img.Cb[ci] = uint8(rnd.Intn(256) [all...] |
/build/make/target/board/generic_x86/ |
BoardConfig.mk | 39 # Build partitioned system.img and vendor.img (if applicable) 74 # Builds a special vbmeta.img that disables AVB verification. 75 # Otherwise, AVB will prevent the device from booting the generic system.img. 77 # metadata into system.img.
|
/device/generic/goldfish/camera/qemu-pipeline3/ |
QemuSensor.cpp | 276 " %d x %d, format %x, stride %d, buf %p, img %p", 278 b.buffer, b.img); 281 captureRGB(b.img, b.width, b.height, b.stride, ×tamp); 284 captureRGBA(b.img, b.width, b.height, b.stride, ×tamp); 302 bAux.img = new uint8_t[b.width * b.height * 3]; 307 captureNV21(b.img, b.width, b.height, b.stride, ×tamp); 346 void QemuSensor::captureRGBA(uint8_t *img, uint32_t width, uint32_t height, 404 mCameraQemuClient.queryFrame(nullptr, img, 0, bufferSize, whiteBalance[0], 411 void QemuSensor::captureRGB(uint8_t *img, uint32_t width, uint32_t height, uint32_t stride, int64_t *timestamp) { 415 void QemuSensor::captureNV21(uint8_t *img, uint32_t width, uint32_t height, uint32_t stride, int64_t *timestamp) [all...] |
/external/libvpx/libvpx/test/ |
external_frame_buffer_test.cc | 118 void CheckXImageFrameBuffer(const vpx_image_t *img) { 119 if (img->fb_priv != NULL) { 121 reinterpret_cast<ExternalFrameBuffer *>(img->fb_priv); 123 ASSERT_TRUE(img->planes[0] >= ext_fb->data && 124 img->planes[0] < (ext_fb->data + ext_fb->size)); 234 virtual void DecompressedFrameHook(const vpx_image_t &img, 246 md5_res.Add(&img); 341 const vpx_image_t *img = NULL; local 344 while ((img = dec_iter.Next()) != NULL) { 345 fb_list_.CheckXImageFrameBuffer(img); [all...] |
/packages/apps/LegacyCamera/jni/feature_mos/src/mosaic/ |
ImageUtils.cpp | 397 void YUVinfo::mapYUVInfoToImage(YUVinfo *img, unsigned char *position) 400 for (i = 0; i < img->Y.height; i++, position += img->Y.width) 401 img->Y.ptr[i] = position; 402 for (i = 0; i < img->V.height; i++, position += img->V.width) 403 img->V.ptr[i] = position; 404 for (i = 0; i < img->U.height; i++, position += img->U.width) 405 img->U.ptr[i] = position [all...] |
/tools/tradefederation/contrib/src/com/android/media/tests/ |
AudioLoopbackImageAnalyzer.java | 82 BufferedImage img = null; local 85 img = ImageIO.read(f); 91 final int width = img.getWidth(); 92 final int height = img.getHeight(); 132 projectPixelsToXAxis(img, targetColors, horizontal, width, height); 142 projectPixelsToYAxis(img, targetColors, vertical, height, durationBounds); 387 BufferedImage img, 396 final int color = img.getRGB(x, y); 408 BufferedImage img, 423 final int color = img.getRGB(x, y) [all...] |
/external/google-breakpad/src/common/windows/ |
pdb_source_line_writer.cc | 111 explicit AutoImage(PLOADED_IMAGE img) : img_(img) {} 669 AutoImage img(ImageLoad((PSTR)code_file.c_str(), NULL)); 670 if (!img) { 675 &(reinterpret_cast<PIMAGE_NT_HEADERS64>(img->FileHeader))->OptionalHeader; 688 ImageRvaToVa(img->FileHeader, 689 img->MappedAddress, 691 &img->LastRvaSection)); 699 ImageRvaToVa(img->FileHeader, 700 img->MappedAddress [all...] |
/external/mesa3d/src/mesa/main/ |
texgetimage.c | 83 GLint img, row; local 91 for (img = 0; img < depth; img++) { 96 ctx->Driver.MapTextureImage(ctx, texImage, zoffset + img, 104 img, row, 0); 110 ctx->Driver.UnmapTextureImage(ctx, texImage, zoffset + img); 132 GLint img, row; local 138 for (img = 0; img < depth; img++) 181 GLint img, row; local 227 GLint img, row; local 434 GLuint img; local [all...] |
/external/opencv/cvaux/src/ |
cv3dtracker.cpp | 60 static void DrawEtalon(IplImage *img, CvPoint2D32f *corners, 136 IplImage *img; local 154 img = gray_img; 159 img = samples[c]; 171 bool found = cvFindChessBoardCornerGuesses(img, tmp_img, 0, 179 cvFindCornerSubPix(img, points, count, cvSize(5,5), cvSize(-1,-1), 327 static void DrawEtalon(IplImage *img, CvPoint2D32f *corners, 345 const CvScalar* colors = img->nChannels == 3 ? rgb_colors : gray_colors; 358 if (img->origin == IPL_ORIGIN_BL) 359 pt.y = img->height - 1 - pt.y [all...] |
cvhmmobs.cpp | 49 // Parameters: img - pointer to the original image ROI 117 icvImgToObs_DCT_8u32f_C1R( uchar * img, int imgStep, CvSize roi, 139 if( !img || !obs ) 170 for( y = 0; y < Ny; y++, img += delta.height * imgStep ) 184 is = img[x + m1 * imgStep]; 191 float t0 = img[x + j * imgStep]; 192 float t1 = img[x + (n1 - 1 - j) * imgStep]; 312 icvImgToObs_DCT_32f_C1R( float * img, int imgStep, CvSize roi, 334 if( !img || !obs ) 365 imgStep /= sizeof(img[0]) [all...] |
/packages/apps/Launcher3/src/com/android/launcher3/dragndrop/ |
AddItemActivity.java | 137 WidgetImageView img = mWidgetCell.getWidgetView(); local 141 if (img.getBitmap() == null) { 145 Rect bounds = img.getBitmapBounds(); 146 bounds.offset(img.getLeft() - (int) mLastTouchPos.x, img.getTop() - (int) mLastTouchPos.y); 150 img.getBitmap().getWidth(), img.getWidth());
|
/test/framework/harnesses/host_controller/build/ |
build_flasher.py | 141 print("checking to flash bootloader.img and radio.img") 203 if partition.endswith(".img"): 267 " ".join(flasher_args[1:] + [device_images["img"]])) 276 {"img": "path_to_repackaged_image"} 294 for img in device_images: 295 if os.path.dirname(device_images[img]) != tmp_dir_path: 296 os.rename(device_images[img], 297 os.path.join(tmp_dir_path, img)) 298 device_images[img] = os.path.join(tmp_dir_path, img [all...] |
/cts/tests/camera/src/android/hardware/camera2/cts/ |
ImageReaderTest.java | 764 Image img = null; local 792 Image img = null; local 901 Image img = mReader.acquireNextImage(); local 939 Image img = mReader.acquireLatestImage(); local 1018 Image img; local [all...] |
/external/valgrind/coregrind/m_debuginfo/ |
readmacho.c | 136 ML_(img_done)(sli->img); 169 // |sli.img|, so NULL out |mimg| after this point, for the sake of 175 if (ML_(img_size)(sli.img) < sizeof(struct fat_header)) { 181 ML_(img_get)(&fh_be, sli.img, fh_be_ioff, sizeof(fh_be)); 187 if (ML_(img_size)(sli.img) < sizeof(struct fat_header) 212 ML_(img_get)(&arch_be, sli.img, arch_be_ioff, sizeof(arch_be)); 219 if (ML_(img_size)(sli.img) < arch.offset + arch.size) { 239 vg_assert(ML_(img_size)(sli.img) >= sizeof(struct fat_header)); 246 if (sli.szB > ML_(img_size)(sli.img)) { 251 if (sli.ioff >= 0 && sli.ioff + sli.szB <= ML_(img_size)(sli.img)) { [all...] |
/external/libjpeg-turbo/java/ |
TJUnitTest.java | 245 private static void initImg(BufferedImage img, int pf, int flags) 247 WritableRaster wr = img.getRaster(); 248 int imgType = img.getType(); 254 (SinglePixelPackedSampleModel)img.getSampleModel(); 258 initIntBuf(buf, img.getWidth(), pitch, img.getHeight(), pf, flags); 260 ComponentSampleModel sm = (ComponentSampleModel)img.getSampleModel(); 264 initBuf(buf, img.getWidth(), pitch, img.getHeight(), pf, flags); 497 private static int checkImg(BufferedImage img, int pf, int subsamp 632 BufferedImage img = null; local 706 BufferedImage img = null; local [all...] |
/external/opencv/cv/src/ |
cvcontours.cpp | 156 schar *img; /* current image row */ member in struct:_CvContourScanner 200 uchar *img = 0; local 218 img = (uchar*)(mat->data.ptr); 233 scanner->img0 = (schar *) img; 234 scanner->img = (schar *) (img + step); 311 memset( img, 0, size.width ); 312 memset( img + step * (size.height - 1), 0, size.width ); 314 for( y = 1, img += step; y < size.height - 1; y++, img += step 838 schar *img; local [all...] |
/device/generic/goldfish/camera/fake-pipeline2/ |
JpegCompressor.cpp | 169 createThumbnail(static_cast<const unsigned char*>(mAuxBuffer.img), 181 nV21JpegCompressor.compressRawImage((void*)mAuxBuffer.img, 185 nV21JpegCompressor.getCompressedImage((void*)mJpegBuffer.img); 193 memcpy(mJpegBuffer.img + cb->width - sizeof(camera3_jpeg_blob_t), 231 delete[] mAuxBuffer.img;
|
/external/libdrm/tests/exynos/ |
exynos_fimg2d_test.c | 189 struct g2d_image img = {0}; local 197 img.bo[0] = dst->handle; 213 img.width = img_w; 214 img.height = img_h; 215 img.stride = img.width * 4; 216 img.color_mode = G2D_COLOR_FMT_ARGB8888 | G2D_ORDER_AXRGB; 217 img.color = 0xff000000 + (random() & 0xffffff); 219 ret = g2d_solid_fill(ctx, &img, x, y, w, h); 322 struct g2d_image img = {0}, tmp_img = {0} local [all...] |
/external/mesa3d/src/gallium/state_trackers/clover/core/ |
resource.cpp | 125 if (image *img = dynamic_cast<image *>(&obj)) { 126 info.format = translate_format(img->format()); 127 info.width0 = img->width(); 128 info.height0 = img->height(); 129 info.depth0 = img->depth();
|
/external/skia/gm/ |
patch.cpp | 85 static void dopatch(SkCanvas* canvas, const SkColor colors[], sk_sp<SkImage> img = nullptr) { 98 if (img) { 99 SkScalar w = img->width(); 100 SkScalar h = img->height(); 101 shader = img->makeShader(SkShader::kClamp_TileMode, SkShader::kClamp_TileMode);
|
/external/skqp/gm/ |
patch.cpp | 85 static void dopatch(SkCanvas* canvas, const SkColor colors[], sk_sp<SkImage> img = nullptr) { 98 if (img) { 99 SkScalar w = img->width(); 100 SkScalar h = img->height(); 101 shader = img->makeShader(SkShader::kClamp_TileMode, SkShader::kClamp_TileMode);
|
/external/tensorflow/tensorflow/contrib/lite/java/src/test/java/org/tensorflow/lite/ |
InterpreterTest.java | 125 float[][][][] img = new float[1][224][224][3]; local 128 img[0][i][j][0] = 0.5f; 129 img[0][i][j][1] = 0.5f; 130 img[0][i][j][2] = 0.5f; 138 interpreter.run(img, labels);
|