/external/chromium-trace/catapult/telemetry/telemetry/internal/image_processing/ |
video_unittest.py | 12 from telemetry.internal.image_processing import video namespace 41 video_obj = video.Video(vid)
|
/external/opencv3/samples/cpp/ |
phase_corr.cpp | 10 VideoCapture video(0); 16 video >> frame;
|
/external/opencv3/samples/python2/ |
color_histogram.py | 11 import video namespace 34 cam = video.create_capture(fn, fallback='synth:bg=../data/baboon.jpg:class=chess:noise=0.05')
|
lappyr.py | 7 Usage : python lappyr.py [<video source>] 17 import video namespace 47 cap = video.create_capture(fn)
|
/hardware/qcom/media/msm8974/mm-video-v4l2/DivxDrmDecrypt/ |
Android.mk | 8 LOCAL_COPY_HEADERS_TO := mm-video-v4l2/DivxDrmDecrypt
|
/hardware/qcom/media/msmcobalt/ |
Android.mk | 10 include $(QCOM_MEDIA_ROOT)/mm-video-v4l2/Android.mk
|
/external/mesa3d/docs/ |
MESA_swap_control.spec | 35 of color buffer swaps, measured in video frame periods. 79 glXSwapIntervalMESA specifies the minimum number of video frame periods 81 will be swapped at most every other video frame.) A return value 86 A video frame period is the time required by the monitor to display a 87 full frame of video data. In the case of an interlaced monitor, 89 fields of a frame of video data. 92 nized to a video frame. The <interval> value is silently clamped to
|
/external/webrtc/webrtc/tools/rtcbot/bot/browser/ |
bot.js | 97 var video = document.createElement('video'); 98 video.autoplay = autoplay; 99 video.muted = muted; 100 document.body.appendChild(video); 101 video.src = URL.createObjectURL(stream); 102 console.log("Stream " + stream.id + " attached to video element");
|
/frameworks/av/media/libstagefright/codecs/avc/enc/src/ |
residual.cpp | 23 AVCCommonObj *video = encvid->common; local 26 int x_position = (video->mb_x << 4); 27 int y_position = (video->mb_y << 4); 44 pDst = video->currPic->Sl + offset1; 45 offset = video->PicWidthInSamplesL - 16; 77 pDst = video->currPic->Scb + ((offset1 + x_position) >> 2); 111 pDst = video->currPic->Scr + ((offset1 + x_position) >> 2); 148 AVCCommonObj *video = encvid->common; local 241 nC = predict_nnz(video, bindx & 3, bindx >> 2); 245 nC = predict_nnz_chroma(video, bindx & 3, bindx >> 2) [all...] |
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
post_filter.cpp | 32 VideoDecData *video, 42 int nTotalMB = video->nTotalMB; 46 uint8 *decodedFrame = video->videoDecControls->outputFrame; 50 width = video->width; 51 height = video->height; 62 if (PVGetDecBitrate(video->videoDecControls) > (100*video->frameRate*(size >> 12))) // MC_sofDeblock 68 QP_store = video->QPMB; 71 pp_mod = video->pstprcTypCur; 163 pp_mod = video->pstprcTypCur [all...] |
/hardware/qcom/media/msm8974/mm-video-legacy/vidc/venc/src/ |
omx_video_encoder.cpp | 110 if(!strncmp((char *)m_nkind,"OMX.qcom.video.encoder.mpeg4",\ 117 else if(!strncmp((char *)m_nkind, "OMX.qcom.video.encoder.h263",\ 123 else if(!strncmp((char *)m_nkind, "OMX.qcom.video.encoder.avc",\ 129 else if(!strncmp((char *)m_nkind, "OMX.qcom.video.encoder.avc.secure",\ 244 // Initialize the video parameters for input port 251 m_sInPortDef.format.video.cMIMEType = "YUV420"; 252 m_sInPortDef.format.video.nFrameWidth = OMX_CORE_QCIF_WIDTH; 253 m_sInPortDef.format.video.nFrameHeight = OMX_CORE_QCIF_HEIGHT; 254 m_sInPortDef.format.video.nStride = OMX_CORE_QCIF_WIDTH; 255 m_sInPortDef.format.video.nSliceHeight = OMX_CORE_QCIF_HEIGHT [all...] |
/developers/build/prebuilts/androidtv/leanback/app/src/main/java/com/example/android/leanback/ |
VideoProvider.java | 119 "category: " + i + " Name:" + category_name + " video length: " 124 JSONObject video = videos.getJSONObject(j); local 125 String description = video.getString(TAG_DESCRIPTION); 126 JSONArray videoUrls = video.getJSONArray(TAG_SOURCES); 130 title = video.getString(TAG_TITLE); 133 video.getString(TAG_BACKGROUND)); 135 video.getString(TAG_CARD_THUMB)); 136 studio = video.getString(TAG_STUDIO);
|
/developers/build/prebuilts/androidtv/leanback/app/src/main/java/com/example/android/tvleanback/data/ |
VideoProvider.java | 89 "category: " + i + " Name:" + category_name + " video length: " 94 JSONObject video = videos.getJSONObject(j); local 95 String description = video.getString(TAG_DESCRIPTION); 96 JSONArray videoUrls = video.getJSONArray(TAG_SOURCES); 100 title = video.getString(TAG_TITLE); 103 video.getString(TAG_BACKGROUND)); 105 video.getString(TAG_CARD_THUMB)); 106 studio = video.getString(TAG_STUDIO);
|
/external/libvpx/libvpx/test/ |
vp9_arf_freq_test.cc | 154 virtual void PreEncodeFrameHook(::libvpx_test::VideoSource *video, 156 if (video->frame() == 0) { 203 libvpx_test::VideoSource *video; local 205 video = new libvpx_test::Y4mVideoSource(test_video_param_.filename, 208 video = new libvpx_test::YUVVideoSource(test_video_param_.filename, 217 ASSERT_NO_FATAL_FAILURE(RunLoop(video)); 224 delete(video);
|
vp9_thread_test.cc | 163 libvpx_test::WebMVideoSource video(filename); 164 video.Init(); 171 for (video.Begin(); video.cxdata(); video.Next()) { 173 decoder.DecodeFrame(video.cxdata(), video.frame_size());
|
encode_perf_test.cc | 92 virtual void PreEncodeFrameHook(::libvpx_test::VideoSource *video, 94 if (video->frame() == 0) { 159 libvpx_test::I420VideoSource video( 170 ASSERT_NO_FATAL_FAILURE(RunLoop(&video));
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/test/ |
resize_test.cc | 146 ResizingVideoSource video; local 147 ASSERT_NO_FATAL_FAILURE(RunLoop(&video)); 196 virtual void PreEncodeFrameHook(libvpx_test::VideoSource *video, 198 if (video->frame() == kStepDownFrame) { 202 if (video->frame() == kStepUpFrame) { 236 ::libvpx_test::I420VideoSource video("hantro_collage_w352h288.yuv", 352, 288, 248 ASSERT_NO_FATAL_FAILURE(RunLoop(&video));
|
/hardware/qcom/media/msm8996/mm-video-v4l2/vidc/venc/src/ |
omx_swvenc_hevc.cpp | 111 if (!strncmp((char *)m_nkind,"OMX.qti.video.encoder.hevc",\ 230 // Initialize the video parameters for input port 237 m_sInPortDef.format.video.cMIMEType = (char *)"YUV420"; 238 m_sInPortDef.format.video.nFrameWidth = OMX_CORE_QCIF_WIDTH; 239 m_sInPortDef.format.video.nFrameHeight = OMX_CORE_QCIF_HEIGHT; 240 m_sInPortDef.format.video.nStride = OMX_CORE_QCIF_WIDTH; 241 m_sInPortDef.format.video.nSliceHeight = OMX_CORE_QCIF_HEIGHT; 242 m_sInPortDef.format.video.nBitrate = 64000; 243 m_sInPortDef.format.video.xFramerate = 15 << 16; 244 m_sInPortDef.format.video.eColorFormat = (OMX_COLOR_FORMATTYPE [all...] |
/hardware/qcom/media/msmcobalt/mm-video-v4l2/vidc/venc/src/ |
omx_swvenc_hevc.cpp | 111 if (!strncmp((char *)m_nkind,"OMX.qti.video.encoder.hevc",\ 230 // Initialize the video parameters for input port 237 m_sInPortDef.format.video.cMIMEType = (char *)"YUV420"; 238 m_sInPortDef.format.video.nFrameWidth = OMX_CORE_QCIF_WIDTH; 239 m_sInPortDef.format.video.nFrameHeight = OMX_CORE_QCIF_HEIGHT; 240 m_sInPortDef.format.video.nStride = OMX_CORE_QCIF_WIDTH; 241 m_sInPortDef.format.video.nSliceHeight = OMX_CORE_QCIF_HEIGHT; 242 m_sInPortDef.format.video.nBitrate = 64000; 243 m_sInPortDef.format.video.xFramerate = 15 << 16; 244 m_sInPortDef.format.video.eColorFormat = (OMX_COLOR_FORMATTYPE [all...] |
/cts/apps/CtsVerifier/src/com/android/cts/verifier/projection/video/ |
ProjectionVideoActivity.java | 17 package com.android.cts.verifier.projection.video;
|
/external/autotest/client/cros/video/ |
import_screenshot_capturer.py | 8 from autotest_lib.client.cros.video import method_logger
|
sequence_generator.py | 9 from autotest_lib.client.cros.video import method_logger 21 video. We would typically specify an interval of a second and then store
|
video_screenshot_collector.py | 5 from autotest_lib.client.cros.video import method_logger 10 Captures and collects screenshots of a video at specified time points. 25 Get a screenshot of video at a particular time. 46 Loads video and waits for player to be ready.
|
/external/opencv3/modules/java/src/ |
video+BackgroundSubtractor.java | 5 package org.opencv.video;
|
/hardware/qcom/media/ |
Android.mk | 18 include $(QCOM_MEDIA_ROOT)/mm-video-legacy/Android.mk 22 include $(QCOM_MEDIA_ROOT)/mm-video-v4l2/Android.mk
|