HomeSort by relevance Sort by last modified time
    Searched refs:input_frame (Results 1 - 18 of 18) sorted by null

  /frameworks/av/media/libstagefright/codecs/amrwbenc/src/
homing.c 30 Word16 encoder_homing_frame_test(Word16 input_frame[])
38 j = (Word16) (input_frame[i] ^ EHF_MASK);
  /external/webrtc/talk/media/base/
videoframefactory.cc 36 const CapturedFrame* input_frame,
42 input_frame, cropped_input_width, cropped_input_height));
54 (input_frame->rotation == webrtc::kVideoRotation_90 ||
55 input_frame->rotation == webrtc::kVideoRotation_270)) {
videoframefactory.h 50 // The returned frame will be a center crop of |input_frame| with
52 virtual VideoFrame* CreateAliasedFrame(const CapturedFrame* input_frame,
56 // The returned frame will be a center crop of |input_frame| with size
59 virtual VideoFrame* CreateAliasedFrame(const CapturedFrame* input_frame,
  /frameworks/av/media/libstagefright/codecs/amrwbenc/inc/
main.h 42 Word16 encoder_homing_frame_test(Word16 input_frame[]);
  /frameworks/av/media/libstagefright/codecs/amrwb/src/
pvamrwbdecoder.h 133 int16 pvDecoder_AmrWb_homing_frame_test(int16 input_frame[], int16 mode);
135 int16 pvDecoder_AmrWb_homing_frame_test_first(int16 input_frame[], int16 mode);
homing_amr_wb_dec.cpp 51 int16 input_frame[], 16-bit input frame
63 int16 dhf_test(int16 input_frame[], int16 mode, int16 nparms)
64 int16 decoder_homing_frame_test(int16 input_frame[], int16 mode)
65 int16 decoder_homing_frame_test_first(int16 input_frame[], int16 mode)
127 int16 dhf_test(int16 input_frame[], int32 mode, int16 nparms);
248 int16 dhf_test(int16 input_frame[], int32 mode, int16 nparms)
271 prms = input_frame;
355 int16 pvDecoder_AmrWb_homing_frame_test(int16 input_frame[], int16 mode)
358 return dhf_test(input_frame, mode, AMR_WB_COMPRESSED[mode]);
366 int16 pvDecoder_AmrWb_homing_frame_test_first(int16 input_frame[], int16 mode
    [all...]
  /external/webrtc/webrtc/common_video/libyuv/
webrtc_libyuv.cc 126 int ExtractBuffer(const VideoFrame& input_frame, size_t size, uint8_t* buffer) {
128 if (input_frame.IsZeroSize())
131 CalcBufferSize(kI420, input_frame.width(), input_frame.height());
140 int width = (plane ? (input_frame.width() + 1) / 2 :
141 input_frame.width());
142 int height = (plane ? (input_frame.height() + 1) / 2 :
143 input_frame.height());
144 const uint8_t* plane_ptr = input_frame.buffer(
149 plane_ptr += input_frame.stride(static_cast<PlaneType>(plane))
    [all...]
scaler_unittest.cc 354 VideoFrame input_frame; local
370 input_frame.CreateFrame(frame_buffer.get(),
378 EXPECT_EQ(0, test_scaler_.Scale(input_frame, &output_frame));
  /external/webrtc/talk/app/webrtc/
androidvideocapturer.cc 99 const cricket::CapturedFrame* input_frame,
111 apply_rotation_ ? input_frame->rotation :
114 scaled_buffer, input_frame->time_stamp,
115 apply_rotation_ ? webrtc::kVideoRotation_0 : input_frame->rotation);
117 return VideoFrameFactory::CreateAliasedFrame(input_frame,
  /external/tensorflow/tensorflow/core/common_runtime/
executor.cc 1118 FrameState* input_frame = nullptr; member in struct:tensorflow::__anon39475::ExecutorState::TaggedNode
1531 FrameState* input_frame = tagged_node.input_frame; local
1935 FrameState* input_frame = tagged_node.input_frame; local
    [all...]
  /external/webrtc/webrtc/common_video/libyuv/include/
webrtc_libyuv.h 95 int ExtractBuffer(const VideoFrame& input_frame, size_t size, uint8_t* buffer);
  /external/webrtc/webrtc/modules/video_coding/codecs/vp8/
vp8_sequence_coder.cc 148 webrtc::VideoFrame input_frame; local
163 input_frame.CreateEmptyFrame(width, height, width, half_width, half_width);
170 height, 0, webrtc::kVideoRotation_0, &input_frame);
171 encoder->Encode(input_frame, NULL, NULL);
  /external/webrtc/webrtc/modules/audio_processing/
audio_processing_performance_unittest.cc 58 input_frame.resize(2);
59 input_frame[0] = &input_framechannels[0];
60 input_frame[1] = &input_framechannels[max_frame_size];
71 std::vector<float*> input_frame; member in struct:webrtc::__anon44171::AudioFrameData
335 &frame_data_.input_frame[0], frame_data_.input_stream_config,
363 &frame_data_.input_frame[0], frame_data_.input_stream_config,
412 rand_gen_, &frame_data_.input_frame[0]);
audio_processing_impl_locking_unittest.cc 95 input_frame.resize(2);
96 input_frame[0] = &input_framechannels[0];
97 input_frame[1] = &input_framechannels[max_frame_size];
112 std::vector<float*> input_frame; member in struct:webrtc::__anon44170::AudioFrameData
685 PopulateAudioFrame(&frame_data_.input_frame[0], kCaptureInputFloatLevel,
723 &frame_data_.input_frame[0], frame_data_.input_samples_per_channel,
730 &frame_data_.input_frame[0], frame_data_.input_stream_config,
    [all...]
  /external/webrtc/talk/app/webrtc/java/jni/
androidmediaencoder_jni.cc 581 VideoFrame input_frame = frame; local
596 input_frame.set_video_frame_buffer(scaled_buffer);
598 input_frame = quality_scaler_.GetScaledFrame(frame);
603 if (!MaybeReconfigureEncoderOnCodecThread(input_frame)) {
625 if (!input_frame.native_handle()) {
641 encode_status = EncodeByteBufferOnCodecThread(jni, key_frame, input_frame,
644 encode_status = EncodeTextureOnCodecThread(jni, key_frame, input_frame);
648 ALOGE << "Failed encode frame with timestamp: " << input_frame.timestamp();
658 timestamps_.push_back(input_frame.timestamp());
659 render_times_ms_.push_back(input_frame.render_time_ms())
    [all...]
  /hardware/intel/img/psb_video/src/
vsp_vp8.c 422 pic->input_frame.surface_id = surface_id;
423 pic->input_frame.irq = 1;
424 pic->input_frame.height = ctx->vp8_seq_param.frame_height;
425 pic->input_frame.width = ctx->vp8_seq_param.frame_width;
428 //pic->input_frame.stride = (ctx->frame_width + 31) & (~31);
429 pic->input_frame.stride = ctx->obj_context->current_render_target->psb_surface->stride;
430 pic->input_frame.format = 0; /* TODO: Specify NV12 = 0 */
472 vsp_cmdbuf_reloc_pic_param(&(pic->input_frame.base),
475 vsp_cmdbuf_reloc_pic_param(&(pic->input_frame.base_uv),
476 pic->input_frame.stride * ctx->obj_context->current_render_target->height
    [all...]
vsp_fw.h 613 struct VssProcPictureVP8 input_frame; member in struct:VssVp8encPictureParameterBuffer
  /frameworks/base/media/mca/filterfw/jni/
jni_shader_program.cpp 102 const GLFrame* input_frame = ConvertFromJava<GLFrame>(env, input); local
103 if (!input || !input_frame) {
107 input_frames.push_back(input_frame);

Completed in 269 milliseconds