/frameworks/av/media/libstagefright/codecs/amrwbenc/src/ |
homing.c | 30 Word16 encoder_homing_frame_test(Word16 input_frame[]) 38 j = (Word16) (input_frame[i] ^ EHF_MASK);
|
/external/webrtc/talk/media/base/ |
videoframefactory.cc | 36 const CapturedFrame* input_frame, 42 input_frame, cropped_input_width, cropped_input_height)); 54 (input_frame->rotation == webrtc::kVideoRotation_90 || 55 input_frame->rotation == webrtc::kVideoRotation_270)) {
|
videoframefactory.h | 50 // The returned frame will be a center crop of |input_frame| with 52 virtual VideoFrame* CreateAliasedFrame(const CapturedFrame* input_frame, 56 // The returned frame will be a center crop of |input_frame| with size 59 virtual VideoFrame* CreateAliasedFrame(const CapturedFrame* input_frame,
|
/frameworks/av/media/libstagefright/codecs/amrwbenc/inc/ |
main.h | 42 Word16 encoder_homing_frame_test(Word16 input_frame[]);
|
/frameworks/av/media/libstagefright/codecs/amrwb/src/ |
pvamrwbdecoder.h | 133 int16 pvDecoder_AmrWb_homing_frame_test(int16 input_frame[], int16 mode); 135 int16 pvDecoder_AmrWb_homing_frame_test_first(int16 input_frame[], int16 mode);
|
homing_amr_wb_dec.cpp | 51 int16 input_frame[], 16-bit input frame 63 int16 dhf_test(int16 input_frame[], int16 mode, int16 nparms) 64 int16 decoder_homing_frame_test(int16 input_frame[], int16 mode) 65 int16 decoder_homing_frame_test_first(int16 input_frame[], int16 mode) 127 int16 dhf_test(int16 input_frame[], int32 mode, int16 nparms); 248 int16 dhf_test(int16 input_frame[], int32 mode, int16 nparms) 271 prms = input_frame; 355 int16 pvDecoder_AmrWb_homing_frame_test(int16 input_frame[], int16 mode) 358 return dhf_test(input_frame, mode, AMR_WB_COMPRESSED[mode]); 366 int16 pvDecoder_AmrWb_homing_frame_test_first(int16 input_frame[], int16 mode [all...] |
/external/webrtc/webrtc/common_video/libyuv/ |
webrtc_libyuv.cc | 126 int ExtractBuffer(const VideoFrame& input_frame, size_t size, uint8_t* buffer) { 128 if (input_frame.IsZeroSize()) 131 CalcBufferSize(kI420, input_frame.width(), input_frame.height()); 140 int width = (plane ? (input_frame.width() + 1) / 2 : 141 input_frame.width()); 142 int height = (plane ? (input_frame.height() + 1) / 2 : 143 input_frame.height()); 144 const uint8_t* plane_ptr = input_frame.buffer( 149 plane_ptr += input_frame.stride(static_cast<PlaneType>(plane)) [all...] |
scaler_unittest.cc | 354 VideoFrame input_frame; local 370 input_frame.CreateFrame(frame_buffer.get(), 378 EXPECT_EQ(0, test_scaler_.Scale(input_frame, &output_frame));
|
/external/webrtc/talk/app/webrtc/ |
androidvideocapturer.cc | 99 const cricket::CapturedFrame* input_frame, 111 apply_rotation_ ? input_frame->rotation : 114 scaled_buffer, input_frame->time_stamp, 115 apply_rotation_ ? webrtc::kVideoRotation_0 : input_frame->rotation); 117 return VideoFrameFactory::CreateAliasedFrame(input_frame,
|
/external/webrtc/webrtc/common_video/libyuv/include/ |
webrtc_libyuv.h | 95 int ExtractBuffer(const VideoFrame& input_frame, size_t size, uint8_t* buffer);
|
/external/webrtc/webrtc/modules/video_coding/codecs/vp8/ |
vp8_sequence_coder.cc | 148 webrtc::VideoFrame input_frame; local 163 input_frame.CreateEmptyFrame(width, height, width, half_width, half_width); 170 height, 0, webrtc::kVideoRotation_0, &input_frame); 171 encoder->Encode(input_frame, NULL, NULL);
|
/external/webrtc/webrtc/modules/audio_processing/ |
audio_processing_performance_unittest.cc | 58 input_frame.resize(2); 59 input_frame[0] = &input_framechannels[0]; 60 input_frame[1] = &input_framechannels[max_frame_size]; 71 std::vector<float*> input_frame; member in struct:webrtc::__anon27865::AudioFrameData 335 &frame_data_.input_frame[0], frame_data_.input_stream_config, 363 &frame_data_.input_frame[0], frame_data_.input_stream_config, 412 rand_gen_, &frame_data_.input_frame[0]);
|
audio_processing_impl_locking_unittest.cc | 95 input_frame.resize(2); 96 input_frame[0] = &input_framechannels[0]; 97 input_frame[1] = &input_framechannels[max_frame_size]; 112 std::vector<float*> input_frame; member in struct:webrtc::__anon27864::AudioFrameData 685 PopulateAudioFrame(&frame_data_.input_frame[0], kCaptureInputFloatLevel, 723 &frame_data_.input_frame[0], frame_data_.input_samples_per_channel, 730 &frame_data_.input_frame[0], frame_data_.input_stream_config, [all...] |
/external/webrtc/talk/app/webrtc/java/jni/ |
androidmediaencoder_jni.cc | 581 VideoFrame input_frame = frame; local 596 input_frame.set_video_frame_buffer(scaled_buffer); 598 input_frame = quality_scaler_.GetScaledFrame(frame); 603 if (!MaybeReconfigureEncoderOnCodecThread(input_frame)) { 625 if (!input_frame.native_handle()) { 641 encode_status = EncodeByteBufferOnCodecThread(jni, key_frame, input_frame, 644 encode_status = EncodeTextureOnCodecThread(jni, key_frame, input_frame); 648 ALOGE << "Failed encode frame with timestamp: " << input_frame.timestamp(); 658 timestamps_.push_back(input_frame.timestamp()); 659 render_times_ms_.push_back(input_frame.render_time_ms()) [all...] |
/hardware/intel/img/psb_video/src/ |
vsp_vp8.c | 418 pic->input_frame.surface_id = surface_id; 419 pic->input_frame.irq = 1; 420 pic->input_frame.height = ctx->vp8_seq_param.frame_height; 421 pic->input_frame.width = ctx->vp8_seq_param.frame_width; 424 //pic->input_frame.stride = (ctx->frame_width + 31) & (~31); 425 pic->input_frame.stride = ctx->obj_context->current_render_target->psb_surface->stride; 426 pic->input_frame.format = 0; /* TODO: Specify NV12 = 0 */ 468 vsp_cmdbuf_reloc_pic_param(&(pic->input_frame.base), 471 vsp_cmdbuf_reloc_pic_param(&(pic->input_frame.base_uv), 472 pic->input_frame.stride * ctx->obj_context->current_render_target->height [all...] |
vsp_fw.h | 613 struct VssProcPictureVP8 input_frame; member in struct:VssVp8encPictureParameterBuffer
|
/frameworks/base/media/mca/filterfw/jni/ |
jni_shader_program.cpp | 109 const GLFrame* input_frame = ConvertFromJava<GLFrame>(env, input); local 110 if (!input || !input_frame) { 114 input_frames.push_back(input_frame);
|