1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style license that can be 3 // found in the LICENSE file. 4 5 #include "content/common/gpu/media/dxva_video_decode_accelerator.h" 6 7 #if !defined(OS_WIN) 8 #error This file should only be built on Windows. 9 #endif // !defined(OS_WIN) 10 11 #include <ks.h> 12 #include <codecapi.h> 13 #include <mfapi.h> 14 #include <mferror.h> 15 #include <wmcodecdsp.h> 16 17 #include "base/bind.h" 18 #include "base/callback.h" 19 #include "base/command_line.h" 20 #include "base/debug/trace_event.h" 21 #include "base/file_version_info.h" 22 #include "base/logging.h" 23 #include "base/memory/scoped_ptr.h" 24 #include "base/memory/shared_memory.h" 25 #include "base/message_loop/message_loop.h" 26 #include "base/win/windows_version.h" 27 #include "media/video/video_decode_accelerator.h" 28 #include "ui/gl/gl_bindings.h" 29 #include "ui/gl/gl_surface_egl.h" 30 #include "ui/gl/gl_switches.h" 31 32 namespace content { 33 34 // We only request 5 picture buffers from the client which are used to hold the 35 // decoded samples. These buffers are then reused when the client tells us that 36 // it is done with the buffer. 37 static const int kNumPictureBuffers = 5; 38 39 #define RETURN_ON_FAILURE(result, log, ret) \ 40 do { \ 41 if (!(result)) { \ 42 DLOG(ERROR) << log; \ 43 return ret; \ 44 } \ 45 } while (0) 46 47 #define RETURN_ON_HR_FAILURE(result, log, ret) \ 48 RETURN_ON_FAILURE(SUCCEEDED(result), \ 49 log << ", HRESULT: 0x" << std::hex << result, \ 50 ret); 51 52 #define RETURN_AND_NOTIFY_ON_FAILURE(result, log, error_code, ret) \ 53 do { \ 54 if (!(result)) { \ 55 DVLOG(1) << log; \ 56 StopOnError(error_code); \ 57 return ret; \ 58 } \ 59 } while (0) 60 61 #define RETURN_AND_NOTIFY_ON_HR_FAILURE(result, log, error_code, ret) \ 62 RETURN_AND_NOTIFY_ON_FAILURE(SUCCEEDED(result), \ 63 log << ", HRESULT: 0x" << std::hex << result, \ 64 error_code, ret); 65 66 // Maximum number of iterations we allow before aborting the attempt to flush 67 // the batched queries to the driver and allow torn/corrupt frames to be 68 // rendered. 69 enum { kMaxIterationsForD3DFlush = 10 }; 70 71 static IMFSample* CreateEmptySample() { 72 base::win::ScopedComPtr<IMFSample> sample; 73 HRESULT hr = MFCreateSample(sample.Receive()); 74 RETURN_ON_HR_FAILURE(hr, "MFCreateSample failed", NULL); 75 return sample.Detach(); 76 } 77 78 // Creates a Media Foundation sample with one buffer of length |buffer_length| 79 // on a |align|-byte boundary. Alignment must be a perfect power of 2 or 0. 80 static IMFSample* CreateEmptySampleWithBuffer(int buffer_length, int align) { 81 CHECK_GT(buffer_length, 0); 82 83 base::win::ScopedComPtr<IMFSample> sample; 84 sample.Attach(CreateEmptySample()); 85 86 base::win::ScopedComPtr<IMFMediaBuffer> buffer; 87 HRESULT hr = E_FAIL; 88 if (align == 0) { 89 // Note that MFCreateMemoryBuffer is same as MFCreateAlignedMemoryBuffer 90 // with the align argument being 0. 91 hr = MFCreateMemoryBuffer(buffer_length, buffer.Receive()); 92 } else { 93 hr = MFCreateAlignedMemoryBuffer(buffer_length, 94 align - 1, 95 buffer.Receive()); 96 } 97 RETURN_ON_HR_FAILURE(hr, "Failed to create memory buffer for sample", NULL); 98 99 hr = sample->AddBuffer(buffer); 100 RETURN_ON_HR_FAILURE(hr, "Failed to add buffer to sample", NULL); 101 102 return sample.Detach(); 103 } 104 105 // Creates a Media Foundation sample with one buffer containing a copy of the 106 // given Annex B stream data. 107 // If duration and sample time are not known, provide 0. 108 // |min_size| specifies the minimum size of the buffer (might be required by 109 // the decoder for input). If no alignment is required, provide 0. 110 static IMFSample* CreateInputSample(const uint8* stream, int size, 111 int min_size, int alignment) { 112 CHECK(stream); 113 CHECK_GT(size, 0); 114 base::win::ScopedComPtr<IMFSample> sample; 115 sample.Attach(CreateEmptySampleWithBuffer(std::max(min_size, size), 116 alignment)); 117 RETURN_ON_FAILURE(sample, "Failed to create empty sample", NULL); 118 119 base::win::ScopedComPtr<IMFMediaBuffer> buffer; 120 HRESULT hr = sample->GetBufferByIndex(0, buffer.Receive()); 121 RETURN_ON_HR_FAILURE(hr, "Failed to get buffer from sample", NULL); 122 123 DWORD max_length = 0; 124 DWORD current_length = 0; 125 uint8* destination = NULL; 126 hr = buffer->Lock(&destination, &max_length, ¤t_length); 127 RETURN_ON_HR_FAILURE(hr, "Failed to lock buffer", NULL); 128 129 CHECK_EQ(current_length, 0u); 130 CHECK_GE(static_cast<int>(max_length), size); 131 memcpy(destination, stream, size); 132 133 hr = buffer->Unlock(); 134 RETURN_ON_HR_FAILURE(hr, "Failed to unlock buffer", NULL); 135 136 hr = buffer->SetCurrentLength(size); 137 RETURN_ON_HR_FAILURE(hr, "Failed to set buffer length", NULL); 138 139 return sample.Detach(); 140 } 141 142 static IMFSample* CreateSampleFromInputBuffer( 143 const media::BitstreamBuffer& bitstream_buffer, 144 DWORD stream_size, 145 DWORD alignment) { 146 base::SharedMemory shm(bitstream_buffer.handle(), true); 147 RETURN_ON_FAILURE(shm.Map(bitstream_buffer.size()), 148 "Failed in base::SharedMemory::Map", NULL); 149 150 return CreateInputSample(reinterpret_cast<const uint8*>(shm.memory()), 151 bitstream_buffer.size(), 152 stream_size, 153 alignment); 154 } 155 156 // Maintains information about a DXVA picture buffer, i.e. whether it is 157 // available for rendering, the texture information, etc. 158 struct DXVAVideoDecodeAccelerator::DXVAPictureBuffer { 159 public: 160 static linked_ptr<DXVAPictureBuffer> Create( 161 const DXVAVideoDecodeAccelerator& decoder, 162 const media::PictureBuffer& buffer, 163 EGLConfig egl_config); 164 ~DXVAPictureBuffer(); 165 166 void ReusePictureBuffer(); 167 // Copies the output sample data to the picture buffer provided by the 168 // client. 169 // The dest_surface parameter contains the decoded bits. 170 bool CopyOutputSampleDataToPictureBuffer( 171 const DXVAVideoDecodeAccelerator& decoder, 172 IDirect3DSurface9* dest_surface); 173 174 bool available() const { 175 return available_; 176 } 177 178 void set_available(bool available) { 179 available_ = available; 180 } 181 182 int id() const { 183 return picture_buffer_.id(); 184 } 185 186 gfx::Size size() const { 187 return picture_buffer_.size(); 188 } 189 190 private: 191 explicit DXVAPictureBuffer(const media::PictureBuffer& buffer); 192 193 bool available_; 194 media::PictureBuffer picture_buffer_; 195 EGLSurface decoding_surface_; 196 base::win::ScopedComPtr<IDirect3DTexture9> decoding_texture_; 197 // Set to true if RGB is supported by the texture. 198 // Defaults to true. 199 bool use_rgb_; 200 201 DISALLOW_COPY_AND_ASSIGN(DXVAPictureBuffer); 202 }; 203 204 // static 205 linked_ptr<DXVAVideoDecodeAccelerator::DXVAPictureBuffer> 206 DXVAVideoDecodeAccelerator::DXVAPictureBuffer::Create( 207 const DXVAVideoDecodeAccelerator& decoder, 208 const media::PictureBuffer& buffer, 209 EGLConfig egl_config) { 210 linked_ptr<DXVAPictureBuffer> picture_buffer(new DXVAPictureBuffer(buffer)); 211 212 EGLDisplay egl_display = gfx::GLSurfaceEGL::GetHardwareDisplay(); 213 214 EGLint use_rgb = 1; 215 eglGetConfigAttrib(egl_display, egl_config, EGL_BIND_TO_TEXTURE_RGB, 216 &use_rgb); 217 218 EGLint attrib_list[] = { 219 EGL_WIDTH, buffer.size().width(), 220 EGL_HEIGHT, buffer.size().height(), 221 EGL_TEXTURE_FORMAT, use_rgb ? EGL_TEXTURE_RGB : EGL_TEXTURE_RGBA, 222 EGL_TEXTURE_TARGET, EGL_TEXTURE_2D, 223 EGL_NONE 224 }; 225 226 picture_buffer->decoding_surface_ = eglCreatePbufferSurface( 227 egl_display, 228 egl_config, 229 attrib_list); 230 RETURN_ON_FAILURE(picture_buffer->decoding_surface_, 231 "Failed to create surface", 232 linked_ptr<DXVAPictureBuffer>(NULL)); 233 234 HANDLE share_handle = NULL; 235 EGLBoolean ret = eglQuerySurfacePointerANGLE( 236 egl_display, 237 picture_buffer->decoding_surface_, 238 EGL_D3D_TEXTURE_2D_SHARE_HANDLE_ANGLE, 239 &share_handle); 240 241 RETURN_ON_FAILURE(share_handle && ret == EGL_TRUE, 242 "Failed to query ANGLE surface pointer", 243 linked_ptr<DXVAPictureBuffer>(NULL)); 244 245 HRESULT hr = decoder.device_->CreateTexture( 246 buffer.size().width(), 247 buffer.size().height(), 248 1, 249 D3DUSAGE_RENDERTARGET, 250 use_rgb ? D3DFMT_X8R8G8B8 : D3DFMT_A8R8G8B8, 251 D3DPOOL_DEFAULT, 252 picture_buffer->decoding_texture_.Receive(), 253 &share_handle); 254 255 RETURN_ON_HR_FAILURE(hr, "Failed to create texture", 256 linked_ptr<DXVAPictureBuffer>(NULL)); 257 picture_buffer->use_rgb_ = !!use_rgb; 258 return picture_buffer; 259 } 260 261 DXVAVideoDecodeAccelerator::DXVAPictureBuffer::DXVAPictureBuffer( 262 const media::PictureBuffer& buffer) 263 : available_(true), 264 picture_buffer_(buffer), 265 decoding_surface_(NULL), 266 use_rgb_(true) { 267 } 268 269 DXVAVideoDecodeAccelerator::DXVAPictureBuffer::~DXVAPictureBuffer() { 270 if (decoding_surface_) { 271 EGLDisplay egl_display = gfx::GLSurfaceEGL::GetHardwareDisplay(); 272 273 eglReleaseTexImage( 274 egl_display, 275 decoding_surface_, 276 EGL_BACK_BUFFER); 277 278 eglDestroySurface( 279 egl_display, 280 decoding_surface_); 281 decoding_surface_ = NULL; 282 } 283 } 284 285 void DXVAVideoDecodeAccelerator::DXVAPictureBuffer::ReusePictureBuffer() { 286 DCHECK(decoding_surface_); 287 EGLDisplay egl_display = gfx::GLSurfaceEGL::GetHardwareDisplay(); 288 eglReleaseTexImage( 289 egl_display, 290 decoding_surface_, 291 EGL_BACK_BUFFER); 292 set_available(true); 293 } 294 295 bool DXVAVideoDecodeAccelerator::DXVAPictureBuffer:: 296 CopyOutputSampleDataToPictureBuffer( 297 const DXVAVideoDecodeAccelerator& decoder, 298 IDirect3DSurface9* dest_surface) { 299 DCHECK(dest_surface); 300 301 D3DSURFACE_DESC surface_desc; 302 HRESULT hr = dest_surface->GetDesc(&surface_desc); 303 RETURN_ON_HR_FAILURE(hr, "Failed to get surface description", false); 304 305 D3DSURFACE_DESC texture_desc; 306 decoding_texture_->GetLevelDesc(0, &texture_desc); 307 308 if (texture_desc.Width != surface_desc.Width || 309 texture_desc.Height != surface_desc.Height) { 310 NOTREACHED() << "Decode surface of different dimension than texture"; 311 return false; 312 } 313 314 hr = decoder.d3d9_->CheckDeviceFormatConversion( 315 D3DADAPTER_DEFAULT, D3DDEVTYPE_HAL, surface_desc.Format, 316 use_rgb_ ? D3DFMT_X8R8G8B8 : D3DFMT_A8R8G8B8); 317 RETURN_ON_HR_FAILURE(hr, "Device does not support format converision", false); 318 319 // This function currently executes in the context of IPC handlers in the 320 // GPU process which ensures that there is always an OpenGL context. 321 GLint current_texture = 0; 322 glGetIntegerv(GL_TEXTURE_BINDING_2D, ¤t_texture); 323 324 glBindTexture(GL_TEXTURE_2D, picture_buffer_.texture_id()); 325 326 glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR); 327 328 base::win::ScopedComPtr<IDirect3DSurface9> d3d_surface; 329 hr = decoding_texture_->GetSurfaceLevel(0, d3d_surface.Receive()); 330 RETURN_ON_HR_FAILURE(hr, "Failed to get surface from texture", false); 331 332 hr = decoder.device_->StretchRect( 333 dest_surface, NULL, d3d_surface, NULL, D3DTEXF_NONE); 334 RETURN_ON_HR_FAILURE(hr, "Colorspace conversion via StretchRect failed", 335 false); 336 337 // Ideally, this should be done immediately before the draw call that uses 338 // the texture. Flush it once here though. 339 hr = decoder.query_->Issue(D3DISSUE_END); 340 RETURN_ON_HR_FAILURE(hr, "Failed to issue END", false); 341 342 // The DXVA decoder has its own device which it uses for decoding. ANGLE 343 // has its own device which we don't have access to. 344 // The above code attempts to copy the decoded picture into a surface 345 // which is owned by ANGLE. As there are multiple devices involved in 346 // this, the StretchRect call above is not synchronous. 347 // We attempt to flush the batched operations to ensure that the picture is 348 // copied to the surface owned by ANGLE. 349 // We need to do this in a loop and call flush multiple times. 350 // We have seen the GetData call for flushing the command buffer fail to 351 // return success occassionally on multi core machines, leading to an 352 // infinite loop. 353 // Workaround is to have an upper limit of 10 on the number of iterations to 354 // wait for the Flush to finish. 355 int iterations = 0; 356 while ((decoder.query_->GetData(NULL, 0, D3DGETDATA_FLUSH) == S_FALSE) && 357 ++iterations < kMaxIterationsForD3DFlush) { 358 Sleep(1); // Poor-man's Yield(). 359 } 360 EGLDisplay egl_display = gfx::GLSurfaceEGL::GetHardwareDisplay(); 361 eglBindTexImage( 362 egl_display, 363 decoding_surface_, 364 EGL_BACK_BUFFER); 365 glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR); 366 glBindTexture(GL_TEXTURE_2D, current_texture); 367 return true; 368 } 369 370 DXVAVideoDecodeAccelerator::PendingSampleInfo::PendingSampleInfo( 371 int32 buffer_id, IMFSample* sample) 372 : input_buffer_id(buffer_id) { 373 output_sample.Attach(sample); 374 } 375 376 DXVAVideoDecodeAccelerator::PendingSampleInfo::~PendingSampleInfo() {} 377 378 // static 379 bool DXVAVideoDecodeAccelerator::CreateD3DDevManager() { 380 TRACE_EVENT0("gpu", "DXVAVideoDecodeAccelerator_CreateD3DDevManager"); 381 382 HRESULT hr = Direct3DCreate9Ex(D3D_SDK_VERSION, d3d9_.Receive()); 383 RETURN_ON_HR_FAILURE(hr, "Direct3DCreate9Ex failed", false); 384 385 D3DPRESENT_PARAMETERS present_params = {0}; 386 present_params.BackBufferWidth = 1; 387 present_params.BackBufferHeight = 1; 388 present_params.BackBufferFormat = D3DFMT_UNKNOWN; 389 present_params.BackBufferCount = 1; 390 present_params.SwapEffect = D3DSWAPEFFECT_DISCARD; 391 present_params.hDeviceWindow = ::GetShellWindow(); 392 present_params.Windowed = TRUE; 393 present_params.Flags = D3DPRESENTFLAG_VIDEO; 394 present_params.FullScreen_RefreshRateInHz = 0; 395 present_params.PresentationInterval = 0; 396 397 hr = d3d9_->CreateDeviceEx(D3DADAPTER_DEFAULT, 398 D3DDEVTYPE_HAL, 399 ::GetShellWindow(), 400 D3DCREATE_FPU_PRESERVE | 401 D3DCREATE_SOFTWARE_VERTEXPROCESSING | 402 D3DCREATE_DISABLE_PSGP_THREADING | 403 D3DCREATE_MULTITHREADED, 404 &present_params, 405 NULL, 406 device_.Receive()); 407 RETURN_ON_HR_FAILURE(hr, "Failed to create D3D device", false); 408 409 hr = DXVA2CreateDirect3DDeviceManager9(&dev_manager_reset_token_, 410 device_manager_.Receive()); 411 RETURN_ON_HR_FAILURE(hr, "DXVA2CreateDirect3DDeviceManager9 failed", false); 412 413 hr = device_manager_->ResetDevice(device_, dev_manager_reset_token_); 414 RETURN_ON_HR_FAILURE(hr, "Failed to reset device", false); 415 416 hr = device_->CreateQuery(D3DQUERYTYPE_EVENT, query_.Receive()); 417 RETURN_ON_HR_FAILURE(hr, "Failed to create D3D device query", false); 418 // Ensure query_ API works (to avoid an infinite loop later in 419 // CopyOutputSampleDataToPictureBuffer). 420 hr = query_->Issue(D3DISSUE_END); 421 RETURN_ON_HR_FAILURE(hr, "Failed to issue END test query", false); 422 return true; 423 } 424 425 DXVAVideoDecodeAccelerator::DXVAVideoDecodeAccelerator( 426 const base::Callback<bool(void)>& make_context_current) 427 : client_(NULL), 428 dev_manager_reset_token_(0), 429 egl_config_(NULL), 430 state_(kUninitialized), 431 pictures_requested_(false), 432 inputs_before_decode_(0), 433 make_context_current_(make_context_current), 434 weak_this_factory_(this) { 435 memset(&input_stream_info_, 0, sizeof(input_stream_info_)); 436 memset(&output_stream_info_, 0, sizeof(output_stream_info_)); 437 } 438 439 DXVAVideoDecodeAccelerator::~DXVAVideoDecodeAccelerator() { 440 client_ = NULL; 441 } 442 443 bool DXVAVideoDecodeAccelerator::Initialize(media::VideoCodecProfile profile, 444 Client* client) { 445 DCHECK(CalledOnValidThread()); 446 447 client_ = client; 448 449 // Not all versions of Windows 7 and later include Media Foundation DLLs. 450 // Instead of crashing while delay loading the DLL when calling MFStartup() 451 // below, probe whether we can successfully load the DLL now. 452 // 453 // See http://crbug.com/339678 for details. 454 HMODULE mfplat_dll = ::LoadLibrary(L"MFPlat.dll"); 455 RETURN_ON_FAILURE(mfplat_dll, "MFPlat.dll is required for decoding", false); 456 457 // TODO(ananta) 458 // H264PROFILE_HIGH video decoding is janky at times. Needs more 459 // investigation. http://crbug.com/426707 460 if (profile != media::H264PROFILE_BASELINE && 461 profile != media::H264PROFILE_MAIN) { 462 RETURN_AND_NOTIFY_ON_FAILURE(false, 463 "Unsupported h264 profile", PLATFORM_FAILURE, false); 464 } 465 466 RETURN_AND_NOTIFY_ON_FAILURE( 467 gfx::g_driver_egl.ext.b_EGL_ANGLE_surface_d3d_texture_2d_share_handle, 468 "EGL_ANGLE_surface_d3d_texture_2d_share_handle unavailable", 469 PLATFORM_FAILURE, 470 false); 471 472 RETURN_AND_NOTIFY_ON_FAILURE((state_ == kUninitialized), 473 "Initialize: invalid state: " << state_, ILLEGAL_STATE, false); 474 475 HRESULT hr = MFStartup(MF_VERSION, MFSTARTUP_FULL); 476 RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "MFStartup failed.", PLATFORM_FAILURE, 477 false); 478 479 RETURN_AND_NOTIFY_ON_FAILURE(CreateD3DDevManager(), 480 "Failed to initialize D3D device and manager", 481 PLATFORM_FAILURE, 482 false); 483 484 RETURN_AND_NOTIFY_ON_FAILURE(InitDecoder(profile), 485 "Failed to initialize decoder", PLATFORM_FAILURE, false); 486 487 RETURN_AND_NOTIFY_ON_FAILURE(GetStreamsInfoAndBufferReqs(), 488 "Failed to get input/output stream info.", PLATFORM_FAILURE, false); 489 490 RETURN_AND_NOTIFY_ON_FAILURE( 491 SendMFTMessage(MFT_MESSAGE_NOTIFY_BEGIN_STREAMING, 0), 492 "Send MFT_MESSAGE_NOTIFY_BEGIN_STREAMING notification failed", 493 PLATFORM_FAILURE, false); 494 495 RETURN_AND_NOTIFY_ON_FAILURE( 496 SendMFTMessage(MFT_MESSAGE_NOTIFY_START_OF_STREAM, 0), 497 "Send MFT_MESSAGE_NOTIFY_START_OF_STREAM notification failed", 498 PLATFORM_FAILURE, false); 499 500 state_ = kNormal; 501 return true; 502 } 503 504 void DXVAVideoDecodeAccelerator::Decode( 505 const media::BitstreamBuffer& bitstream_buffer) { 506 DCHECK(CalledOnValidThread()); 507 508 RETURN_AND_NOTIFY_ON_FAILURE((state_ == kNormal || state_ == kStopped || 509 state_ == kFlushing), 510 "Invalid state: " << state_, ILLEGAL_STATE,); 511 512 base::win::ScopedComPtr<IMFSample> sample; 513 sample.Attach(CreateSampleFromInputBuffer(bitstream_buffer, 514 input_stream_info_.cbSize, 515 input_stream_info_.cbAlignment)); 516 RETURN_AND_NOTIFY_ON_FAILURE(sample, "Failed to create input sample", 517 PLATFORM_FAILURE,); 518 519 RETURN_AND_NOTIFY_ON_HR_FAILURE(sample->SetSampleTime(bitstream_buffer.id()), 520 "Failed to associate input buffer id with sample", PLATFORM_FAILURE,); 521 522 DecodeInternal(sample); 523 } 524 525 void DXVAVideoDecodeAccelerator::AssignPictureBuffers( 526 const std::vector<media::PictureBuffer>& buffers) { 527 DCHECK(CalledOnValidThread()); 528 529 RETURN_AND_NOTIFY_ON_FAILURE((state_ != kUninitialized), 530 "Invalid state: " << state_, ILLEGAL_STATE,); 531 RETURN_AND_NOTIFY_ON_FAILURE((kNumPictureBuffers == buffers.size()), 532 "Failed to provide requested picture buffers. (Got " << buffers.size() << 533 ", requested " << kNumPictureBuffers << ")", INVALID_ARGUMENT,); 534 535 // Copy the picture buffers provided by the client to the available list, 536 // and mark these buffers as available for use. 537 for (size_t buffer_index = 0; buffer_index < buffers.size(); 538 ++buffer_index) { 539 linked_ptr<DXVAPictureBuffer> picture_buffer = 540 DXVAPictureBuffer::Create(*this, buffers[buffer_index], egl_config_); 541 RETURN_AND_NOTIFY_ON_FAILURE(picture_buffer.get(), 542 "Failed to allocate picture buffer", PLATFORM_FAILURE,); 543 544 bool inserted = output_picture_buffers_.insert(std::make_pair( 545 buffers[buffer_index].id(), picture_buffer)).second; 546 DCHECK(inserted); 547 } 548 ProcessPendingSamples(); 549 if (state_ == kFlushing && pending_output_samples_.empty()) 550 FlushInternal(); 551 } 552 553 void DXVAVideoDecodeAccelerator::ReusePictureBuffer( 554 int32 picture_buffer_id) { 555 DCHECK(CalledOnValidThread()); 556 557 RETURN_AND_NOTIFY_ON_FAILURE((state_ != kUninitialized), 558 "Invalid state: " << state_, ILLEGAL_STATE,); 559 560 if (output_picture_buffers_.empty() && stale_output_picture_buffers_.empty()) 561 return; 562 563 OutputBuffers::iterator it = output_picture_buffers_.find(picture_buffer_id); 564 // If we didn't find the picture id in the |output_picture_buffers_| map we 565 // try the |stale_output_picture_buffers_| map, as this may have been an 566 // output picture buffer from before a resolution change, that at resolution 567 // change time had yet to be displayed. The client is calling us back to tell 568 // us that we can now recycle this picture buffer, so if we were waiting to 569 // dispose of it we now can. 570 if (it == output_picture_buffers_.end()) { 571 it = stale_output_picture_buffers_.find(picture_buffer_id); 572 RETURN_AND_NOTIFY_ON_FAILURE(it != stale_output_picture_buffers_.end(), 573 "Invalid picture id: " << picture_buffer_id, INVALID_ARGUMENT,); 574 base::MessageLoop::current()->PostTask(FROM_HERE, 575 base::Bind(&DXVAVideoDecodeAccelerator::DeferredDismissStaleBuffer, 576 weak_this_factory_.GetWeakPtr(), picture_buffer_id)); 577 return; 578 } 579 580 it->second->ReusePictureBuffer(); 581 ProcessPendingSamples(); 582 583 if (state_ == kFlushing && pending_output_samples_.empty()) 584 FlushInternal(); 585 } 586 587 void DXVAVideoDecodeAccelerator::Flush() { 588 DCHECK(CalledOnValidThread()); 589 590 DVLOG(1) << "DXVAVideoDecodeAccelerator::Flush"; 591 592 RETURN_AND_NOTIFY_ON_FAILURE((state_ == kNormal || state_ == kStopped), 593 "Unexpected decoder state: " << state_, ILLEGAL_STATE,); 594 595 state_ = kFlushing; 596 597 RETURN_AND_NOTIFY_ON_FAILURE(SendMFTMessage(MFT_MESSAGE_COMMAND_DRAIN, 0), 598 "Failed to send drain message", PLATFORM_FAILURE,); 599 600 if (!pending_output_samples_.empty()) 601 return; 602 603 FlushInternal(); 604 } 605 606 void DXVAVideoDecodeAccelerator::Reset() { 607 DCHECK(CalledOnValidThread()); 608 609 DVLOG(1) << "DXVAVideoDecodeAccelerator::Reset"; 610 611 RETURN_AND_NOTIFY_ON_FAILURE((state_ == kNormal || state_ == kStopped), 612 "Reset: invalid state: " << state_, ILLEGAL_STATE,); 613 614 state_ = kResetting; 615 616 pending_output_samples_.clear(); 617 618 NotifyInputBuffersDropped(); 619 620 RETURN_AND_NOTIFY_ON_FAILURE(SendMFTMessage(MFT_MESSAGE_COMMAND_FLUSH, 0), 621 "Reset: Failed to send message.", PLATFORM_FAILURE,); 622 623 base::MessageLoop::current()->PostTask( 624 FROM_HERE, 625 base::Bind(&DXVAVideoDecodeAccelerator::NotifyResetDone, 626 weak_this_factory_.GetWeakPtr())); 627 628 state_ = DXVAVideoDecodeAccelerator::kNormal; 629 } 630 631 void DXVAVideoDecodeAccelerator::Destroy() { 632 DCHECK(CalledOnValidThread()); 633 Invalidate(); 634 delete this; 635 } 636 637 bool DXVAVideoDecodeAccelerator::CanDecodeOnIOThread() { 638 return false; 639 } 640 641 bool DXVAVideoDecodeAccelerator::InitDecoder(media::VideoCodecProfile profile) { 642 if (profile < media::H264PROFILE_MIN || profile > media::H264PROFILE_MAX) 643 return false; 644 645 // We mimic the steps CoCreateInstance uses to instantiate the object. This 646 // was previously done because it failed inside the sandbox, and now is done 647 // as a more minimal approach to avoid other side-effects CCI might have (as 648 // we are still in a reduced sandbox). 649 HMODULE decoder_dll = ::LoadLibrary(L"msmpeg2vdec.dll"); 650 RETURN_ON_FAILURE(decoder_dll, 651 "msmpeg2vdec.dll required for decoding is not loaded", 652 false); 653 654 // Check version of DLL, version 6.7.7140 is blacklisted due to high crash 655 // rates in browsers loading that DLL. If that is the version installed we 656 // fall back to software decoding. See crbug/403440. 657 FileVersionInfo* version_info = 658 FileVersionInfo::CreateFileVersionInfoForModule(decoder_dll); 659 RETURN_ON_FAILURE(version_info, 660 "unable to get version of msmpeg2vdec.dll", 661 false); 662 base::string16 file_version = version_info->file_version(); 663 RETURN_ON_FAILURE(file_version.find(L"6.1.7140") == base::string16::npos, 664 "blacklisted version of msmpeg2vdec.dll 6.7.7140", 665 false); 666 667 typedef HRESULT(WINAPI * GetClassObject)( 668 const CLSID & clsid, const IID & iid, void * *object); 669 670 GetClassObject get_class_object = reinterpret_cast<GetClassObject>( 671 GetProcAddress(decoder_dll, "DllGetClassObject")); 672 RETURN_ON_FAILURE( 673 get_class_object, "Failed to get DllGetClassObject pointer", false); 674 675 base::win::ScopedComPtr<IClassFactory> factory; 676 HRESULT hr = get_class_object(__uuidof(CMSH264DecoderMFT), 677 __uuidof(IClassFactory), 678 reinterpret_cast<void**>(factory.Receive())); 679 RETURN_ON_HR_FAILURE(hr, "DllGetClassObject for decoder failed", false); 680 681 hr = factory->CreateInstance(NULL, 682 __uuidof(IMFTransform), 683 reinterpret_cast<void**>(decoder_.Receive())); 684 RETURN_ON_HR_FAILURE(hr, "Failed to create decoder instance", false); 685 686 RETURN_ON_FAILURE(CheckDecoderDxvaSupport(), 687 "Failed to check decoder DXVA support", false); 688 689 hr = decoder_->ProcessMessage( 690 MFT_MESSAGE_SET_D3D_MANAGER, 691 reinterpret_cast<ULONG_PTR>(device_manager_.get())); 692 RETURN_ON_HR_FAILURE(hr, "Failed to pass D3D manager to decoder", false); 693 694 EGLDisplay egl_display = gfx::GLSurfaceEGL::GetHardwareDisplay(); 695 696 EGLint config_attribs[] = { 697 EGL_BUFFER_SIZE, 32, 698 EGL_RED_SIZE, 8, 699 EGL_GREEN_SIZE, 8, 700 EGL_BLUE_SIZE, 8, 701 EGL_SURFACE_TYPE, EGL_PBUFFER_BIT, 702 EGL_ALPHA_SIZE, 0, 703 EGL_NONE 704 }; 705 706 EGLint num_configs; 707 708 if (!eglChooseConfig( 709 egl_display, 710 config_attribs, 711 &egl_config_, 712 1, 713 &num_configs)) 714 return false; 715 716 return SetDecoderMediaTypes(); 717 } 718 719 bool DXVAVideoDecodeAccelerator::CheckDecoderDxvaSupport() { 720 base::win::ScopedComPtr<IMFAttributes> attributes; 721 HRESULT hr = decoder_->GetAttributes(attributes.Receive()); 722 RETURN_ON_HR_FAILURE(hr, "Failed to get decoder attributes", false); 723 724 UINT32 dxva = 0; 725 hr = attributes->GetUINT32(MF_SA_D3D_AWARE, &dxva); 726 RETURN_ON_HR_FAILURE(hr, "Failed to check if decoder supports DXVA", false); 727 728 hr = attributes->SetUINT32(CODECAPI_AVDecVideoAcceleration_H264, TRUE); 729 RETURN_ON_HR_FAILURE(hr, "Failed to enable DXVA H/W decoding", false); 730 return true; 731 } 732 733 bool DXVAVideoDecodeAccelerator::SetDecoderMediaTypes() { 734 RETURN_ON_FAILURE(SetDecoderInputMediaType(), 735 "Failed to set decoder input media type", false); 736 return SetDecoderOutputMediaType(MFVideoFormat_NV12); 737 } 738 739 bool DXVAVideoDecodeAccelerator::SetDecoderInputMediaType() { 740 base::win::ScopedComPtr<IMFMediaType> media_type; 741 HRESULT hr = MFCreateMediaType(media_type.Receive()); 742 RETURN_ON_HR_FAILURE(hr, "MFCreateMediaType failed", false); 743 744 hr = media_type->SetGUID(MF_MT_MAJOR_TYPE, MFMediaType_Video); 745 RETURN_ON_HR_FAILURE(hr, "Failed to set major input type", false); 746 747 hr = media_type->SetGUID(MF_MT_SUBTYPE, MFVideoFormat_H264); 748 RETURN_ON_HR_FAILURE(hr, "Failed to set subtype", false); 749 750 // Not sure about this. msdn recommends setting this value on the input 751 // media type. 752 hr = media_type->SetUINT32(MF_MT_INTERLACE_MODE, 753 MFVideoInterlace_MixedInterlaceOrProgressive); 754 RETURN_ON_HR_FAILURE(hr, "Failed to set interlace mode", false); 755 756 hr = decoder_->SetInputType(0, media_type, 0); // No flags 757 RETURN_ON_HR_FAILURE(hr, "Failed to set decoder input type", false); 758 return true; 759 } 760 761 bool DXVAVideoDecodeAccelerator::SetDecoderOutputMediaType( 762 const GUID& subtype) { 763 base::win::ScopedComPtr<IMFMediaType> out_media_type; 764 765 for (uint32 i = 0; 766 SUCCEEDED(decoder_->GetOutputAvailableType(0, i, 767 out_media_type.Receive())); 768 ++i) { 769 GUID out_subtype = {0}; 770 HRESULT hr = out_media_type->GetGUID(MF_MT_SUBTYPE, &out_subtype); 771 RETURN_ON_HR_FAILURE(hr, "Failed to get output major type", false); 772 773 if (out_subtype == subtype) { 774 hr = decoder_->SetOutputType(0, out_media_type, 0); // No flags 775 RETURN_ON_HR_FAILURE(hr, "Failed to set decoder output type", false); 776 return true; 777 } 778 out_media_type.Release(); 779 } 780 return false; 781 } 782 783 bool DXVAVideoDecodeAccelerator::SendMFTMessage(MFT_MESSAGE_TYPE msg, 784 int32 param) { 785 HRESULT hr = decoder_->ProcessMessage(msg, param); 786 return SUCCEEDED(hr); 787 } 788 789 // Gets the minimum buffer sizes for input and output samples. The MFT will not 790 // allocate buffer for input nor output, so we have to do it ourselves and make 791 // sure they're the correct size. We only provide decoding if DXVA is enabled. 792 bool DXVAVideoDecodeAccelerator::GetStreamsInfoAndBufferReqs() { 793 HRESULT hr = decoder_->GetInputStreamInfo(0, &input_stream_info_); 794 RETURN_ON_HR_FAILURE(hr, "Failed to get input stream info", false); 795 796 hr = decoder_->GetOutputStreamInfo(0, &output_stream_info_); 797 RETURN_ON_HR_FAILURE(hr, "Failed to get decoder output stream info", false); 798 799 DVLOG(1) << "Input stream info: "; 800 DVLOG(1) << "Max latency: " << input_stream_info_.hnsMaxLatency; 801 // There should be three flags, one for requiring a whole frame be in a 802 // single sample, one for requiring there be one buffer only in a single 803 // sample, and one that specifies a fixed sample size. (as in cbSize) 804 CHECK_EQ(input_stream_info_.dwFlags, 0x7u); 805 806 DVLOG(1) << "Min buffer size: " << input_stream_info_.cbSize; 807 DVLOG(1) << "Max lookahead: " << input_stream_info_.cbMaxLookahead; 808 DVLOG(1) << "Alignment: " << input_stream_info_.cbAlignment; 809 810 DVLOG(1) << "Output stream info: "; 811 // The flags here should be the same and mean the same thing, except when 812 // DXVA is enabled, there is an extra 0x100 flag meaning decoder will 813 // allocate its own sample. 814 DVLOG(1) << "Flags: " 815 << std::hex << std::showbase << output_stream_info_.dwFlags; 816 CHECK_EQ(output_stream_info_.dwFlags, 0x107u); 817 DVLOG(1) << "Min buffer size: " << output_stream_info_.cbSize; 818 DVLOG(1) << "Alignment: " << output_stream_info_.cbAlignment; 819 return true; 820 } 821 822 void DXVAVideoDecodeAccelerator::DoDecode() { 823 // This function is also called from FlushInternal in a loop which could 824 // result in the state transitioning to kStopped due to no decoded output. 825 RETURN_AND_NOTIFY_ON_FAILURE((state_ == kNormal || state_ == kFlushing || 826 state_ == kStopped), 827 "DoDecode: not in normal/flushing/stopped state", ILLEGAL_STATE,); 828 829 MFT_OUTPUT_DATA_BUFFER output_data_buffer = {0}; 830 DWORD status = 0; 831 832 HRESULT hr = decoder_->ProcessOutput(0, // No flags 833 1, // # of out streams to pull from 834 &output_data_buffer, 835 &status); 836 IMFCollection* events = output_data_buffer.pEvents; 837 if (events != NULL) { 838 VLOG(1) << "Got events from ProcessOuput, but discarding"; 839 events->Release(); 840 } 841 if (FAILED(hr)) { 842 // A stream change needs further ProcessInput calls to get back decoder 843 // output which is why we need to set the state to stopped. 844 if (hr == MF_E_TRANSFORM_STREAM_CHANGE) { 845 if (!SetDecoderOutputMediaType(MFVideoFormat_NV12)) { 846 // Decoder didn't let us set NV12 output format. Not sure as to why 847 // this can happen. Give up in disgust. 848 NOTREACHED() << "Failed to set decoder output media type to NV12"; 849 state_ = kStopped; 850 } else { 851 DVLOG(1) << "Received output format change from the decoder." 852 " Recursively invoking DoDecode"; 853 DoDecode(); 854 } 855 return; 856 } else if (hr == MF_E_TRANSFORM_NEED_MORE_INPUT) { 857 // No more output from the decoder. Stop playback. 858 state_ = kStopped; 859 return; 860 } else { 861 NOTREACHED() << "Unhandled error in DoDecode()"; 862 return; 863 } 864 } 865 TRACE_EVENT_END_ETW("DXVAVideoDecodeAccelerator.Decoding", this, ""); 866 867 TRACE_COUNTER1("DXVA Decoding", "TotalPacketsBeforeDecode", 868 inputs_before_decode_); 869 870 inputs_before_decode_ = 0; 871 872 RETURN_AND_NOTIFY_ON_FAILURE(ProcessOutputSample(output_data_buffer.pSample), 873 "Failed to process output sample.", PLATFORM_FAILURE,); 874 } 875 876 bool DXVAVideoDecodeAccelerator::ProcessOutputSample(IMFSample* sample) { 877 RETURN_ON_FAILURE(sample, "Decode succeeded with NULL output sample", false); 878 879 base::win::ScopedComPtr<IMFMediaBuffer> output_buffer; 880 HRESULT hr = sample->GetBufferByIndex(0, output_buffer.Receive()); 881 RETURN_ON_HR_FAILURE(hr, "Failed to get buffer from output sample", false); 882 883 base::win::ScopedComPtr<IDirect3DSurface9> surface; 884 hr = MFGetService(output_buffer, MR_BUFFER_SERVICE, 885 IID_PPV_ARGS(surface.Receive())); 886 RETURN_ON_HR_FAILURE(hr, "Failed to get D3D surface from output sample", 887 false); 888 889 LONGLONG input_buffer_id = 0; 890 RETURN_ON_HR_FAILURE(sample->GetSampleTime(&input_buffer_id), 891 "Failed to get input buffer id associated with sample", 892 false); 893 894 pending_output_samples_.push_back( 895 PendingSampleInfo(input_buffer_id, sample)); 896 897 // If we have available picture buffers to copy the output data then use the 898 // first one and then flag it as not being available for use. 899 if (output_picture_buffers_.size()) { 900 ProcessPendingSamples(); 901 return true; 902 } 903 if (pictures_requested_) { 904 DVLOG(1) << "Waiting for picture slots from the client."; 905 return true; 906 } 907 908 // We only read the surface description, which contains its width/height when 909 // we need the picture buffers from the client. Once we have those, then they 910 // are reused. 911 D3DSURFACE_DESC surface_desc; 912 hr = surface->GetDesc(&surface_desc); 913 RETURN_ON_HR_FAILURE(hr, "Failed to get surface description", false); 914 915 // Go ahead and request picture buffers. 916 base::MessageLoop::current()->PostTask( 917 FROM_HERE, 918 base::Bind(&DXVAVideoDecodeAccelerator::RequestPictureBuffers, 919 weak_this_factory_.GetWeakPtr(), 920 surface_desc.Width, 921 surface_desc.Height)); 922 923 pictures_requested_ = true; 924 return true; 925 } 926 927 void DXVAVideoDecodeAccelerator::ProcessPendingSamples() { 928 RETURN_AND_NOTIFY_ON_FAILURE(make_context_current_.Run(), 929 "Failed to make context current", PLATFORM_FAILURE,); 930 931 OutputBuffers::iterator index; 932 933 for (index = output_picture_buffers_.begin(); 934 index != output_picture_buffers_.end() && 935 !pending_output_samples_.empty(); 936 ++index) { 937 if (index->second->available()) { 938 PendingSampleInfo sample_info = pending_output_samples_.front(); 939 940 base::win::ScopedComPtr<IMFMediaBuffer> output_buffer; 941 HRESULT hr = sample_info.output_sample->GetBufferByIndex( 942 0, output_buffer.Receive()); 943 RETURN_AND_NOTIFY_ON_HR_FAILURE( 944 hr, "Failed to get buffer from output sample", PLATFORM_FAILURE,); 945 946 base::win::ScopedComPtr<IDirect3DSurface9> surface; 947 hr = MFGetService(output_buffer, MR_BUFFER_SERVICE, 948 IID_PPV_ARGS(surface.Receive())); 949 RETURN_AND_NOTIFY_ON_HR_FAILURE( 950 hr, "Failed to get D3D surface from output sample", 951 PLATFORM_FAILURE,); 952 953 D3DSURFACE_DESC surface_desc; 954 hr = surface->GetDesc(&surface_desc); 955 RETURN_AND_NOTIFY_ON_HR_FAILURE( 956 hr, "Failed to get surface description", PLATFORM_FAILURE,); 957 958 if (surface_desc.Width != 959 static_cast<uint32>(index->second->size().width()) || 960 surface_desc.Height != 961 static_cast<uint32>(index->second->size().height())) { 962 HandleResolutionChanged(surface_desc.Width, surface_desc.Height); 963 return; 964 } 965 966 RETURN_AND_NOTIFY_ON_FAILURE( 967 index->second->CopyOutputSampleDataToPictureBuffer(*this, surface), 968 "Failed to copy output sample", 969 PLATFORM_FAILURE, ); 970 971 media::Picture output_picture(index->second->id(), 972 sample_info.input_buffer_id, 973 gfx::Rect(index->second->size())); 974 base::MessageLoop::current()->PostTask( 975 FROM_HERE, 976 base::Bind(&DXVAVideoDecodeAccelerator::NotifyPictureReady, 977 weak_this_factory_.GetWeakPtr(), 978 output_picture)); 979 980 index->second->set_available(false); 981 pending_output_samples_.pop_front(); 982 } 983 } 984 985 if (!pending_input_buffers_.empty() && pending_output_samples_.empty()) { 986 base::MessageLoop::current()->PostTask( 987 FROM_HERE, 988 base::Bind(&DXVAVideoDecodeAccelerator::DecodePendingInputBuffers, 989 weak_this_factory_.GetWeakPtr())); 990 } 991 } 992 993 void DXVAVideoDecodeAccelerator::StopOnError( 994 media::VideoDecodeAccelerator::Error error) { 995 DCHECK(CalledOnValidThread()); 996 997 if (client_) 998 client_->NotifyError(error); 999 client_ = NULL; 1000 1001 if (state_ != kUninitialized) { 1002 Invalidate(); 1003 } 1004 } 1005 1006 void DXVAVideoDecodeAccelerator::Invalidate() { 1007 if (state_ == kUninitialized) 1008 return; 1009 weak_this_factory_.InvalidateWeakPtrs(); 1010 output_picture_buffers_.clear(); 1011 stale_output_picture_buffers_.clear(); 1012 pending_output_samples_.clear(); 1013 pending_input_buffers_.clear(); 1014 decoder_.Release(); 1015 MFShutdown(); 1016 state_ = kUninitialized; 1017 } 1018 1019 void DXVAVideoDecodeAccelerator::NotifyInputBufferRead(int input_buffer_id) { 1020 if (client_) 1021 client_->NotifyEndOfBitstreamBuffer(input_buffer_id); 1022 } 1023 1024 void DXVAVideoDecodeAccelerator::NotifyFlushDone() { 1025 if (client_) 1026 client_->NotifyFlushDone(); 1027 } 1028 1029 void DXVAVideoDecodeAccelerator::NotifyResetDone() { 1030 if (client_) 1031 client_->NotifyResetDone(); 1032 } 1033 1034 void DXVAVideoDecodeAccelerator::RequestPictureBuffers(int width, int height) { 1035 // This task could execute after the decoder has been torn down. 1036 if (state_ != kUninitialized && client_) { 1037 client_->ProvidePictureBuffers( 1038 kNumPictureBuffers, 1039 gfx::Size(width, height), 1040 GL_TEXTURE_2D); 1041 } 1042 } 1043 1044 void DXVAVideoDecodeAccelerator::NotifyPictureReady( 1045 const media::Picture& picture) { 1046 // This task could execute after the decoder has been torn down. 1047 if (state_ != kUninitialized && client_) 1048 client_->PictureReady(picture); 1049 } 1050 1051 void DXVAVideoDecodeAccelerator::NotifyInputBuffersDropped() { 1052 if (!client_ || !pending_output_samples_.empty()) 1053 return; 1054 1055 for (PendingInputs::iterator it = pending_input_buffers_.begin(); 1056 it != pending_input_buffers_.end(); ++it) { 1057 LONGLONG input_buffer_id = 0; 1058 RETURN_ON_HR_FAILURE((*it)->GetSampleTime(&input_buffer_id), 1059 "Failed to get buffer id associated with sample",); 1060 client_->NotifyEndOfBitstreamBuffer(input_buffer_id); 1061 } 1062 pending_input_buffers_.clear(); 1063 } 1064 1065 void DXVAVideoDecodeAccelerator::DecodePendingInputBuffers() { 1066 RETURN_AND_NOTIFY_ON_FAILURE((state_ != kUninitialized), 1067 "Invalid state: " << state_, ILLEGAL_STATE,); 1068 1069 if (pending_input_buffers_.empty() || !pending_output_samples_.empty()) 1070 return; 1071 1072 PendingInputs pending_input_buffers_copy; 1073 std::swap(pending_input_buffers_, pending_input_buffers_copy); 1074 1075 for (PendingInputs::iterator it = pending_input_buffers_copy.begin(); 1076 it != pending_input_buffers_copy.end(); ++it) { 1077 DecodeInternal(*it); 1078 } 1079 } 1080 1081 void DXVAVideoDecodeAccelerator::FlushInternal() { 1082 // The DoDecode function sets the state to kStopped when the decoder returns 1083 // MF_E_TRANSFORM_NEED_MORE_INPUT. 1084 // The MFT decoder can buffer upto 30 frames worth of input before returning 1085 // an output frame. This loop here attempts to retrieve as many output frames 1086 // as possible from the buffered set. 1087 while (state_ != kStopped) { 1088 DoDecode(); 1089 if (!pending_output_samples_.empty()) 1090 return; 1091 } 1092 1093 base::MessageLoop::current()->PostTask( 1094 FROM_HERE, 1095 base::Bind(&DXVAVideoDecodeAccelerator::NotifyFlushDone, 1096 weak_this_factory_.GetWeakPtr())); 1097 1098 state_ = kNormal; 1099 } 1100 1101 void DXVAVideoDecodeAccelerator::DecodeInternal( 1102 const base::win::ScopedComPtr<IMFSample>& sample) { 1103 DCHECK(CalledOnValidThread()); 1104 1105 if (state_ == kUninitialized) 1106 return; 1107 1108 if (!pending_output_samples_.empty() || !pending_input_buffers_.empty()) { 1109 pending_input_buffers_.push_back(sample); 1110 return; 1111 } 1112 1113 if (!inputs_before_decode_) { 1114 TRACE_EVENT_BEGIN_ETW("DXVAVideoDecodeAccelerator.Decoding", this, ""); 1115 } 1116 inputs_before_decode_++; 1117 1118 HRESULT hr = decoder_->ProcessInput(0, sample, 0); 1119 // As per msdn if the decoder returns MF_E_NOTACCEPTING then it means that it 1120 // has enough data to produce one or more output samples. In this case the 1121 // recommended options are to 1122 // 1. Generate new output by calling IMFTransform::ProcessOutput until it 1123 // returns MF_E_TRANSFORM_NEED_MORE_INPUT. 1124 // 2. Flush the input data 1125 // We implement the first option, i.e to retrieve the output sample and then 1126 // process the input again. Failure in either of these steps is treated as a 1127 // decoder failure. 1128 if (hr == MF_E_NOTACCEPTING) { 1129 DoDecode(); 1130 RETURN_AND_NOTIFY_ON_FAILURE((state_ == kStopped || state_ == kNormal), 1131 "Failed to process output. Unexpected decoder state: " << state_, 1132 PLATFORM_FAILURE,); 1133 hr = decoder_->ProcessInput(0, sample, 0); 1134 // If we continue to get the MF_E_NOTACCEPTING error we do the following:- 1135 // 1. Add the input sample to the pending queue. 1136 // 2. If we don't have any output samples we post the 1137 // DecodePendingInputBuffers task to process the pending input samples. 1138 // If we have an output sample then the above task is posted when the 1139 // output samples are sent to the client. 1140 // This is because we only support 1 pending output sample at any 1141 // given time due to the limitation with the Microsoft media foundation 1142 // decoder where it recycles the output Decoder surfaces. 1143 if (hr == MF_E_NOTACCEPTING) { 1144 pending_input_buffers_.push_back(sample); 1145 if (pending_output_samples_.empty()) { 1146 base::MessageLoop::current()->PostTask( 1147 FROM_HERE, 1148 base::Bind(&DXVAVideoDecodeAccelerator::DecodePendingInputBuffers, 1149 weak_this_factory_.GetWeakPtr())); 1150 } 1151 return; 1152 } 1153 } 1154 RETURN_AND_NOTIFY_ON_HR_FAILURE(hr, "Failed to process input sample", 1155 PLATFORM_FAILURE,); 1156 1157 DoDecode(); 1158 1159 RETURN_AND_NOTIFY_ON_FAILURE((state_ == kStopped || state_ == kNormal), 1160 "Failed to process output. Unexpected decoder state: " << state_, 1161 ILLEGAL_STATE,); 1162 1163 LONGLONG input_buffer_id = 0; 1164 RETURN_ON_HR_FAILURE(sample->GetSampleTime(&input_buffer_id), 1165 "Failed to get input buffer id associated with sample",); 1166 // The Microsoft Media foundation decoder internally buffers up to 30 frames 1167 // before returning a decoded frame. We need to inform the client that this 1168 // input buffer is processed as it may stop sending us further input. 1169 // Note: This may break clients which expect every input buffer to be 1170 // associated with a decoded output buffer. 1171 // TODO(ananta) 1172 // Do some more investigation into whether it is possible to get the MFT 1173 // decoder to emit an output packet for every input packet. 1174 // http://code.google.com/p/chromium/issues/detail?id=108121 1175 // http://code.google.com/p/chromium/issues/detail?id=150925 1176 base::MessageLoop::current()->PostTask( 1177 FROM_HERE, 1178 base::Bind(&DXVAVideoDecodeAccelerator::NotifyInputBufferRead, 1179 weak_this_factory_.GetWeakPtr(), 1180 input_buffer_id)); 1181 } 1182 1183 void DXVAVideoDecodeAccelerator::HandleResolutionChanged(int width, 1184 int height) { 1185 base::MessageLoop::current()->PostTask( 1186 FROM_HERE, 1187 base::Bind(&DXVAVideoDecodeAccelerator::DismissStaleBuffers, 1188 weak_this_factory_.GetWeakPtr())); 1189 1190 base::MessageLoop::current()->PostTask( 1191 FROM_HERE, 1192 base::Bind(&DXVAVideoDecodeAccelerator::RequestPictureBuffers, 1193 weak_this_factory_.GetWeakPtr(), 1194 width, 1195 height)); 1196 } 1197 1198 void DXVAVideoDecodeAccelerator::DismissStaleBuffers() { 1199 OutputBuffers::iterator index; 1200 1201 for (index = output_picture_buffers_.begin(); 1202 index != output_picture_buffers_.end(); 1203 ++index) { 1204 if (index->second->available()) { 1205 DVLOG(1) << "Dismissing picture id: " << index->second->id(); 1206 client_->DismissPictureBuffer(index->second->id()); 1207 } else { 1208 // Move to |stale_output_picture_buffers_| for deferred deletion. 1209 stale_output_picture_buffers_.insert( 1210 std::make_pair(index->first, index->second)); 1211 } 1212 } 1213 1214 output_picture_buffers_.clear(); 1215 } 1216 1217 void DXVAVideoDecodeAccelerator::DeferredDismissStaleBuffer( 1218 int32 picture_buffer_id) { 1219 OutputBuffers::iterator it = stale_output_picture_buffers_.find( 1220 picture_buffer_id); 1221 DCHECK(it != stale_output_picture_buffers_.end()); 1222 DVLOG(1) << "Dismissing picture id: " << it->second->id(); 1223 client_->DismissPictureBuffer(it->second->id()); 1224 stale_output_picture_buffers_.erase(it); 1225 } 1226 1227 } // namespace content 1228