1 // Copyright (c) 2012 The Chromium Authors. All rights reserved. 2 // Use of this source code is governed by a BSD-style license that can be 3 // found in the LICENSE file. 4 5 // Utility methods for the Core Audio API on Windows. 6 // Always ensure that Core Audio is supported before using these methods. 7 // Use media::CoreAudioIsSupported() for this purpose. 8 // Also, all methods must be called on a valid COM thread. This can be done 9 // by using the base::win::ScopedCOMInitializer helper class. 10 11 #ifndef MEDIA_AUDIO_WIN_CORE_AUDIO_UTIL_WIN_H_ 12 #define MEDIA_AUDIO_WIN_CORE_AUDIO_UTIL_WIN_H_ 13 14 #include <audioclient.h> 15 #include <mmdeviceapi.h> 16 #include <string> 17 18 #include "base/basictypes.h" 19 #include "base/time/time.h" 20 #include "base/win/scoped_comptr.h" 21 #include "media/audio/audio_device_name.h" 22 #include "media/audio/audio_parameters.h" 23 #include "media/base/media_export.h" 24 25 using base::win::ScopedComPtr; 26 27 namespace media { 28 29 30 // Represents audio channel configuration constants as understood by Windows. 31 // E.g. KSAUDIO_SPEAKER_MONO. For a list of possible values see: 32 // http://msdn.microsoft.com/en-us/library/windows/hardware/ff537083(v=vs.85).aspx 33 typedef uint32 ChannelConfig; 34 35 class MEDIA_EXPORT CoreAudioUtil { 36 public: 37 // Returns true if Windows Core Audio is supported. 38 // Always verify that this method returns true before using any of the 39 // methods in this class. 40 static bool IsSupported(); 41 42 // Converts between reference time to base::TimeDelta. 43 // One reference-time unit is 100 nanoseconds. 44 // Example: double s = RefererenceTimeToTimeDelta(t).InMillisecondsF(); 45 static base::TimeDelta RefererenceTimeToTimeDelta(REFERENCE_TIME time); 46 47 // Returns AUDCLNT_SHAREMODE_EXCLUSIVE if --enable-exclusive-mode is used 48 // as command-line flag and AUDCLNT_SHAREMODE_SHARED otherwise (default). 49 static AUDCLNT_SHAREMODE GetShareMode(); 50 51 // The Windows Multimedia Device (MMDevice) API enables audio clients to 52 // discover audio endpoint devices and determine their capabilities. 53 54 // Number of active audio devices in the specified flow data flow direction. 55 // Set |data_flow| to eAll to retrieve the total number of active audio 56 // devices. 57 static int NumberOfActiveDevices(EDataFlow data_flow); 58 59 // Creates an IMMDeviceEnumerator interface which provides methods for 60 // enumerating audio endpoint devices. 61 static ScopedComPtr<IMMDeviceEnumerator> CreateDeviceEnumerator(); 62 63 // Creates a default endpoint device that is specified by a data-flow 64 // direction and role, e.g. default render device. 65 static ScopedComPtr<IMMDevice> CreateDefaultDevice( 66 EDataFlow data_flow, ERole role); 67 68 // Returns the device id of the default output device or an empty string 69 // if no such device exists or if the default device has been disabled. 70 static std::string GetDefaultOutputDeviceID(); 71 72 // Creates an endpoint device that is specified by a unique endpoint device- 73 // identification string. 74 static ScopedComPtr<IMMDevice> CreateDevice(const std::string& device_id); 75 76 // Returns the unique ID and user-friendly name of a given endpoint device. 77 // Example: "{0.0.1.00000000}.{8db6020f-18e3-4f25-b6f5-7726c9122574}", and 78 // "Microphone (Realtek High Definition Audio)". 79 static HRESULT GetDeviceName(IMMDevice* device, AudioDeviceName* name); 80 81 // Returns the device ID/path of the controller (a.k.a. physical device that 82 // |device| is connected to. This ID will be the same for all devices from 83 // the same controller so it is useful for doing things like determining 84 // whether a set of output and input devices belong to the same controller. 85 // The device enumerator is required as well as the device itself since 86 // looking at the device topology is required and we need to open up 87 // associated devices to determine the controller id. 88 // If the ID could not be determined for some reason, an empty string is 89 // returned. 90 static std::string GetAudioControllerID(IMMDevice* device, 91 IMMDeviceEnumerator* enumerator); 92 93 // Accepts an id of an input device and finds a matching output device id. 94 // If the associated hardware does not have an audio output device (e.g. 95 // a webcam with a mic), an empty string is returned. 96 static std::string GetMatchingOutputDeviceID( 97 const std::string& input_device_id); 98 99 // Gets the user-friendly name of the endpoint device which is represented 100 // by a unique id in |device_id|. 101 static std::string GetFriendlyName(const std::string& device_id); 102 103 // Returns true if the provided unique |device_id| corresponds to the current 104 // default device for the specified by a data-flow direction and role. 105 static bool DeviceIsDefault( 106 EDataFlow flow, ERole role, const std::string& device_id); 107 108 // Query if the audio device is a rendering device or a capture device. 109 static EDataFlow GetDataFlow(IMMDevice* device); 110 111 // The Windows Audio Session API (WASAPI) enables client applications to 112 // manage the flow of audio data between the application and an audio endpoint 113 // device. 114 115 // Create an IAudioClient instance for the default IMMDevice where 116 // flow direction and role is define by |data_flow| and |role|. 117 // The IAudioClient interface enables a client to create and initialize an 118 // audio stream between an audio application and the audio engine (for a 119 // shared-mode stream) or the hardware buffer of an audio endpoint device 120 // (for an exclusive-mode stream). 121 static ScopedComPtr<IAudioClient> CreateDefaultClient(EDataFlow data_flow, 122 ERole role); 123 124 // Create an IAudioClient instance for a specific device _or_ the default 125 // device if |device_id| is empty. 126 static ScopedComPtr<IAudioClient> CreateClient(const std::string& device_id, 127 EDataFlow data_flow, 128 ERole role); 129 130 // Create an IAudioClient interface for an existing IMMDevice given by 131 // |audio_device|. Flow direction and role is define by the |audio_device|. 132 static ScopedComPtr<IAudioClient> CreateClient(IMMDevice* audio_device); 133 134 // Get the mix format that the audio engine uses internally for processing 135 // of shared-mode streams. This format is not necessarily a format that the 136 // audio endpoint device supports. Thus, the caller might not succeed in 137 // creating an exclusive-mode stream with a format obtained by this method. 138 static HRESULT GetSharedModeMixFormat(IAudioClient* client, 139 WAVEFORMATPCMEX* format); 140 141 // Returns true if the specified |client| supports the format in |format| 142 // for the given |share_mode| (shared or exclusive). 143 static bool IsFormatSupported(IAudioClient* client, 144 AUDCLNT_SHAREMODE share_mode, 145 const WAVEFORMATPCMEX* format); 146 147 // Returns true if the specified |channel_layout| is supported for the 148 // default IMMDevice where flow direction and role is define by |data_flow| 149 // and |role|. If this method returns true for a certain channel layout, it 150 // means that SharedModeInitialize() will succeed using a format based on 151 // the preferred format where the channel layout has been modified. 152 static bool IsChannelLayoutSupported(const std::string& device_id, 153 EDataFlow data_flow, 154 ERole role, 155 ChannelLayout channel_layout); 156 157 // For a shared-mode stream, the audio engine periodically processes the 158 // data in the endpoint buffer at the period obtained in |device_period|. 159 // For an exclusive mode stream, |device_period| corresponds to the minimum 160 // time interval between successive processing by the endpoint device. 161 // This period plus the stream latency between the buffer and endpoint device 162 // represents the minimum possible latency that an audio application can 163 // achieve. The time in |device_period| is expressed in 100-nanosecond units. 164 static HRESULT GetDevicePeriod(IAudioClient* client, 165 AUDCLNT_SHAREMODE share_mode, 166 REFERENCE_TIME* device_period); 167 168 // Get the preferred audio parameters for the specified |client| or the 169 // given direction and role is define by |data_flow| and |role|, or the 170 // unique device id given by |device_id|. 171 // The acquired values should only be utilized for shared mode streamed since 172 // there are no preferred settings for an exclusive mode stream. 173 static HRESULT GetPreferredAudioParameters(IAudioClient* client, 174 AudioParameters* params); 175 static HRESULT GetPreferredAudioParameters(EDataFlow data_flow, ERole role, 176 AudioParameters* params); 177 static HRESULT GetPreferredAudioParameters(const std::string& device_id, 178 AudioParameters* params); 179 180 // Retrieves an integer mask which corresponds to the channel layout the 181 // audio engine uses for its internal processing/mixing of shared-mode 182 // streams. This mask indicates which channels are present in the multi- 183 // channel stream. The least significant bit corresponds with the Front Left 184 // speaker, the next least significant bit corresponds to the Front Right 185 // speaker, and so on, continuing in the order defined in KsMedia.h. 186 // See http://msdn.microsoft.com/en-us/library/windows/hardware/ff537083(v=vs.85).aspx 187 // for more details. 188 // To get the channel config of the default device, pass an empty string 189 // for |device_id|. 190 static ChannelConfig GetChannelConfig(const std::string& device_id, 191 EDataFlow data_flow); 192 193 // After activating an IAudioClient interface on an audio endpoint device, 194 // the client must initialize it once, and only once, to initialize the audio 195 // stream between the client and the device. In shared mode, the client 196 // connects indirectly through the audio engine which does the mixing. 197 // In exclusive mode, the client connects directly to the audio hardware. 198 // If a valid event is provided in |event_handle|, the client will be 199 // initialized for event-driven buffer handling. If |event_handle| is set to 200 // NULL, event-driven buffer handling is not utilized. 201 static HRESULT SharedModeInitialize(IAudioClient* client, 202 const WAVEFORMATPCMEX* format, 203 HANDLE event_handle, 204 uint32* endpoint_buffer_size); 205 // TODO(henrika): add ExclusiveModeInitialize(...) 206 207 // Create an IAudioRenderClient client for an existing IAudioClient given by 208 // |client|. The IAudioRenderClient interface enables a client to write 209 // output data to a rendering endpoint buffer. 210 static ScopedComPtr<IAudioRenderClient> CreateRenderClient( 211 IAudioClient* client); 212 213 // Create an IAudioCaptureClient client for an existing IAudioClient given by 214 // |client|. The IAudioCaptureClient interface enables a client to read 215 // input data from a capture endpoint buffer. 216 static ScopedComPtr<IAudioCaptureClient> CreateCaptureClient( 217 IAudioClient* client); 218 219 // Fills up the endpoint rendering buffer with silence for an existing 220 // IAudioClient given by |client| and a corresponding IAudioRenderClient 221 // given by |render_client|. 222 static bool FillRenderEndpointBufferWithSilence( 223 IAudioClient* client, IAudioRenderClient* render_client); 224 225 private: 226 CoreAudioUtil() {} 227 ~CoreAudioUtil() {} 228 DISALLOW_COPY_AND_ASSIGN(CoreAudioUtil); 229 }; 230 231 } // namespace media 232 233 #endif // MEDIA_AUDIO_WIN_CORE_AUDIO_UTIL_WIN_H_ 234