| Index: webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| diff --git a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| index ff776358436699ef976b692a8cea71d98cfdbeea..7b31e08eed2a4817a652ad8eddb112e1c14379b0 100644
|
| --- a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| +++ b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| @@ -192,6 +192,10 @@ class WebRtcAudioRecord {
|
| Math.max(BUFFER_SIZE_FACTOR * minBufferSize, byteBuffer.capacity());
|
| Logging.d(TAG, "bufferSizeInBytes: " + bufferSizeInBytes);
|
| try {
|
| + // TODO(henrika): the only supported audio source for input is currently
|
| + // AudioSource.VOICE_COMMUNICATION. Is there any reason why we should
|
| + // support other types, e.g. DEFAULT or MIC? Only reason I can think of
|
| + // is if the device does not support VOICE_COMMUNICATION.
|
| audioRecord = new AudioRecord(AudioSource.VOICE_COMMUNICATION,
|
| sampleRate,
|
| AudioFormat.CHANNEL_IN_MONO,
|
|
|