| Index: webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| diff --git a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| index 7b31e08eed2a4817a652ad8eddb112e1c14379b0..ff776358436699ef976b692a8cea71d98cfdbeea 100644
|
| --- a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| +++ b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java
|
| @@ -192,10 +192,6 @@
|
| Math.max(BUFFER_SIZE_FACTOR * minBufferSize, byteBuffer.capacity());
|
| Logging.d(TAG, "bufferSizeInBytes: " + bufferSizeInBytes);
|
| try {
|
| - // TODO(henrika): the only supported audio source for input is currently
|
| - // AudioSource.VOICE_COMMUNICATION. Is there any reason why we should
|
| - // support other types, e.g. DEFAULT or MIC? Only reason I can think of
|
| - // is if the device does not support VOICE_COMMUNICATION.
|
| audioRecord = new AudioRecord(AudioSource.VOICE_COMMUNICATION,
|
| sampleRate,
|
| AudioFormat.CHANNEL_IN_MONO,
|
|
|