Index: webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java |
diff --git a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java |
index 7b31e08eed2a4817a652ad8eddb112e1c14379b0..ff776358436699ef976b692a8cea71d98cfdbeea 100644 |
--- a/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java |
+++ b/webrtc/modules/audio_device/android/java/src/org/webrtc/voiceengine/WebRtcAudioRecord.java |
@@ -192,10 +192,6 @@ |
Math.max(BUFFER_SIZE_FACTOR * minBufferSize, byteBuffer.capacity()); |
Logging.d(TAG, "bufferSizeInBytes: " + bufferSizeInBytes); |
try { |
- // TODO(henrika): the only supported audio source for input is currently |
- // AudioSource.VOICE_COMMUNICATION. Is there any reason why we should |
- // support other types, e.g. DEFAULT or MIC? Only reason I can think of |
- // is if the device does not support VOICE_COMMUNICATION. |
audioRecord = new AudioRecord(AudioSource.VOICE_COMMUNICATION, |
sampleRate, |
AudioFormat.CHANNEL_IN_MONO, |