Index: webrtc/voice_engine/channel.cc |
diff --git a/webrtc/voice_engine/channel.cc b/webrtc/voice_engine/channel.cc |
index b1b55e8accd9e3d4bf2edafef8ee1c36cb2a0e9e..136c12f195c2776339e80bed04bda887d3f19919 100644 |
--- a/webrtc/voice_engine/channel.cc |
+++ b/webrtc/voice_engine/channel.cc |
@@ -14,6 +14,7 @@ |
#include <utility> |
#include "webrtc/base/checks.h" |
+#include "webrtc/base/criticalsection.h" |
#include "webrtc/base/format_macros.h" |
#include "webrtc/base/logging.h" |
#include "webrtc/base/thread_checker.h" |
@@ -30,7 +31,6 @@ |
#include "webrtc/modules/rtp_rtcp/source/rtp_receiver_strategy.h" |
#include "webrtc/modules/utility/include/audio_frame_operations.h" |
#include "webrtc/modules/utility/include/process_thread.h" |
-#include "webrtc/system_wrappers/include/critical_section_wrapper.h" |
#include "webrtc/system_wrappers/include/trace.h" |
#include "webrtc/voice_engine/include/voe_base.h" |
#include "webrtc/voice_engine/include/voe_external_media.h" |
@@ -157,9 +157,7 @@ struct ChannelStatistics : public RtcpStatistics { |
// Statistics callback, called at each generation of a new RTCP report block. |
class StatisticsProxy : public RtcpStatisticsCallback { |
public: |
- StatisticsProxy(uint32_t ssrc) |
- : stats_lock_(CriticalSectionWrapper::CreateCriticalSection()), |
- ssrc_(ssrc) {} |
+ StatisticsProxy(uint32_t ssrc) : ssrc_(ssrc) {} |
virtual ~StatisticsProxy() {} |
void StatisticsUpdated(const RtcpStatistics& statistics, |
@@ -167,7 +165,7 @@ class StatisticsProxy : public RtcpStatisticsCallback { |
if (ssrc != ssrc_) |
return; |
- CriticalSectionScoped cs(stats_lock_.get()); |
+ rtc::CritScope cs(&stats_lock_); |
stats_.rtcp = statistics; |
if (statistics.jitter > stats_.max_jitter) { |
stats_.max_jitter = statistics.jitter; |
@@ -177,7 +175,7 @@ class StatisticsProxy : public RtcpStatisticsCallback { |
void CNameChanged(const char* cname, uint32_t ssrc) override {} |
ChannelStatistics GetStats() { |
- CriticalSectionScoped cs(stats_lock_.get()); |
+ rtc::CritScope cs(&stats_lock_); |
return stats_; |
} |
@@ -185,7 +183,7 @@ class StatisticsProxy : public RtcpStatisticsCallback { |
// StatisticsUpdated calls are triggered from threads in the RTP module, |
// while GetStats calls can be triggered from the public voice engine API, |
// hence synchronization is needed. |
- rtc::scoped_ptr<CriticalSectionWrapper> stats_lock_; |
+ mutable rtc::CriticalSection stats_lock_; |
the sun
2016/01/21 13:07:28
nit: No need for mutable, unless you make GetStats
tommi
2016/01/21 15:29:22
Done.
|
const uint32_t ssrc_; |
ChannelStatistics stats_; |
}; |
@@ -298,7 +296,7 @@ Channel::InFrameType(FrameType frame_type) |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::InFrameType(frame_type=%d)", frame_type); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
_sendFrameType = (frame_type == kAudioFrameSpeech); |
return 0; |
} |
@@ -306,7 +304,7 @@ Channel::InFrameType(FrameType frame_type) |
int32_t |
Channel::OnRxVadDetected(int vadDecision) |
{ |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (_rxVadObserverPtr) |
{ |
_rxVadObserverPtr->OnRxVad(_channelId, vadDecision); |
@@ -321,7 +319,7 @@ bool Channel::SendRtp(const uint8_t* data, |
WEBRTC_TRACE(kTraceStream, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::SendPacket(channel=%d, len=%" PRIuS ")", len); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (_transportPtr == NULL) |
{ |
@@ -352,7 +350,7 @@ Channel::SendRtcp(const uint8_t *data, size_t len) |
WEBRTC_TRACE(kTraceStream, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::SendRtcp(len=%" PRIuS ")", len); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (_transportPtr == NULL) |
{ |
WEBRTC_TRACE(kTraceError, kTraceVoice, |
@@ -566,7 +564,7 @@ int32_t Channel::GetAudioFrame(int32_t id, AudioFrame* audioFrame) |
// scaling/panning, as that applies to the mix operation. |
// External recipients of the audio (e.g. via AudioTrack), will do their |
// own mixing/dynamic processing. |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (audio_sink_) { |
AudioSinkInterface::Data data( |
&audioFrame->data_[0], |
@@ -580,7 +578,7 @@ int32_t Channel::GetAudioFrame(int32_t id, AudioFrame* audioFrame) |
float left_pan = 1.0f; |
float right_pan = 1.0f; |
{ |
- CriticalSectionScoped cs(&volume_settings_critsect_); |
+ rtc::CritScope cs(&volume_settings_critsect_); |
output_gain = _outputGain; |
left_pan = _panLeft; |
right_pan= _panRight; |
@@ -620,7 +618,7 @@ int32_t Channel::GetAudioFrame(int32_t id, AudioFrame* audioFrame) |
// External media |
if (_outputExternalMedia) |
{ |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
const bool isStereo = (audioFrame->num_channels_ == 2); |
if (_outputExternalMediaCallbackPtr) |
{ |
@@ -633,7 +631,7 @@ int32_t Channel::GetAudioFrame(int32_t id, AudioFrame* audioFrame) |
// Record playout if enabled |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_outputFileRecording && _outputFileRecorderPtr) |
{ |
@@ -660,7 +658,7 @@ int32_t Channel::GetAudioFrame(int32_t id, AudioFrame* audioFrame) |
(GetPlayoutFrequency() / 1000); |
{ |
- CriticalSectionScoped lock(ts_stats_lock_.get()); |
+ rtc::CritScope lock(&ts_stats_lock_); |
// Compute ntp time. |
audioFrame->ntp_time_ms_ = ntp_estimator_.Estimate( |
audioFrame->timestamp_); |
@@ -704,7 +702,7 @@ Channel::NeededFrequency(int32_t id) const |
// limit the spectrum anyway. |
if (channel_state_.Get().output_file_playing) |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_outputFilePlayerPtr) |
{ |
if(_outputFilePlayerPtr->Frequency()>highestNeeded) |
@@ -790,7 +788,7 @@ Channel::RecordFileEnded(int32_t id) |
assert(id == _outputFileRecorderId); |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
_outputFileRecording = false; |
WEBRTC_TRACE(kTraceStateInfo, kTraceVoice, |
@@ -803,11 +801,7 @@ Channel::Channel(int32_t channelId, |
uint32_t instanceId, |
RtcEventLog* const event_log, |
const Config& config) |
- : _fileCritSect(*CriticalSectionWrapper::CreateCriticalSection()), |
- _callbackCritSect(*CriticalSectionWrapper::CreateCriticalSection()), |
- volume_settings_critsect_( |
- *CriticalSectionWrapper::CreateCriticalSection()), |
- _instanceId(instanceId), |
+ : _instanceId(instanceId), |
_channelId(channelId), |
event_log_(event_log), |
rtp_header_parser_(RtpHeaderParser::Create()), |
@@ -848,7 +842,6 @@ Channel::Channel(int32_t channelId, |
playout_delay_ms_(0), |
_numberOfDiscardedPackets(0), |
send_sequence_number_(0), |
- ts_stats_lock_(CriticalSectionWrapper::CreateCriticalSection()), |
rtp_ts_wraparound_handler_(new rtc::TimestampWrapAroundHandler()), |
capture_start_rtp_time_stamp_(-1), |
capture_start_ntp_time_ms_(-1), |
@@ -875,7 +868,6 @@ Channel::Channel(int32_t channelId, |
_lastPayloadType(0), |
_includeAudioLevelIndication(false), |
_outputSpeechType(AudioFrame::kNormalSpeech), |
- video_sync_lock_(CriticalSectionWrapper::CreateCriticalSection()), |
_average_jitter_buffer_delay_us(0), |
_previousTimestamp(0), |
_recPacketDelayMs(20), |
@@ -885,7 +877,6 @@ Channel::Channel(int32_t channelId, |
restored_packet_in_use_(false), |
rtcp_observer_(new VoERtcpObserver(this)), |
network_predictor_(new NetworkPredictor(Clock::GetRealTimeClock())), |
- assoc_send_channel_lock_(CriticalSectionWrapper::CreateCriticalSection()), |
associate_send_channel_(ChannelOwner(nullptr)), |
pacing_enabled_(config.Get<VoicePacing>().enabled), |
feedback_observer_proxy_(pacing_enabled_ ? new TransportFeedbackProxy() |
@@ -952,7 +943,7 @@ Channel::~Channel() |
StopPlayout(); |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_inputFilePlayerPtr) |
{ |
_inputFilePlayerPtr->RegisterModuleFileCallback(NULL); |
@@ -998,11 +989,6 @@ Channel::~Channel() |
_moduleProcessThreadPtr->DeRegisterModule(_rtpRtcpModule.get()); |
// End of modules shutdown |
- |
- // Delete other objects |
- delete &_callbackCritSect; |
- delete &_fileCritSect; |
- delete &volume_settings_critsect_; |
} |
int32_t |
@@ -1163,7 +1149,7 @@ Channel::SetEngineInformation(Statistics& engineStatistics, |
ProcessThread& moduleProcessThread, |
AudioDeviceModule& audioDeviceModule, |
VoiceEngineObserver* voiceEngineObserver, |
- CriticalSectionWrapper* callbackCritSect) |
+ rtc::CriticalSection* callbackCritSect) |
{ |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::SetEngineInformation()"); |
@@ -1186,7 +1172,7 @@ Channel::UpdateLocalTimeStamp() |
} |
void Channel::SetSink(rtc::scoped_ptr<AudioSinkInterface> sink) { |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
audio_sink_ = std::move(sink); |
} |
@@ -1266,7 +1252,7 @@ Channel::StartSend() |
_engineStatisticsPtr->SetLastError( |
VE_RTP_RTCP_MODULE_ERROR, kTraceError, |
"StartSend() RTP/RTCP failed to start sending"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
channel_state_.SetSending(false); |
return -1; |
} |
@@ -1338,7 +1324,7 @@ Channel::RegisterVoiceEngineObserver(VoiceEngineObserver& observer) |
{ |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::RegisterVoiceEngineObserver()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (_voiceEngineObserverPtr) |
{ |
@@ -1356,7 +1342,7 @@ Channel::DeRegisterVoiceEngineObserver() |
{ |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::DeRegisterVoiceEngineObserver()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (!_voiceEngineObserverPtr) |
{ |
@@ -1663,7 +1649,7 @@ int32_t Channel::RegisterExternalTransport(Transport& transport) |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId, _channelId), |
"Channel::RegisterExternalTransport()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (_externalTransport) |
{ |
@@ -1683,7 +1669,7 @@ Channel::DeRegisterExternalTransport() |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::DeRegisterExternalTransport()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (!_transportPtr) |
{ |
@@ -1827,7 +1813,7 @@ int32_t Channel::ReceivedRTCPPacket(const int8_t* data, size_t length) { |
} |
{ |
- CriticalSectionScoped lock(ts_stats_lock_.get()); |
+ rtc::CritScope lock(&ts_stats_lock_); |
ntp_estimator_.UpdateRtcpTimestamp(rtt, ntp_secs, ntp_frac, rtp_timestamp); |
} |
return 0; |
@@ -1856,7 +1842,7 @@ int Channel::StartPlayingFileLocally(const char* fileName, |
} |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_outputFilePlayerPtr) |
{ |
@@ -1935,7 +1921,7 @@ int Channel::StartPlayingFileLocally(InStream* stream, |
} |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
// Destroy the old instance |
if (_outputFilePlayerPtr) |
@@ -1994,7 +1980,7 @@ int Channel::StopPlayingFileLocally() |
} |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_outputFilePlayerPtr->StopPlayingFile() != 0) |
{ |
@@ -2046,7 +2032,7 @@ int Channel::RegisterFilePlayingToMixer() |
if (_outputMixerPtr->SetAnonymousMixabilityStatus(*this, true) != 0) |
{ |
channel_state_.SetOutputFilePlaying(false); |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
_engineStatisticsPtr->SetLastError( |
VE_AUDIO_CONF_MIX_MODULE_ERROR, kTraceError, |
"StartPlayingFile() failed to add participant as file to mixer"); |
@@ -2073,7 +2059,7 @@ int Channel::StartPlayingFileAsMicrophone(const char* fileName, |
"stopPosition=%d)", fileName, loop, format, volumeScaling, |
startPosition, stopPosition); |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (channel_state_.Get().input_file_playing) |
{ |
@@ -2148,7 +2134,7 @@ int Channel::StartPlayingFileAsMicrophone(InStream* stream, |
return -1; |
} |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (channel_state_.Get().input_file_playing) |
{ |
@@ -2204,7 +2190,7 @@ int Channel::StopPlayingFileAsMicrophone() |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::StopPlayingFileAsMicrophone()"); |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (!channel_state_.Get().input_file_playing) |
{ |
@@ -2272,7 +2258,7 @@ int Channel::StartRecordingPlayout(const char* fileName, |
format = kFileFormatCompressedFile; |
} |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
// Destroy the old instance |
if (_outputFileRecorderPtr) |
@@ -2349,7 +2335,7 @@ int Channel::StartRecordingPlayout(OutStream* stream, |
format = kFileFormatCompressedFile; |
} |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
// Destroy the old instance |
if (_outputFileRecorderPtr) |
@@ -2400,7 +2386,7 @@ int Channel::StopRecordingPlayout() |
} |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_outputFileRecorderPtr->StopRecording() != 0) |
{ |
@@ -2420,7 +2406,7 @@ int Channel::StopRecordingPlayout() |
void |
Channel::SetMixWithMicStatus(bool mix) |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
_mixFileWithMicrophone=mix; |
} |
@@ -2443,7 +2429,7 @@ Channel::GetSpeechOutputLevelFullRange(uint32_t& level) const |
int |
Channel::SetMute(bool enable) |
{ |
- CriticalSectionScoped cs(&volume_settings_critsect_); |
+ rtc::CritScope cs(&volume_settings_critsect_); |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::SetMute(enable=%d)", enable); |
_mute = enable; |
@@ -2453,14 +2439,14 @@ Channel::SetMute(bool enable) |
bool |
Channel::Mute() const |
{ |
- CriticalSectionScoped cs(&volume_settings_critsect_); |
+ rtc::CritScope cs(&volume_settings_critsect_); |
return _mute; |
} |
int |
Channel::SetOutputVolumePan(float left, float right) |
{ |
- CriticalSectionScoped cs(&volume_settings_critsect_); |
+ rtc::CritScope cs(&volume_settings_critsect_); |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::SetOutputVolumePan()"); |
_panLeft = left; |
@@ -2471,7 +2457,7 @@ Channel::SetOutputVolumePan(float left, float right) |
int |
Channel::GetOutputVolumePan(float& left, float& right) const |
{ |
- CriticalSectionScoped cs(&volume_settings_critsect_); |
+ rtc::CritScope cs(&volume_settings_critsect_); |
left = _panLeft; |
right = _panRight; |
return 0; |
@@ -2480,7 +2466,7 @@ Channel::GetOutputVolumePan(float& left, float& right) const |
int |
Channel::SetChannelOutputVolumeScaling(float scaling) |
{ |
- CriticalSectionScoped cs(&volume_settings_critsect_); |
+ rtc::CritScope cs(&volume_settings_critsect_); |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::SetChannelOutputVolumeScaling()"); |
_outputGain = scaling; |
@@ -2490,7 +2476,7 @@ Channel::SetChannelOutputVolumeScaling(float scaling) |
int |
Channel::GetChannelOutputVolumeScaling(float& scaling) const |
{ |
- CriticalSectionScoped cs(&volume_settings_critsect_); |
+ rtc::CritScope cs(&volume_settings_critsect_); |
scaling = _outputGain; |
return 0; |
} |
@@ -2600,7 +2586,7 @@ Channel::RegisterRxVadObserver(VoERxVadCallback &observer) |
{ |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::RegisterRxVadObserver()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (_rxVadObserverPtr) |
{ |
@@ -2619,7 +2605,7 @@ Channel::DeRegisterRxVadObserver() |
{ |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::DeRegisterRxVadObserver()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (!_rxVadObserverPtr) |
{ |
@@ -3251,7 +3237,7 @@ Channel::GetRTPStatistics(CallStatistics& stats) |
// --- Timestamps |
{ |
- CriticalSectionScoped lock(ts_stats_lock_.get()); |
+ rtc::CritScope lock(&ts_stats_lock_); |
stats.capture_start_ntp_time_ms_ = capture_start_ntp_time_ms_; |
} |
return 0; |
@@ -3392,7 +3378,7 @@ Channel::PrepareEncodeAndSend(int mixingFrequency) |
if (channel_state_.Get().input_external_media) |
{ |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
const bool isStereo = (_audioFrame.num_channels_ == 2); |
if (_inputExternalMediaCallbackPtr) |
{ |
@@ -3456,7 +3442,7 @@ Channel::EncodeAndSend() |
} |
void Channel::DisassociateSendChannel(int channel_id) { |
- CriticalSectionScoped lock(assoc_send_channel_lock_.get()); |
+ rtc::CritScope lock(&assoc_send_channel_lock_); |
Channel* channel = associate_send_channel_.channel(); |
if (channel && channel->ChannelId() == channel_id) { |
// If this channel is associated with a send channel of the specified |
@@ -3473,7 +3459,7 @@ int Channel::RegisterExternalMediaProcessing( |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::RegisterExternalMediaProcessing()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (kPlaybackPerChannel == type) |
{ |
@@ -3509,7 +3495,7 @@ int Channel::DeRegisterExternalMediaProcessing(ProcessingTypes type) |
WEBRTC_TRACE(kTraceInfo, kTraceVoice, VoEId(_instanceId,_channelId), |
"Channel::DeRegisterExternalMediaProcessing()"); |
- CriticalSectionScoped cs(&_callbackCritSect); |
+ rtc::CritScope cs(&_callbackCritSect); |
if (kPlaybackPerChannel == type) |
{ |
@@ -3571,7 +3557,7 @@ void Channel::GetDecodingCallStatistics(AudioDecodingCallStats* stats) const { |
bool Channel::GetDelayEstimate(int* jitter_buffer_delay_ms, |
int* playout_buffer_delay_ms) const { |
- CriticalSectionScoped cs(video_sync_lock_.get()); |
+ rtc::CritScope lock(&video_sync_lock_); |
if (_average_jitter_buffer_delay_us == 0) { |
return false; |
} |
@@ -3618,7 +3604,7 @@ Channel::SetMinimumPlayoutDelay(int delayMs) |
int Channel::GetPlayoutTimestamp(unsigned int& timestamp) { |
uint32_t playout_timestamp_rtp = 0; |
{ |
- CriticalSectionScoped cs(video_sync_lock_.get()); |
+ rtc::CritScope lock(&video_sync_lock_); |
playout_timestamp_rtp = playout_timestamp_rtp_; |
} |
if (playout_timestamp_rtp == 0) { |
@@ -3672,7 +3658,7 @@ Channel::MixOrReplaceAudioWithFile(int mixingFrequency) |
size_t fileSamples(0); |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_inputFilePlayerPtr == NULL) |
{ |
@@ -3742,7 +3728,7 @@ Channel::MixAudioWithFile(AudioFrame& audioFrame, |
size_t fileSamples(0); |
{ |
- CriticalSectionScoped cs(&_fileCritSect); |
+ rtc::CritScope cs(&_fileCritSect); |
if (_outputFilePlayerPtr == NULL) |
{ |
@@ -3891,7 +3877,7 @@ void Channel::UpdatePlayoutTimestamp(bool rtcp) { |
playout_timestamp); |
{ |
- CriticalSectionScoped cs(video_sync_lock_.get()); |
+ rtc::CritScope lock(&video_sync_lock_); |
if (rtcp) { |
playout_timestamp_rtcp_ = playout_timestamp; |
} else { |
@@ -3932,7 +3918,7 @@ void Channel::UpdatePacketDelay(uint32_t rtp_timestamp, |
if (timestamp_diff_ms == 0) return; |
{ |
- CriticalSectionScoped cs(video_sync_lock_.get()); |
+ rtc::CritScope lock(&video_sync_lock_); |
if (packet_delay_ms >= 10 && packet_delay_ms <= 60) { |
_recPacketDelayMs = packet_delay_ms; |
@@ -4076,7 +4062,7 @@ int64_t Channel::GetRTT(bool allow_associate_channel) const { |
int64_t rtt = 0; |
if (report_blocks.empty()) { |
if (allow_associate_channel) { |
- CriticalSectionScoped lock(assoc_send_channel_lock_.get()); |
+ rtc::CritScope lock(&assoc_send_channel_lock_); |
Channel* channel = associate_send_channel_.channel(); |
// Tries to get RTT from an associated channel. This is important for |
// receive-only channels. |