Index: talk/app/webrtc/mediastreamsignaling.cc |
diff --git a/talk/app/webrtc/mediastreamsignaling.cc b/talk/app/webrtc/mediastreamsignaling.cc |
index b405273902248a8060ef32422d2943496f9c20fa..c12471c778b8d321a3b8b66a5e6ac3f72d5f0b18 100644 |
--- a/talk/app/webrtc/mediastreamsignaling.cc |
+++ b/talk/app/webrtc/mediastreamsignaling.cc |
@@ -27,4 +27,1000 @@ |
#include "talk/app/webrtc/mediastreamsignaling.h" |
-// TODO(deadbeef): Remove this file once Chrome build files don't reference it. |
+#include <vector> |
+ |
+#include "talk/app/webrtc/audiotrack.h" |
+#include "talk/app/webrtc/mediaconstraintsinterface.h" |
+#include "talk/app/webrtc/mediastreamproxy.h" |
+#include "talk/app/webrtc/mediastreamtrackproxy.h" |
+#include "talk/app/webrtc/remoteaudiosource.h" |
+#include "talk/app/webrtc/remotevideocapturer.h" |
+#include "talk/app/webrtc/sctputils.h" |
+#include "talk/app/webrtc/videosource.h" |
+#include "talk/app/webrtc/videotrack.h" |
+#include "talk/media/sctp/sctpdataengine.h" |
+#include "webrtc/base/bytebuffer.h" |
+#include "webrtc/base/stringutils.h" |
+ |
+static const char kDefaultStreamLabel[] = "default"; |
+static const char kDefaultAudioTrackLabel[] = "defaulta0"; |
+static const char kDefaultVideoTrackLabel[] = "defaultv0"; |
+ |
+namespace webrtc { |
+ |
+using rtc::scoped_ptr; |
+using rtc::scoped_refptr; |
+ |
+static bool ParseConstraintsForAnswer( |
+ const MediaConstraintsInterface* constraints, |
+ cricket::MediaSessionOptions* options) { |
+ bool value = false; |
+ size_t mandatory_constraints_satisfied = 0; |
+ |
+ // kOfferToReceiveAudio defaults to true according to spec. |
+ if (!FindConstraint(constraints, |
+ MediaConstraintsInterface::kOfferToReceiveAudio, |
+ &value, &mandatory_constraints_satisfied) || value) { |
+ options->recv_audio = true; |
+ } |
+ |
+ // kOfferToReceiveVideo defaults to false according to spec. But |
+ // if it is an answer and video is offered, we should still accept video |
+ // per default. |
+ value = false; |
+ if (!FindConstraint(constraints, |
+ MediaConstraintsInterface::kOfferToReceiveVideo, |
+ &value, &mandatory_constraints_satisfied) || value) { |
+ options->recv_video = true; |
+ } |
+ |
+ if (FindConstraint(constraints, |
+ MediaConstraintsInterface::kVoiceActivityDetection, |
+ &value, &mandatory_constraints_satisfied)) { |
+ options->vad_enabled = value; |
+ } |
+ |
+ if (FindConstraint(constraints, |
+ MediaConstraintsInterface::kUseRtpMux, |
+ &value, &mandatory_constraints_satisfied)) { |
+ options->bundle_enabled = value; |
+ } else { |
+ // kUseRtpMux defaults to true according to spec. |
+ options->bundle_enabled = true; |
+ } |
+ if (FindConstraint(constraints, |
+ MediaConstraintsInterface::kIceRestart, |
+ &value, &mandatory_constraints_satisfied)) { |
+ options->transport_options.ice_restart = value; |
+ } else { |
+ // kIceRestart defaults to false according to spec. |
+ options->transport_options.ice_restart = false; |
+ } |
+ |
+ if (!constraints) { |
+ return true; |
+ } |
+ return mandatory_constraints_satisfied == constraints->GetMandatory().size(); |
+} |
+ |
+// Returns true if if at least one media content is present and |
+// |options.bundle_enabled| is true. |
+// Bundle will be enabled by default if at least one media content is present |
+// and the constraint kUseRtpMux has not disabled bundle. |
+static bool EvaluateNeedForBundle(const cricket::MediaSessionOptions& options) { |
+ return options.bundle_enabled && |
+ (options.has_audio() || options.has_video() || options.has_data()); |
+} |
+ |
+static bool MediaContentDirectionHasSend(cricket::MediaContentDirection dir) { |
+ return dir == cricket::MD_SENDONLY || dir == cricket::MD_SENDRECV; |
+} |
+ |
+static bool IsValidOfferToReceiveMedia(int value) { |
+ typedef PeerConnectionInterface::RTCOfferAnswerOptions Options; |
+ return (value >= Options::kUndefined) && |
+ (value <= Options::kMaxOfferToReceiveMedia); |
+} |
+ |
+// Add the stream and RTP data channel info to |session_options|. |
+static void SetStreams( |
+ cricket::MediaSessionOptions* session_options, |
+ rtc::scoped_refptr<StreamCollection> streams, |
+ const MediaStreamSignaling::RtpDataChannels& rtp_data_channels) { |
+ session_options->streams.clear(); |
+ if (streams != NULL) { |
+ for (size_t i = 0; i < streams->count(); ++i) { |
+ MediaStreamInterface* stream = streams->at(i); |
+ |
+ AudioTrackVector audio_tracks(stream->GetAudioTracks()); |
+ |
+ // For each audio track in the stream, add it to the MediaSessionOptions. |
+ for (size_t j = 0; j < audio_tracks.size(); ++j) { |
+ scoped_refptr<MediaStreamTrackInterface> track(audio_tracks[j]); |
+ session_options->AddSendStream( |
+ cricket::MEDIA_TYPE_AUDIO, track->id(), stream->label()); |
+ } |
+ |
+ VideoTrackVector video_tracks(stream->GetVideoTracks()); |
+ |
+ // For each video track in the stream, add it to the MediaSessionOptions. |
+ for (size_t j = 0; j < video_tracks.size(); ++j) { |
+ scoped_refptr<MediaStreamTrackInterface> track(video_tracks[j]); |
+ session_options->AddSendStream( |
+ cricket::MEDIA_TYPE_VIDEO, track->id(), stream->label()); |
+ } |
+ } |
+ } |
+ |
+ // Check for data channels. |
+ MediaStreamSignaling::RtpDataChannels::const_iterator data_channel_it = |
+ rtp_data_channels.begin(); |
+ for (; data_channel_it != rtp_data_channels.end(); ++data_channel_it) { |
+ const DataChannel* channel = data_channel_it->second; |
+ if (channel->state() == DataChannel::kConnecting || |
+ channel->state() == DataChannel::kOpen) { |
+ // |streamid| and |sync_label| are both set to the DataChannel label |
+ // here so they can be signaled the same way as MediaStreams and Tracks. |
+ // For MediaStreams, the sync_label is the MediaStream label and the |
+ // track label is the same as |streamid|. |
+ const std::string& streamid = channel->label(); |
+ const std::string& sync_label = channel->label(); |
+ session_options->AddSendStream( |
+ cricket::MEDIA_TYPE_DATA, streamid, sync_label); |
+ } |
+ } |
+} |
+ |
+// Factory class for creating remote MediaStreams and MediaStreamTracks. |
+class RemoteMediaStreamFactory { |
+ public: |
+ explicit RemoteMediaStreamFactory(rtc::Thread* signaling_thread, |
+ cricket::ChannelManager* channel_manager) |
+ : signaling_thread_(signaling_thread), |
+ channel_manager_(channel_manager) { |
+ } |
+ |
+ rtc::scoped_refptr<MediaStreamInterface> CreateMediaStream( |
+ const std::string& stream_label) { |
+ return MediaStreamProxy::Create( |
+ signaling_thread_, MediaStream::Create(stream_label)); |
+ } |
+ |
+ AudioTrackInterface* AddAudioTrack(webrtc::MediaStreamInterface* stream, |
+ const std::string& track_id) { |
+ return AddTrack<AudioTrackInterface, AudioTrack, AudioTrackProxy>( |
+ stream, track_id, RemoteAudioSource::Create().get()); |
+ } |
+ |
+ VideoTrackInterface* AddVideoTrack(webrtc::MediaStreamInterface* stream, |
+ const std::string& track_id) { |
+ return AddTrack<VideoTrackInterface, VideoTrack, VideoTrackProxy>( |
+ stream, track_id, VideoSource::Create(channel_manager_, |
+ new RemoteVideoCapturer(), |
+ NULL).get()); |
+ } |
+ |
+ private: |
+ template <typename TI, typename T, typename TP, typename S> |
+ TI* AddTrack(MediaStreamInterface* stream, const std::string& track_id, |
+ S* source) { |
+ rtc::scoped_refptr<TI> track( |
+ TP::Create(signaling_thread_, T::Create(track_id, source))); |
+ track->set_state(webrtc::MediaStreamTrackInterface::kLive); |
+ if (stream->AddTrack(track)) { |
+ return track; |
+ } |
+ return NULL; |
+ } |
+ |
+ rtc::Thread* signaling_thread_; |
+ cricket::ChannelManager* channel_manager_; |
+}; |
+ |
+MediaStreamSignaling::MediaStreamSignaling( |
+ rtc::Thread* signaling_thread, |
+ MediaStreamSignalingObserver* stream_observer, |
+ cricket::ChannelManager* channel_manager) |
+ : signaling_thread_(signaling_thread), |
+ data_channel_factory_(NULL), |
+ stream_observer_(stream_observer), |
+ local_streams_(StreamCollection::Create()), |
+ remote_streams_(StreamCollection::Create()), |
+ remote_stream_factory_(new RemoteMediaStreamFactory(signaling_thread, |
+ channel_manager)), |
+ last_allocated_sctp_even_sid_(-2), |
+ last_allocated_sctp_odd_sid_(-1) { |
+} |
+ |
+MediaStreamSignaling::~MediaStreamSignaling() { |
+} |
+ |
+void MediaStreamSignaling::TearDown() { |
+ OnAudioChannelClose(); |
+ OnVideoChannelClose(); |
+ OnDataChannelClose(); |
+} |
+ |
+bool MediaStreamSignaling::IsSctpSidAvailable(int sid) const { |
+ if (sid < 0 || sid > static_cast<int>(cricket::kMaxSctpSid)) |
+ return false; |
+ |
+ return FindDataChannelBySid(sid) < 0; |
+} |
+ |
+// Gets the first unused odd/even id based on the DTLS role. If |role| is |
+// SSL_CLIENT, the allocated id starts from 0 and takes even numbers; otherwise, |
+// the id starts from 1 and takes odd numbers. Returns false if no id can be |
+// allocated. |
+bool MediaStreamSignaling::AllocateSctpSid(rtc::SSLRole role, int* sid) { |
+ int& last_id = (role == rtc::SSL_CLIENT) ? |
+ last_allocated_sctp_even_sid_ : last_allocated_sctp_odd_sid_; |
+ |
+ do { |
+ last_id += 2; |
+ } while (last_id <= static_cast<int>(cricket::kMaxSctpSid) && |
+ !IsSctpSidAvailable(last_id)); |
+ |
+ if (last_id > static_cast<int>(cricket::kMaxSctpSid)) { |
+ return false; |
+ } |
+ |
+ *sid = last_id; |
+ return true; |
+} |
+ |
+bool MediaStreamSignaling::HasDataChannels() const { |
+ return !rtp_data_channels_.empty() || !sctp_data_channels_.empty(); |
+} |
+ |
+bool MediaStreamSignaling::AddDataChannel(DataChannel* data_channel) { |
+ ASSERT(data_channel != NULL); |
+ if (data_channel->data_channel_type() == cricket::DCT_RTP) { |
+ if (rtp_data_channels_.find(data_channel->label()) != |
+ rtp_data_channels_.end()) { |
+ LOG(LS_ERROR) << "DataChannel with label " << data_channel->label() |
+ << " already exists."; |
+ return false; |
+ } |
+ rtp_data_channels_[data_channel->label()] = data_channel; |
+ } else { |
+ ASSERT(data_channel->data_channel_type() == cricket::DCT_SCTP); |
+ sctp_data_channels_.push_back(data_channel); |
+ } |
+ return true; |
+} |
+ |
+bool MediaStreamSignaling::AddDataChannelFromOpenMessage( |
+ const cricket::ReceiveDataParams& params, |
+ const rtc::Buffer& payload) { |
+ if (!data_channel_factory_) { |
+ LOG(LS_WARNING) << "Remote peer requested a DataChannel but DataChannels " |
+ << "are not supported."; |
+ return false; |
+ } |
+ |
+ std::string label; |
+ InternalDataChannelInit config; |
+ config.id = params.ssrc; |
+ if (!ParseDataChannelOpenMessage(payload, &label, &config)) { |
+ LOG(LS_WARNING) << "Failed to parse the OPEN message for sid " |
+ << params.ssrc; |
+ return false; |
+ } |
+ config.open_handshake_role = InternalDataChannelInit::kAcker; |
+ |
+ scoped_refptr<DataChannel> channel( |
+ data_channel_factory_->CreateDataChannel(label, &config)); |
+ if (!channel.get()) { |
+ LOG(LS_ERROR) << "Failed to create DataChannel from the OPEN message."; |
+ return false; |
+ } |
+ |
+ stream_observer_->OnAddDataChannel(channel); |
+ return true; |
+} |
+ |
+void MediaStreamSignaling::RemoveSctpDataChannel(int sid) { |
+ ASSERT(sid >= 0); |
+ for (SctpDataChannels::iterator iter = sctp_data_channels_.begin(); |
+ iter != sctp_data_channels_.end(); |
+ ++iter) { |
+ if ((*iter)->id() == sid) { |
+ sctp_data_channels_.erase(iter); |
+ |
+ if (rtc::IsEven(sid) && sid <= last_allocated_sctp_even_sid_) { |
+ last_allocated_sctp_even_sid_ = sid - 2; |
+ } else if (rtc::IsOdd(sid) && sid <= last_allocated_sctp_odd_sid_) { |
+ last_allocated_sctp_odd_sid_ = sid - 2; |
+ } |
+ return; |
+ } |
+ } |
+} |
+ |
+bool MediaStreamSignaling::AddLocalStream(MediaStreamInterface* local_stream) { |
+ if (local_streams_->find(local_stream->label()) != NULL) { |
+ LOG(LS_WARNING) << "MediaStream with label " << local_stream->label() |
+ << "already exist."; |
+ return false; |
+ } |
+ local_streams_->AddStream(local_stream); |
+ |
+ // Find tracks that has already been configured in SDP. This can occur if a |
+ // local session description that contains the MSID of these tracks is set |
+ // before AddLocalStream is called. It can also occur if the local session |
+ // description is not changed and RemoveLocalStream |
+ // is called and later AddLocalStream is called again with the same stream. |
+ AudioTrackVector audio_tracks = local_stream->GetAudioTracks(); |
+ for (AudioTrackVector::const_iterator it = audio_tracks.begin(); |
+ it != audio_tracks.end(); ++it) { |
+ const TrackInfo* track_info = FindTrackInfo(local_audio_tracks_, |
+ local_stream->label(), |
+ (*it)->id()); |
+ if (track_info) { |
+ OnLocalTrackSeen(track_info->stream_label, track_info->track_id, |
+ track_info->ssrc, cricket::MEDIA_TYPE_AUDIO); |
+ } |
+ } |
+ |
+ VideoTrackVector video_tracks = local_stream->GetVideoTracks(); |
+ for (VideoTrackVector::const_iterator it = video_tracks.begin(); |
+ it != video_tracks.end(); ++it) { |
+ const TrackInfo* track_info = FindTrackInfo(local_video_tracks_, |
+ local_stream->label(), |
+ (*it)->id()); |
+ if (track_info) { |
+ OnLocalTrackSeen(track_info->stream_label, track_info->track_id, |
+ track_info->ssrc, cricket::MEDIA_TYPE_VIDEO); |
+ } |
+ } |
+ return true; |
+} |
+ |
+void MediaStreamSignaling::RemoveLocalStream( |
+ MediaStreamInterface* local_stream) { |
+ AudioTrackVector audio_tracks = local_stream->GetAudioTracks(); |
+ for (AudioTrackVector::const_iterator it = audio_tracks.begin(); |
+ it != audio_tracks.end(); ++it) { |
+ const TrackInfo* track_info = FindTrackInfo(local_audio_tracks_, |
+ local_stream->label(), |
+ (*it)->id()); |
+ if (track_info) { |
+ stream_observer_->OnRemoveLocalAudioTrack(local_stream, *it, |
+ track_info->ssrc); |
+ } |
+ } |
+ VideoTrackVector video_tracks = local_stream->GetVideoTracks(); |
+ for (VideoTrackVector::const_iterator it = video_tracks.begin(); |
+ it != video_tracks.end(); ++it) { |
+ const TrackInfo* track_info = FindTrackInfo(local_video_tracks_, |
+ local_stream->label(), |
+ (*it)->id()); |
+ if (track_info) { |
+ stream_observer_->OnRemoveLocalVideoTrack(local_stream, *it); |
+ } |
+ } |
+ |
+ local_streams_->RemoveStream(local_stream); |
+ stream_observer_->OnRemoveLocalStream(local_stream); |
+} |
+ |
+bool MediaStreamSignaling::GetOptionsForOffer( |
+ const PeerConnectionInterface::RTCOfferAnswerOptions& rtc_options, |
+ cricket::MediaSessionOptions* session_options) { |
+ typedef PeerConnectionInterface::RTCOfferAnswerOptions RTCOfferAnswerOptions; |
+ if (!IsValidOfferToReceiveMedia(rtc_options.offer_to_receive_audio) || |
+ !IsValidOfferToReceiveMedia(rtc_options.offer_to_receive_video)) { |
+ return false; |
+ } |
+ |
+ SetStreams(session_options, local_streams_, rtp_data_channels_); |
+ |
+ // According to the spec, offer to receive audio/video if the constraint is |
+ // not set and there are send streams. |
+ if (rtc_options.offer_to_receive_audio == RTCOfferAnswerOptions::kUndefined) { |
+ session_options->recv_audio = |
+ session_options->HasSendMediaStream(cricket::MEDIA_TYPE_AUDIO); |
+ } else { |
+ session_options->recv_audio = (rtc_options.offer_to_receive_audio > 0); |
+ } |
+ if (rtc_options.offer_to_receive_video == RTCOfferAnswerOptions::kUndefined) { |
+ session_options->recv_video = |
+ session_options->HasSendMediaStream(cricket::MEDIA_TYPE_VIDEO); |
+ } else { |
+ session_options->recv_video = (rtc_options.offer_to_receive_video > 0); |
+ } |
+ |
+ session_options->vad_enabled = rtc_options.voice_activity_detection; |
+ session_options->transport_options.ice_restart = rtc_options.ice_restart; |
+ session_options->bundle_enabled = rtc_options.use_rtp_mux; |
+ |
+ session_options->bundle_enabled = EvaluateNeedForBundle(*session_options); |
+ return true; |
+} |
+ |
+bool MediaStreamSignaling::GetOptionsForAnswer( |
+ const MediaConstraintsInterface* constraints, |
+ cricket::MediaSessionOptions* options) { |
+ SetStreams(options, local_streams_, rtp_data_channels_); |
+ |
+ options->recv_audio = false; |
+ options->recv_video = false; |
+ if (!ParseConstraintsForAnswer(constraints, options)) { |
+ return false; |
+ } |
+ options->bundle_enabled = EvaluateNeedForBundle(*options); |
+ return true; |
+} |
+ |
+// Updates or creates remote MediaStream objects given a |
+// remote SessionDesription. |
+// If the remote SessionDesription contains new remote MediaStreams |
+// the observer OnAddStream method is called. If a remote MediaStream is missing |
+// from the remote SessionDescription OnRemoveStream is called. |
+void MediaStreamSignaling::OnRemoteDescriptionChanged( |
+ const SessionDescriptionInterface* desc) { |
+ const cricket::SessionDescription* remote_desc = desc->description(); |
+ rtc::scoped_refptr<StreamCollection> new_streams( |
+ StreamCollection::Create()); |
+ |
+ // Find all audio rtp streams and create corresponding remote AudioTracks |
+ // and MediaStreams. |
+ const cricket::ContentInfo* audio_content = GetFirstAudioContent(remote_desc); |
+ if (audio_content) { |
+ const cricket::AudioContentDescription* desc = |
+ static_cast<const cricket::AudioContentDescription*>( |
+ audio_content->description); |
+ UpdateRemoteStreamsList(desc->streams(), desc->type(), new_streams); |
+ remote_info_.default_audio_track_needed = |
+ MediaContentDirectionHasSend(desc->direction()) && |
+ desc->streams().empty(); |
+ } |
+ |
+ // Find all video rtp streams and create corresponding remote VideoTracks |
+ // and MediaStreams. |
+ const cricket::ContentInfo* video_content = GetFirstVideoContent(remote_desc); |
+ if (video_content) { |
+ const cricket::VideoContentDescription* desc = |
+ static_cast<const cricket::VideoContentDescription*>( |
+ video_content->description); |
+ UpdateRemoteStreamsList(desc->streams(), desc->type(), new_streams); |
+ remote_info_.default_video_track_needed = |
+ MediaContentDirectionHasSend(desc->direction()) && |
+ desc->streams().empty(); |
+ } |
+ |
+ // Update the DataChannels with the information from the remote peer. |
+ const cricket::ContentInfo* data_content = GetFirstDataContent(remote_desc); |
+ if (data_content) { |
+ const cricket::DataContentDescription* data_desc = |
+ static_cast<const cricket::DataContentDescription*>( |
+ data_content->description); |
+ if (rtc::starts_with( |
+ data_desc->protocol().data(), cricket::kMediaProtocolRtpPrefix)) { |
+ UpdateRemoteRtpDataChannels(data_desc->streams()); |
+ } |
+ } |
+ |
+ // Iterate new_streams and notify the observer about new MediaStreams. |
+ for (size_t i = 0; i < new_streams->count(); ++i) { |
+ MediaStreamInterface* new_stream = new_streams->at(i); |
+ stream_observer_->OnAddRemoteStream(new_stream); |
+ } |
+ |
+ // Find removed MediaStreams. |
+ if (remote_info_.IsDefaultMediaStreamNeeded() && |
+ remote_streams_->find(kDefaultStreamLabel) != NULL) { |
+ // The default media stream already exists. No need to do anything. |
+ } else { |
+ UpdateEndedRemoteMediaStreams(); |
+ remote_info_.msid_supported |= remote_streams_->count() > 0; |
+ } |
+ MaybeCreateDefaultStream(); |
+} |
+ |
+void MediaStreamSignaling::OnLocalDescriptionChanged( |
+ const SessionDescriptionInterface* desc) { |
+ const cricket::ContentInfo* audio_content = |
+ GetFirstAudioContent(desc->description()); |
+ if (audio_content) { |
+ if (audio_content->rejected) { |
+ RejectRemoteTracks(cricket::MEDIA_TYPE_AUDIO); |
+ } |
+ const cricket::AudioContentDescription* audio_desc = |
+ static_cast<const cricket::AudioContentDescription*>( |
+ audio_content->description); |
+ UpdateLocalTracks(audio_desc->streams(), audio_desc->type()); |
+ } |
+ |
+ const cricket::ContentInfo* video_content = |
+ GetFirstVideoContent(desc->description()); |
+ if (video_content) { |
+ if (video_content->rejected) { |
+ RejectRemoteTracks(cricket::MEDIA_TYPE_VIDEO); |
+ } |
+ const cricket::VideoContentDescription* video_desc = |
+ static_cast<const cricket::VideoContentDescription*>( |
+ video_content->description); |
+ UpdateLocalTracks(video_desc->streams(), video_desc->type()); |
+ } |
+ |
+ const cricket::ContentInfo* data_content = |
+ GetFirstDataContent(desc->description()); |
+ if (data_content) { |
+ const cricket::DataContentDescription* data_desc = |
+ static_cast<const cricket::DataContentDescription*>( |
+ data_content->description); |
+ if (rtc::starts_with( |
+ data_desc->protocol().data(), cricket::kMediaProtocolRtpPrefix)) { |
+ UpdateLocalRtpDataChannels(data_desc->streams()); |
+ } |
+ } |
+} |
+ |
+void MediaStreamSignaling::OnAudioChannelClose() { |
+ RejectRemoteTracks(cricket::MEDIA_TYPE_AUDIO); |
+} |
+ |
+void MediaStreamSignaling::OnVideoChannelClose() { |
+ RejectRemoteTracks(cricket::MEDIA_TYPE_VIDEO); |
+} |
+ |
+void MediaStreamSignaling::OnDataChannelClose() { |
+ // Use a temporary copy of the RTP/SCTP DataChannel list because the |
+ // DataChannel may callback to us and try to modify the list. |
+ RtpDataChannels temp_rtp_dcs; |
+ temp_rtp_dcs.swap(rtp_data_channels_); |
+ RtpDataChannels::iterator it1 = temp_rtp_dcs.begin(); |
+ for (; it1 != temp_rtp_dcs.end(); ++it1) { |
+ it1->second->OnDataEngineClose(); |
+ } |
+ |
+ SctpDataChannels temp_sctp_dcs; |
+ temp_sctp_dcs.swap(sctp_data_channels_); |
+ SctpDataChannels::iterator it2 = temp_sctp_dcs.begin(); |
+ for (; it2 != temp_sctp_dcs.end(); ++it2) { |
+ (*it2)->OnDataEngineClose(); |
+ } |
+} |
+ |
+void MediaStreamSignaling::UpdateRemoteStreamsList( |
+ const cricket::StreamParamsVec& streams, |
+ cricket::MediaType media_type, |
+ StreamCollection* new_streams) { |
+ TrackInfos* current_tracks = GetRemoteTracks(media_type); |
+ |
+ // Find removed tracks. Ie tracks where the track id or ssrc don't match the |
+ // new StreamParam. |
+ TrackInfos::iterator track_it = current_tracks->begin(); |
+ while (track_it != current_tracks->end()) { |
+ const TrackInfo& info = *track_it; |
+ const cricket::StreamParams* params = |
+ cricket::GetStreamBySsrc(streams, info.ssrc); |
+ if (!params || params->id != info.track_id) { |
+ OnRemoteTrackRemoved(info.stream_label, info.track_id, media_type); |
+ track_it = current_tracks->erase(track_it); |
+ } else { |
+ ++track_it; |
+ } |
+ } |
+ |
+ // Find new and active tracks. |
+ for (cricket::StreamParamsVec::const_iterator it = streams.begin(); |
+ it != streams.end(); ++it) { |
+ // The sync_label is the MediaStream label and the |stream.id| is the |
+ // track id. |
+ const std::string& stream_label = it->sync_label; |
+ const std::string& track_id = it->id; |
+ uint32_t ssrc = it->first_ssrc(); |
+ |
+ rtc::scoped_refptr<MediaStreamInterface> stream = |
+ remote_streams_->find(stream_label); |
+ if (!stream) { |
+ // This is a new MediaStream. Create a new remote MediaStream. |
+ stream = remote_stream_factory_->CreateMediaStream(stream_label); |
+ remote_streams_->AddStream(stream); |
+ new_streams->AddStream(stream); |
+ } |
+ |
+ const TrackInfo* track_info = FindTrackInfo(*current_tracks, stream_label, |
+ track_id); |
+ if (!track_info) { |
+ current_tracks->push_back(TrackInfo(stream_label, track_id, ssrc)); |
+ OnRemoteTrackSeen(stream_label, track_id, it->first_ssrc(), media_type); |
+ } |
+ } |
+} |
+ |
+void MediaStreamSignaling::OnRemoteTrackSeen(const std::string& stream_label, |
+ const std::string& track_id, |
+ uint32_t ssrc, |
+ cricket::MediaType media_type) { |
+ MediaStreamInterface* stream = remote_streams_->find(stream_label); |
+ |
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) { |
+ AudioTrackInterface* audio_track = |
+ remote_stream_factory_->AddAudioTrack(stream, track_id); |
+ stream_observer_->OnAddRemoteAudioTrack(stream, audio_track, ssrc); |
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { |
+ VideoTrackInterface* video_track = |
+ remote_stream_factory_->AddVideoTrack(stream, track_id); |
+ stream_observer_->OnAddRemoteVideoTrack(stream, video_track, ssrc); |
+ } else { |
+ ASSERT(false && "Invalid media type"); |
+ } |
+} |
+ |
+void MediaStreamSignaling::OnRemoteTrackRemoved( |
+ const std::string& stream_label, |
+ const std::string& track_id, |
+ cricket::MediaType media_type) { |
+ MediaStreamInterface* stream = remote_streams_->find(stream_label); |
+ |
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) { |
+ rtc::scoped_refptr<AudioTrackInterface> audio_track = |
+ stream->FindAudioTrack(track_id); |
+ if (audio_track) { |
+ audio_track->set_state(webrtc::MediaStreamTrackInterface::kEnded); |
+ stream->RemoveTrack(audio_track); |
+ stream_observer_->OnRemoveRemoteAudioTrack(stream, audio_track); |
+ } |
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { |
+ rtc::scoped_refptr<VideoTrackInterface> video_track = |
+ stream->FindVideoTrack(track_id); |
+ if (video_track) { |
+ video_track->set_state(webrtc::MediaStreamTrackInterface::kEnded); |
+ stream->RemoveTrack(video_track); |
+ stream_observer_->OnRemoveRemoteVideoTrack(stream, video_track); |
+ } |
+ } else { |
+ ASSERT(false && "Invalid media type"); |
+ } |
+} |
+ |
+void MediaStreamSignaling::RejectRemoteTracks(cricket::MediaType media_type) { |
+ TrackInfos* current_tracks = GetRemoteTracks(media_type); |
+ for (TrackInfos::iterator track_it = current_tracks->begin(); |
+ track_it != current_tracks->end(); ++track_it) { |
+ const TrackInfo& info = *track_it; |
+ MediaStreamInterface* stream = remote_streams_->find(info.stream_label); |
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) { |
+ AudioTrackInterface* track = stream->FindAudioTrack(info.track_id); |
+ // There's no guarantee the track is still available, e.g. the track may |
+ // have been removed from the stream by javascript. |
+ if (track) { |
+ track->set_state(webrtc::MediaStreamTrackInterface::kEnded); |
+ } |
+ } |
+ if (media_type == cricket::MEDIA_TYPE_VIDEO) { |
+ VideoTrackInterface* track = stream->FindVideoTrack(info.track_id); |
+ // There's no guarantee the track is still available, e.g. the track may |
+ // have been removed from the stream by javascript. |
+ if (track) { |
+ track->set_state(webrtc::MediaStreamTrackInterface::kEnded); |
+ } |
+ } |
+ } |
+} |
+ |
+void MediaStreamSignaling::UpdateEndedRemoteMediaStreams() { |
+ std::vector<scoped_refptr<MediaStreamInterface> > streams_to_remove; |
+ for (size_t i = 0; i < remote_streams_->count(); ++i) { |
+ MediaStreamInterface*stream = remote_streams_->at(i); |
+ if (stream->GetAudioTracks().empty() && stream->GetVideoTracks().empty()) { |
+ streams_to_remove.push_back(stream); |
+ } |
+ } |
+ |
+ std::vector<scoped_refptr<MediaStreamInterface> >::const_iterator it; |
+ for (it = streams_to_remove.begin(); it != streams_to_remove.end(); ++it) { |
+ remote_streams_->RemoveStream(*it); |
+ stream_observer_->OnRemoveRemoteStream(*it); |
+ } |
+} |
+ |
+void MediaStreamSignaling::MaybeCreateDefaultStream() { |
+ if (!remote_info_.IsDefaultMediaStreamNeeded()) |
+ return; |
+ |
+ bool default_created = false; |
+ |
+ scoped_refptr<MediaStreamInterface> default_remote_stream = |
+ remote_streams_->find(kDefaultStreamLabel); |
+ if (default_remote_stream == NULL) { |
+ default_created = true; |
+ default_remote_stream = |
+ remote_stream_factory_->CreateMediaStream(kDefaultStreamLabel); |
+ remote_streams_->AddStream(default_remote_stream); |
+ } |
+ if (remote_info_.default_audio_track_needed && |
+ default_remote_stream->GetAudioTracks().size() == 0) { |
+ remote_audio_tracks_.push_back(TrackInfo(kDefaultStreamLabel, |
+ kDefaultAudioTrackLabel, 0)); |
+ |
+ OnRemoteTrackSeen(kDefaultStreamLabel, kDefaultAudioTrackLabel, 0, |
+ cricket::MEDIA_TYPE_AUDIO); |
+ } |
+ if (remote_info_.default_video_track_needed && |
+ default_remote_stream->GetVideoTracks().size() == 0) { |
+ remote_video_tracks_.push_back(TrackInfo(kDefaultStreamLabel, |
+ kDefaultVideoTrackLabel, 0)); |
+ OnRemoteTrackSeen(kDefaultStreamLabel, kDefaultVideoTrackLabel, 0, |
+ cricket::MEDIA_TYPE_VIDEO); |
+ } |
+ if (default_created) { |
+ stream_observer_->OnAddRemoteStream(default_remote_stream); |
+ } |
+} |
+ |
+MediaStreamSignaling::TrackInfos* MediaStreamSignaling::GetRemoteTracks( |
+ cricket::MediaType type) { |
+ if (type == cricket::MEDIA_TYPE_AUDIO) |
+ return &remote_audio_tracks_; |
+ else if (type == cricket::MEDIA_TYPE_VIDEO) |
+ return &remote_video_tracks_; |
+ ASSERT(false && "Unknown MediaType"); |
+ return NULL; |
+} |
+ |
+MediaStreamSignaling::TrackInfos* MediaStreamSignaling::GetLocalTracks( |
+ cricket::MediaType media_type) { |
+ ASSERT(media_type == cricket::MEDIA_TYPE_AUDIO || |
+ media_type == cricket::MEDIA_TYPE_VIDEO); |
+ |
+ return (media_type == cricket::MEDIA_TYPE_AUDIO) ? |
+ &local_audio_tracks_ : &local_video_tracks_; |
+} |
+ |
+void MediaStreamSignaling::UpdateLocalTracks( |
+ const std::vector<cricket::StreamParams>& streams, |
+ cricket::MediaType media_type) { |
+ TrackInfos* current_tracks = GetLocalTracks(media_type); |
+ |
+ // Find removed tracks. Ie tracks where the track id, stream label or ssrc |
+ // don't match the new StreamParam. |
+ TrackInfos::iterator track_it = current_tracks->begin(); |
+ while (track_it != current_tracks->end()) { |
+ const TrackInfo& info = *track_it; |
+ const cricket::StreamParams* params = |
+ cricket::GetStreamBySsrc(streams, info.ssrc); |
+ if (!params || params->id != info.track_id || |
+ params->sync_label != info.stream_label) { |
+ OnLocalTrackRemoved(info.stream_label, info.track_id, info.ssrc, |
+ media_type); |
+ track_it = current_tracks->erase(track_it); |
+ } else { |
+ ++track_it; |
+ } |
+ } |
+ |
+ // Find new and active tracks. |
+ for (cricket::StreamParamsVec::const_iterator it = streams.begin(); |
+ it != streams.end(); ++it) { |
+ // The sync_label is the MediaStream label and the |stream.id| is the |
+ // track id. |
+ const std::string& stream_label = it->sync_label; |
+ const std::string& track_id = it->id; |
+ uint32_t ssrc = it->first_ssrc(); |
+ const TrackInfo* track_info = FindTrackInfo(*current_tracks, |
+ stream_label, |
+ track_id); |
+ if (!track_info) { |
+ current_tracks->push_back(TrackInfo(stream_label, track_id, ssrc)); |
+ OnLocalTrackSeen(stream_label, track_id, it->first_ssrc(), media_type); |
+ } |
+ } |
+} |
+ |
+void MediaStreamSignaling::OnLocalTrackSeen(const std::string& stream_label, |
+ const std::string& track_id, |
+ uint32_t ssrc, |
+ cricket::MediaType media_type) { |
+ MediaStreamInterface* stream = local_streams_->find(stream_label); |
+ if (!stream) { |
+ LOG(LS_WARNING) << "An unknown local MediaStream with label " |
+ << stream_label << " has been configured."; |
+ return; |
+ } |
+ |
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) { |
+ AudioTrackInterface* audio_track = stream->FindAudioTrack(track_id); |
+ if (!audio_track) { |
+ LOG(LS_WARNING) << "An unknown local AudioTrack with id , " |
+ << track_id << " has been configured."; |
+ return; |
+ } |
+ stream_observer_->OnAddLocalAudioTrack(stream, audio_track, ssrc); |
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { |
+ VideoTrackInterface* video_track = stream->FindVideoTrack(track_id); |
+ if (!video_track) { |
+ LOG(LS_WARNING) << "An unknown local VideoTrack with id , " |
+ << track_id << " has been configured."; |
+ return; |
+ } |
+ stream_observer_->OnAddLocalVideoTrack(stream, video_track, ssrc); |
+ } else { |
+ ASSERT(false && "Invalid media type"); |
+ } |
+} |
+ |
+void MediaStreamSignaling::OnLocalTrackRemoved(const std::string& stream_label, |
+ const std::string& track_id, |
+ uint32_t ssrc, |
+ cricket::MediaType media_type) { |
+ MediaStreamInterface* stream = local_streams_->find(stream_label); |
+ if (!stream) { |
+ // This is the normal case. Ie RemoveLocalStream has been called and the |
+ // SessionDescriptions has been renegotiated. |
+ return; |
+ } |
+ // A track has been removed from the SessionDescription but the MediaStream |
+ // is still associated with MediaStreamSignaling. This only occurs if the SDP |
+ // doesn't match with the calls to AddLocalStream and RemoveLocalStream. |
+ |
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) { |
+ AudioTrackInterface* audio_track = stream->FindAudioTrack(track_id); |
+ if (!audio_track) { |
+ return; |
+ } |
+ stream_observer_->OnRemoveLocalAudioTrack(stream, audio_track, ssrc); |
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) { |
+ VideoTrackInterface* video_track = stream->FindVideoTrack(track_id); |
+ if (!video_track) { |
+ return; |
+ } |
+ stream_observer_->OnRemoveLocalVideoTrack(stream, video_track); |
+ } else { |
+ ASSERT(false && "Invalid media type."); |
+ } |
+} |
+ |
+void MediaStreamSignaling::UpdateLocalRtpDataChannels( |
+ const cricket::StreamParamsVec& streams) { |
+ std::vector<std::string> existing_channels; |
+ |
+ // Find new and active data channels. |
+ for (cricket::StreamParamsVec::const_iterator it =streams.begin(); |
+ it != streams.end(); ++it) { |
+ // |it->sync_label| is actually the data channel label. The reason is that |
+ // we use the same naming of data channels as we do for |
+ // MediaStreams and Tracks. |
+ // For MediaStreams, the sync_label is the MediaStream label and the |
+ // track label is the same as |streamid|. |
+ const std::string& channel_label = it->sync_label; |
+ RtpDataChannels::iterator data_channel_it = |
+ rtp_data_channels_.find(channel_label); |
+ if (!VERIFY(data_channel_it != rtp_data_channels_.end())) { |
+ continue; |
+ } |
+ // Set the SSRC the data channel should use for sending. |
+ data_channel_it->second->SetSendSsrc(it->first_ssrc()); |
+ existing_channels.push_back(data_channel_it->first); |
+ } |
+ |
+ UpdateClosingDataChannels(existing_channels, true); |
+} |
+ |
+void MediaStreamSignaling::UpdateRemoteRtpDataChannels( |
+ const cricket::StreamParamsVec& streams) { |
+ std::vector<std::string> existing_channels; |
+ |
+ // Find new and active data channels. |
+ for (cricket::StreamParamsVec::const_iterator it = streams.begin(); |
+ it != streams.end(); ++it) { |
+ // The data channel label is either the mslabel or the SSRC if the mslabel |
+ // does not exist. Ex a=ssrc:444330170 mslabel:test1. |
+ std::string label = it->sync_label.empty() ? |
+ rtc::ToString(it->first_ssrc()) : it->sync_label; |
+ RtpDataChannels::iterator data_channel_it = |
+ rtp_data_channels_.find(label); |
+ if (data_channel_it == rtp_data_channels_.end()) { |
+ // This is a new data channel. |
+ CreateRemoteDataChannel(label, it->first_ssrc()); |
+ } else { |
+ data_channel_it->second->SetReceiveSsrc(it->first_ssrc()); |
+ } |
+ existing_channels.push_back(label); |
+ } |
+ |
+ UpdateClosingDataChannels(existing_channels, false); |
+} |
+ |
+void MediaStreamSignaling::UpdateClosingDataChannels( |
+ const std::vector<std::string>& active_channels, bool is_local_update) { |
+ RtpDataChannels::iterator it = rtp_data_channels_.begin(); |
+ while (it != rtp_data_channels_.end()) { |
+ DataChannel* data_channel = it->second; |
+ if (std::find(active_channels.begin(), active_channels.end(), |
+ data_channel->label()) != active_channels.end()) { |
+ ++it; |
+ continue; |
+ } |
+ |
+ if (is_local_update) |
+ data_channel->SetSendSsrc(0); |
+ else |
+ data_channel->RemotePeerRequestClose(); |
+ |
+ if (data_channel->state() == DataChannel::kClosed) { |
+ rtp_data_channels_.erase(it); |
+ it = rtp_data_channels_.begin(); |
+ } else { |
+ ++it; |
+ } |
+ } |
+} |
+ |
+void MediaStreamSignaling::CreateRemoteDataChannel(const std::string& label, |
+ uint32_t remote_ssrc) { |
+ if (!data_channel_factory_) { |
+ LOG(LS_WARNING) << "Remote peer requested a DataChannel but DataChannels " |
+ << "are not supported."; |
+ return; |
+ } |
+ scoped_refptr<DataChannel> channel( |
+ data_channel_factory_->CreateDataChannel(label, NULL)); |
+ if (!channel.get()) { |
+ LOG(LS_WARNING) << "Remote peer requested a DataChannel but" |
+ << "CreateDataChannel failed."; |
+ return; |
+ } |
+ channel->SetReceiveSsrc(remote_ssrc); |
+ stream_observer_->OnAddDataChannel(channel); |
+} |
+ |
+void MediaStreamSignaling::OnDataTransportCreatedForSctp() { |
+ SctpDataChannels::iterator it = sctp_data_channels_.begin(); |
+ for (; it != sctp_data_channels_.end(); ++it) { |
+ (*it)->OnTransportChannelCreated(); |
+ } |
+} |
+ |
+void MediaStreamSignaling::OnDtlsRoleReadyForSctp(rtc::SSLRole role) { |
+ SctpDataChannels::iterator it = sctp_data_channels_.begin(); |
+ for (; it != sctp_data_channels_.end(); ++it) { |
+ if ((*it)->id() < 0) { |
+ int sid; |
+ if (!AllocateSctpSid(role, &sid)) { |
+ LOG(LS_ERROR) << "Failed to allocate SCTP sid."; |
+ continue; |
+ } |
+ (*it)->SetSctpSid(sid); |
+ } |
+ } |
+} |
+ |
+void MediaStreamSignaling::OnRemoteSctpDataChannelClosed(uint32_t sid) { |
+ int index = FindDataChannelBySid(sid); |
+ if (index < 0) { |
+ LOG(LS_WARNING) << "Unexpected sid " << sid |
+ << " of the remotely closed DataChannel."; |
+ return; |
+ } |
+ sctp_data_channels_[index]->Close(); |
+} |
+ |
+const MediaStreamSignaling::TrackInfo* |
+MediaStreamSignaling::FindTrackInfo( |
+ const MediaStreamSignaling::TrackInfos& infos, |
+ const std::string& stream_label, |
+ const std::string track_id) const { |
+ |
+ for (TrackInfos::const_iterator it = infos.begin(); |
+ it != infos.end(); ++it) { |
+ if (it->stream_label == stream_label && it->track_id == track_id) |
+ return &*it; |
+ } |
+ return NULL; |
+} |
+ |
+int MediaStreamSignaling::FindDataChannelBySid(int sid) const { |
+ for (size_t i = 0; i < sctp_data_channels_.size(); ++i) { |
+ if (sctp_data_channels_[i]->id() == sid) { |
+ return static_cast<int>(i); |
+ } |
+ } |
+ return -1; |
+} |
+ |
+} // namespace webrtc |