Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(94)

Unified Diff: talk/app/webrtc/peerconnection.cc

Issue 1393563002: Moving MediaStreamSignaling logic into PeerConnection. (Closed) Base URL: https://chromium.googlesource.com/external/webrtc.git@master
Patch Set: Fixing copyright header Created 5 years, 2 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « talk/app/webrtc/peerconnection.h ('k') | talk/app/webrtc/peerconnectionendtoend_unittest.cc » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: talk/app/webrtc/peerconnection.cc
diff --git a/talk/app/webrtc/peerconnection.cc b/talk/app/webrtc/peerconnection.cc
index 86902b03591277b91d4a6106b38e7caf8e04d495..44a231e8c5eb0856ddd81772df4e88018403bf4f 100644
--- a/talk/app/webrtc/peerconnection.cc
+++ b/talk/app/webrtc/peerconnection.cc
@@ -30,22 +30,36 @@
#include <vector>
#include <cctype> // for isdigit
+#include "talk/app/webrtc/audiotrack.h"
#include "talk/app/webrtc/dtmfsender.h"
#include "talk/app/webrtc/jsepicecandidate.h"
#include "talk/app/webrtc/jsepsessiondescription.h"
#include "talk/app/webrtc/mediaconstraintsinterface.h"
+#include "talk/app/webrtc/mediastream.h"
+#include "talk/app/webrtc/mediastreamproxy.h"
+#include "talk/app/webrtc/mediastreamtrackproxy.h"
+#include "talk/app/webrtc/remoteaudiosource.h"
+#include "talk/app/webrtc/remotevideocapturer.h"
#include "talk/app/webrtc/rtpreceiver.h"
#include "talk/app/webrtc/rtpsender.h"
#include "talk/app/webrtc/streamcollection.h"
+#include "talk/app/webrtc/videosource.h"
+#include "talk/app/webrtc/videotrack.h"
+#include "talk/media/sctp/sctpdataengine.h"
#include "webrtc/p2p/client/basicportallocator.h"
#include "talk/session/media/channelmanager.h"
#include "webrtc/base/logging.h"
#include "webrtc/base/stringencode.h"
+#include "webrtc/base/stringutils.h"
#include "webrtc/system_wrappers/interface/field_trial.h"
namespace {
+using webrtc::DataChannel;
+using webrtc::MediaConstraintsInterface;
+using webrtc::MediaStreamInterface;
using webrtc::PeerConnectionInterface;
+using webrtc::StreamCollection;
using webrtc::StunConfigurations;
using webrtc::TurnConfigurations;
typedef webrtc::PortAllocatorFactoryInterface::StunConfiguration
@@ -53,6 +67,10 @@ typedef webrtc::PortAllocatorFactoryInterface::StunConfiguration
typedef webrtc::PortAllocatorFactoryInterface::TurnConfiguration
TurnConfiguration;
+static const char kDefaultStreamLabel[] = "default";
+static const char kDefaultAudioTrackLabel[] = "defaulta0";
+static const char kDefaultVideoTrackLabel[] = "defaultv0";
+
// The min number of tokens must present in Turn host uri.
// e.g. user@turn.example.org
static const size_t kTurnHostTokensNum = 2;
@@ -84,6 +102,7 @@ static_assert(INVALID == ARRAY_SIZE(kValidIceServiceTypes),
enum {
MSG_SET_SESSIONDESCRIPTION_SUCCESS = 0,
MSG_SET_SESSIONDESCRIPTION_FAILED,
+ MSG_CREATE_SESSIONDESCRIPTION_FAILED,
MSG_GETSTATS,
};
@@ -97,6 +116,15 @@ struct SetSessionDescriptionMsg : public rtc::MessageData {
std::string error;
};
+struct CreateSessionDescriptionMsg : public rtc::MessageData {
+ explicit CreateSessionDescriptionMsg(
+ webrtc::CreateSessionDescriptionObserver* observer)
+ : observer(observer) {}
+
+ rtc::scoped_refptr<webrtc::CreateSessionDescriptionObserver> observer;
+ std::string error;
+};
+
struct GetStatsMsg : public rtc::MessageData {
GetStatsMsg(webrtc::StatsObserver* observer,
webrtc::MediaStreamTrackInterface* track)
@@ -302,10 +330,210 @@ bool ParseIceServerUrl(const PeerConnectionInterface::IceServer& server,
return true;
}
+// Check if we can send |new_stream| on a PeerConnection.
+bool CanAddLocalMediaStream(webrtc::StreamCollectionInterface* current_streams,
+ webrtc::MediaStreamInterface* new_stream) {
+ if (!new_stream || !current_streams) {
+ return false;
+ }
+ if (current_streams->find(new_stream->label()) != nullptr) {
+ LOG(LS_ERROR) << "MediaStream with label " << new_stream->label()
+ << " is already added.";
+ return false;
+ }
+ return true;
+}
+
+bool MediaContentDirectionHasSend(cricket::MediaContentDirection dir) {
+ return dir == cricket::MD_SENDONLY || dir == cricket::MD_SENDRECV;
+}
+
+bool IsValidOfferToReceiveMedia(int value) {
+ typedef PeerConnectionInterface::RTCOfferAnswerOptions Options;
+ return (value >= Options::kUndefined) &&
+ (value <= Options::kMaxOfferToReceiveMedia);
+}
+
+// Add the stream and RTP data channel info to |session_options|.
+void SetStreams(cricket::MediaSessionOptions* session_options,
+ rtc::scoped_refptr<StreamCollection> streams,
+ const std::map<std::string, rtc::scoped_refptr<DataChannel>>&
+ rtp_data_channels) {
+ session_options->streams.clear();
+ if (streams != nullptr) {
+ for (size_t i = 0; i < streams->count(); ++i) {
+ MediaStreamInterface* stream = streams->at(i);
+ // For each audio track in the stream, add it to the MediaSessionOptions.
+ for (const auto& track : stream->GetAudioTracks()) {
+ session_options->AddSendStream(cricket::MEDIA_TYPE_AUDIO, track->id(),
+ stream->label());
+ }
+ // For each video track in the stream, add it to the MediaSessionOptions.
+ for (const auto& track : stream->GetVideoTracks()) {
+ session_options->AddSendStream(cricket::MEDIA_TYPE_VIDEO, track->id(),
+ stream->label());
+ }
+ }
+ }
+
+ // Check for data channels.
+ for (const auto& kv : rtp_data_channels) {
+ const DataChannel* channel = kv.second;
+ if (channel->state() == DataChannel::kConnecting ||
+ channel->state() == DataChannel::kOpen) {
+ // |streamid| and |sync_label| are both set to the DataChannel label
+ // here so they can be signaled the same way as MediaStreams and Tracks.
+ // For MediaStreams, the sync_label is the MediaStream label and the
+ // track label is the same as |streamid|.
+ const std::string& streamid = channel->label();
+ const std::string& sync_label = channel->label();
+ session_options->AddSendStream(cricket::MEDIA_TYPE_DATA, streamid,
+ sync_label);
+ }
+ }
+}
+
} // namespace
namespace webrtc {
+// Factory class for creating remote MediaStreams and MediaStreamTracks.
+class RemoteMediaStreamFactory {
+ public:
+ explicit RemoteMediaStreamFactory(rtc::Thread* signaling_thread,
+ cricket::ChannelManager* channel_manager)
+ : signaling_thread_(signaling_thread),
+ channel_manager_(channel_manager) {}
+
+ rtc::scoped_refptr<MediaStreamInterface> CreateMediaStream(
+ const std::string& stream_label) {
+ return MediaStreamProxy::Create(signaling_thread_,
+ MediaStream::Create(stream_label));
+ }
+
+ AudioTrackInterface* AddAudioTrack(webrtc::MediaStreamInterface* stream,
+ const std::string& track_id) {
+ return AddTrack<AudioTrackInterface, AudioTrack, AudioTrackProxy>(
+ stream, track_id, RemoteAudioSource::Create().get());
+ }
+
+ VideoTrackInterface* AddVideoTrack(webrtc::MediaStreamInterface* stream,
+ const std::string& track_id) {
+ return AddTrack<VideoTrackInterface, VideoTrack, VideoTrackProxy>(
+ stream, track_id,
+ VideoSource::Create(channel_manager_, new RemoteVideoCapturer(),
+ nullptr)
+ .get());
+ }
+
+ private:
+ template <typename TI, typename T, typename TP, typename S>
+ TI* AddTrack(MediaStreamInterface* stream,
+ const std::string& track_id,
+ S* source) {
+ rtc::scoped_refptr<TI> track(
+ TP::Create(signaling_thread_, T::Create(track_id, source)));
+ track->set_state(webrtc::MediaStreamTrackInterface::kLive);
+ if (stream->AddTrack(track)) {
+ return track;
+ }
+ return nullptr;
+ }
+
+ rtc::Thread* signaling_thread_;
+ cricket::ChannelManager* channel_manager_;
+};
+
+bool ConvertRtcOptionsForOffer(
+ const PeerConnectionInterface::RTCOfferAnswerOptions& rtc_options,
+ cricket::MediaSessionOptions* session_options) {
+ typedef PeerConnectionInterface::RTCOfferAnswerOptions RTCOfferAnswerOptions;
+ if (!IsValidOfferToReceiveMedia(rtc_options.offer_to_receive_audio) ||
+ !IsValidOfferToReceiveMedia(rtc_options.offer_to_receive_video)) {
+ return false;
+ }
+
+ // According to the spec, offer to receive audio/video if the constraint is
+ // not set and there are send streams.
+ if (rtc_options.offer_to_receive_audio == RTCOfferAnswerOptions::kUndefined) {
+ session_options->recv_audio =
+ session_options->HasSendMediaStream(cricket::MEDIA_TYPE_AUDIO);
+ } else {
+ session_options->recv_audio = (rtc_options.offer_to_receive_audio > 0);
+ }
+ if (rtc_options.offer_to_receive_video == RTCOfferAnswerOptions::kUndefined) {
+ session_options->recv_video =
+ session_options->HasSendMediaStream(cricket::MEDIA_TYPE_VIDEO);
+ } else {
+ session_options->recv_video = (rtc_options.offer_to_receive_video > 0);
+ }
+
+ session_options->vad_enabled = rtc_options.voice_activity_detection;
+ session_options->transport_options.ice_restart = rtc_options.ice_restart;
+ session_options->bundle_enabled =
+ rtc_options.use_rtp_mux &&
+ (session_options->has_audio() || session_options->has_video() ||
+ session_options->has_data());
+
+ return true;
+}
+
+bool ParseConstraintsForAnswer(const MediaConstraintsInterface* constraints,
+ cricket::MediaSessionOptions* session_options) {
+ bool value = false;
+ size_t mandatory_constraints_satisfied = 0;
+
+ // kOfferToReceiveAudio defaults to true according to spec.
+ if (!FindConstraint(constraints,
+ MediaConstraintsInterface::kOfferToReceiveAudio, &value,
+ &mandatory_constraints_satisfied) ||
+ value) {
+ session_options->recv_audio = true;
+ }
+
+ // kOfferToReceiveVideo defaults to false according to spec. But
+ // if it is an answer and video is offered, we should still accept video
+ // per default.
+ value = false;
+ if (!FindConstraint(constraints,
+ MediaConstraintsInterface::kOfferToReceiveVideo, &value,
+ &mandatory_constraints_satisfied) ||
+ value) {
+ session_options->recv_video = true;
+ }
+
+ if (FindConstraint(constraints,
+ MediaConstraintsInterface::kVoiceActivityDetection, &value,
+ &mandatory_constraints_satisfied)) {
+ session_options->vad_enabled = value;
+ }
+
+ if (FindConstraint(constraints, MediaConstraintsInterface::kUseRtpMux, &value,
+ &mandatory_constraints_satisfied)) {
+ session_options->bundle_enabled = value;
+ } else {
+ // kUseRtpMux defaults to true according to spec.
+ session_options->bundle_enabled = true;
+ }
+ session_options->bundle_enabled =
+ session_options->bundle_enabled &&
+ (session_options->has_audio() || session_options->has_video() ||
+ session_options->has_data());
+
+ if (FindConstraint(constraints, MediaConstraintsInterface::kIceRestart,
+ &value, &mandatory_constraints_satisfied)) {
+ session_options->transport_options.ice_restart = value;
+ } else {
+ // kIceRestart defaults to false according to spec.
+ session_options->transport_options.ice_restart = false;
+ }
+
+ if (!constraints) {
+ return true;
+ }
+ return mandatory_constraints_satisfied == constraints->GetMandatory().size();
+}
+
bool ParseIceServers(const PeerConnectionInterface::IceServers& servers,
StunConfigurations* stun_config,
TurnConfigurations* turn_config) {
@@ -333,22 +561,6 @@ bool ParseIceServers(const PeerConnectionInterface::IceServers& servers,
return true;
}
-// Check if we can send |new_stream| on a PeerConnection.
-// Currently only one audio but multiple video track is supported per
-// PeerConnection.
-bool CanAddLocalMediaStream(webrtc::StreamCollectionInterface* current_streams,
- webrtc::MediaStreamInterface* new_stream) {
- if (!new_stream || !current_streams)
- return false;
- if (current_streams->find(new_stream->label()) != NULL) {
- LOG(LS_ERROR) << "MediaStream with label " << new_stream->label()
- << " is already added.";
- return false;
- }
-
- return true;
-}
-
PeerConnection::PeerConnection(PeerConnectionFactory* factory)
: factory_(factory),
observer_(NULL),
@@ -356,14 +568,12 @@ PeerConnection::PeerConnection(PeerConnectionFactory* factory)
signaling_state_(kStable),
ice_state_(kIceNew),
ice_connection_state_(kIceConnectionNew),
- ice_gathering_state_(kIceGatheringNew) {
-}
+ ice_gathering_state_(kIceGatheringNew),
+ local_streams_(StreamCollection::Create()),
+ remote_streams_(StreamCollection::Create()) {}
PeerConnection::~PeerConnection() {
RTC_DCHECK(signaling_thread()->IsCurrent());
- if (mediastream_signaling_) {
- mediastream_signaling_->TearDown();
- }
// Need to detach RTP senders/receivers from WebRtcSession,
// since it's about to be destroyed.
for (const auto& sender : senders_) {
@@ -380,9 +590,10 @@ bool PeerConnection::Initialize(
PortAllocatorFactoryInterface* allocator_factory,
rtc::scoped_ptr<DtlsIdentityStoreInterface> dtls_identity_store,
PeerConnectionObserver* observer) {
- RTC_DCHECK(observer != NULL);
- if (!observer)
+ RTC_DCHECK(observer != nullptr);
+ if (!observer) {
return false;
+ }
observer_ = observer;
std::vector<PortAllocatorFactoryInterface::StunConfiguration> stun_config;
@@ -400,8 +611,8 @@ bool PeerConnection::Initialize(
cricket::PORTALLOCATOR_ENABLE_IPV6;
bool value;
// If IPv6 flag was specified, we'll not override it by experiment.
- if (FindConstraint(
- constraints, MediaConstraintsInterface::kEnableIPv6, &value, NULL)) {
+ if (FindConstraint(constraints, MediaConstraintsInterface::kEnableIPv6,
+ &value, nullptr)) {
if (!value) {
portallocator_flags &= ~(cricket::PORTALLOCATOR_ENABLE_IPV6);
}
@@ -419,36 +630,45 @@ bool PeerConnection::Initialize(
// No step delay is used while allocating ports.
port_allocator_->set_step_delay(cricket::kMinimumStepDelay);
- mediastream_signaling_.reset(new MediaStreamSignaling(
- factory_->signaling_thread(), this, factory_->channel_manager()));
+ remote_stream_factory_.reset(new RemoteMediaStreamFactory(
+ factory_->signaling_thread(), factory_->channel_manager()));
- session_.reset(new WebRtcSession(factory_->channel_manager(),
- factory_->signaling_thread(),
- factory_->worker_thread(),
- port_allocator_.get(),
- mediastream_signaling_.get()));
- stats_.reset(new StatsCollector(session_.get()));
+ session_.reset(new WebRtcSession(
+ factory_->channel_manager(), factory_->signaling_thread(),
+ factory_->worker_thread(), port_allocator_.get()));
+ stats_.reset(new StatsCollector(this));
// Initialize the WebRtcSession. It creates transport channels etc.
if (!session_->Initialize(factory_->options(), constraints,
- dtls_identity_store.Pass(), configuration))
+ dtls_identity_store.Pass(), configuration)) {
return false;
+ }
// Register PeerConnection as receiver of local ice candidates.
// All the callbacks will be posted to the application from PeerConnection.
session_->RegisterIceObserver(this);
session_->SignalState.connect(this, &PeerConnection::OnSessionStateChange);
+ session_->SignalVoiceChannelDestroyed.connect(
+ this, &PeerConnection::OnVoiceChannelDestroyed);
+ session_->SignalVideoChannelDestroyed.connect(
+ this, &PeerConnection::OnVideoChannelDestroyed);
+ session_->SignalDataChannelCreated.connect(
+ this, &PeerConnection::OnDataChannelCreated);
+ session_->SignalDataChannelDestroyed.connect(
+ this, &PeerConnection::OnDataChannelDestroyed);
+ session_->SignalDataChannelOpenMessage.connect(
+ this, &PeerConnection::OnDataChannelOpenMessage);
return true;
}
rtc::scoped_refptr<StreamCollectionInterface>
PeerConnection::local_streams() {
- return mediastream_signaling_->local_streams();
+ return local_streams_;
}
rtc::scoped_refptr<StreamCollectionInterface>
PeerConnection::remote_streams() {
- return mediastream_signaling_->remote_streams();
+ return remote_streams_;
}
// TODO(deadbeef): Create RtpSenders immediately here, even if local
@@ -457,20 +677,57 @@ bool PeerConnection::AddStream(MediaStreamInterface* local_stream) {
if (IsClosed()) {
return false;
}
- if (!CanAddLocalMediaStream(mediastream_signaling_->local_streams(),
- local_stream))
+ if (!CanAddLocalMediaStream(local_streams_, local_stream)) {
return false;
+ }
- if (!mediastream_signaling_->AddLocalStream(local_stream)) {
- return false;
+ local_streams_->AddStream(local_stream);
+
+ // Find tracks that have already been configured in SDP. This can occur if a
+ // local session description that contains the MSID of these tracks is set
+ // before AddLocalStream is called. It can also occur if the local session
+ // description is not changed and RemoveLocalStream is called and later
+ // AddLocalStream is called again with the same stream.
+ for (const auto& track : local_stream->GetAudioTracks()) {
+ const TrackInfo* track_info =
+ FindTrackInfo(local_audio_tracks_, local_stream->label(), track->id());
+ if (track_info) {
+ CreateAudioSender(local_stream, track.get(), track_info->ssrc);
+ }
}
+ for (const auto& track : local_stream->GetVideoTracks()) {
+ const TrackInfo* track_info =
+ FindTrackInfo(local_video_tracks_, local_stream->label(), track->id());
+ if (track_info) {
+ CreateVideoSender(local_stream, track.get(), track_info->ssrc);
+ }
+ }
+
stats_->AddStream(local_stream);
observer_->OnRenegotiationNeeded();
return true;
}
+// TODO(deadbeef): Don't destroy RtpSenders here; they should be kept around
+// indefinitely.
void PeerConnection::RemoveStream(MediaStreamInterface* local_stream) {
- mediastream_signaling_->RemoveLocalStream(local_stream);
+ for (const auto& track : local_stream->GetAudioTracks()) {
+ const TrackInfo* track_info =
+ FindTrackInfo(local_audio_tracks_, local_stream->label(), track->id());
+ if (track_info) {
+ DestroyAudioSender(local_stream, track.get(), track_info->ssrc);
+ }
+ }
+ for (const auto& track : local_stream->GetVideoTracks()) {
+ const TrackInfo* track_info =
+ FindTrackInfo(local_video_tracks_, local_stream->label(), track->id());
+ if (track_info) {
+ DestroyVideoSender(local_stream, track.get());
+ }
+ }
+
+ local_streams_->RemoveStream(local_stream);
+
if (IsClosed()) {
return;
}
@@ -483,7 +740,7 @@ rtc::scoped_refptr<DtmfSenderInterface> PeerConnection::CreateDtmfSender(
LOG(LS_ERROR) << "CreateDtmfSender - track is NULL.";
return NULL;
}
- if (!mediastream_signaling_->local_streams()->FindAudioTrack(track->id())) {
+ if (!local_streams_->FindAudioTrack(track->id())) {
LOG(LS_ERROR) << "CreateDtmfSender is called with a non local audio track.";
return NULL;
}
@@ -553,16 +810,17 @@ rtc::scoped_refptr<DataChannelInterface>
PeerConnection::CreateDataChannel(
const std::string& label,
const DataChannelInit* config) {
- bool first_datachannel = !mediastream_signaling_->HasDataChannels();
+ bool first_datachannel = !HasDataChannels();
rtc::scoped_ptr<InternalDataChannelInit> internal_config;
if (config) {
internal_config.reset(new InternalDataChannelInit(*config));
}
rtc::scoped_refptr<DataChannelInterface> channel(
- session_->CreateDataChannel(label, internal_config.get()));
- if (!channel.get())
- return NULL;
+ InternalCreateDataChannel(label, internal_config.get()));
+ if (!channel.get()) {
+ return nullptr;
+ }
// Trigger the onRenegotiationNeeded event for every new RTP DataChannel, or
// the first SCTP DataChannel.
@@ -575,7 +833,7 @@ PeerConnection::CreateDataChannel(
void PeerConnection::CreateOffer(CreateSessionDescriptionObserver* observer,
const MediaConstraintsInterface* constraints) {
- if (!VERIFY(observer != NULL)) {
+ if (!VERIFY(observer != nullptr)) {
LOG(LS_ERROR) << "CreateOffer - observer is NULL.";
return;
}
@@ -626,27 +884,45 @@ void PeerConnection::CreateOffer(CreateSessionDescriptionObserver* observer,
void PeerConnection::CreateOffer(CreateSessionDescriptionObserver* observer,
const RTCOfferAnswerOptions& options) {
- if (!VERIFY(observer != NULL)) {
+ if (!VERIFY(observer != nullptr)) {
LOG(LS_ERROR) << "CreateOffer - observer is NULL.";
return;
}
- session_->CreateOffer(observer, options);
+
+ cricket::MediaSessionOptions session_options;
+ if (!GetOptionsForOffer(options, &session_options)) {
+ std::string error = "CreateOffer called with invalid options.";
+ LOG(LS_ERROR) << error;
+ PostCreateSessionDescriptionFailure(observer, error);
+ return;
+ }
+
+ session_->CreateOffer(observer, options, session_options);
}
void PeerConnection::CreateAnswer(
CreateSessionDescriptionObserver* observer,
const MediaConstraintsInterface* constraints) {
- if (!VERIFY(observer != NULL)) {
+ if (!VERIFY(observer != nullptr)) {
LOG(LS_ERROR) << "CreateAnswer - observer is NULL.";
return;
}
- session_->CreateAnswer(observer, constraints);
+
+ cricket::MediaSessionOptions session_options;
+ if (!GetOptionsForAnswer(constraints, &session_options)) {
+ std::string error = "CreateAnswer called with invalid constraints.";
+ LOG(LS_ERROR) << error;
+ PostCreateSessionDescriptionFailure(observer, error);
+ return;
+ }
+
+ session_->CreateAnswer(observer, constraints, session_options);
}
void PeerConnection::SetLocalDescription(
SetSessionDescriptionObserver* observer,
SessionDescriptionInterface* desc) {
- if (!VERIFY(observer != NULL)) {
+ if (!VERIFY(observer != nullptr)) {
LOG(LS_ERROR) << "SetLocalDescription - observer is NULL.";
return;
}
@@ -662,8 +938,50 @@ void PeerConnection::SetLocalDescription(
PostSetSessionDescriptionFailure(observer, error);
return;
}
- SetSessionDescriptionMsg* msg = new SetSessionDescriptionMsg(observer);
+
+ // If setting the description decided our SSL role, allocate any necessary
+ // SCTP sids.
+ rtc::SSLRole role;
+ if (session_->data_channel_type() == cricket::DCT_SCTP &&
+ session_->GetSslRole(&role)) {
+ AllocateSctpSids(role);
+ }
+
+ // Update state and SSRC of local MediaStreams and DataChannels based on the
+ // local session description.
+ const cricket::ContentInfo* audio_content =
+ GetFirstAudioContent(desc->description());
+ if (audio_content) {
+ const cricket::AudioContentDescription* audio_desc =
+ static_cast<const cricket::AudioContentDescription*>(
+ audio_content->description);
+ UpdateLocalTracks(audio_desc->streams(), audio_desc->type());
+ }
+
+ const cricket::ContentInfo* video_content =
+ GetFirstVideoContent(desc->description());
+ if (video_content) {
+ const cricket::VideoContentDescription* video_desc =
+ static_cast<const cricket::VideoContentDescription*>(
+ video_content->description);
+ UpdateLocalTracks(video_desc->streams(), video_desc->type());
+ }
+
+ const cricket::ContentInfo* data_content =
+ GetFirstDataContent(desc->description());
+ if (data_content) {
+ const cricket::DataContentDescription* data_desc =
+ static_cast<const cricket::DataContentDescription*>(
+ data_content->description);
+ if (rtc::starts_with(data_desc->protocol().data(),
+ cricket::kMediaProtocolRtpPrefix)) {
+ UpdateLocalRtpDataChannels(data_desc->streams());
+ }
+ }
+
+ SetSessionDescriptionMsg* msg = new SetSessionDescriptionMsg(observer);
signaling_thread()->Post(this, MSG_SET_SESSIONDESCRIPTION_SUCCESS, msg);
+
// MaybeStartGathering needs to be called after posting
// MSG_SET_SESSIONDESCRIPTION_SUCCESS, so that we don't signal any candidates
// before signaling that SetLocalDescription completed.
@@ -673,7 +991,7 @@ void PeerConnection::SetLocalDescription(
void PeerConnection::SetRemoteDescription(
SetSessionDescriptionObserver* observer,
SessionDescriptionInterface* desc) {
- if (!VERIFY(observer != NULL)) {
+ if (!VERIFY(observer != nullptr)) {
LOG(LS_ERROR) << "SetRemoteDescription - observer is NULL.";
return;
}
@@ -689,16 +1007,78 @@ void PeerConnection::SetRemoteDescription(
PostSetSessionDescriptionFailure(observer, error);
return;
}
- SetSessionDescriptionMsg* msg = new SetSessionDescriptionMsg(observer);
- signaling_thread()->Post(this, MSG_SET_SESSIONDESCRIPTION_SUCCESS, msg);
-}
-void PeerConnection::PostSetSessionDescriptionFailure(
- SetSessionDescriptionObserver* observer,
- const std::string& error) {
- SetSessionDescriptionMsg* msg = new SetSessionDescriptionMsg(observer);
- msg->error = error;
- signaling_thread()->Post(this, MSG_SET_SESSIONDESCRIPTION_FAILED, msg);
+ // If setting the description decided our SSL role, allocate any necessary
+ // SCTP sids.
+ rtc::SSLRole role;
+ if (session_->data_channel_type() == cricket::DCT_SCTP &&
+ session_->GetSslRole(&role)) {
+ AllocateSctpSids(role);
+ }
+
+ const cricket::SessionDescription* remote_desc = desc->description();
+
+ // We wait to signal new streams until we finish processing the description,
+ // since only at that point will new streams have all their tracks.
+ rtc::scoped_refptr<StreamCollection> new_streams(StreamCollection::Create());
+
+ // Find all audio rtp streams and create corresponding remote AudioTracks
+ // and MediaStreams.
+ const cricket::ContentInfo* audio_content = GetFirstAudioContent(remote_desc);
+ if (audio_content) {
+ const cricket::AudioContentDescription* desc =
+ static_cast<const cricket::AudioContentDescription*>(
+ audio_content->description);
+ UpdateRemoteStreamsList(desc->streams(), desc->type(), new_streams);
+ remote_info_.default_audio_track_needed =
+ MediaContentDirectionHasSend(desc->direction()) &&
+ desc->streams().empty();
+ }
+
+ // Find all video rtp streams and create corresponding remote VideoTracks
+ // and MediaStreams.
+ const cricket::ContentInfo* video_content = GetFirstVideoContent(remote_desc);
+ if (video_content) {
+ const cricket::VideoContentDescription* desc =
+ static_cast<const cricket::VideoContentDescription*>(
+ video_content->description);
+ UpdateRemoteStreamsList(desc->streams(), desc->type(), new_streams);
+ remote_info_.default_video_track_needed =
+ MediaContentDirectionHasSend(desc->direction()) &&
+ desc->streams().empty();
+ }
+
+ // Update the DataChannels with the information from the remote peer.
+ const cricket::ContentInfo* data_content = GetFirstDataContent(remote_desc);
+ if (data_content) {
+ const cricket::DataContentDescription* data_desc =
+ static_cast<const cricket::DataContentDescription*>(
+ data_content->description);
+ if (rtc::starts_with(data_desc->protocol().data(),
+ cricket::kMediaProtocolRtpPrefix)) {
+ UpdateRemoteRtpDataChannels(data_desc->streams());
+ }
+ }
+
+ // Iterate new_streams and notify the observer about new MediaStreams.
+ for (size_t i = 0; i < new_streams->count(); ++i) {
+ MediaStreamInterface* new_stream = new_streams->at(i);
+ stats_->AddStream(new_stream);
+ observer_->OnAddStream(new_stream);
+ }
+
+ // Find removed MediaStreams.
+ if (remote_info_.IsDefaultMediaStreamNeeded() &&
+ remote_streams_->find(kDefaultStreamLabel) != nullptr) {
+ // The default media stream already exists. No need to do anything.
+ } else {
+ UpdateEndedRemoteMediaStreams();
+ remote_info_.msid_supported |= remote_streams_->count() > 0;
+ }
+ MaybeCreateDefaultStream();
+
+ SetSessionDescriptionMsg* msg = new SetSessionDescriptionMsg(observer);
+ signaling_thread()->Post(this, MSG_SET_SESSIONDESCRIPTION_SUCCESS, msg);
}
bool PeerConnection::SetConfiguration(const RTCConfiguration& config) {
@@ -832,6 +1212,13 @@ void PeerConnection::OnMessage(rtc::Message* msg) {
delete param;
break;
}
+ case MSG_CREATE_SESSIONDESCRIPTION_FAILED: {
+ CreateSessionDescriptionMsg* param =
+ static_cast<CreateSessionDescriptionMsg*>(msg->pdata);
+ param->observer->OnFailure(param->error);
+ delete param;
+ break;
+ }
case MSG_GETSTATS: {
GetStatsMsg* param = static_cast<GetStatsMsg*>(msg->pdata);
StatsReports reports;
@@ -846,37 +1233,22 @@ void PeerConnection::OnMessage(rtc::Message* msg) {
}
}
-void PeerConnection::OnAddRemoteStream(MediaStreamInterface* stream) {
- stats_->AddStream(stream);
- observer_->OnAddStream(stream);
-}
-
-void PeerConnection::OnRemoveRemoteStream(MediaStreamInterface* stream) {
- observer_->OnRemoveStream(stream);
-}
-
-void PeerConnection::OnAddDataChannel(DataChannelInterface* data_channel) {
- observer_->OnDataChannel(DataChannelProxy::Create(signaling_thread(),
- data_channel));
-}
-
-void PeerConnection::OnAddRemoteAudioTrack(MediaStreamInterface* stream,
- AudioTrackInterface* audio_track,
- uint32_t ssrc) {
+void PeerConnection::CreateAudioReceiver(MediaStreamInterface* stream,
+ AudioTrackInterface* audio_track,
+ uint32_t ssrc) {
receivers_.push_back(new AudioRtpReceiver(audio_track, ssrc, session_.get()));
}
-void PeerConnection::OnAddRemoteVideoTrack(MediaStreamInterface* stream,
- VideoTrackInterface* video_track,
- uint32_t ssrc) {
+void PeerConnection::CreateVideoReceiver(MediaStreamInterface* stream,
+ VideoTrackInterface* video_track,
+ uint32_t ssrc) {
receivers_.push_back(new VideoRtpReceiver(video_track, ssrc, session_.get()));
}
// TODO(deadbeef): Keep RtpReceivers around even if track goes away in remote
// description.
-void PeerConnection::OnRemoveRemoteAudioTrack(
- MediaStreamInterface* stream,
- AudioTrackInterface* audio_track) {
+void PeerConnection::DestroyAudioReceiver(MediaStreamInterface* stream,
+ AudioTrackInterface* audio_track) {
auto it = FindReceiverForTrack(audio_track);
if (it == receivers_.end()) {
LOG(LS_WARNING) << "RtpReceiver for track with id " << audio_track->id()
@@ -887,9 +1259,8 @@ void PeerConnection::OnRemoveRemoteAudioTrack(
}
}
-void PeerConnection::OnRemoveRemoteVideoTrack(
- MediaStreamInterface* stream,
- VideoTrackInterface* video_track) {
+void PeerConnection::DestroyVideoReceiver(MediaStreamInterface* stream,
+ VideoTrackInterface* video_track) {
auto it = FindReceiverForTrack(video_track);
if (it == receivers_.end()) {
LOG(LS_WARNING) << "RtpReceiver for track with id " << video_track->id()
@@ -900,24 +1271,24 @@ void PeerConnection::OnRemoveRemoteVideoTrack(
}
}
-void PeerConnection::OnAddLocalAudioTrack(MediaStreamInterface* stream,
- AudioTrackInterface* audio_track,
- uint32_t ssrc) {
+void PeerConnection::CreateAudioSender(MediaStreamInterface* stream,
+ AudioTrackInterface* audio_track,
+ uint32_t ssrc) {
senders_.push_back(new AudioRtpSender(audio_track, ssrc, session_.get()));
stats_->AddLocalAudioTrack(audio_track, ssrc);
}
-void PeerConnection::OnAddLocalVideoTrack(MediaStreamInterface* stream,
- VideoTrackInterface* video_track,
- uint32_t ssrc) {
+void PeerConnection::CreateVideoSender(MediaStreamInterface* stream,
+ VideoTrackInterface* video_track,
+ uint32_t ssrc) {
senders_.push_back(new VideoRtpSender(video_track, ssrc, session_.get()));
}
// TODO(deadbeef): Keep RtpSenders around even if track goes away in local
// description.
-void PeerConnection::OnRemoveLocalAudioTrack(MediaStreamInterface* stream,
- AudioTrackInterface* audio_track,
- uint32_t ssrc) {
+void PeerConnection::DestroyAudioSender(MediaStreamInterface* stream,
+ AudioTrackInterface* audio_track,
+ uint32_t ssrc) {
auto it = FindSenderForTrack(audio_track);
if (it == senders_.end()) {
LOG(LS_WARNING) << "RtpSender for track with id " << audio_track->id()
@@ -930,8 +1301,8 @@ void PeerConnection::OnRemoveLocalAudioTrack(MediaStreamInterface* stream,
stats_->RemoveLocalAudioTrack(audio_track, ssrc);
}
-void PeerConnection::OnRemoveLocalVideoTrack(MediaStreamInterface* stream,
- VideoTrackInterface* video_track) {
+void PeerConnection::DestroyVideoSender(MediaStreamInterface* stream,
+ VideoTrackInterface* video_track) {
auto it = FindSenderForTrack(video_track);
if (it == senders_.end()) {
LOG(LS_WARNING) << "RtpSender for track with id " << video_track->id()
@@ -943,15 +1314,12 @@ void PeerConnection::OnRemoveLocalVideoTrack(MediaStreamInterface* stream,
}
}
-void PeerConnection::OnRemoveLocalStream(MediaStreamInterface* stream) {
-}
-
void PeerConnection::OnIceConnectionChange(
PeerConnectionInterface::IceConnectionState new_state) {
RTC_DCHECK(signaling_thread()->IsCurrent());
// After transitioning to "closed", ignore any additional states from
// WebRtcSession (such as "disconnected").
- if (ice_connection_state_ == kIceConnectionClosed) {
+ if (IsClosed()) {
return;
}
ice_connection_state_ = new_state;
@@ -998,6 +1366,540 @@ void PeerConnection::ChangeSignalingState(
observer_->OnStateChange(PeerConnectionObserver::kSignalingState);
}
+void PeerConnection::PostSetSessionDescriptionFailure(
+ SetSessionDescriptionObserver* observer,
+ const std::string& error) {
+ SetSessionDescriptionMsg* msg = new SetSessionDescriptionMsg(observer);
+ msg->error = error;
+ signaling_thread()->Post(this, MSG_SET_SESSIONDESCRIPTION_FAILED, msg);
+}
+
+void PeerConnection::PostCreateSessionDescriptionFailure(
+ CreateSessionDescriptionObserver* observer,
+ const std::string& error) {
+ CreateSessionDescriptionMsg* msg = new CreateSessionDescriptionMsg(observer);
+ msg->error = error;
+ signaling_thread()->Post(this, MSG_CREATE_SESSIONDESCRIPTION_FAILED, msg);
+}
+
+bool PeerConnection::GetOptionsForOffer(
+ const PeerConnectionInterface::RTCOfferAnswerOptions& rtc_options,
+ cricket::MediaSessionOptions* session_options) {
+ SetStreams(session_options, local_streams_, rtp_data_channels_);
+
+ if (!ConvertRtcOptionsForOffer(rtc_options, session_options)) {
+ return false;
+ }
+
+ if (session_->data_channel_type() == cricket::DCT_SCTP && HasDataChannels()) {
+ session_options->data_channel_type = cricket::DCT_SCTP;
+ }
+ return true;
+}
+
+bool PeerConnection::GetOptionsForAnswer(
+ const MediaConstraintsInterface* constraints,
+ cricket::MediaSessionOptions* session_options) {
+ SetStreams(session_options, local_streams_, rtp_data_channels_);
+ session_options->recv_audio = false;
+ session_options->recv_video = false;
+
+ if (!ParseConstraintsForAnswer(constraints, session_options)) {
+ return false;
+ }
+
+ // RTP data channel is handled in MediaSessionOptions::AddStream. SCTP streams
+ // are not signaled in the SDP so does not go through that path and must be
+ // handled here.
+ if (session_->data_channel_type() == cricket::DCT_SCTP) {
+ session_options->data_channel_type = cricket::DCT_SCTP;
+ }
+ return true;
+}
+
+void PeerConnection::UpdateRemoteStreamsList(
+ const cricket::StreamParamsVec& streams,
+ cricket::MediaType media_type,
+ StreamCollection* new_streams) {
+ TrackInfos* current_tracks = GetRemoteTracks(media_type);
+
+ // Find removed tracks. I.e., tracks where the track id or ssrc don't match
+ // the
+ // new StreamParam.
+ auto track_it = current_tracks->begin();
+ while (track_it != current_tracks->end()) {
+ const TrackInfo& info = *track_it;
+ const cricket::StreamParams* params =
+ cricket::GetStreamBySsrc(streams, info.ssrc);
+ if (!params || params->id != info.track_id) {
+ OnRemoteTrackRemoved(info.stream_label, info.track_id, media_type);
+ track_it = current_tracks->erase(track_it);
+ } else {
+ ++track_it;
+ }
+ }
+
+ // Find new and active tracks.
+ for (const cricket::StreamParams& params : streams) {
+ // The sync_label is the MediaStream label and the |stream.id| is the
+ // track id.
+ const std::string& stream_label = params.sync_label;
+ const std::string& track_id = params.id;
+ uint32_t ssrc = params.first_ssrc();
+
+ rtc::scoped_refptr<MediaStreamInterface> stream =
+ remote_streams_->find(stream_label);
+ if (!stream) {
+ // This is a new MediaStream. Create a new remote MediaStream.
+ stream = remote_stream_factory_->CreateMediaStream(stream_label);
+ remote_streams_->AddStream(stream);
+ new_streams->AddStream(stream);
+ }
+
+ const TrackInfo* track_info =
+ FindTrackInfo(*current_tracks, stream_label, track_id);
+ if (!track_info) {
+ current_tracks->push_back(TrackInfo(stream_label, track_id, ssrc));
+ OnRemoteTrackSeen(stream_label, track_id, ssrc, media_type);
+ }
+ }
+}
+
+void PeerConnection::OnRemoteTrackSeen(const std::string& stream_label,
+ const std::string& track_id,
+ uint32_t ssrc,
+ cricket::MediaType media_type) {
+ MediaStreamInterface* stream = remote_streams_->find(stream_label);
+
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) {
+ AudioTrackInterface* audio_track =
+ remote_stream_factory_->AddAudioTrack(stream, track_id);
+ CreateAudioReceiver(stream, audio_track, ssrc);
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) {
+ VideoTrackInterface* video_track =
+ remote_stream_factory_->AddVideoTrack(stream, track_id);
+ CreateVideoReceiver(stream, video_track, ssrc);
+ } else {
+ RTC_DCHECK(false && "Invalid media type");
+ }
+}
+
+void PeerConnection::OnRemoteTrackRemoved(const std::string& stream_label,
+ const std::string& track_id,
+ cricket::MediaType media_type) {
+ MediaStreamInterface* stream = remote_streams_->find(stream_label);
+
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) {
+ rtc::scoped_refptr<AudioTrackInterface> audio_track =
+ stream->FindAudioTrack(track_id);
+ if (audio_track) {
+ audio_track->set_state(webrtc::MediaStreamTrackInterface::kEnded);
+ stream->RemoveTrack(audio_track);
+ DestroyAudioReceiver(stream, audio_track);
+ }
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) {
+ rtc::scoped_refptr<VideoTrackInterface> video_track =
+ stream->FindVideoTrack(track_id);
+ if (video_track) {
+ video_track->set_state(webrtc::MediaStreamTrackInterface::kEnded);
+ stream->RemoveTrack(video_track);
+ DestroyVideoReceiver(stream, video_track);
+ }
+ } else {
+ ASSERT(false && "Invalid media type");
+ }
+}
+
+void PeerConnection::UpdateEndedRemoteMediaStreams() {
+ std::vector<rtc::scoped_refptr<MediaStreamInterface>> streams_to_remove;
+ for (size_t i = 0; i < remote_streams_->count(); ++i) {
+ MediaStreamInterface* stream = remote_streams_->at(i);
+ if (stream->GetAudioTracks().empty() && stream->GetVideoTracks().empty()) {
+ streams_to_remove.push_back(stream);
+ }
+ }
+
+ for (const auto& stream : streams_to_remove) {
+ remote_streams_->RemoveStream(stream);
+ observer_->OnRemoveStream(stream);
+ }
+}
+
+void PeerConnection::MaybeCreateDefaultStream() {
+ if (!remote_info_.IsDefaultMediaStreamNeeded()) {
+ return;
+ }
+
+ bool default_created = false;
+
+ rtc::scoped_refptr<MediaStreamInterface> default_remote_stream =
+ remote_streams_->find(kDefaultStreamLabel);
+ if (default_remote_stream == nullptr) {
+ default_created = true;
+ default_remote_stream =
+ remote_stream_factory_->CreateMediaStream(kDefaultStreamLabel);
+ remote_streams_->AddStream(default_remote_stream);
+ }
+ if (remote_info_.default_audio_track_needed &&
+ default_remote_stream->GetAudioTracks().size() == 0) {
+ remote_audio_tracks_.push_back(
+ TrackInfo(kDefaultStreamLabel, kDefaultAudioTrackLabel, 0));
+ OnRemoteTrackSeen(kDefaultStreamLabel, kDefaultAudioTrackLabel, 0,
+ cricket::MEDIA_TYPE_AUDIO);
+ }
+ if (remote_info_.default_video_track_needed &&
+ default_remote_stream->GetVideoTracks().size() == 0) {
+ remote_video_tracks_.push_back(
+ TrackInfo(kDefaultStreamLabel, kDefaultVideoTrackLabel, 0));
+ OnRemoteTrackSeen(kDefaultStreamLabel, kDefaultVideoTrackLabel, 0,
+ cricket::MEDIA_TYPE_VIDEO);
+ }
+ if (default_created) {
+ stats_->AddStream(default_remote_stream);
+ observer_->OnAddStream(default_remote_stream);
+ }
+}
+
+void PeerConnection::EndRemoteTracks(cricket::MediaType media_type) {
+ TrackInfos* current_tracks = GetRemoteTracks(media_type);
+ for (TrackInfos::iterator track_it = current_tracks->begin();
+ track_it != current_tracks->end(); ++track_it) {
+ const TrackInfo& info = *track_it;
+ MediaStreamInterface* stream = remote_streams_->find(info.stream_label);
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) {
+ AudioTrackInterface* track = stream->FindAudioTrack(info.track_id);
+ // There's no guarantee the track is still available, e.g. the track may
+ // have been removed from the stream by javascript.
+ if (track) {
+ track->set_state(webrtc::MediaStreamTrackInterface::kEnded);
+ }
+ }
+ if (media_type == cricket::MEDIA_TYPE_VIDEO) {
+ VideoTrackInterface* track = stream->FindVideoTrack(info.track_id);
+ // There's no guarantee the track is still available, e.g. the track may
+ // have been removed from the stream by javascript.
+ if (track) {
+ track->set_state(webrtc::MediaStreamTrackInterface::kEnded);
+ }
+ }
+ }
+}
+
+void PeerConnection::UpdateLocalTracks(
+ const std::vector<cricket::StreamParams>& streams,
+ cricket::MediaType media_type) {
+ TrackInfos* current_tracks = GetLocalTracks(media_type);
+
+ // Find removed tracks. I.e., tracks where the track id, stream label or ssrc
+ // don't match the new StreamParam.
+ TrackInfos::iterator track_it = current_tracks->begin();
+ while (track_it != current_tracks->end()) {
+ const TrackInfo& info = *track_it;
+ const cricket::StreamParams* params =
+ cricket::GetStreamBySsrc(streams, info.ssrc);
+ if (!params || params->id != info.track_id ||
+ params->sync_label != info.stream_label) {
+ OnLocalTrackRemoved(info.stream_label, info.track_id, info.ssrc,
+ media_type);
+ track_it = current_tracks->erase(track_it);
+ } else {
+ ++track_it;
+ }
+ }
+
+ // Find new and active tracks.
+ for (const cricket::StreamParams& params : streams) {
+ // The sync_label is the MediaStream label and the |stream.id| is the
+ // track id.
+ const std::string& stream_label = params.sync_label;
+ const std::string& track_id = params.id;
+ uint32_t ssrc = params.first_ssrc();
+ const TrackInfo* track_info =
+ FindTrackInfo(*current_tracks, stream_label, track_id);
+ if (!track_info) {
+ current_tracks->push_back(TrackInfo(stream_label, track_id, ssrc));
+ OnLocalTrackSeen(stream_label, track_id, params.first_ssrc(), media_type);
+ }
+ }
+}
+
+void PeerConnection::OnLocalTrackSeen(const std::string& stream_label,
+ const std::string& track_id,
+ uint32_t ssrc,
+ cricket::MediaType media_type) {
+ MediaStreamInterface* stream = local_streams_->find(stream_label);
+ if (!stream) {
+ LOG(LS_WARNING) << "An unknown local MediaStream with label "
+ << stream_label << " has been configured.";
+ return;
+ }
+
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) {
+ AudioTrackInterface* audio_track = stream->FindAudioTrack(track_id);
+ if (!audio_track) {
+ LOG(LS_WARNING) << "An unknown local AudioTrack with id , " << track_id
+ << " has been configured.";
+ return;
+ }
+ CreateAudioSender(stream, audio_track, ssrc);
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) {
+ VideoTrackInterface* video_track = stream->FindVideoTrack(track_id);
+ if (!video_track) {
+ LOG(LS_WARNING) << "An unknown local VideoTrack with id , " << track_id
+ << " has been configured.";
+ return;
+ }
+ CreateVideoSender(stream, video_track, ssrc);
+ } else {
+ RTC_DCHECK(false && "Invalid media type");
+ }
+}
+
+void PeerConnection::OnLocalTrackRemoved(const std::string& stream_label,
+ const std::string& track_id,
+ uint32_t ssrc,
+ cricket::MediaType media_type) {
+ MediaStreamInterface* stream = local_streams_->find(stream_label);
+ if (!stream) {
+ // This is the normal case. I.e., RemoveLocalStream has been called and the
+ // SessionDescriptions has been renegotiated.
+ return;
+ }
+ // A track has been removed from the SessionDescription but the MediaStream
+ // is still associated with PeerConnection. This only occurs if the SDP
+ // doesn't match with the calls to AddLocalStream and RemoveLocalStream.
+ if (media_type == cricket::MEDIA_TYPE_AUDIO) {
+ AudioTrackInterface* audio_track = stream->FindAudioTrack(track_id);
+ if (!audio_track) {
+ return;
+ }
+ DestroyAudioSender(stream, audio_track, ssrc);
+ } else if (media_type == cricket::MEDIA_TYPE_VIDEO) {
+ VideoTrackInterface* video_track = stream->FindVideoTrack(track_id);
+ if (!video_track) {
+ return;
+ }
+ DestroyVideoSender(stream, video_track);
+ } else {
+ RTC_DCHECK(false && "Invalid media type.");
+ }
+}
+
+void PeerConnection::UpdateLocalRtpDataChannels(
+ const cricket::StreamParamsVec& streams) {
+ std::vector<std::string> existing_channels;
+
+ // Find new and active data channels.
+ for (const cricket::StreamParams& params : streams) {
+ // |it->sync_label| is actually the data channel label. The reason is that
+ // we use the same naming of data channels as we do for
+ // MediaStreams and Tracks.
+ // For MediaStreams, the sync_label is the MediaStream label and the
+ // track label is the same as |streamid|.
+ const std::string& channel_label = params.sync_label;
+ auto data_channel_it = rtp_data_channels_.find(channel_label);
+ if (!VERIFY(data_channel_it != rtp_data_channels_.end())) {
+ continue;
+ }
+ // Set the SSRC the data channel should use for sending.
+ data_channel_it->second->SetSendSsrc(params.first_ssrc());
+ existing_channels.push_back(data_channel_it->first);
+ }
+
+ UpdateClosingRtpDataChannels(existing_channels, true);
+}
+
+void PeerConnection::UpdateRemoteRtpDataChannels(
+ const cricket::StreamParamsVec& streams) {
+ std::vector<std::string> existing_channels;
+
+ // Find new and active data channels.
+ for (const cricket::StreamParams& params : streams) {
+ // The data channel label is either the mslabel or the SSRC if the mslabel
+ // does not exist. Ex a=ssrc:444330170 mslabel:test1.
+ std::string label = params.sync_label.empty()
+ ? rtc::ToString(params.first_ssrc())
+ : params.sync_label;
+ auto data_channel_it = rtp_data_channels_.find(label);
+ if (data_channel_it == rtp_data_channels_.end()) {
+ // This is a new data channel.
+ CreateRemoteRtpDataChannel(label, params.first_ssrc());
+ } else {
+ data_channel_it->second->SetReceiveSsrc(params.first_ssrc());
+ }
+ existing_channels.push_back(label);
+ }
+
+ UpdateClosingRtpDataChannels(existing_channels, false);
+}
+
+void PeerConnection::UpdateClosingRtpDataChannels(
+ const std::vector<std::string>& active_channels,
+ bool is_local_update) {
+ auto it = rtp_data_channels_.begin();
+ while (it != rtp_data_channels_.end()) {
+ DataChannel* data_channel = it->second;
+ if (std::find(active_channels.begin(), active_channels.end(),
+ data_channel->label()) != active_channels.end()) {
+ ++it;
+ continue;
+ }
+
+ if (is_local_update) {
+ data_channel->SetSendSsrc(0);
+ } else {
+ data_channel->RemotePeerRequestClose();
+ }
+
+ if (data_channel->state() == DataChannel::kClosed) {
+ rtp_data_channels_.erase(it);
+ it = rtp_data_channels_.begin();
+ } else {
+ ++it;
+ }
+ }
+}
+
+void PeerConnection::CreateRemoteRtpDataChannel(const std::string& label,
+ uint32_t remote_ssrc) {
+ rtc::scoped_refptr<DataChannel> channel(
+ InternalCreateDataChannel(label, nullptr));
+ if (!channel.get()) {
+ LOG(LS_WARNING) << "Remote peer requested a DataChannel but"
+ << "CreateDataChannel failed.";
+ return;
+ }
+ channel->SetReceiveSsrc(remote_ssrc);
+ observer_->OnDataChannel(
+ DataChannelProxy::Create(signaling_thread(), channel));
+}
+
+rtc::scoped_refptr<DataChannel> PeerConnection::InternalCreateDataChannel(
+ const std::string& label,
+ const InternalDataChannelInit* config) {
+ if (IsClosed()) {
+ return nullptr;
+ }
+ if (session_->data_channel_type() == cricket::DCT_NONE) {
+ LOG(LS_ERROR)
+ << "InternalCreateDataChannel: Data is not supported in this call.";
+ return nullptr;
+ }
+ InternalDataChannelInit new_config =
+ config ? (*config) : InternalDataChannelInit();
+ if (session_->data_channel_type() == cricket::DCT_SCTP) {
+ if (new_config.id < 0) {
+ rtc::SSLRole role;
+ if (session_->GetSslRole(&role) &&
+ !sid_allocator_.AllocateSid(role, &new_config.id)) {
+ LOG(LS_ERROR) << "No id can be allocated for the SCTP data channel.";
+ return nullptr;
+ }
+ } else if (!sid_allocator_.ReserveSid(new_config.id)) {
+ LOG(LS_ERROR) << "Failed to create a SCTP data channel "
+ << "because the id is already in use or out of range.";
+ return nullptr;
+ }
+ }
+
+ rtc::scoped_refptr<DataChannel> channel(DataChannel::Create(
+ session_.get(), session_->data_channel_type(), label, new_config));
+ if (!channel) {
+ sid_allocator_.ReleaseSid(new_config.id);
+ return nullptr;
+ }
+
+ if (channel->data_channel_type() == cricket::DCT_RTP) {
+ if (rtp_data_channels_.find(channel->label()) != rtp_data_channels_.end()) {
+ LOG(LS_ERROR) << "DataChannel with label " << channel->label()
+ << " already exists.";
+ return nullptr;
+ }
+ rtp_data_channels_[channel->label()] = channel;
+ } else {
+ RTC_DCHECK(channel->data_channel_type() == cricket::DCT_SCTP);
+ sctp_data_channels_.push_back(channel);
+ channel->SignalClosed.connect(this,
+ &PeerConnection::OnSctpDataChannelClosed);
+ }
+
+ return channel;
+}
+
+bool PeerConnection::HasDataChannels() const {
+ return !rtp_data_channels_.empty() || !sctp_data_channels_.empty();
+}
+
+void PeerConnection::AllocateSctpSids(rtc::SSLRole role) {
+ for (const auto& channel : sctp_data_channels_) {
+ if (channel->id() < 0) {
+ int sid;
+ if (!sid_allocator_.AllocateSid(role, &sid)) {
+ LOG(LS_ERROR) << "Failed to allocate SCTP sid.";
+ continue;
+ }
+ channel->SetSctpSid(sid);
+ }
+ }
+}
+
+void PeerConnection::OnSctpDataChannelClosed(DataChannel* channel) {
+ for (auto it = sctp_data_channels_.begin(); it != sctp_data_channels_.end();
+ ++it) {
+ if (it->get() == channel) {
+ int sid = channel->id();
+ RTC_DCHECK(sid >= 0);
+ sid_allocator_.ReleaseSid(sid);
+ sctp_data_channels_.erase(it);
+ return;
+ }
+ }
+}
+
+void PeerConnection::OnVoiceChannelDestroyed() {
+ EndRemoteTracks(cricket::MEDIA_TYPE_AUDIO);
+}
+
+void PeerConnection::OnVideoChannelDestroyed() {
+ EndRemoteTracks(cricket::MEDIA_TYPE_VIDEO);
+}
+
+void PeerConnection::OnDataChannelCreated() {
+ for (const auto& channel : sctp_data_channels_) {
+ channel->OnTransportChannelCreated();
+ }
+}
+
+void PeerConnection::OnDataChannelDestroyed() {
+ // Use a temporary copy of the RTP/SCTP DataChannel list because the
+ // DataChannel may callback to us and try to modify the list.
+ std::map<std::string, rtc::scoped_refptr<DataChannel>> temp_rtp_dcs;
+ temp_rtp_dcs.swap(rtp_data_channels_);
+ for (const auto& kv : temp_rtp_dcs) {
+ kv.second->OnTransportChannelDestroyed();
+ }
+
+ std::vector<rtc::scoped_refptr<DataChannel>> temp_sctp_dcs;
+ temp_sctp_dcs.swap(sctp_data_channels_);
+ for (const auto& channel : temp_sctp_dcs) {
+ channel->OnTransportChannelDestroyed();
+ }
+}
+
+void PeerConnection::OnDataChannelOpenMessage(
+ const std::string& label,
+ const InternalDataChannelInit& config) {
+ rtc::scoped_refptr<DataChannel> channel(
+ InternalCreateDataChannel(label, &config));
+ if (!channel.get()) {
+ LOG(LS_ERROR) << "Failed to create DataChannel from the OPEN message.";
+ return;
+ }
+
+ observer_->OnDataChannel(
+ DataChannelProxy::Create(signaling_thread(), channel));
+}
+
std::vector<rtc::scoped_refptr<RtpSenderInterface>>::iterator
PeerConnection::FindSenderForTrack(MediaStreamTrackInterface* track) {
return std::find_if(
@@ -1016,4 +1918,42 @@ PeerConnection::FindReceiverForTrack(MediaStreamTrackInterface* track) {
});
}
+PeerConnection::TrackInfos* PeerConnection::GetRemoteTracks(
+ cricket::MediaType media_type) {
+ RTC_DCHECK(media_type == cricket::MEDIA_TYPE_AUDIO ||
+ media_type == cricket::MEDIA_TYPE_VIDEO);
+ return (media_type == cricket::MEDIA_TYPE_AUDIO) ? &remote_audio_tracks_
+ : &remote_video_tracks_;
+}
+
+PeerConnection::TrackInfos* PeerConnection::GetLocalTracks(
+ cricket::MediaType media_type) {
+ RTC_DCHECK(media_type == cricket::MEDIA_TYPE_AUDIO ||
+ media_type == cricket::MEDIA_TYPE_VIDEO);
+ return (media_type == cricket::MEDIA_TYPE_AUDIO) ? &local_audio_tracks_
+ : &local_video_tracks_;
+}
+
+const PeerConnection::TrackInfo* PeerConnection::FindTrackInfo(
+ const PeerConnection::TrackInfos& infos,
+ const std::string& stream_label,
+ const std::string track_id) const {
+ for (const TrackInfo& track_info : infos) {
+ if (track_info.stream_label == stream_label &&
+ track_info.track_id == track_id) {
+ return &track_info;
+ }
+ }
+ return nullptr;
+}
+
+DataChannel* PeerConnection::FindDataChannelBySid(int sid) const {
+ for (const auto& channel : sctp_data_channels_) {
+ if (channel->id() == sid) {
+ return channel;
+ }
+ }
+ return nullptr;
+}
+
} // namespace webrtc
« no previous file with comments | « talk/app/webrtc/peerconnection.h ('k') | talk/app/webrtc/peerconnectionendtoend_unittest.cc » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698