<?xmlversion="1.0" encoding="US-ASCII"?>version='1.0' encoding='utf-8'?> <!DOCTYPE rfc SYSTEM"rfc2629.dtd"> <?rfc toc="yes"?> <?rfc tocompact="yes"?> <?rfc tocdepth="3"?> <?rfc tocindent="yes"?> <?rfc symrefs="yes"?> <?rfc sortrefs="yes"?> <?rfc comments="yes"?> <?rfc inline="yes"?> <?rfc compact="yes"?> <?rfc subcompact="no"?>"rfc2629-xhtml.ent"> <rfc xmlns:xi="http://www.w3.org/2001/XInclude" category="std"docName="draft-ietf-rtcweb-rtp-usage-26" ipr="trust200902">number="8834" submissionType="IETF" consensus="true" obsoletes="" updates="" xml:lang="en" tocInclude="true" symRefs="true" sortRefs="true" version="3" ipr="trust200902" docName="draft-ietf-rtcweb-rtp-usage-26"> <!-- xml2rfc v2v3 conversion 2.34.0 --> <front> <title abbrev="RTP forWebRTC">Web Real-Time Communication (WebRTC): MediaWebRTC">Media Transport and Use ofRTP</title>RTP in WebRTC</title> <seriesInfo name="RFC" value="8834"/> <author fullname="Colin Perkins"initials="C. S."initials="C." surname="Perkins"> <organization>University of Glasgow</organization> <address> <postal> <street>School of Computing Science</street> <city>Glasgow</city> <code>G12 8QQ</code> <country>United Kingdom</country> </postal> <email>csp@csperkins.org</email> <uri>https://csperkins.org/</uri> </address> </author> <author fullname="Magnus Westerlund" initials="M." surname="Westerlund"> <organization>Ericsson</organization> <address> <postal> <street>Farogatan 6</street><city>SE-164 80 Kista</city><city>Kista</city> <code>164 80</code> <country>Sweden</country> </postal><phone>+46 10 714 82 87</phone><email>magnus.westerlund@ericsson.com</email> </address> </author> <authorfullname="Joergfullname="Jörg Ott" initials="J." surname="Ott"> <organization>Aalto University</organization> <address> <postal> <street>School of Electrical Engineering</street> <city>Espoo</city> <code>02150</code> <country>Finland</country> </postal> <email>jorg.ott@aalto.fi</email> </address> </author> <dateday="12" month="June" year="2015" />month="October" year="2020"/> <workgroup>RTCWEB Working Group</workgroup> <abstract> <t>The framework for Web Real-Time Communication (WebRTC)frameworkprovides support for direct interactive rich communication using audio, video, text, collaboration, games, etc. between two peers'web-browsers.web browsers. This memo describes the media transport aspects of the WebRTC framework. It specifies how the Real-time Transport Protocol (RTP) is used in the WebRTCcontext,context and gives requirements for which RTP features, profiles, and extensions need to be supported.</t> </abstract> </front> <middle> <sectiontitle="Introduction">numbered="true" toc="default"> <name>Introduction</name> <t>The <xreftarget="RFC3550">Real-timetarget="RFC3550" format="default">Real-time Transport Protocol (RTP)</xref> provides a framework for delivery of audio and video teleconferencing data and other real-time media applications. Previous work has defined the RTP protocol, along with numerous profiles, payload formats, and other extensions. When combined with appropriatesignalling,signaling, these form the basis for many teleconferencing systems.</t> <t>The Web Real-TimecommunicationCommunication (WebRTC) framework provides the protocol building blocks to support direct, interactive, real-time communication using audio, video, collaboration, games,etc.,etc. between two peers'web-browsers.web browsers. This memo describes how the RTP framework is to be used in the WebRTC context. It proposes a baseline set of RTP features that are to be implemented by all WebRTCEndpoints,endpoints, along with suggested extensions for enhanced functionality.</t> <t>This memo specifies a protocol intended for use within the WebRTCframework,framework but is not restricted to that context. An overview of the WebRTC framework is given in <xreftarget="I-D.ietf-rtcweb-overview"></xref>.</t>target="RFC8825" format="default"/>.</t> <t>The structure of this memo is as follows. <xreftarget="sec-rationale"></xref>target="sec-rationale" format="default"/> outlines our rationaleinfor preparing this memo and choosing these RTP features. <xreftarget="sec-terminology"></xref>target="sec-terminology" format="default"/> defines terminology. Requirements for core RTP protocols are described in <xreftarget="sec-rtp-core"></xref>target="sec-rtp-core" format="default"/>, and suggested RTP extensions are described in <xreftarget="sec-rtp-extn"></xref>.target="sec-rtp-extn" format="default"/>. <xreftarget="sec-rtp-robust"></xref>target="sec-rtp-robust" format="default"/> outlines mechanisms that can increase robustness to network problems, while <xreftarget="sec-rate-control"></xref>target="sec-rate-control" format="default"/> describes congestion control and rate adaptation mechanisms. The discussion of mandated RTP mechanisms concludes in <xreftarget="sec-perf"></xref>target="sec-perf" format="default"/> with a review of performance monitoring and network management tools. <xreftarget="sec-extn"></xref>target="sec-extn" format="default"/> gives some guidelines for future incorporation of other RTP and RTP Control Protocol (RTCP) extensions into this framework. <xreftarget="sec-signalling"></xref>target="sec-signalling" format="default"/> describes requirements placed on thesignallingsignaling channel. <xreftarget="sec-webrtc-api"></xref>target="sec-webrtc-api" format="default"/> discusses the relationship between features of the RTP framework and the WebRTC application programming interface (API), and <xreftarget="sec-rtp-func"></xref>target="sec-rtp-func" format="default"/> discusses RTP implementation considerations. The memo concludes with <xreftarget="sec-security">securitytarget="sec-security" format="default">security considerations</xref> and <xreftarget="sec-iana">IANAtarget="sec-iana" format="default">IANA considerations</xref>.</t> </section> <section anchor="sec-rationale"title="Rationale">numbered="true" toc="default"> <name>Rationale</name> <t>The RTP framework comprises the RTP data transfer protocol, the RTP control protocol, and numerous RTP payload formats, profiles, and extensions. This range of add-ons has allowed RTP to meet various needs that were not envisaged by the original protocoldesigners,designers andtosupport many new media encodings, but it raises the question of what extensions are to be supported by new implementations. The development of the WebRTC framework provides an opportunity to review the available RTP features andextensions,extensions andtodefine a common baseline RTP feature set for all WebRTCEndpoints.endpoints. This builds on the past 20 years of RTP development to mandate the use of extensions that have shown widespread utility, while still remaining compatible with the wide installed base of RTP implementations where possible.</t> <t>RTP and RTCP extensions that are not discussed in this document can be implemented by WebRTCEndpointsendpoints if they are beneficial for new use cases. However, they are not necessary to address the WebRTC use cases and requirements identified in <xreftarget="RFC7478"></xref>.</t>target="RFC7478" format="default"/>.</t> <t>While the baseline set of RTP features and extensions defined in this memo is targeted at the requirements of the WebRTC framework, it is expected to be broadly useful for other conferencing-related uses of RTP. In particular, it is likely that this set of RTP features and extensions will be appropriate for other desktop or mobilevideo conferencingvideo-conferencing systems, or for room-based high-quality telepresence applications.</t> </section> <section anchor="sec-terminology"title="Terminology"> <t>Thenumbered="true" toc="default"> <name>Terminology</name> <t> The key words"MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY","<bcp14>MUST</bcp14>", "<bcp14>MUST NOT</bcp14>", "<bcp14>REQUIRED</bcp14>", "<bcp14>SHALL</bcp14>", "<bcp14>SHALL NOT</bcp14>", "<bcp14>SHOULD</bcp14>", "<bcp14>SHOULD NOT</bcp14>", "<bcp14>RECOMMENDED</bcp14>", "<bcp14>NOT RECOMMENDED</bcp14>", "<bcp14>MAY</bcp14>", and"OPTIONAL""<bcp14>OPTIONAL</bcp14>" in this document are to be interpreted as described in BCP 14 <xreftarget="RFC2119"></xref>. The RFC 2119 interpretation of these key words appliestarget="RFC2119"/> <xref target="RFC8174"/> when, and onlywhen writtenwhen, they appear inALL CAPS.all capitals, as shown here. Lower- or mixed-case uses of these key words are not to be interpreted as carrying special significance in thismemo.</t>memo. </t> <t>We define the following additionalterms:<list style="hanging"> <t hangText="WebRTC MediaStream:">Theterms:</t> <dl newline="false" spacing="normal"> <dt>WebRTC MediaStream:</dt> <dd>The MediaStream concept defined by the W3C in the <xreftarget="W3C.WD-mediacapture-streams-20130903">WebRTCtarget="W3C-MEDIA-CAPTURE" format="default">WebRTC API</xref>. A MediaStream consists of zero or moreMediaStreamTracks.</t> <t hangText="MediaStreamTrack:">PartMediaStreamTracks.</dd> <dt>MediaStreamTrack:</dt> <dd>Part of the MediaStream concept defined by the W3C in the <xreftarget="W3C.WD-mediacapture-streams-20130903">WebRTCtarget="W3C-MEDIA-CAPTURE" format="default">WebRTC API</xref>. A MediaStreamTrack is an individual stream of media from any type of media sourcelikesuch as a microphone or a camera, butalsoconceptualsources, like asources such as an audio mix or a videocomposition,composition arepossible.</t> <t hangText="Transport-layer Flow:">A uni-directionalalso possible.</dd> <dt>Transport-layer flow:</dt> <dd>A unidirectional flow of transport packets that are identified byhavinga particular 5-tuple of source IP address, source port, destination IP address, destination port, and transportprotocol used.</t> <t hangText="Bi-directional Transport-layer Flow:">A bi-directionalprotocol.</dd> <dt>Bidirectional transport-layer flow:</dt> <dd>A bidirectional transport-layer flow is a transport-layer flow that is symmetric. That is, the transport-layer flow in the reverse direction has a 5-tuple where the source and destination address and ports are swapped compared to the forward path transport-layer flow, and the transport protocol is thesame.</t> </list></t>same.</dd> </dl> <t>This document uses the terminology from <xreftarget="I-D.ietf-avtext-rtp-grouping-taxonomy"></xref>target="RFC7656" format="default"/> and <xreftarget="I-D.ietf-rtcweb-overview"></xref>.target="RFC8825" format="default"/>. Other terms are used according to their definitions from the <xreftarget="RFC3550">RTP Specification</xref>. Especiallytarget="RFC3550" format="default">RTP specification</xref>. In particular, note the following frequently used terms: RTPStream,stream, RTPSession,session, andEndpoint.</t>endpoint.</t> </section> <section anchor="sec-rtp-core"title="WebRTCnumbered="true" toc="default"> <name>WebRTC Use of RTP: CoreProtocols">Protocols</name> <t>The following sections describe the core features of RTP and RTCP that need to be implemented, along with the mandated RTP profiles. Also described are the core extensions providing essential features that all WebRTCEndpointsendpoints need to implement to function effectively on today's networks.</t> <section anchor="sec-rtp-rtcp"title="RTPnumbered="true" toc="default"> <name>RTP andRTCP">RTCP</name> <t>The <xreftarget="RFC3550">Real-timetarget="RFC3550" format="default">Real-time Transport Protocol (RTP) </xref> isREQUIRED<bcp14>REQUIRED</bcp14> to be implemented as the media transport protocol for WebRTC. RTP itself comprises two parts: the RTP data transferprotocol,protocol and the RTPcontrol protocolControl Protocol (RTCP). RTCP is a fundamental and integral part ofRTP,RTP andMUST<bcp14>MUST</bcp14> be implemented and used in all WebRTCEndpoints.</t>endpoints.</t> <t>The following RTP and RTCP features are sometimes omitted inlimited functionalitylimited-functionality implementations of RTP, but they areREQUIRED<bcp14>REQUIRED</bcp14> in all WebRTCEndpoints: <list style="symbols"> <t>Supportendpoints: </t> <ul spacing="normal"> <li>Support for use of multiple simultaneousSSRCsynchronization source (SSRC) values in a single RTP session, including support for RTP endpoints that send many SSRC values simultaneously, following <xreftarget="RFC3550"></xref>target="RFC3550" format="default"/> and <xreftarget="I-D.ietf-avtcore-rtp-multi-stream"></xref>.target="RFC8108" format="default"/>. The RTCPoptimisationsoptimizations for multi-SSRC sessions defined in <xreftarget="I-D.ietf-avtcore-rtp-multi-stream-optimisation"></xref> MAYtarget="RFC8861" format="default"/> <bcp14>MAY</bcp14> be supported; ifsupportedsupported, the usageMUST<bcp14>MUST</bcp14> besignalled.</t> <t>Randomsignaled.</li> <li>Random choice of SSRC on joining a session; collision detection and resolution for SSRC values (see also <xreftarget="sec-ssrc"></xref>).</t> <t>Supporttarget="sec-ssrc" format="default"/>).</li> <li>Support for reception of RTP data packets containingCSRCcontributing source (CSRC) lists, as generated by RTP mixers, and RTCP packets relating toCSRCs.</t> <t>SendingCSRCs.</li> <li>Sending correctsynchronisationsynchronization information in the RTCP Sender Reports, to allow receivers to implementlip-synchronisation;lip synchronization; see <xreftarget="rapid-sync"></xref>target="rapid-sync" format="default"/> regarding support for the rapid RTPsynchronisation extensions.</t> <t>Supportsynchronization extensions.</li> <li>Support for multiplesynchronisationsynchronization contexts. Participants that send multiple simultaneous RTP packet streamsSHOULD<bcp14>SHOULD</bcp14> do so as part of a singlesynchronisationsynchronization context, using a single RTCP CNAME for all streams and allowing receivers to play the streams out in asynchronisedsynchronized manner. For compatibility with potential future versions of this specification, or for interoperability with non-WebRTC devices through a gateway, receiversMUST<bcp14>MUST</bcp14> support multiplesynchronisationsynchronization contexts, indicated by the use of multiple RTCP CNAMEs in an RTP session. This specification mandates the usage of a single CNAME when sending RTPStreamsstreams in somecircumstances,circumstances; see <xreftarget="sec-cname"></xref>.</t> <t>Supporttarget="sec-cname" format="default"/>.</li> <li>Support for sending and receiving RTCP SR, RR,SDES,Source Description (SDES), and BYE packet types. Note that support for other RTCP packet types isOPTIONAL,<bcp14>OPTIONAL</bcp14> unless mandated by other parts of this specification. Note that additional RTCPPacketpacket types are used by the <xreftarget="sec-profile">RTP/SAVPF Profile</xref>target="sec-profile" format="default">RTP/SAVPF profile</xref> and the other <xreftarget="sec-rtp-extn">RTCPtarget="sec-rtp-extn" format="default">RTCP extensions</xref>. WebRTC endpoints that implement theSDPSession Description Protocol (SDP) bundle negotiation extension will use the SDPgrouping framework 'mid'Grouping Framework "mid" attribute to identify media streams. Such endpointsMUST<bcp14>MUST</bcp14> implement the RTCP SDESMIDmedia identification (MID) item described in <xreftarget="I-D.ietf-mmusic-sdp-bundle-negotiation"></xref>.</t> <t>Supporttarget="RFC8843" format="default"/>.</li> <li>Support for multiple endpoints in a single RTP session, and for scaling the RTCP transmission interval according to the number of participants in the session; support forrandomisedrandomized RTCP transmission intervals to avoidsynchronisationsynchronization of RTCP reports; support for RTCP timer reconsideration(Section 6.3.6 of <xref target="RFC3550"></xref>)(<xref target="RFC3550" section="6.3.6" sectionFormat="of"/>) and reverse reconsideration(Section 6.3.4 of <xref target="RFC3550"></xref>).</t> <t>Support(<xref target="RFC3550" sectionFormat="of" section="6.3.4"/>).</li> <li>Support for configuring the RTCP bandwidth as a fraction of the media bandwidth, and for configuring the fraction of the RTCP bandwidth allocated tosenders,senders -- e.g., using the SDP "b=" line <xreftarget="RFC4566"></xref><xref target="RFC3556"></xref>.</t> <t>Supporttarget="RFC4566" format="default"/> <xref target="RFC3556" format="default"/>.</li> <li>Support for the reduced minimum RTCP reporting interval described inSection 6.2 of<xreftarget="RFC3550"></xref>.target="RFC3550" sectionFormat="of" section="6.2"/>. When using the reduced minimum RTCP reporting interval, the fixed(non-reduced)(nonreduced) minimum intervalMUST<bcp14>MUST</bcp14> be used when calculating the participant timeout interval (see Sections6.2<xref target="RFC3550" section="6.2" sectionFormat="bare"/> and6.3.5<xref target="RFC3550" section="6.3.5" sectionFormat="bare"/> of <xreftarget="RFC3550"></xref>).target="RFC3550" format="default"/>). The delay before sending the initial compound RTCP packet can be set to zero (seeSection 6.2 of<xreftarget="RFC3550"></xref>target="RFC3550" section="6.2" sectionFormat="of"/> as updated by <xreftarget="I-D.ietf-avtcore-rtp-multi-stream"></xref>).</t> <t>Supporttarget="RFC8108" format="default"/>).</li> <li>Support for discontinuous transmission. RTP allows endpoints to pause and resume transmission at any time. When resuming, the RTP sequence number will increase by one, as usual, while the increase in the RTP timestamp value will depend on the duration of the pause. Discontinuous transmission is most commonly used with some audio payload formats, but it is not audiospecific,specific and can be used with any RTP payloadformat.</t> <t>Ignoreformat.</li> <li>Ignore unknown RTCP packet types and RTP header extensions. This is to ensure robust handling of future extensions, middleboxbehaviours,behaviors, etc., that can result innot signalled RTCP packet types orreceiving RTP header extensionsbeing received.or RTCP packet types that were not signaled. If a compound RTCP packetis receivedthat contains a mixture of known and unknown RTCP packettypes,types is received, the knownpacketspacket types need to be processed as usual, with only the unknown packet types beingdiscarded.</t> </list></t>discarded.</li> </ul> <t>It is known that a significant number of legacy RTP implementations, especially those targeted atVoIP-only systems,systems with only Voice over IP (VoIP), do not support all of the abovefeatures,features and in some cases do not support RTCP at all. Implementers are advised to consider the requirements for graceful degradation when interoperating with legacy implementations.</t> <t>Other implementation considerations are discussed in <xreftarget="sec-rtp-func"></xref>.</t>target="sec-rtp-func" format="default"/>.</t> </section> <section anchor="sec-profile"title="Choicenumbered="true" toc="default"> <name>Choice of the RTPProfile">Profile</name> <t>The complete specification of RTP for a particular application domain requires the choice of an RTPProfile.profile. For WebRTC use, the <xreftarget="RFC5124">Extended Securetarget="RFC5124" format="default">extended secure RTPProfileprofile forRTCP-Based FeedbackRTCP-based feedback (RTP/SAVPF)</xref>, as extended by <xreftarget="RFC7007"></xref>, MUSTtarget="RFC7007" format="default"/>, <bcp14>MUST</bcp14> be implemented. The RTP/SAVPF profile is the combination of the basic <xreftarget="RFC3551">RTP/AVPtarget="RFC3551" format="default">RTP/AVP profile</xref>, the <xreftarget="RFC4585">RTPtarget="RFC4585" format="default">RTP profile for RTCP-based feedback (RTP/AVPF)</xref>, and the <xreftarget="RFC3711">securetarget="RFC3711" format="default">secure RTP profile (RTP/SAVP)</xref>.</t> <t>The RTCP-based feedback extensions <xreftarget="RFC4585"></xref>target="RFC4585" format="default"/> are needed for the improved RTCP timer model. This allows more flexible transmission of RTCP packets in response to events, rather than strictly according to bandwidth, and is vital for being able to report congestion signals as well as media events. These extensions also allow saving RTCP bandwidth, and an endpoint will commonly only use the full RTCP bandwidth allocation if there are many events that require feedback. The timer rules are also needed to make use of the RTP conferencing extensions discussed in <xreftarget="conf-ext"></xref>.</t> <t><list style="empty"> <t>Note:target="conf-ext" format="default"/>.</t> <aside><t>Note: The enhanced RTCP timer model defined in the RTP/AVPF profile is backwards compatible with legacy systems that implement only the RTP/AVP or RTP/SAVP profile, given some constraints on parameter configuration such as the RTCP bandwidth value and"trr-int" (the"trr&nbhy;int". The most important factor for interworking with RTP/(S)AVP endpoints via a gateway is to set thetrr-int"trr-int" parameter to a value representing 4seconds,seconds; seeSection 6.1 in<xreftarget="I-D.ietf-avtcore-rtp-multi-stream"></xref>).</t> </list></t>target="RFC8108" section="7.1.3" sectionFormat="of"/>.</t> </aside> <t>The secure RTP (SRTP) profile extensions <xreftarget="RFC3711"></xref>target="RFC3711" format="default"/> are needed to provide media encryption, integrity protection, replayprotectionprotection, and a limited form of source authentication. WebRTCEndpoints MUST NOTendpoints <bcp14>MUST NOT</bcp14> send packets using the basic RTP/AVP profile or the RTP/AVPF profile; theyMUST<bcp14>MUST</bcp14> employ the full RTP/SAVPF profile to protect all RTP and RTCP packets that aregenerated (i.e.,generated. In other words, implementationsMUST<bcp14>MUST</bcp14> use SRTP andSRTCP).SRTCP. The RTP/SAVPF profileMUST<bcp14>MUST</bcp14> be configured using the cipher suites, DTLS-SRTP protection profiles, keying mechanisms, and other parameters described in <xreftarget="I-D.ietf-rtcweb-security-arch"></xref>.</t>target="RFC8827" format="default"/>.</t> </section> <section anchor="sec.codecs"title="Choicenumbered="true" toc="default"> <name>Choice of RTP PayloadFormats"> <t>Mandatory to implementFormats</name> <t>Mandatory-to-implement audio codecs and RTP payload formats for WebRTC endpoints are defined in <xreftarget="I-D.ietf-rtcweb-audio"></xref>. Mandatory to implementtarget="RFC7874" format="default"/>. Mandatory-to-implement video codecs and RTP payload formats for WebRTC endpoints are defined in <xreftarget="I-D.ietf-rtcweb-video"></xref>.target="RFC7742" format="default"/>. WebRTC endpointsMAY<bcp14>MAY</bcp14> additionally implement any other codec for which an RTP payload format and associatedsignallingsignaling has been defined.</t> <t>WebRTCEndpointsendpoints cannot assume that the other participants in an RTP session understand any RTP payload format, no matter how common. The mapping between RTP payload type numbers and specific configurations of particular RTP payload formatsMUST<bcp14>MUST</bcp14> be agreed before those payload types/formats can be used. In an SDP context, this can be done using the "a=rtpmap:" and "a=fmtp:" attributes associated with an "m=" line, along with any other SDP attributes needed to configure the RTP payload format.</t> <t>Endpoints can signal support for multiple RTP payloadformats,formats or multiple configurations of a single RTP payload format, as long as each unique RTP payload format configuration uses a different RTP payload type number. As outlined in <xreftarget="sec-ssrc"></xref>,target="sec-ssrc" format="default"/>, the RTP payload type number is sometimes used to associate an RTP packet stream with asignallingsignaling context. This association is possible provided unique RTP payload type numbers are used in each context. For example, an RTP packet stream can be associated with an SDP "m=" line by comparing the RTP payload type numbers used by the RTP packet stream with payload typessignalledsignaled in the "a=rtpmap:" lines in the media sections of the SDP. This leads to the followingconsiderations:<list style="empty"> <t>Ifconsiderations:</t> <ul empty="true" spacing="normal"> <li>If RTP packet streams are being associated withsignallingsignaling contexts based on the RTP payload type, then the assignment of RTP payload type numbersMUST<bcp14>MUST</bcp14> be unique acrosssignalling contexts.</t> <t>Ifsignaling contexts.</li> <li>If the same RTP payload format configuration is used in multiple contexts, then a different RTP payload type number has to be assigned in each context to ensureuniqueness.</t> <t>Ifuniqueness.</li> <li>If the RTP payload type number is not being used to associate RTP packet streams with asignallingsignaling context, then the same RTP payload type number can be used to indicate the exact same RTP payload format configuration in multiplecontexts.</t> </list>Acontexts.</li> </ul> <t>A single RTP payload type numberMUST NOT<bcp14>MUST NOT</bcp14> be assigned to different RTP payload formats, or different configurations of the same RTP payload format, within a single RTP session (note that the "m=" lines in an <xreftarget="I-D.ietf-mmusic-sdp-bundle-negotiation">SDP bundletarget="RFC8843" format="default">SDP BUNDLE group</xref> form a single RTP session).</t> <t>An endpoint that hassignalledsignaled support for multiple RTP payload formatsMUST<bcp14>MUST</bcp14> be able to accept data in any of those payload formats at any time, unless it has previouslysignalledsignaled limitations on its decoding capability. This requirement is constrained if several types of media (e.g., audio and video) are sent in the same RTP session. In such a case, a source (SSRC) is restricted to switching only between the RTP payload formatssignalledsignaled for the type of media that is being sent by that source; see <xreftarget="sec.session-mux"></xref>.target="sec.session-mux" format="default"/>. To support rapid rate adaptation by changingcodec,codecs, RTP does not require advancesignallingsignaling for changes between RTP payload formats used by a single SSRC that weresignalledsignaled during sessionset-up.</t>setup.</t> <t>If performing changes between two RTP payload types that use different RTP clock rates, an RTP senderMUST<bcp14>MUST</bcp14> follow the recommendations inSection 4.1 of<xreftarget="RFC7160"></xref>.target="RFC7160" section="4.1" sectionFormat="of"/>. RTP receiversMUST<bcp14>MUST</bcp14> follow the recommendations in Section 4.3 of <xreftarget="RFC7160"></xref>target="RFC7160" format="default"/> in order to support sources that switch between clock rates in an RTPsession (thesesession. These recommendations for receivers are backwards compatible with the case where senders use only a single clockrate).</t>rate.</t> </section> <section anchor="sec.session-mux"title="Usenumbered="true" toc="default"> <name>Use of RTPSessions">Sessions</name> <t>An association amongst a set of endpoints communicating using RTP is known as an RTP session <xreftarget="RFC3550"></xref>.target="RFC3550" format="default"/>. An endpoint can be involved in several RTP sessions at the same time. In a multimedia session, each type of media has typically been carried in a separate RTP session (e.g., using one RTP session for theaudio,audio and a separate RTP session using a different transport-layer flow for the video). WebRTCEndpointsendpoints areREQUIRED<bcp14>REQUIRED</bcp14> to implement support for multimedia sessions in this way, separating each RTP session using different transport-layer flows for compatibility with legacy systems (this is sometimes called session multiplexing).</t> <t>Inmodern daymodern-day networks, however, with the widespread use of network address/port translators (NAT/NAPT) and firewalls, it is desirable to reduce the number of transport-layer flows used by RTP applications. This can be done by sending all the RTP packet streams in a single RTP session, which will comprise a single transport-layerflow (thisflow. This will prevent the use of some quality-of-service mechanisms, as discussed in <xreftarget="sec-differentiated"></xref>).target="sec-differentiated" format="default"/>. Implementations are therefore alsoREQUIRED<bcp14>REQUIRED</bcp14> to support transport of all RTP packet streams, independent of media type, in a single RTP session using a singletransport layertransport-layer flow, according to <xreftarget="I-D.ietf-avtcore-multi-media-rtp-session"></xref>target="RFC8860" format="default"/> (this is sometimes called SSRC multiplexing). If multiple types of media are to be used in a single RTP session, all participants in that RTP sessionMUST<bcp14>MUST</bcp14> agree to this usage. In an SDP context, the mechanisms described in <xreftarget="I-D.ietf-mmusic-sdp-bundle-negotiation"></xref>target="RFC8843" format="default"/> can be used to signal such a bundle of RTP packet streams forming a single RTP session.</t> <t>Further discussion about the suitability of different RTP session structures and multiplexing methods to different scenarios can be found in <xreftarget="I-D.ietf-avtcore-multiplex-guidelines"></xref>.</t>target="I-D.ietf-avtcore-multiplex-guidelines" format="default"/>.</t> </section> <section anchor="sec.rtcp-mux"title="RTPnumbered="true" toc="default"> <name>RTP and RTCPMultiplexing">Multiplexing</name> <t>Historically, RTP and RTCP have been run on separatetransport layertransport-layer flows (e.g., two UDP ports for each RTP session, oneportfor RTP and oneportfor RTCP). With the increased use of Network Address/Port Translation(NAT/NAPT)(NAT/NAPT), this has become problematic, since maintaining multiple NAT bindings can be costly. It also complicates firewall administration, since multiple ports need to be opened to allow RTP traffic. To reduce these costs and sessionset-upsetup times, implementations areREQUIRED<bcp14>REQUIRED</bcp14> to support multiplexing RTP data packets and RTCP control packets on a single transport-layer flow <xreftarget="RFC5761"></xref>.target="RFC5761" format="default"/>. Such RTP and RTCP multiplexingMUST<bcp14>MUST</bcp14> be negotiated in thesignallingsignaling channel before it is used. If SDP is used forsignalling,signaling, this negotiationMUST<bcp14>MUST</bcp14> use the mechanism defined in <xreftarget="RFC5761"/>.target="RFC5761" format="default"/>. Implementations can also support sending RTP and RTCP on separate transport-layer flows, but this isOPTIONAL<bcp14>OPTIONAL</bcp14> to implement. If an implementation does not support RTP and RTCP sent on separate transport-layer flows, itMUST<bcp14>MUST</bcp14> indicate that using the mechanism defined in <xreftarget="I-D.ietf-mmusic-mux-exclusive"/>.target="RFC8858" format="default"/>. </t> <t>Note that the use of RTP and RTCP multiplexed onto a single transport-layer flow ensures that there is occasional traffic sent on that port, even if there is no active media traffic. This can be useful to keep NAT bindings alive <xreftarget="RFC6263"></xref>.</t>target="RFC6263" format="default"/>.</t> </section> <sectiontitle="Reducednumbered="true" toc="default"> <name>Reduced SizeRTCP">RTCP</name> <t>RTCP packets are usually sent as compound RTCP packets, and <xreftarget="RFC3550"></xref>target="RFC3550" format="default"/> requires that those compound packets start withana Sender Report (SR) or Receiver Report (RR) packet. When using frequent RTCP feedback messages under the RTP/AVPFProfileprofile <xreftarget="RFC4585"></xref>target="RFC4585" format="default"/>, these statistics are not needed in every packet, and they unnecessarily increase the mean RTCP packet size. This can limit the frequency at which RTCP packets can be sent within the RTCP bandwidth share.</t> <t>To avoid this problem, <xreftarget="RFC5506"></xref>target="RFC5506" format="default"/> specifies how to reduce the mean RTCP message size and allow for more frequent feedback. Frequent feedback, in turn, is essential to make real-time applications quickly aware of changing networkconditions,conditions and to allow them to adapt their transmission and encodingbehaviour.behavior. ImplementationsMUST<bcp14>MUST</bcp14> support sending and receivingnon-compoundnoncompound RTCP feedback packets <xreftarget="RFC5506"></xref>.target="RFC5506" format="default"/>. Use ofnon-compoundnoncompound RTCP packetsMUST<bcp14>MUST</bcp14> be negotiated using thesignallingsignaling channel. If SDP is used forsignalling,signaling, this negotiationMUST<bcp14>MUST</bcp14> use the attributes defined in <xreftarget="RFC5506"></xref>.target="RFC5506" format="default"/>. For backwards compatibility, implementations are alsoREQUIRED<bcp14>REQUIRED</bcp14> to support the use of compound RTCP feedback packets if the remote endpoint does not agree to the use ofnon-compoundnoncompound RTCP in thesignallingsignaling exchange.</t> </section> <sectiontitle="Symmetric RTP/RTCP">numbered="true" toc="default"> <name>Symmetric RTP/RTCP</name> <t>To ease traversal of NAT and firewall devices, implementations areREQUIRED<bcp14>REQUIRED</bcp14> to implement and use <xreftarget="RFC4961">Symmetrictarget="RFC4961" format="default">symmetric RTP</xref>. The reason for using symmetric RTP is primarily to avoid issues with NATs andFirewallsfirewalls by ensuring that the send and receive RTP packet streams, as well as RTCP, are actuallybi-directionalbidirectional transport-layer flows. This will keep alive the NAT and firewallpinholes,pinholes and help indicate consent that the receive direction is a transport-layer flow the intended recipient actually wants. In addition, it saves resources, specifically ports at the endpoints, but also in thenetwork asnetwork, because the NAT mappings or firewall state is notunnecessaryunnecessarily bloated. The amount ofper flowper-flow QoS state kept in the network is also reduced.</t> </section> <section anchor="sec-ssrc"title="Choicenumbered="true" toc="default"> <name>Choice of RTPSynchronisationSynchronization Source(SSRC)">(SSRC)</name> <t>Implementations areREQUIRED<bcp14>REQUIRED</bcp14> to supportsignalledsignaled RTPsynchronisationsynchronization source (SSRC) identifiers. If SDP is used, thisMUST<bcp14>MUST</bcp14> be done using the "a=ssrc:" SDP attribute defined inSection 4.1Sections <xref target="RFC5576" sectionFormat="bare" section="4.1" format="default"/> andSection 5<xref target="RFC5576" sectionFormat="bare" section="5" format="default"/> of <xreftarget="RFC5576"></xref>target="RFC5576"/> and the "previous-ssrc" source attribute defined inSection 6.2 of<xreftarget="RFC5576"></xref>;target="RFC5576" sectionFormat="of" section="6.2" format="default"/>; other per-SSRC attributes defined in <xreftarget="RFC5576"></xref> MAYtarget="RFC5576" format="default"/> <bcp14>MAY</bcp14> be supported.</t> <t>While support forsignalledsignaled SSRC identifiers is mandated, their use in an RTP session isOPTIONAL.<bcp14>OPTIONAL</bcp14>. ImplementationsMUST<bcp14>MUST</bcp14> be prepared to accept RTP and RTCP packets using SSRCs that have not been explicitlysignalledsignaled ahead of time. ImplementationsMUST<bcp14>MUST</bcp14> support random SSRCassignment,assignment andMUST<bcp14>MUST</bcp14> support SSRC collision detection and resolution, according to <xreftarget="RFC3550"></xref>.target="RFC3550" format="default"/>. When usingsignalledsignaled SSRC values, collision detectionMUST<bcp14>MUST</bcp14> be performed as described inSection 5 of<xreftarget="RFC5576"></xref>.</t>target="RFC5576" sectionFormat="of" section="5" format="default"/>.</t> <t>It is often desirable to associate an RTP packet stream with a non-RTP context. For users of the WebRTCAPIAPI, a mapping between SSRCs and MediaStreamTracksareis provided per <xreftarget="sec-webrtc-api"></xref>.target="sec-webrtc-api" format="default"/>. For gateways or otherusagesusages, it is possible to associate an RTP packet stream with an "m=" line in a session description formatted using SDP. If SSRCs aresignalledsignaled, this is straightforward (inSDPSDP, the "a=ssrc:" line will be at the media level, allowing a direct association with an "m=" line). If SSRCs are notsignalled,signaled, the RTP payload type numbers used in an RTP packet stream are often sufficient to associate that packet stream with asignalling context (e.g.,signaling context. For example, if RTP payload type numbers are assigned as described in <xreftarget="sec.codecs"></xref>target="sec.codecs" format="default"/> of this memo, the RTP payload types used by an RTP packet stream can be compared with values in SDP "a=rtpmap:" lines, which are at the media level inSDP,SDP and so map to an "m="line).</t>line.</t> </section> <section anchor="sec-cname"title="Generationnumbered="true" toc="default"> <name>Generation of the RTCP Canonical Name(CNAME)">(CNAME)</name> <t>The RTCP Canonical Name (CNAME) provides a persistent transport-level identifier for an RTP endpoint. While theSynchronisation Source (SSRC)SSRC identifier for an RTP endpoint can change if a collision isdetected,detected or when the RTP application is restarted, its RTCP CNAME is meant to stay unchanged for the duration ofaan <xreftarget="W3C.WD-webrtc-20130910">RTCPeerConnection</xref>,target="W3C.WebRTC" format="default">RTCPeerConnection</xref>, so that RTP endpoints can be uniquely identified and associated with their RTP packet streams within a set of related RTP sessions.</t> <t>Each RTP endpointMUST<bcp14>MUST</bcp14> have at least one RTCP CNAME, and that RTCP CNAMEMUST<bcp14>MUST</bcp14> be unique within the RTCPeerConnection. RTCP CNAMEs identify a particularsynchronisation context,synchronization context -- i.e., all SSRCs associated with a single RTCP CNAME share a common reference clock. If an endpoint has SSRCs that are associated with severalunsynchronisedunsynchronized reference clocks, and hence differentsynchronisationsynchronization contexts, it will need to use multiple RTCP CNAMEs, one for eachsynchronisationsynchronization context.</t> <t>Taking the discussion in <xreftarget="sec-webrtc-api"></xref>target="sec-webrtc-api" format="default"/> into account, a WebRTCEndpoint MUST NOTendpoint <bcp14>MUST NOT</bcp14> use more than one RTCP CNAME in the RTP sessions belonging to a single RTCPeerConnection (that is, an RTCPeerConnection forms asynchronisationsynchronization context). RTP middleboxesMAY<bcp14>MAY</bcp14> generate RTP packet streams associated with more than one RTCP CNAME, to allow them to avoid having to resynchronize media from multiple different endpoints that are part of amulti-partymultiparty RTP session.</t> <t>The <xreftarget="RFC3550">RTPtarget="RFC3550" format="default">RTP specification</xref> includes guidelines for choosing a unique RTP CNAME, but these are not sufficient in the presence of NAT devices. In addition, long-term persistent identifiers can be problematic from a <xreftarget="sec-security">privacytarget="sec-security" format="default">privacy viewpoint</xref>. Accordingly, a WebRTCEndpoint MUSTendpoint <bcp14>MUST</bcp14> generate a new, unique, short-term persistent RTCP CNAME for each RTCPeerConnection, following <xreftarget="RFC7022"></xref>,target="RFC7022" format="default"/>, with a single exception; if explicitly requested atcreationcreation, an RTCPeerConnectionMAY<bcp14>MAY</bcp14> use the same CNAME asasan existing RTCPeerConnection within their common same-origin context.</t><t>An<t>A WebRTCEndpoint MUSTendpoint <bcp14>MUST</bcp14> support reception of any CNAME that matches the syntax limitations specified by the <xreftarget="RFC3550">RTPtarget="RFC3550" format="default">RTP specification</xref> and cannot assume that any CNAME will be chosen according to the form suggested above.</t> </section> <section anchor="sec-leap-sec"title="Handlingnumbered="true" toc="default"> <name>Handling of LeapSeconds">Seconds</name> <t>The guidelines given in <xref target="RFC7164" format="default"/> regarding handling of leap seconds to limit their impact on RTP media play-out and synchronizationgiven in <xref target="RFC7164"></xref> SHOULD<bcp14>SHOULD</bcp14> be followed.</t> </section> </section> <section anchor="sec-rtp-extn"title="WebRTCnumbered="true" toc="default"> <name>WebRTC Use of RTP:Extensions">Extensions</name> <t>There are a number of RTP extensions that are either needed to obtain full functionality, or extremely useful to improve on the baseline performance, in the WebRTC context. One set of these extensions is related to conferencing, while others are more generic in nature. The following subsections describe the various RTP extensions mandated or suggested for use within WebRTC.</t> <section anchor="conf-ext"title="Conferencingnumbered="true" toc="default"> <name>Conferencing Extensions andTopologies">Topologies</name> <t>RTP is a protocol that inherently supports group communication. Groups can be implemented by having each endpoint send its RTP packet streams to an RTP middlebox that redistributes the traffic, by using a mesh of unicast RTP packet streams between endpoints, or by using an IP multicast group to distribute the RTP packet streams. These topologies can be implemented in a number of ways as discussed in <xreftarget="I-D.ietf-avtcore-rtp-topologies-update"></xref>.</t>target="RFC7667" format="default"/>.</t> <t>While the use of IP multicast groups is popular in IPTV systems, the topologies based on RTP middleboxes are dominant in interactivevideo conferencingvideo-conferencing environments. Topologies based on a mesh of unicast transport-layer flows to create a common RTP session have not seen widespread deployment to date. Accordingly, WebRTCEndpointsendpoints are not expected to support topologies based on IP multicast groups orto supportmesh-based topologies, such as a point-to-multipoint mesh configured as a single RTP session(Topo-Mesh("Topo-Mesh" in the terminology of <xreftarget="I-D.ietf-avtcore-rtp-topologies-update"></xref>).target="RFC7667" format="default"/>). However, a point-to-multipoint mesh constructed using several RTP sessions, implemented in WebRTC using independent <xreftarget="W3C.WD-webrtc-20130910">RTCPeerConnections</xref>,target="W3C.WebRTC" format="default">RTCPeerConnections</xref>, can be expected to be used inWebRTC,WebRTC and needs to be supported.</t> <t>WebRTCEndpointsendpoints implemented according to this memo are expected to support all the topologies described in <xreftarget="I-D.ietf-avtcore-rtp-topologies-update"></xref>target="RFC7667" format="default"/> where the RTP endpoints send and receive unicast RTP packet streams to and from some peer device, provided that peer can participate in performing congestion control on the RTP packet streams. The peer device could be another RTP endpoint, or it could be an RTP middlebox that redistributes the RTP packet streams to other RTP endpoints. This limitation means that some of the RTP middlebox-based topologies are not suitable for use in WebRTC. Specifically:<list style="symbols"> <t>Video switching MCUs</t> <ul spacing="normal"> <li>Video-switching Multipoint Control Units (MCUs) (Topo-Video-switch-MCU)SHOULD NOT<bcp14>SHOULD NOT</bcp14> be used, since they make the use of RTCP for congestion control andquality of servicequality-of-service reports problematic (seeSection 3.8 of<xreftarget="I-D.ietf-avtcore-rtp-topologies-update"></xref>).</t> <t>Thetarget="RFC7667" section="3.8" sectionFormat="of"/>).</li> <li>The Relay-Transport Translator (Topo-PtM-Trn-Translator) topologySHOULD NOT<bcp14>SHOULD NOT</bcp14> beusedused, because its safe use requires a congestion control algorithm or RTP circuit breaker that handles point to multipoint, which has not yet beenstandardised.</t> </list></t>standardized.</li> </ul> <t>The following topology can be used, however it has some issues worthnoting:<list style="symbols"> <t>Content modifyingnoting:</t> <ul spacing="normal"> <li>Content-modifying MCUs with RTCP termination (Topo-RTCP-terminating-MCU)MAY<bcp14>MAY</bcp14> be used. Note that in this RTPTopology,topology, RTP loop detection and identification of active senders is the responsibility of the WebRTC application; since the clients are isolated from each other at the RTP layer, RTP cannot assist with these functions (seesection 3.9 of<xreftarget="I-D.ietf-avtcore-rtp-topologies-update"></xref>).</t> </list></t>target="RFC7667" section="3.9" sectionFormat="of"/>).</li> </ul> <t>The RTP extensions described in Sections <xreftarget="sec-fir"></xref>target="sec-fir" format="counter"/> to <xreftarget="sec.tmmbr"></xref>target="sec.tmmbr" format="counter"/> are designed to be used withcentralisedcentralized conferencing, where an RTP middlebox (e.g., a conference bridge) receives a participant's RTP packet streams and distributes them to the other participants. These extensions are not necessary for interoperability; an RTP endpoint that does not implement these extensions will workcorrectly,correctly but might offer poor performance. Support for the listed extensions will greatly improve the quality ofexperience and,experience; to provide a reasonable baseline quality, some of these extensions are mandatory to be supported by WebRTCEndpoints.</t>endpoints.</t> <t>The RTCP conferencing extensions are defined in <xreftarget="RFC4585">Extendedtarget="RFC4585" format="default">"Extended RTP Profile for Real-time Transport Control Protocol (RTCP)-Based Feedback(RTP/AVPF)</xref>(RTP/AVPF)"</xref> andthe memo on<xreftarget="RFC5104">Codectarget="RFC5104" format="default">"Codec Control Messages(CCM)inRTP/AVPF</xref>;the RTP Audio-Visual Profile with Feedback (AVPF)"</xref>; they are fully usable by the <xreftarget="RFC5124">Securetarget="RFC5124" format="default"> secure variant of this profile (RTP/SAVPF)</xref>.</t> <section anchor="sec-fir"title="Fullnumbered="true" toc="default"> <name>Full Intra Request(FIR)">(FIR)</name> <t>The Full Intra Request message is defined in Sections3.5.1<xref target="RFC5104" section="3.5.1" sectionFormat="bare"/> and4.3.1<xref target="RFC5104" section="4.3.1" sectionFormat="bare"/> ofthe<xreftarget="RFC5104">Codectarget="RFC5104" format="default">Codec Control Messages</xref>. It is used to make the mixer request a new Intra picture from a participant in the session. This is used when switching between sources to ensure that the receivers can decode the video or other predictive media encoding with long prediction chains. WebRTCEndpointsendpoints that are sending mediaMUST<bcp14>MUST</bcp14> understand and react to FIR feedback messages they receive, since this greatly improves the user experience when usingcentralisedcentralized mixer-based conferencing. Support for sending FIR messages isOPTIONAL.</t><bcp14>OPTIONAL</bcp14>.</t> </section> <sectiontitle="Picturenumbered="true" toc="default"> <name>Picture Loss Indication(PLI)">(PLI)</name> <t>The Picture Loss Indication message is defined inSection 6.3.1 of the<xreftarget="RFC4585">RTP/AVPFtarget="RFC4585" section="6.3.1" sectionFormat="of">the RTP/AVPF profile</xref>. It is used by a receiver to tell the sending encoder that it lost the decoder context and would like to have it repaired somehow. This is semantically different from the Full Intra Requestaboveabove, as there could be multiple ways tofulfilfulfill the request. WebRTCEndpointsendpoints that are sending mediaMUST<bcp14>MUST</bcp14> understand and react to PLI feedback messages as aloss toleranceloss-tolerance mechanism. ReceiversMAY<bcp14>MAY</bcp14> send PLI messages.</t> </section> <sectiontitle="Slicenumbered="true" toc="default"> <name>Slice Loss Indication(SLI)">(SLI)</name> <t>The Slice Loss Indication message is defined inSection 6.3.2 of the<xreftarget="RFC4585">RTP/AVPFtarget="RFC4585" section="6.3.2" sectionFormat="of">the RTP/AVPF profile</xref>. It is used by a receiver to tell the encoder that it has detected the loss or corruption of one or more consecutive macroblocks,blocks and would like to have these repaired somehow. It isRECOMMENDED<bcp14>RECOMMENDED</bcp14> that receivers generate SLI feedback messages if slices are lost when using a codec that supports the concept of macro blocks. A sender that receives an SLI feedback messageSHOULD<bcp14>SHOULD</bcp14> attempt to repair the lost slice(s).</t> </section> <sectiontitle="Referencenumbered="true" toc="default"> <name>Reference Picture Selection Indication(RPSI)">(RPSI)</name> <t>Reference Picture Selection Indication (RPSI) messages are defined inSection 6.3.3 of the<xreftarget="RFC4585">RTP/AVPFtarget="RFC4585" section="6.3.3" sectionFormat="of">the RTP/AVPF profile </xref>. Somevideo encodingvideo-encoding standards allow the use of older reference pictures than the most recent one for predictive coding. If such a codec is in use, and if the encoder haslearntlearned that encoder-decodersynchronisationsynchronization has been lost, then aknown as correctknown-as-correct reference picture can be used as a base for future coding. The RPSI message allows this to besignalled.signaled. Receivers that detect that encoder-decodersynchronisationsynchronization has been lostSHOULD<bcp14>SHOULD</bcp14> generate an RPSI feedback message if the codec being used supportsreference picturereference-picture selection.AAn RTPpacket streampacket-stream sender that receives such an RPSI messageSHOULD<bcp14>SHOULD</bcp14> act on that messages to change the reference picture, if it is possible to do so within the available bandwidthconstraints,constraints and with the codec being used.</t> </section> <sectiontitle="Temporal-Spatial Trade-offnumbered="true" toc="default"> <name>Temporal-Spatial Trade-Off Request(TSTR)">(TSTR)</name> <t>The temporal-spatial trade-off request and notification are defined in Sections3.5.2<xref target="RFC5104" section="3.5.2" sectionFormat="bare"/> and4.3.2<xref target="RFC5104" section="4.3.2" sectionFormat="bare"/> of <xreftarget="RFC5104"></xref>.target="RFC5104" format="default"/>. This request can be used to ask the video encoder to change the trade-off it makes between temporal and spatialresolution,resolution -- forexampleexample, to prefer high spatial image quality but low frame rate. Support for TSTR requests and notifications isOPTIONAL.</t><bcp14>OPTIONAL</bcp14>.</t> </section> <section anchor="sec.tmmbr"title="Temporarynumbered="true" toc="default"> <name>Temporary Maximum Media Stream Bit Rate Request(TMMBR)">(TMMBR)</name> <t>TheTMMBRTemporary Maximum Media Stream Bit Rate Request (TMMBR) feedback message is defined in Sections3.5.4<xref target="RFC5104" section="3.5.4" sectionFormat="bare"/> and4.2.1<xref target="RFC5104" section="4.2.1" sectionFormat="bare"/> ofthe<xreftarget="RFC5104">Codectarget="RFC5104" format="default">Codec Control Messages</xref>. This request and itsnotificationcorresponding Temporary Maximum Media Stream Bit Rate Notification (TMMBN) message <xref target="RFC5104"/> are used by a media receiver to inform the sending party that there is a current limitation on the amount of bandwidth available to this receiver. There can be various reasons for this: for example, an RTP mixer can use this message to limit the media rate of the sender being forwarded by the mixer (without doing media transcoding) to fit the bottlenecks existing towards the other session participants. WebRTCEndpointsendpoints that are sending media areREQUIRED<bcp14>REQUIRED</bcp14> to implement support for TMMBRmessages,messages andMUST<bcp14>MUST</bcp14> follow bandwidth limitations set by a TMMBR message received for their SSRC. The sending of TMMBRrequestsmessages isOPTIONAL.</t><bcp14>OPTIONAL</bcp14>.</t> </section> </section> <sectiontitle="Header Extensions">numbered="true" toc="default"> <name>Header Extensions</name> <t>The <xreftarget="RFC3550">RTPtarget="RFC3550" format="default">RTP specification</xref> provides the capability to include RTP header extensions containing in-band data, but the format and semantics of the extensions are poorly specified. The use of header extensions isOPTIONAL<bcp14>OPTIONAL</bcp14> in WebRTC, but if they are used, theyMUST<bcp14>MUST</bcp14> be formatted andsignalledsignaled following the general mechanism for RTP header extensions defined in <xreftarget="RFC5285"></xref>,target="RFC8285" format="default"/>, since this gives well-defined semantics to RTP header extensions.</t> <t>As noted in <xreftarget="RFC5285"></xref>,target="RFC8285" format="default"/>, the requirement from the RTP specification that header extensions are "designed so that the header extension may be ignored" <xreftarget="RFC3550"></xref>target="RFC3550" format="default"/> stands. To be specific, header extensionsMUST<bcp14>MUST</bcp14> only be used for data that can safely be ignored by the recipient without affectinginteroperability,interoperability andMUST NOT<bcp14>MUST NOT</bcp14> be used when the presence of the extension has changed the form or nature of the rest of the packet in a way that is not compatible with the way the stream issignalledsignaled (e.g., as defined by the payload type). Valid examples of RTP header extensions might include metadata that is additional to the usual RTPinformation,information but that can safely be ignored without compromising interoperability.</t> <section anchor="rapid-sync"title="Rapid Synchronisation">numbered="true" toc="default"> <name>Rapid Synchronization</name> <t>Many RTP sessions requiresynchronisationsynchronization between audio, video, and other content. Thissynchronisationsynchronization is performed by receivers, using information contained in RTCP SR packets, as described in the <xreftarget="RFC3550">RTPtarget="RFC3550" format="default">RTP specification</xref>. This basic mechanism can be slow, however, so it isRECOMMENDED<bcp14>RECOMMENDED</bcp14> that the rapid RTPsynchronisationsynchronization extensions described in <xreftarget="RFC6051"></xref>target="RFC6051" format="default"/> be implemented in addition to RTCP SR-basedsynchronisation.</t>synchronization.</t> <t>This header extension uses the<xref target="RFC5285"></xref>generic header extensionframework,framework described in <xref target="RFC8285" format="default"/> and so needs to be negotiated before it can be used.</t> </section> <section anchor="sec-client-to-mixer"title="Client-to-Mixernumbered="true" toc="default"> <name>Client-to-Mixer AudioLevel">Level</name> <t>The <xreftarget="RFC6464">Client to Mixer Audio Leveltarget="RFC6464" format="default">client-to-mixer audio level extension</xref> is an RTP header extension used by an endpoint to inform a mixer about the level of audio activity in the packet to which the header is attached. This enables an RTP middlebox to make mixing or selection decisions without decoding or detailed inspection of the payload, reducing the complexity in some types of mixers. It can also save decoding resources in receivers, which can choose to decode only the most relevant RTP packet streams based on audio activity levels.</t> <t>The <xreftarget="RFC6464">Client-to-Mixer Audio Level</xref>target="RFC6464" format="default">client-to-mixer audio level header extensionMUST</xref> <bcp14>MUST</bcp14> be implemented. It isREQUIRED<bcp14>REQUIRED</bcp14> that implementationsarebe capable of encrypting the header extension according to <xreftarget="RFC6904"></xref>target="RFC6904" format="default"/>, since the information contained in these header extensions can be considered sensitive. The use of this encryption isRECOMMENDED, however<bcp14>RECOMMENDED</bcp14>; however, usage of the encryption can be explicitly disabled through API orsignalling.</t>signaling.</t> <t>This header extension uses the<xref target="RFC5285"></xref>generic header extensionframework,framework described in <xref target="RFC8285" format="default"/> and so needs to be negotiated before it can be used.</t> </section> <section anchor="sec-mixer-to-client"title="Mixer-to-Clientnumbered="true" toc="default"> <name>Mixer-to-Client AudioLevel">Level</name> <t>The <xreftarget="RFC6465">Mixer to Client Audio Leveltarget="RFC6465" format="default">mixer-to-client audio level header extension</xref> provides an endpoint with the audio level of the different sources mixed into a common source stream byaan RTP mixer. This enables a user interface to indicate the relative activity level of each session participant, rather than just being included or not based on the CSRC field. This is a pureoptimisationoptimization ofnon critical functions,non-critical functions and is henceOPTIONAL<bcp14>OPTIONAL</bcp14> to implement. If this header extension is implemented, it isREQUIRED<bcp14>REQUIRED</bcp14> that implementationsarebe capable of encrypting the header extension according to <xreftarget="RFC6904"></xref>target="RFC6904" format="default"/>, since the information contained in these header extensions can be considered sensitive. It is furtherRECOMMENDED<bcp14>RECOMMENDED</bcp14> that this encryptionisbe used, unless the encryption has been explicitly disabled through API orsignalling.</t>signaling.</t> <t>This header extension uses the<xref target="RFC5285"></xref>generic header extensionframework,framework described in <xref target="RFC8285" format="default"/> and so needs to be negotiated before it can be used.</t> </section> <section anchor="sec-mid"title="Medianumbered="true" toc="default"> <name>Media StreamIdentification">Identification</name> <t>WebRTC endpoints that implement the SDP bundle negotiation extension will use the SDPgrouping framework 'mid'Grouping Framework "mid" attribute to identify media streams. Such endpointsMUST<bcp14>MUST</bcp14> implement the RTP MID header extension described in <xreftarget="I-D.ietf-mmusic-sdp-bundle-negotiation"></xref>.</t>target="RFC8843" format="default"/>.</t> <t>This header extension uses the<xref target="RFC5285"></xref>generic header extensionframework,framework described in <xref target="RFC8285" format="default"/> and so needs to be negotiated before it can be used.</t> </section> <section anchor="sec-cvo"title="Coordinationnumbered="true" toc="default"> <name>Coordination of VideoOrientation">Orientation</name> <t>WebRTC endpoints that send or receive videoMUST<bcp14>MUST</bcp14> implement the coordination of video orientation (CVO) RTP header extension as described inSection 4 of<xreftarget="I-D.ietf-rtcweb-video"></xref>.</t>target="RFC7742" section="4" sectionFormat="of"/>.</t> <t>This header extension uses the<xref target="RFC5285"></xref>generic header extensionframework,framework described in <xref target="RFC8285" format="default"/> and so needs to be negotiated before it can be used.</t> </section> </section> </section> <section anchor="sec-rtp-robust"title="WebRTCnumbered="true" toc="default"> <name>WebRTC Use of RTP: Improving TransportRobustness">Robustness</name> <t>There are tools that can make RTP packet streams robust against packet loss and reduce the impact of loss on media quality. However, they generally add some overhead compared to a non-robust stream. The overhead needs to be considered, and the aggregatebit-rate MUSTbitrate <bcp14>MUST</bcp14> be rate controlled to avoid causing network congestion (see <xreftarget="sec-rate-control"></xref>).target="sec-rate-control" format="default"/>). As a result, improving robustness might require a lower base encodingquality,quality but has the potential to deliver that quality with fewer errors. The mechanisms described in the followingsub-sectionssubsections can be used to improve tolerance to packet loss.</t> <section anchor="sec-rtx"title="Negativenumbered="true" toc="default"> <name>Negative Acknowledgements and RTPRetransmission">Retransmission</name> <t>As a consequence of supporting the RTP/SAVPF profile, implementations can send negative acknowledgements (NACKs) for RTP data packets <xreftarget="RFC4585"></xref>.target="RFC4585" format="default"/>. This feedback can be used to inform a sender of the loss of particular RTP packets, subject to the capacity limitations of the RTCP feedback channel. A sender can use this information tooptimiseoptimize the user experience by adapting the media encoding to compensate for known lost packets.</t> <t>RTP packet stream senders areREQUIRED<bcp14>REQUIRED</bcp14> to understand theGenericgeneric NACK message defined inSection 6.2.1 of<xreftarget="RFC4585"></xref>,target="RFC4585" sectionFormat="of" section="6.2.1"/>, butMAYthey <bcp14>MAY</bcp14> choose to ignore some or all of this feedback (followingSection 4.2 of<xreftarget="RFC4585"></xref>).target="RFC4585" sectionFormat="of" section="4.2"/>). ReceiversMAY<bcp14>MAY</bcp14> send NACKs for missing RTP packets. Guidelines on when to send NACKs are provided in <xreftarget="RFC4585"></xref>.target="RFC4585" format="default"/>. It is not expected that a receiver will send a NACK for every lost RTPpacket, ratherpacket; rather, it needs to consider the cost of sending NACKfeedback,feedback and the importance of the lostpacket,packet to make an informed decision on whether it is worth telling the sender about apacket losspacket-loss event.</t> <t>The <xreftarget="RFC4588">RTP Retransmission Payload Format</xref>target="RFC4588" format="default">RTP retransmission payload format</xref> offers the ability to retransmit lost packets based on NACK feedback. Retransmission needs to be used with care in interactive real-time applications to ensure that the retransmitted packet arrives in time to be useful, but it can be effective in environments with relatively low networkRTT (anRTT. (An RTP sender can estimate the RTT to the receivers using the information in RTCP SR and RR packets, as described at the end ofSection 6.4.1 of<xreftarget="RFC3550"></xref>).target="RFC3550" section="6.4.1" sectionFormat="of"/>). The use of retransmissions can also increase the forward RTPbandwidth,bandwidth and can potentiallycausedcause increased packet loss if the original packet loss was caused by network congestion. Note, however, that retransmission of an important lost packet to repair decoder state can have lower cost than sending a full intra frame. It is not appropriate to blindly retransmit RTP packets in response to a NACK. The importance of lost packets and the likelihood of them arriving in time to be usefulneedsneed to be considered before RTP retransmission is used.</t> <t>Receivers areREQUIRED<bcp14>REQUIRED</bcp14> to implement support for RTP retransmission packets <xreftarget="RFC4588"></xref>target="RFC4588" format="default"/> sent using SSRCmultiplexing,multiplexing andMAY<bcp14>MAY</bcp14> also support RTP retransmission packets sent using session multiplexing. SendersMAY<bcp14>MAY</bcp14> send RTP retransmission packets in response to NACKs if support for the RTP retransmission payload format has beennegotiated,negotiated andifthe sender believes it is useful to send a retransmission of the packet(s) referenced in the NACK. Senders do not need to retransmit every NACKed packet.</t> </section> <section anchor="sec-FEC"title="Forwardnumbered="true" toc="default"> <name>Forward Error Correction(FEC)">(FEC)</name> <t>The use of Forward Error Correction (FEC) can provide an effective protection against some degree of packet loss, at the cost of steady bandwidth overhead. There are several FEC schemes that are defined for use with RTP. Some of these schemes are specific to a particular RTP payload format, and others operate across RTP packets and can be used with any payload format.It needs to be notedNote that using redundant encoding or FEC will lead to increasedplay outplay-out delay, which needs to be considered when choosing FEC schemes and their parameters.</t> <t>WebRTC endpointsMUST<bcp14>MUST</bcp14> follow the recommendations for FEC use given in <xreftarget="I-D.ietf-rtcweb-fec"></xref>.target="RFC8854" format="default"/>. WebRTC endpointsMAY<bcp14>MAY</bcp14> support other types of FEC, but theseMUST<bcp14>MUST</bcp14> be negotiated before they are used.</t> </section> </section> <section anchor="sec-rate-control"title="WebRTCnumbered="true" toc="default"> <name>WebRTC Use of RTP: Rate Control and MediaAdaptation">Adaptation</name> <t>WebRTC will be used in heterogeneous network environments using a variety of link technologies, including both wired and wireless links, to interconnect potentially large groups of users around the world. As a result, the network paths between users can have widely varying one-way delays, availablebit-rates,bitrates, load levels, and traffic mixtures. Individual endpoints can send one or more RTP packet streams to each participant, and there can be several participants. Each of these RTP packet streams can contain different types of media, and the type of media,bit rate,bitrate, and number of RTP packet streams as well as transport-layer flows can be highly asymmetric. Non-RTP traffic can share the network paths with RTP transport-layer flows. Since the network environment is not predictable or stable, WebRTCEndpoints MUSTendpoints <bcp14>MUST</bcp14> ensure that the RTP traffic they generate can adapt to match changes in the available network capacity.</t> <t>The quality of experience for users of WebRTC is very dependent on effective adaptation of the media to the limitations of the network. Endpoints have to be designed so they do not transmit significantly more data than the network path can support, except for very short timeperiods, otherwiseperiods; otherwise, high levels of network packet loss or delay spikes will occur, causing media quality degradation. The limiting factor on the capacity of the network path might be the link bandwidth, or it might be competition with other traffic on the link (this can be non-WebRTC traffic, traffic due to other WebRTC flows, or even competition with other WebRTC flows in the same session).</t> <t>An effective media congestion control algorithm is therefore an essential part of the WebRTC framework. However, at the time of this writing, there is no standard congestion control algorithm that can be used for interactive media applications such as WebRTC's flows. Some requirements for congestion control algorithms for RTCPeerConnections are discussed in <xreftarget="I-D.ietf-rmcat-cc-requirements"></xref>.target="RFC8836" format="default"/>. If a standardized congestion control algorithm that satisfies these requirements is developed in the future, this memo will need to bebeupdated to mandate its use.</t> <sectiontitle="Boundarynumbered="true" toc="default"> <name>Boundary Conditions and CircuitBreakers">Breakers</name> <t>WebRTCEndpoints MUSTendpoints <bcp14>MUST</bcp14> implement the RTP circuit breaker algorithm that is described in <xreftarget="I-D.ietf-avtcore-rtp-circuit-breakers"></xref>.target="RFC8083" format="default"/>. The RTP circuit breaker is designed to enable applications torecogniserecognize and react to situations of extreme network congestion. However, since the RTP circuit breaker might not be triggered until congestion becomes extreme, it cannot be considered a substitute for congestion control, and applicationsMUST<bcp14>MUST</bcp14> also implement congestion control to allow them to adapt to changes in network capacity. The congestion control algorithm will have to be proprietary until a standardized congestion control algorithm is available. Any future RTP congestion control algorithms are expected to operate within the envelope allowed by the circuit breaker.</t> <t>Thesession establishment signallingsession-establishment signaling will also necessarily establish boundaries to which the mediabit-ratebitrate will conform. The choice of media codecs providesupper-upper andlower-boundslower bounds on the supportedbit-ratesbitrates that the application canutiliseutilize to provide useful quality, and thepacketisationpacketization choices that exist. In addition, thesignallingsignaling channel can establish maximum mediabit-ratebitrate boundaries using, for example, the SDP "b=AS:" or "b=CT:" lines and the RTP/AVPFTemporary Maximum Media Stream Bit Rate (TMMBR) RequestsTMMBR messages (see <xreftarget="sec.tmmbr"></xref>target="sec.tmmbr" format="default"/> of this memo).SignalledSignaled bandwidth limitations, such as SDP "b=AS:" or "b=CT:" lines received from the peer,MUST<bcp14>MUST</bcp14> be followed when sending RTP packet streams. A WebRTCEndpointendpoint receiving mediaSHOULD<bcp14>SHOULD</bcp14> signal its bandwidth limitations. These limitations have to be based on known bandwidth limitations, for example the capacity of the edge links.</t> </section> <sectiontitle="Congestionnumbered="true" toc="default"> <name>Congestion Control Interoperability and LegacySystems">Systems</name> <t>All endpoints that wish to interwork with WebRTCMUST<bcp14>MUST</bcp14> implement RTCP and provide congestion feedback via the defined RTCP reporting mechanisms.</t> <t>When interworking with legacy implementations that support RTCP using the <xreftarget="RFC3551">RTP/AVPtarget="RFC3551" format="default">RTP/AVP profile</xref>, congestion feedback is provided in RTCP RR packets every few seconds. Implementations that have to interwork with such endpointsMUST<bcp14>MUST</bcp14> ensure that they keep within the <xreftarget="I-D.ietf-avtcore-rtp-circuit-breakers"> RTPtarget="RFC8083" format="default">RTP circuit breaker</xref> constraints to limit the congestion they can cause.</t> <t>If a legacy endpoint supports RTP/AVPF, this enables negotiation of important parameters for frequent reporting, such as the "trr-int" parameter, and the possibility that the endpoint supports some useful feedback format for congestion controlpurposepurposes such as <xreftarget="RFC5104">target="RFC5104" format="default"> TMMBR</xref>. Implementations that have to interwork with such endpointsMUST<bcp14>MUST</bcp14> ensure that they stay within the <xreftarget="I-D.ietf-avtcore-rtp-circuit-breakers">target="RFC8083" format="default"> RTP circuit breaker</xref> constraints to limit the congestion they can cause, but they might find that they can achieve better congestion response depending on the amount of feedback that is available.</t> <t>With proprietary congestion controlalgorithmsalgorithms, issues can arise when different algorithms and implementations interact in a communication session. If the different implementations have made different choices in regards to the type of adaptation, for example one sender based, and one receiver based, then one could end up in a situation where one direction is dualcontrolled,controlled when the other direction is not controlled. This memo cannot mandatebehaviourbehavior for proprietary congestion control algorithms, but implementations that use such algorithms ought to be aware of thisissue,issue and try to ensure that effective congestion control is negotiated for media flowing in both directions. If the IETF were tostandardisestandardize both sender- and receiver-based congestion control algorithms for WebRTC traffic in the future, the issues of interoperability, control, and ensuring that both directions of media flow are congestion controlled would also need to be considered.</t> </section> </section> <section anchor="sec-perf"title="WebRTCnumbered="true" toc="default"> <name>WebRTC Use of RTP: PerformanceMonitoring">Monitoring</name> <t>As described in <xreftarget="sec-rtp-rtcp"></xref>,target="sec-rtp-rtcp" format="default"/>, implementations areREQUIRED<bcp14>REQUIRED</bcp14> to generate RTCP Sender Report (SR) and Reception Report (RR) packets relating to the RTP packet streams they send and receive. These RTCP reports can be used for performance monitoring purposes, since they include basicpacket losspacket-loss and jitter statistics.</t> <t>A large number of additional performance metrics are supported by the RTCP Extended Reports (XR)framework,framework; see <xreftarget="RFC3611"></xref><xref target="RFC6792"></xref>.target="RFC3611" format="default"/> and <xref target="RFC6792" format="default"/>. At the time of this writing, it is not clear what extended metrics are suitable for use in WebRTC, so there is no requirement that implementations generate RTCP XR packets. However, implementations that can use detailed performance monitoring dataMAY<bcp14>MAY</bcp14> generate RTCP XR packets as appropriate. The use of RTCP XR packetsSHOULD<bcp14>SHOULD</bcp14> besignalled;signaled; implementationsMUST<bcp14>MUST</bcp14> ignore RTCP XR packets that are unexpected or not understood.</t> </section> <section anchor="sec-extn"title="WebRTCnumbered="true" toc="default"> <name>WebRTC Use of RTP: FutureExtensions">Extensions</name> <t>It is possible that the core set of RTP protocols and RTP extensions specified in this memo will prove insufficient for the future needs of WebRTC. In this case, future updates to this memo have to be made followingthe<xreftarget="RFC2736"> Guidelinestarget="RFC2736" format="default">"Guidelines for Writers of RTP Payload FormatSpecifications </xref>,Specifications"</xref>, <xreftarget="I-D.ietf-payload-rtp-howto">Howtarget="RFC8088" format="default">"How to Write an RTP PayloadFormat</xref>Format"</xref>, and <xreftarget="RFC5968"> Guidelinestarget="RFC5968" format="default">"Guidelines for Extending the RTP ControlProtocol</xref>, and SHOULDProtocol (RTCP)"</xref>. They also <bcp14>SHOULD</bcp14> take into account any future guidelines for extending RTP and related protocols that have been developed.</t> <t>Authors of future extensions are urged to consider the wide range of environments in which RTP is used when recommending extensions, since extensions that are applicable in some scenarios can be problematic in others. Where possible, the WebRTC framework will adopt RTP extensions that are of general utility, to enable easy implementation of a gateway to other applications using RTP, rather than adopt mechanisms that are narrowly targeted at specific WebRTC use cases.</t> </section> <section anchor="sec-signalling"title="Signalling Considerations">numbered="true" toc="default"> <name>Signaling Considerations</name> <t>RTP is built with the assumption that an externalsignallingsignaling channelexists,exists and can be used to configure RTP sessions and their features. The basic configuration of an RTP session consists of the following parameters:</t><t><list style="hanging"> <t hangText="RTP Profile:">The<dl newline="false" spacing="normal"> <dt>RTP profile:</dt> <dd>The name of the RTP profile to be used in the session. The <xreftarget="RFC3551">RTP/AVP</xref>target="RFC3551" format="default">RTP/AVP</xref> and <xreftarget="RFC4585">RTP/AVPF</xref>target="RFC4585" format="default">RTP/AVPF</xref> profiles can interoperate on a basic level, as can their securevariantsvariants, <xreftarget="RFC3711">RTP/SAVP</xref>target="RFC3711" format="default">RTP/SAVP</xref> and <xreftarget="RFC5124">RTP/SAVPF</xref>.target="RFC5124" format="default">RTP/SAVPF</xref>. The secure variants of the profiles do not directly interoperate with thenon-securenonsecure variants, due to the presence of additional header fields for authentication in SRTP packets and cryptographic transformation of the payload. WebRTC requires the use of the RTP/SAVPF profile, and thisMUST<bcp14>MUST</bcp14> besignalled.signaled. Interworking functions might transform this into the RTP/SAVP profile for a legacy usecase,case by indicating to the WebRTCEndpointendpoint that the RTP/SAVPF is used and configuring atrr-int"trr-int" value of 4seconds.</t> <t hangText="Transport Information:">Sourceseconds.</dd> <dt>Transport information:</dt> <dd>Source and destination IPaddress(s)address(es) and ports for RTP and RTCPMUST<bcp14>MUST</bcp14> besignalledsignaled for each RTP session. InWebRTCWebRTC, these transport addresses will be provided by <xreftarget="RFC5245">ICE</xref>target="RFC8445" format="default">Interactive Connectivity Establishment (ICE)</xref> that signals candidates and arrives at nominated candidate address pairs. If <xreftarget="RFC5761">RTPtarget="RFC5761" format="default">RTP and RTCP multiplexing</xref> is to beused,used such that a singleport, i.e.port -- i.e., transport-layerflow,flow -- is used for RTP and RTCP flows, thisMUST<bcp14>MUST</bcp14> besignalledsignaled (see <xreftarget="sec.rtcp-mux"></xref>).</t> <t hangText="RTP Payload Types,target="sec.rtcp-mux" format="default"/>).</dd> <dt>RTP payload types, media formats, and formatparameters:">Theparameters:</dt> <dd>The mapping between media type names (and hence the RTP payload formats to beused),used) and the RTP payload type numbersMUST<bcp14>MUST</bcp14> besignalled.signaled. Each media typeMAY<bcp14>MAY</bcp14> also have a number of media type parameters thatMUST<bcp14>MUST</bcp14> also besignalledsignaled to configure the codec and RTP payload format (the "a=fmtp:" line from SDP). <xreftarget="sec.codecs"></xref>target="sec.codecs" format="default"/> of this memo discusses requirements for uniqueness of payloadtypes.</t> <t hangText="RTP Extensions:">Thetypes.</dd> <dt>RTP extensions:</dt> <dd>The use of any additional RTP header extensions and RTCP packet types, including any necessary parameters,MUST<bcp14>MUST</bcp14> besignalled.signaled. Thissignalling is to ensuresignaling ensures that a WebRTCEndpoint's behaviour,endpoint’s behavior, especially when sending,of any extensionsis predictable and consistent. Forrobustness,robustness andforcompatibility with non-WebRTC systems that might be connected to a WebRTC session via a gateway, implementations areREQUIRED<bcp14>REQUIRED</bcp14> to ignore unknown RTCP packets and RTP header extensions (see also <xreftarget="sec-rtp-rtcp"></xref>).</t> <t hangText="RTCP Bandwidth:">Supporttarget="sec-rtp-rtcp" format="default"/>).</dd> <dt>RTCP bandwidth:</dt> <dd>Support for exchanging RTCPBandwidthbandwidth valuestowith the endpoints will be necessary. ThisSHALL<bcp14>SHALL</bcp14> be done as described in <xreftarget="RFC3556">"Sessiontarget="RFC3556" format="default">"Session Description Protocol (SDP) Bandwidth Modifiers for RTP Control Protocol (RTCP) Bandwidth"</xref> if using SDP, or something semantically equivalent. This also ensures that the endpoints have a common view of the RTCP bandwidth. A common view of the RTCP bandwidth among different endpoints isimportant,important to prevent differences in RTCP packet timing and timeout intervals causing interoperabilityproblems.</t> </list></t>problems.</dd> </dl> <t>These parameters are often expressed in SDP messages conveyed within an offer/answer exchange. RTP does not depend on SDP oronthe offer/answermodel,model but does require all the necessary parameters to be agreedupon,upon and provided to the RTP implementation. Note that inWebRTCWebRTC, it will depend on thesignallingsignaling model and API how these parameters need to beconfiguredconfigured, but they willbeneed to either be set in the API or explicitlysignalledsignaled between the peers.</t> </section> <section anchor="sec-webrtc-api"title="WebRTCnumbered="true" toc="default"> <name>WebRTC APIConsiderations">Considerations</name> <t>The <xreftarget="W3C.WD-webrtc-20130910">WebRTCtarget="W3C.WebRTC" format="default">WebRTC API</xref> and the <xreftarget="W3C.WD-mediacapture-streams-20130903">Mediatarget="W3C-MEDIA-CAPTURE" format="default">Media Capture and Streams API</xref>definesdefine andusesuse the concept of a MediaStream that consists of zero or more MediaStreamTracks. A MediaStreamTrack is an individual stream of media from any type of mediasource likesource, such as a microphone or a camera, butalsoconceptual sources, likeaan audio mix or a video composition, are also possible. The MediaStreamTracks within a MediaStream might need to be synchronized duringplay back.</t>playback.</t> <t>A MediaStreamTrack'srealisationrealization inRTPRTP, in the context of anRTCPeerConnectionRTCPeerConnection, consists of a source packetstreamstream, identifiedwithby anSSRCSSRC, sent within an RTP session that is part of the RTCPeerConnection. The MediaStreamTrack can also result in additional packet streams, and thus SSRCs, in the same RTP session. These can be dependent packet streams from scalable encoding of the source stream associated with the MediaStreamTrack, if such a media encoder is used. They can also be redundancy packetstreams,streams; these are created when applying <xreftarget="sec-FEC">Forwardtarget="sec-FEC" format="default">Forward Error Correction</xref> or <xreftarget="sec-rtx">RTPtarget="sec-rtx" format="default">RTP retransmission</xref> to the source packet stream.</t> <t>It is important to note that the same media source can be feeding multiple MediaStreamTracks. As different sets of constraints or other parameters can be applied to the MediaStreamTrack, each MediaStreamTrack instance added toaan RTCPeerConnectionSHALL<bcp14>SHALL</bcp14> result in an independent source packetstream,stream with its own set of associated packetstreams,streams and thus different SSRC(s). It will depend on applied constraints and parameters if the source stream and the encoding configuration will be identical between different MediaStreamTracks sharing the same media source. If the encoding parameters and constraints are the same, an implementation could choose to use only one encoded stream to create the different RTP packet streams. Note that suchoptimisationsoptimizations would need to take into account that the constraints for one of the MediaStreamTracks can change at anymoment change,moment, meaning that the encoding configurations might no longer beidenticalidentical, and two different encoder instances would then be needed.</t> <t>The same MediaStreamTrack can also be included in multiple MediaStreams, thus multiple sets of MediaStreams can implicitly need to use the samesynchronisationsynchronization base. To ensure that this works in allcases,cases and does not force an endpoint to disrupt the media by changingsynchronisationsynchronization base and CNAME during delivery of any ongoing packet streams, all MediaStreamTracks and their associated SSRCs originating from the same endpoint need to be sent using the same CNAME within one RTCPeerConnection. This is motivating the use of a single CNAME in <xreftarget="sec-cname"></xref>. <list style="empty"> <t>Thetarget="sec-cname" format="default"/>. </t> <aside><t>The requirementon usingto use the same CNAME for all SSRCs that originate from the sameendpoint,endpoint does not require a middlebox that forwards traffic from multiple endpoints to only use a single CNAME.</t></list></t></aside> <t>Different CNAMEs normally need to be used for different RTCPeerConnection instances, as specified in <xreftarget="sec-cname"></xref>.target="sec-cname" format="default"/>. Having two communication sessions with the same CNAME could enable tracking of a user or device across different services (seeSection 4.4.1 of<xreftarget="I-D.ietf-rtcweb-security"></xref>target="RFC8826" section="4.4.1" sectionFormat="of"/> for details). A web application can request that the CNAMEs used in different RTCPeerConnections (within asame-orignsame-origin context) be thesame,same; this allows for synchronization of the endpoint's RTP packet streams across the differentRTCPeerConnections.<list style="empty"> <t>Note: thisRTCPeerConnections.</t> <aside><t>Note: This doesn't result in a tracking issue, since the creation of matching CNAMEs depends on existing tracking within a single origin.</t></list>The</aside> <t>The above will currently force a WebRTCEndpointendpoint that receives a MediaStreamTrack on one RTCPeerConnection and adds it asanoutgoing one on any RTCPeerConnection to performresynchronisationresynchronization of the stream. Since the sending party needs to change the CNAME to the one it uses, this implies it has to use a local system clock as the timebase for thesynchronisation.synchronization. Thus, the relative relation between the timebase of the incoming stream and the system sending out needs to be defined. This relation also needs monitoring for clock drift and likely adjustments of thesynchronisation.synchronization. The sending entity is also responsible for congestion control for its sent streams. In cases of packetlossloss, the loss of incoming data also needs to be handled. This leads to the observation that the method that is least likely to cause issues or interruptions in the outgoing source packet stream is a model of full decoding, includingrepair etc.,repair, followed by encoding of the media again into the outgoing packet stream.OptimisationsOptimizations of this method are clearly possible and implementation specific.</t> <t>A WebRTCEndpoint MUSTendpoint <bcp14>MUST</bcp14> support receiving multiple MediaStreamTracks, where each of the different MediaStreamTracks (andtheirits sets of associated packet streams) uses different CNAMEs. However, MediaStreamTracks that are received with different CNAMEs have no definedsynchronisation.<list style="empty"> <t>Note:synchronization.</t> <aside><t>Note: The motivation for supporting reception of multiple CNAMEs is to allow for forward compatibility with any future changes that enable more efficient stream handling when endpoints relay/forward streams. It also ensures that endpoints can interoperate with certain types ofmulti-streammultistream middleboxes or endpoints that are notWebRTC.</t> </list></t>WebRTC.</t></aside> <t><xreftarget="I-D.ietf-rtcweb-jsep">Javascripttarget="RFC8829" format="default">"JavaScript Session EstablishmentProtocol</xref>Protocol (JSEP)"</xref> specifies that the binding between the WebRTC MediaStreams,MediaStreamTracksMediaStreamTracks, and the SSRC is done as specified in <xreftarget="I-D.ietf-mmusic-msid">"Cross Session Streamtarget="RFC8830" format="default">"WebRTC MediaStream Identification in the Session Description Protocol"</xref>. Section 4.1 of <xreftarget="I-D.ietf-mmusic-msid">The MSIDtarget="RFC8830" format="default">the MediaStream Identification (MSID) document</xref> alsodefines, in section 4.1,defines how to mapunknownsource packetstreamstreams with unknown SSRCs to MediaStreamTracks and MediaStreams. This later is relevant to handle some cases of legacy interoperability.CommonlyCommonly, the RTPPayload Typepayload type of any incoming packets will reveal if the packet stream is a source stream or a redundancy or dependent packet stream. The association to the correct source packet stream depends on the payload format in use for the packet stream.</t><t>Finally<t>Finally, this specification puts a requirement on the WebRTC API to realize a method for determining the <xreftarget="sec-rtp-rtcp">CSRCtarget="sec-rtp-rtcp" format="default">CSRC list</xref> as well as the <xreftarget="sec-mixer-to-client">Mixer-to-Clienttarget="sec-mixer-to-client" format="default">mixer-to-client audio levels</xref> (whensupported) andsupported); the basic requirements for this is further discussed in <xreftarget="sec-media-stream-id"></xref>.</t>target="sec-media-stream-id" format="default"/>.</t> </section> <section anchor="sec-rtp-func"title="RTPnumbered="true" toc="default"> <name>RTP ImplementationConsiderations">Considerations</name> <t>The following discussion provides some guidance on the implementation of the RTP features described in this memo. The focus is on a WebRTCEndpointendpoint implementation perspective, and while some mention is made of thebehaviourbehavior of middleboxes, that is not the focus of this memo.</t> <sectiontitle="Configurationnumbered="true" toc="default"> <name>Configuration and Use of RTPSessions">Sessions</name> <t>A WebRTCEndpointendpoint will be a simultaneous participant in one or more RTP sessions. Each RTP session can convey multiple mediasources,sources andcaninclude media data from multiple endpoints. In the following, some ways in which WebRTCEndpointsendpoints can configure and use RTP sessions are outlined.</t> <section anchor="sec.multiple-flows"title="Usenumbered="true" toc="default"> <name>Use of Multiple Media SourcesWithinwithin an RTPSession">Session</name> <t>RTP is a group communication protocol, and every RTP session can potentially contain multiple RTP packet streams. There are several reasons why this might be desirable:<list style="hanging"> <t hangText="Multiple</t> <ul> <li><t>Multiple mediatypes:">Outsidetypes:</t> <t>Outside of WebRTC, it is common to use one RTP session for each type of media source (e.g., one RTP session for audio sources and one for video sources, each sent over differenttransport layertransport-layer flows). However, to reduce the number of UDP ports used, the default in WebRTC is to send all types of media in a single RTP session, as described in <xreftarget="sec.session-mux"></xref>,target="sec.session-mux" format="default"/>, using RTP and RTCP multiplexing (<xreftarget="sec.rtcp-mux"></xref>)target="sec.rtcp-mux" format="default"/>) to further reduce the number of UDP ports needed. This RTP session then uses only onebi-directionalbidirectional transport-layerflow,flow but will contain multiple RTP packet streams, each containing a different type of media. A common example might be an endpoint with a camera and microphone that sends two RTP packet streams, one video and one audio, into a single RTP session.</t><t hangText="Multiple Capture Devices:">A</li> <li> <t>Multiple capture devices:</t> <t>A WebRTCEndpointendpoint might have multiple cameras, microphones, or other media capture devices, and so it might want to generate several RTP packet streams of the same media type. Alternatively, it might want to send media from a single capture device in several different formats or quality settings at once. Both can result in a single endpoint sending multiple RTP packet streams of the same media type into a single RTP session at the same time.</t><t hangText="Associated Repair Data:">An</li> <li> <t>Associated repair data:</t> <t>An endpoint might sendaan RTP packet stream that is somehow associated with another stream. For example, it might send an RTP packet stream that contains FEC or retransmission data relating to another stream. Some RTP payload formats send this sort of associated repair data as part of the source packet stream, while others send it as a separate packet stream.</t><t hangText="Layered or Multiple Description Coding:">An</li> <li> <t>Layered or multiple-description coding:</t> <t>Within a single RTP session, an endpoint can use a layered mediacodec,codec -- forexampleexample, H.264SVC,SVC -- or amultiple description codec,multiple-description codec that generates multiple RTP packet streams, each with a distinct RTPSSRC, within a single RTP session.</t> <t hangText="RTP Mixers, Translators,SSRC.</t> </li> <li> <t>RTP mixers, translators, andOther Middleboxes:">Another middleboxes:</t> <t>An RTP session, in the WebRTC context, is a point-to-point association between an endpoint and some other peer device, where those devices share a common SSRC space. The peer device might be another WebRTCEndpoint,endpoint, or it might be an RTP mixer, translator, or some other form ofmedia processingmedia-processing middlebox. In the latter cases, the middlebox might send mixed or relayed RTP streams from several participants,thatwhich the WebRTCEndpointendpoint will need to render. Thus, even though a WebRTCEndpointendpoint might only be a member of a single RTP session, the peer device might be extending that RTP session to incorporate other endpoints. WebRTC is a group communicationenvironmentenvironment, and endpoints need to be capable of receiving, decoding, and playing out multiple RTP packet streams at once, even in a single RTP session.</t></list></t></li> </ul> </section> <section anchor="sec.multiple-sessions"title="Usenumbered="true" toc="default"> <name>Use of Multiple RTPSessions">Sessions</name> <t>In addition to sending and receiving multiple RTP packet streams within a single RTP session, a WebRTCEndpointendpoint might participate in multiple RTP sessions. There are several reasons why a WebRTCEndpointendpoint might choose to do this:<list style="hanging"> <t hangText="To</t> <ul> <li><t>To interoperate with legacydevices:">Thedevices:</t> <t>The common practice in the non-WebRTC world is to send different types of media in separate RTPsessions,sessions -- forexampleexample, using one RTP session for audio and another RTP session, on a separatetransport layertransport-layer flow, for video. All WebRTCEndpointsendpoints need to support the option of sending different types of media on different RTPsessions,sessions so they can interwork with such legacy devices. This is discussed further in <xreftarget="sec.session-mux"></xref>.</t> <t hangText="Totarget="sec.session-mux" format="default"/>.</t></li> <li><t>To provide enhanced quality ofservice:">Someservice:</t> <t>Some network-basedquality of servicequality-of-service mechanisms operate on the granularity oftransport layertransport-layer flows. Ifit is desired touse of these mechanisms to provide differentiated quality of service for some RTP packetstreams,streams is desired, then those RTP packet streams need to be sent in a separate RTP session using a different transport-layer flow, and with appropriatequality of servicequality-of-service marking. This is discussed further in <xreftarget="sec-differentiated"></xref>.</t> <t hangText="Totarget="sec-differentiated" format="default"/>.</t></li> <li><t>To separate media with differentpurposes:">Anpurposes:</t> <t>An endpoint might want to send RTP packet streams that have different purposes on different RTP sessions, to make it easy for the peer device to distinguish them. For example, somecentralisedcentralized multiparty conferencing systems display the active speaker in highresolution,resolution but showlow resolutionlow-resolution "thumbnails" of other participants. Such systems might configure the endpoints to send simulcast high- and low-resolution versions of their video using separate RTPsessions,sessions to simplify the operation of the RTP middlebox. In the WebRTCcontextcontext, this is currently possible by establishing multiple WebRTC MediaStreamTracks that have the same media source in one (or more) RTCPeerConnection. Each MediaStreamTrack is then configured to deliver a particular media quality and thus mediabit-rate,bitrate, and it will produce an independently encoded version with the codec parameters agreed specifically in the context of that RTCPeerConnection. The RTP middlebox can distinguish packets corresponding to the low- and high-resolution streams by inspecting their SSRC, RTP payload type, or some other information contained in RTP payload, RTP headerextensionextension, or RTCPpackets, butpackets. However, it can be easier to distinguish the RTP packet streams if they arrive on separate RTP sessions on separate transport-layerflows.</t> <t hangText="Toflows.</t></li> <li><t>To directly connect with multiplepeers:">A multi-partypeers:</t> <t>A multiparty conference does not need to use an RTP middlebox. Rather, a multi-unicast mesh can be created, comprising several distinct RTP sessions, with each participant sending RTP traffic over a separate RTP session (that is, using an independent RTCPeerConnection object) to every other participant, as shown in <xreftarget="fig-mesh"></xref>.target="fig-mesh" format="default"/>. This topology has the benefit of not requiring an RTP middlebox node that is trusted to access and manipulate the media data. The downside is that it increases the used bandwidth at each sender by requiring one copy of the RTP packet streams for each participant thatareis part of the same session beyond the sender itself.</t></list></t><figurealign="center" anchor="fig-mesh" title="Multi-unicast using severalanchor="fig-mesh"> <name>Multi-unicast Using Several RTPsessions"> <artwork><![CDATA[Sessions</name> <artwork name="" type="" align="left" alt=""> <![CDATA[ +---+ +---+ | A |<--->| B | +---+ +---+ ^ ^ \ / \ / v v +---+ | C | +---+ ]]></artwork> </figure><t><list style="hanging"><t>The multi-unicast topology could also be implemented as a single RTP session, spanning multiple peer-to-peertransport layertransport-layer connections, or as several pairwise RTP sessions, one between each pair of peers. To maintain a coherent mapping of the relationship between RTP sessions and RTCPeerConnectionobjectsobjects, it isrecommendRECOMMENDED that thisisbe implemented as several individual RTP sessions. The only downside is that endpoint A will not learn of the quality of any transmission happening between B and C, since it will not see RTCP reports for the RTP session between B and C, whereas it would if all three participants were part of a single RTP session. Experience with the Mbone tools (experimental RTP-based multicast conferencing tools from the late 1990s) hasshowedshown that RTCP reception quality reports for third parties can be presented to users in a way that helps them understand asymmetric network problems, and the approach of using separate RTP sessions prevents this. However, an advantage of using separate RTP sessions is that it enables using different mediabit-ratesbitrates and RTP session configurations between the different peers, thus not forcing B to endure the same quality reductions as C will if there are limitations in the transport from A toC as C will.C. It is believed that these advantages outweigh the limitations in debugging power.</t><t hangText="To</li> <li><t>To indirectly connect with multiplepeers:">Apeers:</t> <t>A common scenario inmulti-partymultiparty conferencing is to create indirect connections to multiple peers, using an RTP mixer, translator, or some other type of RTP middlebox. <xreftarget="fig-mixerFirst"></xref>target="fig-mixerFirst" format="default"/> outlines a simple topology that might be used in a four-personcentralisedcentralized conference. The middlebox acts tooptimiseoptimize the transmission of RTP packet streams from certain perspectives, either by only sending some of the received RTP packet stream to any given receiver, or by providing a combined RTP packet stream out of a set of contributing streams.</t></list></t><figurealign="center" anchor="fig-mixerFirst" title="RTP mixeranchor="fig-mixerFirst"> <name>RTP Mixer withonly unicast paths"> <artwork><![CDATA[Only Unicast Paths</name> <artwork name="" type="" align="left" alt=""> <![CDATA[ +---+ +-------------+ +---+ | A |<---->| |<---->| B | +---+ | RTP mixer, | +---+ | translator, | | or other | +---+ | middlebox | +---+ | C |<---->| |<---->| D | +---+ +-------------+ +---+ ]]></artwork> </figure><t><list style="hanging"><t>There are various methods of implementation for the middlebox. If implemented as a standard RTP mixer or translator, a single RTP session will extend across the middlebox and encompass all the endpoints in onemulti-partymultiparty session. Other types ofmiddleboxmiddleboxes might use separate RTP sessions between each endpoint and the middlebox. A common aspect is that these RTP middleboxes can use a number of tools to control the media encoding provided by a WebRTCEndpoint.endpoint. This includes functions like requesting the breaking of the encoding chain andhavehaving the encoder produce aso calledso-called Intra frame. Another common aspect is limiting thebit-ratebitrate of agivenstream to bettersuit the mixer view ofmatch themultiple down-streams. Othersmixed output. Other aspects are controlling the most suitableframe-rate,frame rate, picture resolution, and the trade-off betweenframe-rateframe rate and spatial quality. The middlebox has the responsibility to correctly perform congestion control,source identification,identify sources, and managesynchronisationsynchronization while providing the application with suitable mediaoptimisations.optimizations. The middlebox also has to be a trusted node when it comes to security, since it manipulates either the RTP header or the media itself (or both) received from oneendpoint,endpoint before sendingitthem on towards theendpoint(s),endpoint(s); thus they need to be able to decrypt and then re-encrypt the RTP packet stream before sending it out.</t><t>RTP Mixers can create a situation where an endpoint experiences a situation in-between a session with only two endpoints and multiple RTP sessions. Mixers<t>Mixers are expected to not forward RTCP reports regarding RTP packet streams across themselves. This is due to the differenceinbetween the RTP packet streams provided to the different endpoints. The original media source lacks information about a mixer's manipulations prior tosending itbeing sent to the different receivers. This scenario also results inthatan endpoint's feedback or requestsgogoing to the mixer. When the mixer can't act on this by itself, it is forced to go to the original media source tofulfilfulfill thereceiversreceiver's request. This will not necessarily be explicitly visible to any RTP and RTCP traffic, but the interactions and the time to complete them will indicate such dependencies.</t> <t>Providing source authentication inmulti-partymultiparty scenarios is a challenge. In the mixer-based topologies, endpoints source authentication is based on, firstly, verifying that media comes from the mixer by cryptographic verification and, secondly, trust in the mixer to correctly identify any source towards the endpoint. In RTP sessions where multiple endpoints are directly visible to an endpoint, all endpoints will have knowledge about each others' masterkeys,keys and can thus inject packetsclaimedclaiming to come from another endpoint in the session. Any node performing relay can performnon-cryptographicnoncryptographic mitigation by preventing forwarding of packets that have SSRC fields that came from other endpoints before. For cryptographic verification of the source, SRTP would require additional securitymechanisms,mechanisms -- forexampleexample, <xreftarget="RFC4383">TESLAtarget="RFC4383" format="default"> Timed Efficient Stream Loss-Tolerant Authentication (TESLA) forSRTP</xref>,SRTP</xref> -- that are not part of the base WebRTC standards.</t><t hangText="To</li> <li><t>To forward media between multiplepeers:">Itpeers:</t> <t>It is sometimes desirable for an endpoint that receives an RTP packet stream to be able to forward that RTP packet stream to a third party. The are some obvious security and privacy implications in supporting this, but also potential uses. This is supported in the W3C API by taking the received and decoded media and using it as a media source that isre-encodingre-encoded and transmitted as a new stream.</t> <t>At the RTP layer, media forwarding acts as a back-to-back RTP receiver and RTP sender. The receiving side terminates the RTP session and decodes the media, while the sender side re-encodes and transmits the media using an entirely separate RTP session. The original sender will only see a single receiver of the media, and will not be able to tell that forwarding is happening based on RTP-layerinformationinformation, since the RTP session that is used to send the forwarded media is not connected to the RTP session on which the media was received by the node doing the forwarding.</t> <t>The endpoint that is performing the forwarding is responsible for producing an RTP packet stream suitable for onwards transmission. The outgoing RTP session that is used to send the forwarded media is entirely separatetofrom the RTP session on which the media was received. This will require media transcoding for congestion controlpurposepurposes to produce a suitablebit-ratebitrate for the outgoing RTP session, reducing media quality and forcing the forwarding endpoint to spend the resource on the transcoding. The media transcoding does result in a separation of the two differentlegslegs, removing almost all dependencies, and allowing the forwarding endpoint tooptimiseoptimize its media transcoding operation. The cost is greatly increased computational complexity on the forwarding node. Receivers of the forwarded stream will see the forwarding device as the sender of thestream,stream and will not be able to tell from the RTP layer that they are receiving a forwarded stream rather than an entirely new RTP packet stream generated by the forwarding device.</t></list></t></li> </ul> </section> <section anchor="sec-differentiated"title="Differentiatednumbered="true" toc="default"> <name>Differentiated Treatment of RTPStreams">Streams</name> <t>There are use cases for differentiated treatment of RTP packet streams. Such differentiation can happen at several places in the system. First of all is the prioritization within the endpoint sending the media, whichcontrols,controls both which RTP packet streamsthatwill besent,sent and their allocation ofbit-ratebitrate out of the current availableaggregateaggregate, as determined by the congestion control.</t> <t>It is expected that the <xreftarget="W3C.WD-webrtc-20130910">WebRTCtarget="W3C.WebRTC" format="default">WebRTC API</xref> will allow the application to indicate relative priorities for different MediaStreamTracks. These priorities can then be used to influence the local RTP processing, especially when it comes tocongestion control response indetermining how to divide the available bandwidth between the RTP packetstreams.streams for the sake of congestion control. Any changes in relative priority will also need to be considered for RTP packet streams that are associated with the main RTP packet streams, such as redundant streams for RTP retransmission and FEC. The importance of such redundant RTP packet streams is dependent on the media type and codec used,in regardswith regard to how robust that codec istoagainst packet loss. However, a default policy mighttobe to use the same priority for a redundant RTP packet stream as for the source RTP packet stream.</t> <t>Secondly, the network can prioritize transport-layer flows andsub-flows,subflows, including RTP packet streams. Typically, differential treatment includes two steps, the first being identifying whether an IP packet belongs to a class that has to be treated differently, the second consisting of the actual mechanismto prioritizefor prioritizing packets. Three common methods for classifying IP packets are:<list style="hanging"> <t hangText="DiffServ:">The</t> <dl> <dt>DiffServ:</dt> <dd>The endpoint marks a packet with a DiffServ code point to indicate to the network that the packet belongs to a particularclass.</t> <t hangText="Flow based:">Packetsclass.</dd> <dt>Flow based:</dt> <dd>Packets that need to be given a particular treatment are identified using a combination of IP and portaddress.</t> <t hangText="Deep Packet Inspection:">Aaddress.</dd> <dt>Deep packet inspection:</dt> <dd>A network classifier (DPI) inspects the packet and tries to determine if the packet represents a particular application and type that is to beprioritized.</t> </list></t>prioritized.</dd> </dl> <t>Flow-based differentiation will provide the same treatment to all packets within a transport-layer flow, i.e., relative prioritization is not possible. Moreover, if the resources arelimitedlimited, it might not be possible to provide differential treatment compared tobest-effortbest effort for all the RTP packet streams used in a WebRTC session. The use of flow-based differentiation needs to be coordinated between the WebRTC system and the network(s). The WebRTC endpoint needs to know that flow-based differentiation might be used to provide the separation of the RTP packet streams onto different UDP flows to enable a more granular usage offlow basedflow-based differentiation. The used flows, their5-tuples5-tuples, and prioritization will need to be communicated to the network so that it can identify the flows correctly to enable prioritization. No specific protocol support for this is specified.</t> <t>DiffServ assumes that either the endpoint or a classifier can mark the packets with an appropriateDSCPDifferentiated Services Code Point (DSCP) so that the packets are treated according to that marking. If the endpoint is to mark thetraffictraffic, two requirements arise in the WebRTC context: 1) The WebRTCEndpointendpoint has to know whichDSCPDSCPs to use and know that it can use them on some set of RTP packet streams. 2) The information needs to be propagated to the operating system when transmitting the packet. Details of this process are outside the scope of this memo and are further discussed in <xreftarget="I-D.ietf-tsvwg-rtcweb-qos">"DSCP and other packet markingstarget="RFC8837" format="default">"Differentiated Services Code Point (DSCP) Packet Markings forRTCWebWebRTC QoS"</xref>.</t><t>Deep Packet Inspectors will, despite<t>Despite the SRTP media encryption, deep packet inspectors will still be fairly capableatof classifying the RTP streams. The reason is that SRTP leaves the first 12 bytes of the RTP header unencrypted. This enables easy RTP stream identification using the SSRC and provides the classifier with useful information that can be correlated todeterminedetermine, forexampleexample, the stream's media type. Using packet sizes, reception times, packet inter-spacing, RTP timestampincrementsincrements, and sequence numbers, fairly reliable classifications are achieved.</t> <t>Forpacket basedpacket-based markingschemesschemes, it might be possible to mark individual RTP packets differently based on the relative priority of the RTP payload. Forexampleexample, video codecs that have I, P, and B pictures couldprioritiseprioritize any payloads carrying only B frames less, as these are less damaging toloose.lose. However, depending on the QoS mechanism and what markingsthatare applied, this can result in not only differentpacket droppacket-drop probabilities but also packetreordering,reordering; see <xreftarget="I-D.ietf-tsvwg-rtcweb-qos"></xref>target="RFC8837" format="default"/> and <xreftarget="I-D.ietf-dart-dscp-rtp"></xref>target="RFC7657" format="default"/> for further discussion. As a defaultpolicypolicy, all RTP packets related toaan RTP packet stream ought to be provided with the same prioritization; per-packet prioritization is outside the scope of thismemo,memo but might be specified elsewhere in future.</t> <t>It is also important to consider how RTCP packets associated with a particular RTP packet stream need to be marked. RTCP compound packets with Sender Reports(SR),(SRs) ought to be marked with the same priority as the RTP packet stream itself, so the RTCP-based round-trip time (RTT) measurements are done using the same transport-layer flow priority as the RTP packet stream experiences. RTCP compound packets containing an RR packet ought to be sent with the priority used by the majority of the RTP packet streams reported on. RTCP packets containing time-critical feedback packets can use higher priority to improve the timeliness and likelihood of delivery of such feedback.</t> </section> </section> <sectiontitle="Medianumbered="true" toc="default"> <name>Media Source, RTP Streams, and ParticipantIdentification">Identification</name> <section anchor="sec-media-stream-id"title="Medianumbered="true" toc="default"> <name>Media SourceIdentification">Identification</name> <t>Each RTP packet stream is identified by a uniquesynchronisationsynchronization source (SSRC) identifier. The SSRC identifier is carried in each of the RTP packets comprisingaan RTP packet stream, and is also used to identify that stream in the corresponding RTCP reports. The SSRC is chosen as discussed in <xreftarget="sec-ssrc"></xref>.target="sec-ssrc" format="default"/>. The first stage in demultiplexing RTP and RTCP packets received on a singletransport layertransport-layer flow at a WebRTCEndpointendpoint is to separate the RTP packet streams based on their SSRC value; once that is done, additional demultiplexing steps can determine how and where to render the media.</t> <t>RTP allows a mixer, or other RTP-layer middlebox, to combine encoded streams from multiple media sources to form a new encoded stream from a new media source (the mixer). The RTP packets in that new RTP packet stream can include aContributing Sourcecontributing source (CSRC) list, indicating which original SSRCs contributed to the combined source stream. As described in <xreftarget="sec-rtp-rtcp"></xref>,target="sec-rtp-rtcp" format="default"/>, implementations need to support reception of RTP data packets containing a CSRC list and RTCP packets that relate to sources present in the CSRC list. The CSRC list can change on a packet-by-packet basis, depending on the mixing operation being performed. Knowledge of what media sources contributed to a particular RTP packet can be important if the user interface indicates which participants are active in the session. Changes in the CSRC list included in packetsneedsneed to be exposed to the WebRTC application using someAPI,API if the application is to be able to track changes in session participation. It is desirable to map CSRC values back into WebRTC MediaStream identities as they cross this API, to avoid exposing the SSRC/CSRCname spacenamespace to WebRTC applications.</t> <t>If the mixer-to-client audio level extension <xreftarget="RFC6465"></xref>target="RFC6465" format="default"/> is being used in the session (see <xreftarget="sec-mixer-to-client"></xref>),target="sec-mixer-to-client" format="default"/>), the information in the CSRC list is augmented byaudio levelaudio-level information for each contributing source. It is desirable to expose this information to the WebRTC application using some API, after mapping the CSRC values to WebRTC MediaStream identities, so it can be exposed in the user interface.</t> </section> <sectiontitle="SSRCnumbered="true" toc="default"> <name>SSRC CollisionDetection">Detection</name> <t>The RTP standard requires RTP implementations to have support for detecting and handling SSRCcollisions,collisions -- i.e., be able to resolve the conflict when two different endpoints use the same SSRC value (seesection 8.2 of<xreftarget="RFC3550"></xref>).target="RFC3550" section="8.2" sectionFormat="of"/>). This requirement also applies to WebRTCEndpoints.endpoints. There are several scenarios where SSRC collisions can occur:<list style="symbols"> <t>In</t> <ul spacing="normal"> <li>In a point-to-point session where each SSRC is associated with either of the two endpoints andwherethe mainmedia carryingmedia-carrying SSRC identifier will be announced in thesignallingsignaling channel, a collision is less likely to occur due to the information about used SSRCs. If SDP is used, this information is provided by <xreftarget="RFC5576">Source-Specifictarget="RFC5576" format="default">source-specific SDPAttributes</xref>.attributes</xref>. Still, collisions can occur if both endpoints start using a new SSRC identifier prior to havingsignalledsignaled it to the peer and received acknowledgement on thesignallingsignaling message.The<xreftarget="RFC5576">Source-Specific SDP Attributes</xref>target="RFC5576" format="default">"Source-Specific Media Attributes in the Session Description Protocol (SDP)"</xref> contains a mechanism to signal how the endpoint resolved the SSRCcollision.</t> <t>SSRCcollision.</li> <li>SSRC values that have not beensignalledsignaled could also appear in an RTP session. This is more likely than it appears, since some RTP functions use extra SSRCs to provide their functionality. For example, retransmission data might be transmitted using a separate RTP packet stream that requires its own SSRC, separatetofrom the SSRC of the source RTP packet stream <xreftarget="RFC4588"></xref>.target="RFC4588" format="default"/>. In those cases, an endpoint can create a new SSRC that strictly doesn't need to be announced over thesignallingsignaling channel to function correctly on both RTP and RTCPeerConnectionlevel.</t> <t>Multiplelevel.</li> <li>Multiple endpoints in a multiparty conference can create new sources and signal those towards the RTP middlebox. In cases where the SSRC/CSRC are propagated between the different endpoints from the RTPmiddleboxmiddlebox, collisions canoccur.</t> <t>Anoccur.</li> <li>An RTP middlebox could connect an endpoint's RTCPeerConnection to another RTCPeerConnection from the same endpoint, thus forming a loop where the endpoint will receive its own traffic. While it is clearly considered a bug, it is important that the endpointisbe able torecogniserecognize and handle the case when it occurs. This case becomes even more problematic when mediamixers,mixers andso on,such are involved, where the stream received is a different stream but still contains this client'sinput.</t> </list></t>input.</li> </ul> <t>These SSRC/CSRC collisions can only be handled on the RTP levelas long aswhen the same RTP session is extended across multiple RTCPeerConnections byaan RTP middlebox. To resolve the more generic case where multiple RTCPeerConnections are interconnected, identification of the mediasource(s)source or sources that are part of a MediaStreamTrack being propagated across multiple interconnected RTCPeerConnection needs to be preserved across these interconnections.</t> </section> <sectiontitle="Media Synchronisation Context">numbered="true" toc="default"> <name>Media Synchronization Context</name> <t>When an endpoint sends media from more than one media source, it needs to consider if (and which of) these media sources are to be synchronized. In RTP/RTCP,synchronisationsynchronization is provided by having a set of RTP packet streams be indicated as coming from the samesynchronisationsynchronization context and logical endpoint by using the same RTCP CNAME identifier.</t> <t>The next provision is that the internal clocks of all mediasources,sources -- i.e., what drives the RTPtimestamp,timestamp -- can be correlated to a system clock that is provided in RTCP Sender Reports encoded in an NTP format. By correlating all RTP timestamps to a common system clock for all sources, the timing relation of the different RTP packet streams, also across multiple RTPsessionssessions, can be derived at the receiver and, if desired, the streams can be synchronized. The requirement is for the media sender to provide the correlation information;itwhether or not the information is used is up to thereceiver to use it or not.</t>receiver.</t> </section> </section> </section> <section anchor="sec-security"title="Security Considerations">numbered="true" toc="default"> <name>Security Considerations</name> <t>The overall security architecture for WebRTC is described in <xreftarget="I-D.ietf-rtcweb-security-arch"></xref>,target="RFC8827" format="default"/>, and security considerations for the WebRTC framework are described in <xreftarget="I-D.ietf-rtcweb-security"></xref>.target="RFC8826" format="default"/>. These considerations also apply to this memo.</t> <t>The security considerations of the RTP specification, the RTP/SAVPF profile, and the various RTP/RTCP extensions and RTP payload formats that form the complete protocol suite described in this memo apply. It isnotbelieved that there areanyno new security considerations resulting from the combination of these various protocol extensions.</t><t>The <xref target="RFC5124">Extended<t><xref target="RFC5124" format="default">"Extended Secure RTP Profile for Real-time Transport Control Protocol (RTCP)-BasedFeedback</xref> (RTP/SAVPF)Feedback (RTP/SAVPF)"</xref> provides handling of fundamental issues by offering confidentiality,integrityintegrity, and partial source authentication. A media-security solution that is mandatory to implement and usemedia security solutionis created by combining this secured RTP profile and <xreftarget="RFC5764">DTLS-SRTP keying</xref>target="RFC5764" format="default">DTLS-SRTP keying</xref>, as defined by <xreftarget="I-D.ietf-rtcweb-security-arch">Section 5.5 of</xref>.</t>target="RFC8827" section="5.5" sectionFormat="of"/>.</t> <t>RTCP packets convey a Canonical Name (CNAME) identifier that is used to associate RTP packet streams that need to besynchronisedsynchronized across related RTP sessions. Inappropriate choice of CNAME values can be a privacy concern, since long-term persistent CNAME identifiers can be used to track users across multiple WebRTC calls. <xreftarget="sec-cname"></xref>target="sec-cname" format="default"/> of this memo mandates generation of short-term persistent RTCP CNAMES, as specified inRFC7022,RFC 7022, resulting in untraceable CNAME values that alleviate this risk.</t> <t>Some potentialdenial of servicedenial-of-service attacks exist if the RTCP reporting interval is configured to an inappropriate value. This could be done by configuring the RTCP bandwidth fraction to an excessively large or small value using the SDP "b=RR:" or "b=RS:" lines <xreftarget="RFC3556"></xref>,target="RFC3556" format="default"/> or some similar mechanism, or by choosing an excessively large or small value for the RTP/AVPF minimal receiver report interval (if using SDP, this is the "a=rtcp-fb:... trr-int" parameter) <xreftarget="RFC4585"></xref>.target="RFC4585" format="default"/>. The risks are asfollows:<list style="numbers"> <t>thefollows:</t> <ol spacing="normal" type="1"> <li>the RTCP bandwidth could be configured to make the regular reporting interval so large that effective congestion control cannot be maintained, potentially leading to denial of service due to congestion caused by the mediatraffic;</t> <t>thetraffic;</li> <li>the RTCP interval could be configured to a very small value, causing endpoints to generatehigh ratehigh-rate RTCP traffic, potentially leading to denial of service due to thenon-congestion controlledRTCPtraffic; and</t> <t>RTCPtraffic not being congestion controlled; and</li> <li>RTCP parameters could be configured differently for each endpoint, with some of the endpoints using a large reporting interval and some using a smaller interval, leading to denial of service due to premature participant timeouts due to mismatched timeout periodswhichthat are based on the reportinginterval (thisinterval. This is a particular concern if endpoints use a small butnon-zerononzero value for the RTP/AVPF minimal receiver report interval (trr-int) <xreftarget="RFC4585"></xref>,target="RFC4585" format="default"/>, as discussed inSection 6.1 of<xreftarget="I-D.ietf-avtcore-rtp-multi-stream"></xref>).</t> </list>Prematuretarget="RFC8108" section="6.1" sectionFormat="of"/>.</li> </ol> <t>Premature participant timeout can be avoided by using the fixed(non-reduced)(nonreduced) minimum interval when calculating the participant timeout (see <xreftarget="sec-rtp-rtcp"></xref>target="sec-rtp-rtcp" format="default"/> of this memo andSection 6.1 of<xreftarget="I-D.ietf-avtcore-rtp-multi-stream"></xref>).target="RFC8108" section="7.1.2" sectionFormat="of"/>). To address the other concerns, endpointsSHOULD<bcp14>SHOULD</bcp14> ignore parameters that configure the RTCP reporting interval to be significantly longer than the defaultfive secondfive-second interval specified in <xreftarget="RFC3550"></xref>target="RFC3550" format="default"/> (unless the media data rate is so low that the longer reporting interval roughly corresponds to 5% of the media data rate), or that configure the RTCP reporting interval small enough that the RTCP bandwidth would exceed the media bandwidth.</t> <t>The guidelines in <xreftarget="RFC6562"></xref>target="RFC6562" format="default"/> apply when using variablebit ratebitrate (VBR) audio codecs such as Opus (see <xreftarget="sec.codecs"></xref>target="sec.codecs" format="default"/> for discussion of mandated audio codecs). The guidelines in <xreftarget="RFC6562"></xref>target="RFC6562" format="default"/> also apply, but are of lesser importance, when using the client-to-mixer audio level header extensions (<xreftarget="sec-client-to-mixer"></xref>)target="sec-client-to-mixer" format="default"/>) or the mixer-to-client audio level header extensions (<xreftarget="sec-mixer-to-client"></xref>).target="sec-mixer-to-client" format="default"/>). The use of the encryption of the header extensions areRECOMMENDED,<bcp14>RECOMMENDED</bcp14>, unless there are known reasons, like RTP middleboxes performingvoice activity basedvoice-activity-based source selection orthird partythird-party monitoring that will greatly benefit from the information, and this has been expressed using API orsignalling.signaling. If further evidenceareis produced to show that information leakage is significant fromaudio levelaudio-level indications, then use of encryption needs to be mandated at that time.</t> <t>Inmulti-partymultiparty communication scenarios using RTPMiddleboxes,middleboxes, a lot of trust is placed on these middleboxes to preserve thesessionssession's security. The middlebox needs to maintainthe confidentiality,confidentiality and integrity and perform source authentication. As discussed in <xreftarget="sec.multiple-flows"></xref>target="sec.multiple-flows" format="default"/>, the middlebox can perform checks thatpreventsprevent any endpoint participating in a conferenceto impersonatefrom impersonating another. Some additional security considerations regardingmulti-partymultiparty topologies can be found in <xreftarget="I-D.ietf-avtcore-rtp-topologies-update"></xref>.</t>target="RFC7667" format="default"/>.</t> </section> <section anchor="sec-iana"title="IANA Considerations">numbered="true" toc="default"> <name>IANA Considerations</name> <t>Thismemo makesdocument has norequest of IANA.</t> <t>Note to RFC Editor: this section is to be removed on publication as an RFC.</t> </section> <section anchor="Acknowledgements" title="Acknowledgements"> <t>The authors would like to thank Bernard Aboba, Harald Alvestrand, Cary Bran, Ben Campbell, Alissa Cooper, Spencer Dawkins, Charles Eckel, Alex Eleftheriadis, Christian Groves, Chris Inacio, Cullen Jennings, Olle Johansson, Suhas Nandakumar, Dan Romascanu, Jim Spring, Martin Thomson, and the other members of the IETF RTCWEB working group for their valuable feedback.</t>IANA actions.</t> </section> </middle> <back><references title="Normative References"> <?rfc include="reference.RFC.3550"?> <?rfc include='reference.RFC.2119'?> <?rfc include='reference.RFC.2736'?> <?rfc include='reference.RFC.3551'?> <?rfc include='reference.RFC.3556'?> <?rfc include='reference.RFC.3711'?> <?rfc include='reference.RFC.4566'?> <?rfc include='reference.RFC.4585'?> <?rfc include='reference.RFC.4588'?> <?rfc include='reference.RFC.4961'?> <?rfc include='reference.RFC.5104'?> <?rfc include='reference.RFC.5124'?> <?rfc include='reference.RFC.5285'?> <?rfc include='reference.RFC.5506'?> <?rfc include='reference.RFC.5761'?> <?rfc include='reference.RFC.5764'?> <?rfc include='reference.RFC.6051'?> <?rfc include='reference.RFC.6464'?> <?rfc include='reference.RFC.6465'?> <?rfc include='reference.RFC.6562'?> <?rfc include='reference.RFC.6904'?> <?rfc include='reference.RFC.7007'?> <?rfc include='reference.RFC.7022'?> <?rfc include='reference.RFC.7160'?> <?rfc include='reference.RFC.7164'?> <?rfc include='reference.I-D.ietf-avtcore-multi-media-rtp-session'?> <?rfc include='reference.I-D.ietf-mmusic-mux-exclusive'?> <?rfc include='reference.I-D.ietf-avtcore-rtp-multi-stream'?> <?rfc include='reference.I-D.ietf-avtcore-rtp-multi-stream-optimisation'?> <?rfc include='reference.I-D.ietf-rtcweb-audio'?> <?rfc include='reference.I-D.ietf-rtcweb-video'?> <?rfc include='reference.I-D.ietf-rtcweb-security'?> <?rfc include='reference.I-D.ietf-avtcore-rtp-circuit-breakers'?> <?rfc include='reference.I-D.ietf-rtcweb-security-arch'?> <?rfc include='reference.I-D.ietf-rtcweb-fec'?> <?rfc include='reference.I-D.ietf-mmusic-sdp-bundle-negotiation'?> <?rfc include='reference.I-D.ietf-rtcweb-overview'?> <?rfc include='reference.I-D.ietf-avtcore-rtp-topologies-update'?><displayreference target="I-D.ietf-avtcore-multiplex-guidelines" to="MULTIPLEX"/> <references> <name>References</name> <references> <name>Normative References</name> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.3550.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.2119.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.2736.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.3551.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.3556.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.3711.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.4566.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.4585.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.4588.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.4961.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.5104.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.5124.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.8285.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.5506.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.5761.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.5764.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.6051.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.6464.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.6465.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.6562.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.6904.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7007.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7022.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7160.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7164.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7742.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7874.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.8083.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.8174.xml"/> <!-- draft-ietf-avtcore-multi-media-rtp-session: RFC 8860 --> <referenceanchor='W3C.WD-webrtc-20130910' target='http://www.w3.org/TR/2013/WD-webrtc-20130910'>anchor="RFC8860" target="https://www.rfc-editor.org/info/rfc8860"> <front><title>WebRTC 1.0: Real-time Communication Between Browsers</title><title>Sending Multiple Types of Media in a Single RTP Session</title> <authorinitials='A.' surname='Bergkvist' fullname='Adam Bergkvist'> <organization />initials="M." surname="Westerlund" fullname="Magnus Westerlund"> <organization/> </author> <authorinitials='D.' surname='Burnett' fullname='Daniel Burnett'> <organization />initials="C." surname="Perkins" fullname="Colin Perkins"> <organization/> </author> <author initials="J." surname="Lennox" fullname="Jonathan Lennox"> <organization/> </author> <date month="October" year="2020"/> </front> <seriesInfo name="RFC" value="8860"/> <seriesInfo name="DOI" value="10.17487/RFC8860"/> </reference> <!-- draft-ietf-avtcore-rtp-multi-stream: RFC 8108 --> <xi:include href="https://xml2rfc.ietf.org/public/rfc/bibxml/reference.RFC.8108.xml"/> <!-- draft-ietf-avtcore-rtp-multi-stream-optimisation: RFC 8861 --> <reference anchor="RFC8861" target="https://www.rfc-editor.org/info/rfc8861"> <front> <title>Sending Multiple RTP Streams in a Single RTP Session: Grouping RTP Control Protocol (RTCP) Reception Statistics and Other Feedback</title> <author initials="J." surname="Lennox" fullname="Jonathan Lennox"> <organization/> </author> <author initials="M." surname="Westerlund" fullname="Magnus Westerlund"> <organization/> </author> <author initials="Q." surname="W" fullname="Qin Wu"> <organization/> </author> <author initials="C." surname="Perkins" fullname="Colin Perkins"> <organization/> </author> <date month="October" year="2020"/> </front> <seriesInfo name="RFC" value="8861"/> <seriesInfo name="DOI" value="10.17487/RFC8861"/> </reference> <!--draft-ietf-mmusic-mux-exclusive-12; part of C238; RFC 8858--> <reference anchor='RFC8858' target="https://www.rfc-editor.org/info/rfc8858"> <front> <title>Indicating Exclusive Support of RTP and RTP Control Protocol (RTCP) Multiplexing Using the Session Description Protocol (SDP)</title> <author initials='C.'surname='Jennings' fullname='Cullen Jennings'>surname='Holmberg' fullname='Christer Holmberg'> <organization /> </author> <date month="October" year='2020' /> </front> <seriesInfo name='RFC' value='8858' /> <seriesInfo name="DOI" value="10.17487/RFC8858"/> </reference> <!-- draft-ietf-rtcweb-fec: RFC 8854 --> <reference anchor="RFC8854" target="https://www.rfc-editor.org/info/rfc8854"> <front> <title>WebRTC Forward Error Correction Requirements</title> <author initials="J." surname="Uberti" fullname="Justin Uberti"> <organization/> </author> <date month="October" year="2020"/> </front> <seriesInfo name="RFC" value="8854"/> <seriesInfo name="DOI" value="10.17487/RFC8854"/> </reference> <!-- draft-ietf-rtcweb-overview: RFC 8825 --> <reference anchor="RFC8825" target="https://www.rfc-editor.org/info/rfc8825"> <front> <title>Overview: Real-Time Protocols for Browser-Based Applications</title> <authorinitials='A.' surname='Narayanan' fullname='Anant Narayanan'>initials="H." surname="Alvestrand" fullname="Harald T. Alvestrand"> <organization /> </author> <datemonth='September' day='10' year='2013'month="October" year="2020" /> </front> <seriesInfoname='World Wide Web Consortium WD' value='WD-webrtc-20130910' /> <format type='HTML' target='http://www.w3.org/TR/2013/WD-webrtc-20130910'name="RFC" value="8825" /> <seriesInfo name="DOI" value="10.17487/RFC8825"/> </reference> <!--draft-ietf-rtcweb-security: RFC 8826 --> <reference anchor="RFC8826" target="https://www.rfc-editor.org/info/rfc8826"> <front> <title>Security Considerations for WebRTC</title> <author initials='E.' surname='Rescorla' fullname='Eric Rescorla'> <organization/> </author> <date month='October' year='2020'/> </front> <seriesInfo name="RFC" value="8826"/> <seriesInfo name="DOI" value="10.17487/RFC8826"/> </reference> <!--draft-ietf-rtcweb-security-arch: RFC 8827 --> <reference anchor="RFC8827" target="https://www.rfc-editor.org/info/rfc8827"> <front> <title>WebRTC Security Architecture</title> <author initials='E.' surname='Rescorla' fullname='Eric Rescorla'> <organization/> </author> <date month='October' year='2020'/> </front> <seriesInfo name="RFC" value="8827"/> <seriesInfo name="DOI" value="10.17487/RFC8827"/> </reference> <!-- draft-ietf-mmusic-sdp-bundle-negotiation (RFC 8843) --> <referenceanchor='W3C.WD-mediacapture-streams-20130903' target='http://www.w3.org/TR/2013/WD-mediacapture-streams-20130903'>anchor="RFC8843" target="https://www.rfc-editor.org/info/rfc8843"> <front> <title>Negotiating Media Multiplexing Using the Session Description Protocol (SDP)</title> <author initials="C" surname="Holmberg" fullname="Christer Holmberg"> <organization/> </author> <author initials="H" surname="Alvestrand" fullname="Harald Alvestrand"> <organization/> </author> <author initials="C" surname="Jennings" fullname="Cullen Jennings"> <organization/> </author> <date month="October" year="2020"/> </front> <seriesInfo name="RFC" value="8843"/> <seriesInfo name="DOI" value="10.17487/RFC8843"/> </reference> <reference anchor="W3C.WebRTC" target="https://www.w3.org/TR/2019/CR-webrtc-20191213/"> <front> <title>WebRTC 1.0: Real-time Communication Between Browsers</title> <author initials="C." surname="Jennings"> <organization/> </author> <author initials="H." surname="Boström"> <organization/> </author> <author initials="J-I." surname="Bruaroey"> <organization/> </author> <date year="2019" month="December" day="13"/> </front> <refcontent>W3C Candidate Recommendation</refcontent> </reference> <reference anchor="W3C-MEDIA-CAPTURE" target="https://www.w3.org/TR/2019/CR-mediacapture-streams-20190702/"> <front> <title>Media Capture and Streams</title> <authorinitials='D.' surname='Burnett' fullname='Daniel Burnett'> <organization />initials="D." surname="Burnett" fullname="Daniel Burnett"> <organization/> </author> <authorinitials='A.' surname='Bergkvist' fullname='Adam Bergkvist'> <organization />initials="A." surname="Bergkvist" fullname="Adam Bergkvist"> <organization/> </author> <authorinitials='C.' surname='Jennings' fullname='Cullen Jennings'> <organization />initials="C." surname="Jennings" fullname="Cullen Jennings"> <organization/> </author> <authorinitials='A.' surname='Narayanan' fullname='Anant Narayanan'> <organization />initials="A." surname="Narayanan" fullname="Anant Narayanan"> <organization/> </author> <author initials="B" surname="Aboba" fullname="Bernard Aboba"> <organization/> </author> <author initials="J-I." surname="Bruaroey" fullname="Jan-Ivar Bruaroey"> <organization/> </author> <author initials="H" surname="Boström" fullname="Henrik Boström"> <organization/> </author> <datemonth='September' day='3' year='2013'month="July" day="2" year="2019"/> </front> <refcontent>W3C Candidate Recommendation</refcontent> </reference> </references> <references> <name>Informative References</name> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.3611.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.4383.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.8445.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.5576.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.5968.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.6263.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.6792.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7478.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7656.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7657.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.7667.xml"/> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml/reference.RFC.8088.xml"/> <!-- draft-ietf-mmusic-msid-17 (RFC 8830) --> <reference anchor="RFC8830" target="https://www.rfc-editor.org/info/rfc8830"> <front> <title>WebRTC MediaStream Identification in the Session Description Protocol</title> <author initials="H" surname="Alvestrand" fullname="Harald Alvestrand"> <organization/> </author> <date month="October" year="2020"/> </front> <seriesInfo name="RFC" value="8830" /> <seriesInfo name="DOI" value="10.17487/RFC8830"/> </reference> <!-- draft-ietf-avtcore-multiplex-guidelines (EDIT) --> <xi:include href="https://xml2rfc.tools.ietf.org/public/rfc/bibxml3/reference.I-D.ietf-avtcore-multiplex-guidelines.xml"/> <!-- draft-ietf-rmcat-cc-requirements-09: RFC 8836 --> <reference anchor="RFC8836" target="https://www.rfc-editor.org/info/rfc8836"> <front> <title>Congestion Control Requirements for Interactive Real-Time Media</title> <author initials="R" surname="Jesup" fullname="Randell Jesup"> <organization/> </author> <author initials="Z" surname="Sarker" fullname="Zaheduzzaman Sarker" role="editor"> <organization/> </author> <date month="October" year="2020"/> </front> <seriesInfoname='World Wide Web Consortium WD' value='WD-mediacapture-streams-20130903'name="RFC" value="8836" /><format type='HTML' target='http://www.w3.org/TR/2013/WD-mediacapture-streams-20130903'<seriesInfo name="DOI" value="10.17487/RFC8836"/> </reference> <!-- draft-ietf-tsvwg-rtcweb-qos-18: RFC 8837 --> <reference anchor="RFC8837" target="https://www.rfc-editor.org/info/rfc8837"> <front> <title>Differentiated Services Code Point (DSCP) Packet Markings for WebRTC QoS</title> <author initials="P." surname="Jones" fullname="Paul Jones"> <organization/> </author> <author initials="S." surname="Dhesikan" fullname="Subha Dhesikan"> <organization/> </author> <author initials="C." surname="Jennings" fullname="Cullen Jennings"> <organization/> </author> <author initials="D." surname="Druta" fullname="Dan Druta"> <organization/> </author> <date month="October" year="2020"/> </front> <seriesInfo name="RFC" value="8837" /> <seriesInfo name="DOI" value="10.17487/RFC8837"/> </reference> <reference anchor="RFC8829" target="https://www.rfc-editor.org/info/rfc8829"> <front> <title>JavaScript Session Establishment Protocol (JSEP)</title> <author initials='J.' surname='Uberti' fullname='Justin Uberti'> <organization/> </author> <author initials="C." surname="Jennings" fullname="Cullen Jennings"> <organization/> </author> <author initials="E." surname="Rescorla" fullname="Eric Rescorla" role="editor"> <organization/> </author> <date month='October' year='2020'/> </front> <seriesInfo name="RFC" value="8829"/> <seriesInfo name="DOI" value="10.17487/RFC8829"/> </reference> </references><references title="Informative References"> <?rfc include='reference.RFC.3611'?> <?rfc include='reference.RFC.4383'?> <?rfc include='reference.RFC.5245'?> <?rfc include='reference.RFC.5576'?> <?rfc include='reference.RFC.5968'?> <?rfc include='reference.RFC.6263'?> <?rfc include='reference.RFC.6792'?> <?rfc include='reference.RFC.7478'?> <?rfc include='reference.I-D.ietf-mmusic-msid'?> <?rfc include='reference.I-D.ietf-avtcore-multiplex-guidelines'?> <?rfc include='reference.I-D.ietf-payload-rtp-howto'?> <?rfc include='reference.I-D.ietf-rmcat-cc-requirements'?> <?rfc include='reference.I-D.ietf-tsvwg-rtcweb-qos'?> <?rfc include='reference.I-D.ietf-avtext-rtp-grouping-taxonomy'?> <?rfc include='reference.I-D.ietf-dart-dscp-rtp'?> <?rfc include='reference.I-D.ietf-rtcweb-jsep'?></references> <section anchor="Acknowledgements" numbered="false" toc="default"> <name>Acknowledgements</name> <t>The authors would like to thank <contact fullname="Bernard Aboba"/>, <contact fullname="Harald Alvestrand"/>, <contact fullname="Cary Bran"/>, <contact fullname="Ben Campbell"/>, <contact fullname="Alissa Cooper"/>, <contact fullname="Spencer Dawkins"/>, <contact fullname="Charles Eckel"/>, <contact fullname="Alex Eleftheriadis"/>, <contact fullname="Christian Groves"/>, <contact fullname="Chris Inacio"/>, <contact fullname="Cullen Jennings"/>, <contact fullname="Olle Johansson"/>, <contact fullname="Suhas Nandakumar"/>, <contact fullname="Dan Romascanu"/>, <contact fullname="Jim Spring"/>, <contact fullname="Martin Thomson"/>, and the other members of the IETF RTCWEB working group for their valuable feedback.</t> </section> </back> </rfc><!-- vim: set ts=2 sw=2 tw=78 et ai: -->