/* * Copyright (c) 2015 The WebRTC project authors. All Rights Reserved. * * Use of this source code is governed by a BSD-style license * that can be found in the LICENSE file in the root of the source * tree. An additional intellectual property rights grant can be found * in the file PATENTS. All contributing project authors may * be found in the AUTHORS file in the root of the source tree. */ #include "video/payload_router.h" #include "modules/rtp_rtcp/include/rtp_rtcp.h" #include "modules/rtp_rtcp/include/rtp_rtcp_defines.h" #include "modules/video_coding/include/video_codec_interface.h" #include "rtc_base/checks.h" #include "rtc_base/random.h" #include "rtc_base/timeutils.h" #include "system_wrappers/include/field_trial.h" namespace webrtc { namespace { // Map information from info into rtp. void CopyCodecSpecific(const CodecSpecificInfo* info, RTPVideoHeader* rtp) { RTC_DCHECK(info); switch (info->codecType) { case kVideoCodecVP8: { rtp->codec = kRtpVideoVp8; rtp->codecHeader.VP8.InitRTPVideoHeaderVP8(); rtp->codecHeader.VP8.pictureId = info->codecSpecific.VP8.pictureId; rtp->codecHeader.VP8.nonReference = info->codecSpecific.VP8.nonReference; rtp->codecHeader.VP8.temporalIdx = info->codecSpecific.VP8.temporalIdx; rtp->codecHeader.VP8.layerSync = info->codecSpecific.VP8.layerSync; rtp->codecHeader.VP8.tl0PicIdx = info->codecSpecific.VP8.tl0PicIdx; rtp->codecHeader.VP8.keyIdx = info->codecSpecific.VP8.keyIdx; rtp->simulcastIdx = info->codecSpecific.VP8.simulcastIdx; return; } case kVideoCodecVP9: { rtp->codec = kRtpVideoVp9; rtp->codecHeader.VP9.InitRTPVideoHeaderVP9(); rtp->codecHeader.VP9.inter_pic_predicted = info->codecSpecific.VP9.inter_pic_predicted; rtp->codecHeader.VP9.flexible_mode = info->codecSpecific.VP9.flexible_mode; rtp->codecHeader.VP9.ss_data_available = info->codecSpecific.VP9.ss_data_available; rtp->codecHeader.VP9.picture_id = info->codecSpecific.VP9.picture_id; rtp->codecHeader.VP9.tl0_pic_idx = info->codecSpecific.VP9.tl0_pic_idx; rtp->codecHeader.VP9.temporal_idx = info->codecSpecific.VP9.temporal_idx; rtp->codecHeader.VP9.spatial_idx = info->codecSpecific.VP9.spatial_idx; rtp->codecHeader.VP9.temporal_up_switch = info->codecSpecific.VP9.temporal_up_switch; rtp->codecHeader.VP9.inter_layer_predicted = info->codecSpecific.VP9.inter_layer_predicted; rtp->codecHeader.VP9.gof_idx = info->codecSpecific.VP9.gof_idx; rtp->codecHeader.VP9.num_spatial_layers = info->codecSpecific.VP9.num_spatial_layers; if (info->codecSpecific.VP9.ss_data_available) { rtp->codecHeader.VP9.spatial_layer_resolution_present = info->codecSpecific.VP9.spatial_layer_resolution_present; if (info->codecSpecific.VP9.spatial_layer_resolution_present) { for (size_t i = 0; i < info->codecSpecific.VP9.num_spatial_layers; ++i) { rtp->codecHeader.VP9.width[i] = info->codecSpecific.VP9.width[i]; rtp->codecHeader.VP9.height[i] = info->codecSpecific.VP9.height[i]; } } rtp->codecHeader.VP9.gof.CopyGofInfoVP9(info->codecSpecific.VP9.gof); } rtp->codecHeader.VP9.num_ref_pics = info->codecSpecific.VP9.num_ref_pics; for (int i = 0; i < info->codecSpecific.VP9.num_ref_pics; ++i) rtp->codecHeader.VP9.pid_diff[i] = info->codecSpecific.VP9.p_diff[i]; return; } case kVideoCodecH264: rtp->codec = kRtpVideoH264; rtp->codecHeader.H264.packetization_mode = info->codecSpecific.H264.packetization_mode; return; case kVideoCodecGeneric: rtp->codec = kRtpVideoGeneric; rtp->simulcastIdx = info->codecSpecific.generic.simulcast_idx; return; default: return; } } } // namespace // Currently only used if forced fallback for VP8 is enabled. // Consider adding tl0idx and set for VP8 and VP9. // Make picture id not codec specific. class PayloadRouter::RtpPayloadParams final { public: RtpPayloadParams(const uint32_t ssrc, const RtpPayloadState* state) : ssrc_(ssrc) { Random random(rtc::TimeMicros()); state_.picture_id = state ? state->picture_id : (random.Rand() & 0x7FFF); } ~RtpPayloadParams() {} void Set(RTPVideoHeader* rtp_video_header) { if (rtp_video_header->codec == kRtpVideoVp8 && rtp_video_header->codecHeader.VP8.pictureId != kNoPictureId) { rtp_video_header->codecHeader.VP8.pictureId = state_.picture_id; state_.picture_id = (state_.picture_id + 1) & 0x7FFF; } } uint32_t ssrc() const { return ssrc_; } RtpPayloadState state() const { return state_; } private: const uint32_t ssrc_; RtpPayloadState state_; }; PayloadRouter::PayloadRouter(const std::vector& rtp_modules, const std::vector& ssrcs, int payload_type, const std::map& states) : active_(false), rtp_modules_(rtp_modules), payload_type_(payload_type), forced_fallback_enabled_((webrtc::field_trial::IsEnabled( "WebRTC-VP8-Forced-Fallback-Encoder-v2"))) { RTC_DCHECK_EQ(ssrcs.size(), rtp_modules.size()); // SSRCs are assumed to be sorted in the same order as |rtp_modules|. for (uint32_t ssrc : ssrcs) { // Restore state if it previously existed. const RtpPayloadState* state = nullptr; auto it = states.find(ssrc); if (it != states.end()) { state = &it->second; } params_.push_back(RtpPayloadParams(ssrc, state)); } } PayloadRouter::~PayloadRouter() {} void PayloadRouter::SetActive(bool active) { rtc::CritScope lock(&crit_); if (active_ == active) return; active_ = active; for (auto& module : rtp_modules_) { module->SetSendingStatus(active_); module->SetSendingMediaStatus(active_); } } bool PayloadRouter::IsActive() { rtc::CritScope lock(&crit_); return active_ && !rtp_modules_.empty(); } std::map PayloadRouter::GetRtpPayloadStates() const { rtc::CritScope lock(&crit_); std::map payload_states; for (const auto& param : params_) { payload_states[param.ssrc()] = param.state(); } return payload_states; } EncodedImageCallback::Result PayloadRouter::OnEncodedImage( const EncodedImage& encoded_image, const CodecSpecificInfo* codec_specific_info, const RTPFragmentationHeader* fragmentation) { rtc::CritScope lock(&crit_); RTC_DCHECK(!rtp_modules_.empty()); if (!active_) return Result(Result::ERROR_SEND_FAILED); RTPVideoHeader rtp_video_header; memset(&rtp_video_header, 0, sizeof(RTPVideoHeader)); if (codec_specific_info) CopyCodecSpecific(codec_specific_info, &rtp_video_header); rtp_video_header.rotation = encoded_image.rotation_; rtp_video_header.content_type = encoded_image.content_type_; if (encoded_image.timing_.flags != TimingFrameFlags::kInvalid && encoded_image.timing_.flags != TimingFrameFlags::kNotTriggered) { rtp_video_header.video_timing.encode_start_delta_ms = VideoSendTiming::GetDeltaCappedMs( encoded_image.capture_time_ms_, encoded_image.timing_.encode_start_ms); rtp_video_header.video_timing.encode_finish_delta_ms = VideoSendTiming::GetDeltaCappedMs( encoded_image.capture_time_ms_, encoded_image.timing_.encode_finish_ms); rtp_video_header.video_timing.packetization_finish_delta_ms = 0; rtp_video_header.video_timing.pacer_exit_delta_ms = 0; rtp_video_header.video_timing.network_timestamp_delta_ms = 0; rtp_video_header.video_timing.network2_timestamp_delta_ms = 0; rtp_video_header.video_timing.flags = encoded_image.timing_.flags; } else { rtp_video_header.video_timing.flags = TimingFrameFlags::kInvalid; } rtp_video_header.playout_delay = encoded_image.playout_delay_; int stream_index = rtp_video_header.simulcastIdx; RTC_DCHECK_LT(stream_index, rtp_modules_.size()); if (forced_fallback_enabled_) { // Sets picture id. The SW and HW encoder have separate picture id // sequences, set picture id to not cause sequence discontinuties at encoder // changes. params_[stream_index].Set(&rtp_video_header); } uint32_t frame_id; bool send_result = rtp_modules_[stream_index]->SendOutgoingData( encoded_image._frameType, payload_type_, encoded_image._timeStamp, encoded_image.capture_time_ms_, encoded_image._buffer, encoded_image._length, fragmentation, &rtp_video_header, &frame_id); if (!send_result) return Result(Result::ERROR_SEND_FAILED); return Result(Result::OK, frame_id); } void PayloadRouter::OnBitrateAllocationUpdated( const BitrateAllocation& bitrate) { rtc::CritScope lock(&crit_); if (IsActive()) { if (rtp_modules_.size() == 1) { // If spatial scalability is enabled, it is covered by a single stream. rtp_modules_[0]->SetVideoBitrateAllocation(bitrate); } else { // Simulcast is in use, split the BitrateAllocation into one struct per // rtp stream, moving over the temporal layer allocation. for (size_t si = 0; si < rtp_modules_.size(); ++si) { // Don't send empty TargetBitrate messages on streams not being relayed. if (!bitrate.IsSpatialLayerUsed(si)) break; BitrateAllocation layer_bitrate; for (int tl = 0; tl < kMaxTemporalStreams; ++tl) { if (bitrate.HasBitrate(si, tl)) layer_bitrate.SetBitrate(0, tl, bitrate.GetBitrate(si, tl)); } rtp_modules_[si]->SetVideoBitrateAllocation(layer_bitrate); } } } } } // namespace webrtc