Initial CL for this effort, with a working RTCVideoEncoder/Decoder for H264
(wrapping the VideoToolbox codec).
Some notes / things left to do:
- There are some hard-coded references to codec types that are supported by
webrtc::VideoCodec, cricket::VideoCodec, webrtc::CodecSpecificInfo etc
since we need to convert to/from these types in ObjCVideoEncoder/Decoder.
These types would need to be more codec agnostic to avoid this.
- Most interfaces are borrowed from the design document for injectable
codecs in Android. Some data in the corresponding C++ classes is discarded
when converting to the Obj-C version, since it has fewer fields. I have not
verified whether all data that we do keep is needed, or whether we might be
losing anything useful in these conversions.
- Implement the VideoToolbox codec code directly in the RTCVideoEncoderH264
classes, instead of wrapping webrtc::H264VideoToolboxEncoder / decoder.
Eliminates converting between ObjC/C++ types outside the ObjCVideoEncoder/
Decoder wrapper classes.
- List the injected codec factory's supported codecs in the list of codecs in
AppRTCMobile.
BUG=webrtc:7924
R=magjed@webrtc.org
Review-Url: https://codereview.webrtc.org/2966023002 .
Cr-Commit-Position: refs/heads/master@{#18928}
140 lines
4.1 KiB
Plaintext
140 lines
4.1 KiB
Plaintext
/*
|
|
* Copyright 2015 The WebRTC project authors. All Rights Reserved.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license
|
|
* that can be found in the LICENSE file in the root of the source
|
|
* tree. An additional intellectual property rights grant can be found
|
|
* in the file PATENTS. All contributing project authors may
|
|
* be found in the AUTHORS file in the root of the source tree.
|
|
*/
|
|
|
|
#import "webrtc/sdk/objc/Framework/Headers/WebRTC/RTCVideoFrame.h"
|
|
#import "webrtc/sdk/objc/Framework/Headers/WebRTC/RTCVideoFrameBuffer.h"
|
|
|
|
@implementation RTCVideoFrame {
|
|
RTCVideoRotation _rotation;
|
|
int64_t _timeStampNs;
|
|
}
|
|
|
|
@synthesize buffer = _buffer;
|
|
@synthesize timeStamp;
|
|
|
|
- (int)width {
|
|
return _buffer.width;
|
|
}
|
|
|
|
- (int)height {
|
|
return _buffer.height;
|
|
}
|
|
|
|
- (RTCVideoRotation)rotation {
|
|
return _rotation;
|
|
}
|
|
|
|
- (const uint8_t *)dataY {
|
|
if ([_buffer conformsToProtocol:@protocol(RTCI420Buffer)]) {
|
|
return ((id<RTCI420Buffer>)_buffer).dataY;
|
|
} else {
|
|
return nullptr;
|
|
}
|
|
}
|
|
|
|
- (const uint8_t *)dataU {
|
|
if ([_buffer conformsToProtocol:@protocol(RTCI420Buffer)]) {
|
|
return ((id<RTCI420Buffer>)_buffer).dataU;
|
|
} else {
|
|
return nullptr;
|
|
}
|
|
}
|
|
|
|
- (const uint8_t *)dataV {
|
|
if ([_buffer conformsToProtocol:@protocol(RTCI420Buffer)]) {
|
|
return ((id<RTCI420Buffer>)_buffer).dataV;
|
|
} else {
|
|
return nullptr;
|
|
}
|
|
}
|
|
|
|
- (int)strideY {
|
|
if ([_buffer conformsToProtocol:@protocol(RTCI420Buffer)]) {
|
|
return ((id<RTCI420Buffer>)_buffer).strideY;
|
|
} else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
- (int)strideU {
|
|
if ([_buffer conformsToProtocol:@protocol(RTCI420Buffer)]) {
|
|
return ((id<RTCI420Buffer>)_buffer).strideU;
|
|
} else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
- (int)strideV {
|
|
if ([_buffer conformsToProtocol:@protocol(RTCI420Buffer)]) {
|
|
return ((id<RTCI420Buffer>)_buffer).strideV;
|
|
} else {
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
- (int64_t)timeStampNs {
|
|
return _timeStampNs;
|
|
}
|
|
|
|
- (CVPixelBufferRef)nativeHandle {
|
|
if ([_buffer isKindOfClass:[RTCCVPixelBuffer class]]) {
|
|
return ((RTCCVPixelBuffer *)_buffer).pixelBuffer;
|
|
} else {
|
|
return nullptr;
|
|
}
|
|
}
|
|
|
|
- (RTCVideoFrame *)newI420VideoFrame {
|
|
return [[RTCVideoFrame alloc] initWithBuffer:[_buffer toI420]
|
|
rotation:_rotation
|
|
timeStampNs:_timeStampNs];
|
|
}
|
|
|
|
- (instancetype)initWithPixelBuffer:(CVPixelBufferRef)pixelBuffer
|
|
rotation:(RTCVideoRotation)rotation
|
|
timeStampNs:(int64_t)timeStampNs {
|
|
return [self initWithBuffer:[[RTCCVPixelBuffer alloc] initWithPixelBuffer:pixelBuffer]
|
|
rotation:rotation
|
|
timeStampNs:timeStampNs];
|
|
}
|
|
|
|
- (instancetype)initWithPixelBuffer:(CVPixelBufferRef)pixelBuffer
|
|
scaledWidth:(int)scaledWidth
|
|
scaledHeight:(int)scaledHeight
|
|
cropWidth:(int)cropWidth
|
|
cropHeight:(int)cropHeight
|
|
cropX:(int)cropX
|
|
cropY:(int)cropY
|
|
rotation:(RTCVideoRotation)rotation
|
|
timeStampNs:(int64_t)timeStampNs {
|
|
RTCCVPixelBuffer *rtcPixelBuffer = [[RTCCVPixelBuffer alloc] initWithPixelBuffer:pixelBuffer
|
|
adaptedWidth:scaledWidth
|
|
adaptedHeight:scaledHeight
|
|
cropWidth:cropWidth
|
|
cropHeight:cropHeight
|
|
cropX:cropX
|
|
cropY:cropY];
|
|
return [self initWithBuffer:rtcPixelBuffer rotation:rotation timeStampNs:timeStampNs];
|
|
}
|
|
|
|
- (instancetype)initWithBuffer:(id<RTCVideoFrameBuffer>)buffer
|
|
rotation:(RTCVideoRotation)rotation
|
|
timeStampNs:(int64_t)timeStampNs {
|
|
if (self = [super init]) {
|
|
_buffer = buffer;
|
|
_rotation = rotation;
|
|
_timeStampNs = timeStampNs;
|
|
}
|
|
|
|
return self;
|
|
}
|
|
|
|
@end
|