SSE2 optimizations for spatial content analysis.

Using intrinsics to work on 16 pixels at a time when
computing the spatial metrics.
Review URL: http://webrtc-codereview.appspot.com/92013

git-svn-id: http://webrtc.googlecode.com/svn/trunk@366 4adac7df-926f-26a2-2b94-8c16560cd09d
This commit is contained in:
frkoenig@google.com 2011-08-15 15:56:23 +00:00
parent 3cbbcd98b6
commit 6d171c4f98
4 changed files with 404 additions and 131 deletions

View File

@ -7,20 +7,21 @@
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#include "content_analysis.h"
#include "tick_util.h"
#include "system_wrappers/interface/cpu_features_wrapper.h"
#include <math.h>
#include <stdlib.h>
#include <xmmintrin.h>
namespace webrtc {
VPMContentAnalysis::VPMContentAnalysis():
VPMContentAnalysis::VPMContentAnalysis(bool RTCD):
_origFrame(NULL),
_prevFrame(NULL),
_width(0),
_height(0),
_skipNum(1),
_motionMagnitudeNZ(0.0f),
_spatialPredErr(0.0f),
_spatialPredErrH(0.0f),
@ -33,6 +34,19 @@ _firstFrame(true),
_CAInit(false),
_cMetrics(NULL)
{
ComputeSpatialMetrics = &VPMContentAnalysis::ComputeSpatialMetrics_C;
if (RTCD)
{
if(WebRtc_GetCPUInfo(kSSE2))
{
#if defined(__SSE2__)
ComputeSpatialMetrics =
&VPMContentAnalysis::ComputeSpatialMetrics_SSE2;
#endif
}
}
Release();
}
@ -42,7 +56,6 @@ VPMContentAnalysis::~VPMContentAnalysis()
}
VideoContentMetrics*
VPMContentAnalysis::ComputeContentMetrics(const VideoFrame* inputFrame)
{
@ -51,26 +64,30 @@ VPMContentAnalysis::ComputeContentMetrics(const VideoFrame* inputFrame)
return NULL;
}
//Init if needed (native dimension change)
// Init if needed (native dimension change)
if (_width != inputFrame->Width() || _height != inputFrame->Height())
{
Initialize((WebRtc_UWord16)inputFrame->Width(), (WebRtc_UWord16)inputFrame->Height());
if (VPM_OK != Initialize((WebRtc_UWord16)inputFrame->Width(),
(WebRtc_UWord16)inputFrame->Height()))
{
return NULL;
}
}
_origFrame = inputFrame->Buffer();
//compute spatial metrics: 3 spatial prediction errors
ComputeSpatialMetrics();
// compute spatial metrics: 3 spatial prediction errors
(this->*ComputeSpatialMetrics)();
//compute motion metrics
// compute motion metrics
if (_firstFrame == false)
ComputeMotionMetrics();
// saving current frame as previous one: Y only
memcpy(_prevFrame, _origFrame, _width * _height);
// saving current frame as previous one: Y only
memcpy(_prevFrame, _origFrame, _width * _height);
_firstFrame = false;
_CAInit = true;
_firstFrame = false;
_CAInit = true;
return ContentMetrics();
}
@ -104,20 +121,45 @@ VPMContentAnalysis::Initialize(WebRtc_UWord16 width, WebRtc_UWord16 height)
_height = height;
_firstFrame = true;
// skip parameter: # of skipped rows: for complexity reduction
// temporal also currently uses it for column reduction.
_skipNum = 1;
// use skipNum = 2 for 4CIF, WHD
if ( (_height >= 576) && (_width >= 704) )
{
_skipNum = 2;
}
// use skipNum = 4 for FULLL_HD images
if ( (_height >= 1080) && (_width >= 1920) )
{
_skipNum = 4;
}
if (_cMetrics != NULL)
{
delete _cMetrics;
}
_cMetrics = new VideoContentMetrics();
if (_cMetrics == NULL)
{
return VPM_MEMORY;
}
if (_prevFrame != NULL)
{
delete [] _prevFrame;
}
// Spatial Metrics don't work on a border of 8. Minimum processing
// block size is 16 pixels. So make sure the width and height support this.
if (_width <= 32 || _height <= 32)
{
_CAInit = false;
return VPM_PARAMETER_ERROR;
}
_cMetrics = new VideoContentMetrics();
if (_cMetrics == NULL)
{
return VPM_MEMORY;
}
_prevFrame = new WebRtc_UWord8[_width * _height] ; // Y only
if (_prevFrame == NULL)
{
@ -128,39 +170,44 @@ VPMContentAnalysis::Initialize(WebRtc_UWord16 width, WebRtc_UWord16 height)
}
//Compute motion metrics: magnitude over non-zero motion vectors, and size of zero cluster
// Compute motion metrics: magnitude over non-zero motion vectors,
// and size of zero cluster
WebRtc_Word32
VPMContentAnalysis::ComputeMotionMetrics()
{
//Motion metrics: only one is derived from normalized (MAD) temporal difference
// Motion metrics: only one is derived from normalized
// (MAD) temporal difference
TemporalDiffMetric();
return VPM_OK;
return VPM_OK;
}
//Normalized temporal difference (MAD): used as a motion level metric
//Normalize MAD by spatial contrast: images with more contrast (pixel variance) likely have larger temporal difference
//To reduce complexity, we compute the metric for a reduced set of points.
// Normalized temporal difference (MAD): used as a motion level metric
// Normalize MAD by spatial contrast: images with more contrast
// (pixel variance) likely have larger temporal difference
// To reduce complexity, we compute the metric for a reduced set of points.
WebRtc_Word32
VPMContentAnalysis::TemporalDiffMetric()
{
//size of original frame
// size of original frame
WebRtc_UWord16 sizei = _height;
WebRtc_UWord16 sizej = _width;
//skip parameter: # of skipped pixels along x & y direction: for complexity reduction
WebRtc_UWord8 skipNum = 1; // 1 == all pixels, 2 == 1/4 reduction, 3 == 1/9 reduction
// Use the member variable _skipNum later when this gets
// optimized
// skip parameter: # of skipped rows: for complexity reduction
// temporal also currently uses it for column reduction.
WebRtc_UWord32 skipNum = 1;
//use skipNum = 2 for 4CIF, WHD
if ( (sizei >= 576) && (sizej >= 704) )
// use skipNum = 2 for 4CIF, WHD
if ( (_height >= 576) && (_width >= 704) )
{
skipNum = 2;
}
//use skipNum = 3 for FULLL_HD images
if ( (sizei >= 1080) && (sizej >= 1920) )
// use skipNum = 3 for FULLL_HD images
if ( (_height >= 1080) && (_width >= 1920) )
{
skipNum = 3;
}
@ -174,25 +221,28 @@ VPMContentAnalysis::TemporalDiffMetric()
WebRtc_UWord32 pixelSum = 0;
WebRtc_UWord32 pixelSqSum = 0;
WebRtc_UWord8 bord = 8; //avoid boundary
WebRtc_UWord32 numPixels = 0; //counter for # of pixels
WebRtc_UWord8 bord = 8; // avoid boundary
WebRtc_UWord32 numPixels = 0; // counter for # of pixels
WebRtc_UWord32 ssn;
for(WebRtc_UWord16 i = bord; i < sizei - bord; i += skipNum)
for(WebRtc_UWord16 j = bord; j < sizej - bord; j += skipNum)
{
numPixels += 1;
ssn = i * sizej + j;
for(WebRtc_UWord16 j = bord; j < sizej - bord; j += skipNum)
{
numPixels += 1;
ssn = i * sizej + j;
WebRtc_UWord8 currPixel = _origFrame[ssn];
WebRtc_UWord8 prevPixel = _prevFrame[ssn];
WebRtc_UWord8 currPixel = _origFrame[ssn];
WebRtc_UWord8 prevPixel = _prevFrame[ssn];
tempDiffSum += (WebRtc_UWord32) abs((WebRtc_Word16)(currPixel - prevPixel));
pixelSum += (WebRtc_UWord32) _origFrame[ssn];
pixelSqSum += (WebRtc_UWord32) (_origFrame[ssn] * _origFrame[ssn]);
tempDiffSum += (WebRtc_UWord32)
abs((WebRtc_Word16)(currPixel - prevPixel));
pixelSum += (WebRtc_UWord32) _origFrame[ssn];
pixelSqSum += (WebRtc_UWord32) (_origFrame[ssn] * _origFrame[ssn]);
}
}
//default
// default
_motionMagnitudeNZ = 0.0f;
if (tempDiffSum == 0)
@ -200,7 +250,7 @@ VPMContentAnalysis::TemporalDiffMetric()
return VPM_OK;
}
//normalize over all pixels
// normalize over all pixels
tempDiffAvg = (float)tempDiffSum / (float)(numPixels);
pixelSumAvg = (float)pixelSum / (float)(numPixels);
pixelSqSumAvg = (float)pixelSqSum / (float)(numPixels);
@ -217,96 +267,266 @@ VPMContentAnalysis::TemporalDiffMetric()
}
//Compute spatial metrics:
//To reduce complexity, we compute the metric for a reduced set of points.
//The spatial metrics are rough estimates of the prediction error cost for each QM spatial mode: 2x2,1x2,2x1
//The metrics are a simple estimate of the up-sampling prediction error, estimated assuming sub-sampling for decimation (no filtering),
//and up-sampling back up with simple bilinear interpolation.
// Compute spatial metrics:
// To reduce complexity, we compute the metric for a reduced set of points.
// The spatial metrics are rough estimates of the prediction error cost for
// each QM spatial mode: 2x2,1x2,2x1
// The metrics are a simple estimate of the up-sampling prediction error,
// estimated assuming sub-sampling for decimation (no filtering),
// and up-sampling back up with simple bilinear interpolation.
WebRtc_Word32
VPMContentAnalysis::ComputeSpatialMetrics()
VPMContentAnalysis::ComputeSpatialMetrics_C()
{
//size of original frame
WebRtc_UWord16 sizei = _height;
WebRtc_UWord16 sizej = _width;
//skip parameter: # of skipped pixels along x & y direction: for complexity reduction
WebRtc_UWord8 skipNum = 1; // 1 == all pixels, 2 == 1/4 reduction, 3 == 1/9 reduction
//use skipNum = 2 for 4CIF, WHD
if ( (sizei >= 576) && (sizej >= 704) )
{
skipNum = 2;
}
//use skipNum = 3 for FULLL_HD images
if ( (sizei >= 1080) && (sizej >= 1920) )
{
skipNum = 3;
}
const WebRtc_UWord16 sizei = _height;
const WebRtc_UWord16 sizej = _width;
float spatialErr = 0.0f;
float spatialErrH = 0.0f;
float spatialErrV = 0.0f;
//pixel mean square average: used to normalize the spatial metrics
float pixelMSA = 0;
// pixel mean square average: used to normalize the spatial metrics
WebRtc_UWord32 pixelMSA = 0;
float norm = 1.0f;
WebRtc_UWord8 bord = 8; //avoid boundary
WebRtc_UWord32 numPixels = 0; //counter for # of pixels
WebRtc_UWord32 ssn1,ssn2,ssn3,ssn4,ssn5;
const WebRtc_UWord32 bord = 8; // avoid boundary
WebRtc_UWord32 numPixels = 0; // counter for # of pixels
WebRtc_UWord32 spatialErrSum = 0;
WebRtc_UWord32 spatialErrVSum = 0;
WebRtc_UWord32 spatialErrHSum = 0;
for(WebRtc_UWord16 i = bord; i < sizei - bord; i += skipNum)
for(WebRtc_UWord16 j = bord; j < sizej - bord; j += skipNum)
// make sure work section is a multiple of 16
const WebRtc_UWord32 width_end = ((sizej - 2*bord) & -16) + bord;
for(WebRtc_UWord16 i = bord; i < sizei - bord; i += _skipNum)
{
numPixels += 1;
ssn1= i * sizej + j;
ssn2 = (i + 1) * sizej + j; //bottom
ssn3 = (i - 1) * sizej + j; //top
ssn4 = i * sizej + j + 1; //right
ssn5 = i * sizej + j - 1; //left
for(WebRtc_UWord16 j = bord; j < width_end; j++)
{
WebRtc_UWord32 ssn1,ssn2,ssn3,ssn4,ssn5;
WebRtc_UWord16 refPixel1 = _origFrame[ssn1] << 1;
WebRtc_UWord16 refPixel2 = _origFrame[ssn1] << 2;
numPixels += 1;
ssn1= i * sizej + j;
ssn2 = (i + 1) * sizej + j; // bottom
ssn3 = (i - 1) * sizej + j; // top
ssn4 = i * sizej + j + 1; // right
ssn5 = i * sizej + j - 1; // left
WebRtc_UWord8 bottPixel = _origFrame[ssn2];
WebRtc_UWord8 topPixel = _origFrame[ssn3];
WebRtc_UWord8 rightPixel = _origFrame[ssn4];
WebRtc_UWord8 leftPixel = _origFrame[ssn5];
WebRtc_UWord16 refPixel1 = _origFrame[ssn1] << 1;
WebRtc_UWord16 refPixel2 = _origFrame[ssn1] << 2;
spatialErrSum += (WebRtc_UWord32) abs((WebRtc_Word16)(refPixel2 - (WebRtc_UWord16)(bottPixel + topPixel + leftPixel + rightPixel)));
spatialErrVSum += (WebRtc_UWord32) abs((WebRtc_Word16)(refPixel1 - (WebRtc_UWord16)(bottPixel + topPixel)));
spatialErrHSum += (WebRtc_UWord32) abs((WebRtc_Word16)(refPixel1 - (WebRtc_UWord16)(leftPixel + rightPixel)));
WebRtc_UWord8 bottPixel = _origFrame[ssn2];
WebRtc_UWord8 topPixel = _origFrame[ssn3];
WebRtc_UWord8 rightPixel = _origFrame[ssn4];
WebRtc_UWord8 leftPixel = _origFrame[ssn5];
pixelMSA += (float)_origFrame[ssn1];
spatialErrSum += (WebRtc_UWord32) abs((WebRtc_Word16)(refPixel2
- (WebRtc_UWord16)(bottPixel + topPixel
+ leftPixel + rightPixel)));
spatialErrVSum += (WebRtc_UWord32) abs((WebRtc_Word16)(refPixel1
- (WebRtc_UWord16)(bottPixel + topPixel)));
spatialErrHSum += (WebRtc_UWord32) abs((WebRtc_Word16)(refPixel1
- (WebRtc_UWord16)(leftPixel + rightPixel)));
pixelMSA += _origFrame[ssn1];
}
}
//normalize over all pixels
// normalize over all pixels
spatialErr = (float)spatialErrSum / (float)(4 * numPixels);
spatialErrH = (float)spatialErrHSum / (float)(2 * numPixels);
spatialErrV = (float)spatialErrVSum / (float)(2 * numPixels);
norm = (float)pixelMSA / float(numPixels);
//normalize to RMS pixel level: use avg pixel level for now
// normalize to RMS pixel level: use avg pixel level for now
//2X2:
// 2X2:
_spatialPredErr = spatialErr / (norm);
//1X2:
// 1X2:
_spatialPredErrH = spatialErrH / (norm);
//2X1:
// 2X1:
_spatialPredErrV = spatialErrV / (norm);
return VPM_OK;
}
#if defined(__SSE2__)
WebRtc_Word32
VPMContentAnalysis::ComputeSpatialMetrics_SSE2()
{
// avoid boundary
const WebRtc_Word32 bord = 8;
// counter for # of pixels
const WebRtc_UWord32 numPixels = ((_width - 2*bord) & -16)*
(_height - 2*bord) / _skipNum;
const WebRtc_UWord8* imgBuf = _origFrame + bord*_width;
const WebRtc_Word32 width_end = ((_width - 2*bord) & -16) + bord;
__m128i se_32 = _mm_setzero_si128();
__m128i sev_32 = _mm_setzero_si128();
__m128i seh_32 = _mm_setzero_si128();
__m128i msa_32 = _mm_setzero_si128();
const __m128i z = _mm_setzero_si128();
// Error is accumulated as a 32 bit value. Looking at HD content with a
// height of 1080 lines, or about 67 macro blocks. If the 16 bit row
// value is maxed out at 65529 for every row, 65529*1080 = 70777800, which
// will not roll over a 32 bit accumulator.
// _skipNum is also used to reduce the number of rows
for(WebRtc_Word32 i = 0; i < (_height - 2*bord); i += _skipNum)
{
__m128i se_16 = _mm_setzero_si128();
__m128i sev_16 = _mm_setzero_si128();
__m128i seh_16 = _mm_setzero_si128();
__m128i msa_16 = _mm_setzero_si128();
// Row error is accumulated as a 16 bit value. There are 8
// accumulators. Max value of a 16 bit number is 65529. Looking
// at HD content, 1080p, has a width of 1920, 120 macro blocks.
// A mb at a time is processed at a time. Absolute max error at
// a point would be abs(0-255+255+255+255) which equals 1020.
// 120*1020 = 122400. The probability of hitting this is quite low
// on well behaved content. A specially crafted image could roll over.
// bord could also be adjusted to concentrate on just the center of
// the images for an HD capture in order to reduce the possiblity of
// rollover.
const WebRtc_UWord8 *lineTop = imgBuf - _width + bord;
const WebRtc_UWord8 *lineCen = imgBuf + bord;
const WebRtc_UWord8 *lineBot = imgBuf + _width + bord;
for(WebRtc_Word32 j = 0; j < width_end - bord; j += 16)
{
const __m128i t = _mm_loadu_si128((__m128i*)(lineTop));
const __m128i l = _mm_loadu_si128((__m128i*)(lineCen - 1));
const __m128i c = _mm_loadu_si128((__m128i*)(lineCen));
const __m128i r = _mm_loadu_si128((__m128i*)(lineCen + 1));
const __m128i b = _mm_loadu_si128((__m128i*)(lineBot));
lineTop += 16;
lineCen += 16;
lineBot += 16;
// center pixel unpacked
__m128i clo = _mm_unpacklo_epi8(c,z);
__m128i chi = _mm_unpackhi_epi8(c,z);
// left right pixels unpacked and added together
const __m128i lrlo = _mm_add_epi16(_mm_unpacklo_epi8(l,z),
_mm_unpacklo_epi8(r,z));
const __m128i lrhi = _mm_add_epi16(_mm_unpackhi_epi8(l,z),
_mm_unpackhi_epi8(r,z));
// top & bottom pixels unpacked and added together
const __m128i tblo = _mm_add_epi16(_mm_unpacklo_epi8(t,z),
_mm_unpacklo_epi8(b,z));
const __m128i tbhi = _mm_add_epi16(_mm_unpackhi_epi8(t,z),
_mm_unpackhi_epi8(b,z));
// running sum of all pixels
msa_16 = _mm_add_epi16(msa_16, _mm_add_epi16(chi, clo));
clo = _mm_slli_epi16(clo, 1);
chi = _mm_slli_epi16(chi, 1);
const __m128i sevtlo = _mm_subs_epi16(clo, tblo);
const __m128i sevthi = _mm_subs_epi16(chi, tbhi);
const __m128i sehtlo = _mm_subs_epi16(clo, lrlo);
const __m128i sehthi = _mm_subs_epi16(chi, lrhi);
clo = _mm_slli_epi16(clo, 1);
chi = _mm_slli_epi16(chi, 1);
const __m128i setlo = _mm_subs_epi16(clo,
_mm_add_epi16(lrlo, tblo));
const __m128i sethi = _mm_subs_epi16(chi,
_mm_add_epi16(lrhi, tbhi));
// Add to 16 bit running sum
se_16 = _mm_add_epi16(se_16,
_mm_max_epi16(setlo,
_mm_subs_epi16(z, setlo)));
se_16 = _mm_add_epi16(se_16,
_mm_max_epi16(sethi,
_mm_subs_epi16(z, sethi)));
sev_16 = _mm_add_epi16(sev_16,
_mm_max_epi16(sevtlo,
_mm_subs_epi16(z, sevtlo)));
sev_16 = _mm_add_epi16(sev_16,
_mm_max_epi16(sevthi,
_mm_subs_epi16(z, sevthi)));
seh_16 = _mm_add_epi16(seh_16,
_mm_max_epi16(sehtlo,
_mm_subs_epi16(z, sehtlo)));
seh_16 = _mm_add_epi16(seh_16,
_mm_max_epi16(sehthi,
_mm_subs_epi16(z, sehthi)));
}
// Add to 32 bit running sum as to not roll over.
se_32 = _mm_add_epi32(se_32,
_mm_add_epi32(_mm_unpackhi_epi16(se_16,z),
_mm_unpacklo_epi16(se_16,z)));
sev_32 = _mm_add_epi32(sev_32,
_mm_add_epi32(_mm_unpackhi_epi16(sev_16,z),
_mm_unpacklo_epi16(sev_16,z)));
seh_32 = _mm_add_epi32(seh_32,
_mm_add_epi32(_mm_unpackhi_epi16(seh_16,z),
_mm_unpacklo_epi16(seh_16,z)));
msa_32 = _mm_add_epi32(msa_32,
_mm_add_epi32(_mm_unpackhi_epi16(msa_16,z),
_mm_unpacklo_epi16(msa_16,z)));
imgBuf += _width * _skipNum;
}
WebRtc_Word64 se_64[2];
WebRtc_Word64 sev_64[2];
WebRtc_Word64 seh_64[2];
WebRtc_Word64 msa_64[2];
// bring sums out of vector registers and into integer register
// domain, summing them along the way
_mm_store_si128 ((__m128i*)se_64,
_mm_add_epi64(_mm_unpackhi_epi32(se_32,z),
_mm_unpacklo_epi32(se_32,z)));
_mm_store_si128 ((__m128i*)sev_64,
_mm_add_epi64(_mm_unpackhi_epi32(sev_32,z),
_mm_unpacklo_epi32(sev_32,z)));
_mm_store_si128 ((__m128i*)seh_64,
_mm_add_epi64(_mm_unpackhi_epi32(seh_32,z),
_mm_unpacklo_epi32(seh_32,z)));
_mm_store_si128 ((__m128i*)msa_64,
_mm_add_epi64(_mm_unpackhi_epi32(msa_32,z),
_mm_unpacklo_epi32(msa_32,z)));
const WebRtc_UWord32 spatialErrSum = se_64[0] + se_64[1];;
const WebRtc_UWord32 spatialErrVSum = sev_64[0] + sev_64[1];
const WebRtc_UWord32 spatialErrHSum = seh_64[0] + seh_64[1];
const WebRtc_UWord32 pixelMSA = msa_64[0] + msa_64[1];
// normalize over all pixels
const float spatialErr = (float)spatialErrSum / (float)(4 * numPixels);
const float spatialErrH = (float)spatialErrHSum / (float)(2 * numPixels);
const float spatialErrV = (float)spatialErrVSum / (float)(2 * numPixels);
const float norm = (float)pixelMSA / float(numPixels);
// normalize to RMS pixel level: use avg pixel level for now
// 2X2:
_spatialPredErr = spatialErr / norm;
// 1X2:
_spatialPredErrH = spatialErrH / norm;
// 2X1:
_spatialPredErrV = spatialErrV / norm;
return VPM_OK;
}
#endif
VideoContentMetrics*
VPMContentAnalysis::ContentMetrics()
@ -320,10 +540,10 @@ VPMContentAnalysis::ContentMetrics()
_cMetrics->spatialPredErr = _spatialPredErr;
_cMetrics->spatialPredErrH = _spatialPredErrH;
_cMetrics->spatialPredErrV = _spatialPredErrV;
//normalized temporal difference (MAD)
// normalized temporal difference (MAD)
_cMetrics->motionMagnitudeNZ = _motionMagnitudeNZ;
//Set to zero: not computed
// Set to zero: not computed
_cMetrics->motionPredErr = _motionPredErr;
_cMetrics->sizeZeroMotion = _sizeZeroMotion;
_cMetrics->motionHorizontalness = _motionHorizontalness;
@ -333,4 +553,4 @@ VPMContentAnalysis::ContentMetrics()
}
} //namespace
} // namespace

View File

@ -25,56 +25,65 @@ namespace webrtc {
class VPMContentAnalysis
{
public:
VPMContentAnalysis();
VPMContentAnalysis(bool RTCD = true);
~VPMContentAnalysis();
//Initialize ContentAnalysis - should be called prior to extractContentFeature
//Inputs: width, height
//Return value: 0 if OK, negative value upon error
// Initialize ContentAnalysis - should be called prior to
// extractContentFeature
// Inputs: width, height
// Return value: 0 if OK, negative value upon error
WebRtc_Word32 Initialize( WebRtc_UWord16 width, WebRtc_UWord16 height);
//Extract content Feature - main function of ContentAnalysis
//Input: new frame
//Return value: pointer to structure containing content Analysis metrics or NULL value upon error
// Extract content Feature - main function of ContentAnalysis
// Input: new frame
// Return value: pointer to structure containing content Analysis
// metrics or NULL value upon error
VideoContentMetrics* ComputeContentMetrics(const VideoFrame* inputFrame);
//Release all allocated memory
//Output: 0 if OK, negative value upon error
// Release all allocated memory
// Output: 0 if OK, negative value upon error
WebRtc_Word32 Release();
private:
//return motion metrics
// return motion metrics
VideoContentMetrics* ContentMetrics();
//Normalized temporal difference metric: for motion magnitude
// Normalized temporal difference metric: for motion magnitude
WebRtc_Word32 TemporalDiffMetric();
//Motion metric method: call 2 metrics (magnitude and size)
// Motion metric method: call 2 metrics (magnitude and size)
WebRtc_Word32 ComputeMotionMetrics();
//Spatial metric method: computes the 3 frame-average spatial prediction errors (1x2,2x1,2x2)
WebRtc_Word32 ComputeSpatialMetrics();
// Spatial metric method: computes the 3 frame-average spatial
// prediction errors (1x2,2x1,2x2)
typedef WebRtc_Word32 (VPMContentAnalysis::*ComputeSpatialMetricsFunc)();
ComputeSpatialMetricsFunc ComputeSpatialMetrics;
WebRtc_Word32 ComputeSpatialMetrics_C();
#if defined(__SSE2__)
WebRtc_Word32 ComputeSpatialMetrics_SSE2();
#endif
const WebRtc_UWord8* _origFrame;
WebRtc_UWord8* _prevFrame;
WebRtc_UWord16 _width;
WebRtc_UWord16 _height;
WebRtc_UWord32 _skipNum;
//Content Metrics:
//stores the local average of the metrics
float _motionMagnitudeNZ; //for motion class
float _spatialPredErr; //for spatial class
float _spatialPredErrH; //for spatial class
float _spatialPredErrV; //for spatial class
float _sizeZeroMotion; //for motion class
float _motionPredErr; //for complexity class:
float _motionHorizontalness; //for coherence class
float _motionClusterDistortion; //for coherence class
// Content Metrics:
// stores the local average of the metrics
float _motionMagnitudeNZ; // motion class
float _spatialPredErr; // spatial class
float _spatialPredErrH; // spatial class
float _spatialPredErrV; // spatial class
float _sizeZeroMotion; // motion class
float _motionPredErr; // complexity class:
float _motionHorizontalness; // coherence class
float _motionClusterDistortion; // coherence class
bool _firstFrame;
bool _CAInit;
VideoContentMetrics* _cMetrics;
}; // end of VPMContentAnalysis class definition

View File

@ -0,0 +1,42 @@
/*
* Copyright (c) 2011 The WebRTC project authors. All Rights Reserved.
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#include "unit_test.h"
#include "video_processing.h"
#include "vplib.h"
#include "content_analysis.h"
using namespace webrtc;
TEST_F(VideoProcessingModuleTest, ContentAnalysis)
{
VPMContentAnalysis _ca_c(false);
VPMContentAnalysis _ca_sse;
VideoContentMetrics *_cM_c, *_cM_SSE;
_ca_c.Initialize(_width,_height);
_ca_sse.Initialize(_width,_height);
ASSERT_EQ(352, _width);
ASSERT_EQ(288, _height);
while (fread(_videoFrame.Buffer(), 1, _frameLength, _sourceFile)
== _frameLength)
{
_cM_c = _ca_c.ComputeContentMetrics(&_videoFrame);
_cM_SSE = _ca_sse.ComputeContentMetrics(&_videoFrame);
ASSERT_EQ(_cM_c->spatialPredErr, _cM_SSE->spatialPredErr);
ASSERT_EQ(_cM_c->spatialPredErrV, _cM_SSE->spatialPredErrV);
ASSERT_EQ(_cM_c->spatialPredErrH, _cM_SSE->spatialPredErrH);
ASSERT_EQ(_cM_c->motionMagnitudeNZ, _cM_SSE->motionMagnitudeNZ);
}
ASSERT_NE(0, feof(_sourceFile)) << "Error reading source file";
}

View File

@ -20,29 +20,31 @@
'include_dirs': [
'../../../../system_wrappers/interface',
'../../../../common_video/vplib/main/interface',
'../../../../modules/video_processing/main/source',
],
'sources': [
# headers
'unit_test/unit_test.h',
# sources
'unit_test/brightness_detection_test.cc',
'unit_test/color_enhancement_test.cc',
'unit_test/content_metrics_test.cc',
'unit_test/deflickering_test.cc',
'unit_test/denoising_test.cc',
'unit_test/unit_test.cc',
'unit_test/unit_test.cc',
], # source
'conditions': [
['OS=="linux"', {
'cflags': [
'-fexceptions',
],
}],
], # conditions
},
],