644629c1c7
Adds "grresourcecache_add" and "grresourcecache_find" bench tests to test GrResourceCache::add and GrResourceCache::find. The tests work only with GPU backends, since GrResourceCache needs an GrGpu. Modifies bench tests to override SkBenchmark::isSuitableFor(Backend) function that specifies what kind of backend the test is inteded for. This replaces the previous "fIsRendering" flag that would indicate test that did no rendering. Adds SkCanvas::getGrContext() call to get the GrContext that the canvas ends up drawing to. The member function solves a common use-case that is also used in the benchmark added here. R=mtklein@google.com, bsalomon@google.com Author: kkinnunen@nvidia.com Review URL: https://codereview.chromium.org/73643005 git-svn-id: http://skia.googlecode.com/svn/trunk@12334 2bbb7eff-a529-9590-31e7-b0007b416f81
479 lines
15 KiB
C++
479 lines
15 KiB
C++
|
|
/*
|
|
* Copyright 2011 Google Inc.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license that can be
|
|
* found in the LICENSE file.
|
|
*/
|
|
#include "SkBenchmark.h"
|
|
#include "SkMatrix.h"
|
|
#include "SkMatrixUtils.h"
|
|
#include "SkRandom.h"
|
|
#include "SkString.h"
|
|
|
|
class MatrixBench : public SkBenchmark {
|
|
SkString fName;
|
|
public:
|
|
MatrixBench(const char name[]) {
|
|
fName.printf("matrix_%s", name);
|
|
}
|
|
|
|
virtual bool isSuitableFor(Backend backend) SK_OVERRIDE {
|
|
return backend == kNonRendering_Backend;
|
|
}
|
|
|
|
virtual void performTest() = 0;
|
|
|
|
protected:
|
|
virtual int mulLoopCount() const { return 1; }
|
|
|
|
virtual const char* onGetName() {
|
|
return fName.c_str();
|
|
}
|
|
|
|
virtual void onDraw(SkCanvas*) {
|
|
for (int i = 0; i < this->getLoops(); i++) {
|
|
this->performTest();
|
|
}
|
|
}
|
|
|
|
private:
|
|
typedef SkBenchmark INHERITED;
|
|
};
|
|
|
|
// we want to stop the compiler from eliminating code that it thinks is a no-op
|
|
// so we have a non-static global we increment, hoping that will convince the
|
|
// compiler to execute everything
|
|
int gMatrixBench_NonStaticGlobal;
|
|
|
|
#define always_do(pred) \
|
|
do { \
|
|
if (pred) { \
|
|
++gMatrixBench_NonStaticGlobal; \
|
|
} \
|
|
} while (0)
|
|
|
|
class EqualsMatrixBench : public MatrixBench {
|
|
public:
|
|
EqualsMatrixBench() : INHERITED("equals") {}
|
|
protected:
|
|
virtual void performTest() {
|
|
SkMatrix m0, m1, m2;
|
|
|
|
m0.reset();
|
|
m1.reset();
|
|
m2.reset();
|
|
always_do(m0 == m1);
|
|
always_do(m1 == m2);
|
|
always_do(m2 == m0);
|
|
}
|
|
private:
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
class ScaleMatrixBench : public MatrixBench {
|
|
public:
|
|
ScaleMatrixBench() : INHERITED("scale") {
|
|
fSX = fSY = SkFloatToScalar(1.5f);
|
|
fM0.reset();
|
|
fM1.setScale(fSX, fSY);
|
|
fM2.setTranslate(fSX, fSY);
|
|
}
|
|
protected:
|
|
virtual void performTest() {
|
|
SkMatrix m;
|
|
m = fM0; m.preScale(fSX, fSY);
|
|
m = fM1; m.preScale(fSX, fSY);
|
|
m = fM2; m.preScale(fSX, fSY);
|
|
}
|
|
private:
|
|
SkMatrix fM0, fM1, fM2;
|
|
SkScalar fSX, fSY;
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
// having unknown values in our arrays can throw off the timing a lot, perhaps
|
|
// handling NaN values is a lot slower. Anyway, this guy is just meant to put
|
|
// reasonable values in our arrays.
|
|
template <typename T> void init9(T array[9]) {
|
|
SkRandom rand;
|
|
for (int i = 0; i < 9; i++) {
|
|
array[i] = rand.nextSScalar1();
|
|
}
|
|
}
|
|
|
|
// Test the performance of setConcat() non-perspective case:
|
|
// using floating point precision only.
|
|
class FloatConcatMatrixBench : public MatrixBench {
|
|
public:
|
|
FloatConcatMatrixBench() : INHERITED("concat_floatfloat") {
|
|
init9(mya);
|
|
init9(myb);
|
|
init9(myr);
|
|
}
|
|
protected:
|
|
virtual int mulLoopCount() const { return 4; }
|
|
|
|
static inline void muladdmul(float a, float b, float c, float d,
|
|
float* result) {
|
|
*result = a * b + c * d;
|
|
}
|
|
virtual void performTest() {
|
|
const float* a = mya;
|
|
const float* b = myb;
|
|
float* r = myr;
|
|
muladdmul(a[0], b[0], a[1], b[3], &r[0]);
|
|
muladdmul(a[0], b[1], a[1], b[4], &r[1]);
|
|
muladdmul(a[0], b[2], a[1], b[5], &r[2]);
|
|
r[2] += a[2];
|
|
muladdmul(a[3], b[0], a[4], b[3], &r[3]);
|
|
muladdmul(a[3], b[1], a[4], b[4], &r[4]);
|
|
muladdmul(a[3], b[2], a[4], b[5], &r[5]);
|
|
r[5] += a[5];
|
|
r[6] = r[7] = 0.0f;
|
|
r[8] = 1.0f;
|
|
}
|
|
private:
|
|
float mya [9];
|
|
float myb [9];
|
|
float myr [9];
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
static inline float SkDoubleToFloat(double x) {
|
|
return static_cast<float>(x);
|
|
}
|
|
|
|
// Test the performance of setConcat() non-perspective case:
|
|
// using floating point precision but casting up to float for
|
|
// intermediate results during computations.
|
|
class FloatDoubleConcatMatrixBench : public MatrixBench {
|
|
public:
|
|
FloatDoubleConcatMatrixBench() : INHERITED("concat_floatdouble") {
|
|
init9(mya);
|
|
init9(myb);
|
|
init9(myr);
|
|
}
|
|
protected:
|
|
virtual int mulLoopCount() const { return 4; }
|
|
|
|
static inline void muladdmul(float a, float b, float c, float d,
|
|
float* result) {
|
|
*result = SkDoubleToFloat((double)a * b + (double)c * d);
|
|
}
|
|
virtual void performTest() {
|
|
const float* a = mya;
|
|
const float* b = myb;
|
|
float* r = myr;
|
|
muladdmul(a[0], b[0], a[1], b[3], &r[0]);
|
|
muladdmul(a[0], b[1], a[1], b[4], &r[1]);
|
|
muladdmul(a[0], b[2], a[1], b[5], &r[2]);
|
|
r[2] += a[2];
|
|
muladdmul(a[3], b[0], a[4], b[3], &r[3]);
|
|
muladdmul(a[3], b[1], a[4], b[4], &r[4]);
|
|
muladdmul(a[3], b[2], a[4], b[5], &r[5]);
|
|
r[5] += a[5];
|
|
r[6] = r[7] = 0.0f;
|
|
r[8] = 1.0f;
|
|
}
|
|
private:
|
|
float mya [9];
|
|
float myb [9];
|
|
float myr [9];
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
// Test the performance of setConcat() non-perspective case:
|
|
// using double precision only.
|
|
class DoubleConcatMatrixBench : public MatrixBench {
|
|
public:
|
|
DoubleConcatMatrixBench() : INHERITED("concat_double") {
|
|
init9(mya);
|
|
init9(myb);
|
|
init9(myr);
|
|
}
|
|
protected:
|
|
virtual int mulLoopCount() const { return 4; }
|
|
|
|
static inline void muladdmul(double a, double b, double c, double d,
|
|
double* result) {
|
|
*result = a * b + c * d;
|
|
}
|
|
virtual void performTest() {
|
|
const double* a = mya;
|
|
const double* b = myb;
|
|
double* r = myr;
|
|
muladdmul(a[0], b[0], a[1], b[3], &r[0]);
|
|
muladdmul(a[0], b[1], a[1], b[4], &r[1]);
|
|
muladdmul(a[0], b[2], a[1], b[5], &r[2]);
|
|
r[2] += a[2];
|
|
muladdmul(a[3], b[0], a[4], b[3], &r[3]);
|
|
muladdmul(a[3], b[1], a[4], b[4], &r[4]);
|
|
muladdmul(a[3], b[2], a[4], b[5], &r[5]);
|
|
r[5] += a[5];
|
|
r[6] = r[7] = 0.0;
|
|
r[8] = 1.0;
|
|
}
|
|
private:
|
|
double mya [9];
|
|
double myb [9];
|
|
double myr [9];
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
class GetTypeMatrixBench : public MatrixBench {
|
|
public:
|
|
GetTypeMatrixBench()
|
|
: INHERITED("gettype") {
|
|
fArray[0] = (float) fRnd.nextS();
|
|
fArray[1] = (float) fRnd.nextS();
|
|
fArray[2] = (float) fRnd.nextS();
|
|
fArray[3] = (float) fRnd.nextS();
|
|
fArray[4] = (float) fRnd.nextS();
|
|
fArray[5] = (float) fRnd.nextS();
|
|
fArray[6] = (float) fRnd.nextS();
|
|
fArray[7] = (float) fRnd.nextS();
|
|
fArray[8] = (float) fRnd.nextS();
|
|
}
|
|
protected:
|
|
// Putting random generation of the matrix inside performTest()
|
|
// would help us avoid anomalous runs, but takes up 25% or
|
|
// more of the function time.
|
|
virtual void performTest() {
|
|
fMatrix.setAll(fArray[0], fArray[1], fArray[2],
|
|
fArray[3], fArray[4], fArray[5],
|
|
fArray[6], fArray[7], fArray[8]);
|
|
always_do(fMatrix.getType());
|
|
fMatrix.dirtyMatrixTypeCache();
|
|
always_do(fMatrix.getType());
|
|
fMatrix.dirtyMatrixTypeCache();
|
|
always_do(fMatrix.getType());
|
|
fMatrix.dirtyMatrixTypeCache();
|
|
always_do(fMatrix.getType());
|
|
fMatrix.dirtyMatrixTypeCache();
|
|
always_do(fMatrix.getType());
|
|
fMatrix.dirtyMatrixTypeCache();
|
|
always_do(fMatrix.getType());
|
|
fMatrix.dirtyMatrixTypeCache();
|
|
always_do(fMatrix.getType());
|
|
fMatrix.dirtyMatrixTypeCache();
|
|
always_do(fMatrix.getType());
|
|
}
|
|
private:
|
|
SkMatrix fMatrix;
|
|
float fArray[9];
|
|
SkRandom fRnd;
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
class ScaleTransMixedMatrixBench : public MatrixBench {
|
|
public:
|
|
ScaleTransMixedMatrixBench() : INHERITED("scaletrans_mixed") {
|
|
fMatrix.setAll(fRandom.nextSScalar1(), fRandom.nextSScalar1(), fRandom.nextSScalar1(),
|
|
fRandom.nextSScalar1(), fRandom.nextSScalar1(), fRandom.nextSScalar1(),
|
|
fRandom.nextSScalar1(), fRandom.nextSScalar1(), fRandom.nextSScalar1());
|
|
int i;
|
|
for (i = 0; i < kCount; i++) {
|
|
fSrc[i].fX = fRandom.nextSScalar1();
|
|
fSrc[i].fY = fRandom.nextSScalar1();
|
|
fDst[i].fX = fRandom.nextSScalar1();
|
|
fDst[i].fY = fRandom.nextSScalar1();
|
|
}
|
|
}
|
|
protected:
|
|
virtual void performTest() {
|
|
SkPoint* dst = fDst;
|
|
const SkPoint* src = fSrc;
|
|
int count = kCount;
|
|
float mx = fMatrix[SkMatrix::kMScaleX];
|
|
float my = fMatrix[SkMatrix::kMScaleY];
|
|
float tx = fMatrix[SkMatrix::kMTransX];
|
|
float ty = fMatrix[SkMatrix::kMTransY];
|
|
do {
|
|
dst->fY = SkScalarMulAdd(src->fY, my, ty);
|
|
dst->fX = SkScalarMulAdd(src->fX, mx, tx);
|
|
src += 1;
|
|
dst += 1;
|
|
} while (--count);
|
|
}
|
|
private:
|
|
enum {
|
|
kCount = 16
|
|
};
|
|
SkMatrix fMatrix;
|
|
SkPoint fSrc [kCount];
|
|
SkPoint fDst [kCount];
|
|
SkRandom fRandom;
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
class ScaleTransDoubleMatrixBench : public MatrixBench {
|
|
public:
|
|
ScaleTransDoubleMatrixBench() : INHERITED("scaletrans_double") {
|
|
init9(fMatrix);
|
|
int i;
|
|
for (i = 0; i < kCount; i++) {
|
|
fSrc[i].fX = fRandom.nextSScalar1();
|
|
fSrc[i].fY = fRandom.nextSScalar1();
|
|
fDst[i].fX = fRandom.nextSScalar1();
|
|
fDst[i].fY = fRandom.nextSScalar1();
|
|
}
|
|
}
|
|
protected:
|
|
virtual void performTest() {
|
|
SkPoint* dst = fDst;
|
|
const SkPoint* src = fSrc;
|
|
int count = kCount;
|
|
// As doubles, on Z600 Linux systems this is 2.5x as expensive as mixed mode
|
|
float mx = (float) fMatrix[SkMatrix::kMScaleX];
|
|
float my = (float) fMatrix[SkMatrix::kMScaleY];
|
|
float tx = (float) fMatrix[SkMatrix::kMTransX];
|
|
float ty = (float) fMatrix[SkMatrix::kMTransY];
|
|
do {
|
|
dst->fY = src->fY * my + ty;
|
|
dst->fX = src->fX * mx + tx;
|
|
src += 1;
|
|
dst += 1;
|
|
} while (--count);
|
|
}
|
|
private:
|
|
enum {
|
|
kCount = 16
|
|
};
|
|
double fMatrix [9];
|
|
SkPoint fSrc [kCount];
|
|
SkPoint fDst [kCount];
|
|
SkRandom fRandom;
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
class DecomposeMatrixBench : public MatrixBench {
|
|
public:
|
|
DecomposeMatrixBench() : INHERITED("decompose") {}
|
|
|
|
protected:
|
|
virtual void onPreDraw() {
|
|
for (int i = 0; i < 10; ++i) {
|
|
SkScalar rot0 = (fRandom.nextBool()) ? fRandom.nextRangeF(-180, 180) : 0.0f;
|
|
SkScalar sx = fRandom.nextRangeF(-3000.f, 3000.f);
|
|
SkScalar sy = (fRandom.nextBool()) ? fRandom.nextRangeF(-3000.f, 3000.f) : sx;
|
|
SkScalar rot1 = fRandom.nextRangeF(-180, 180);
|
|
fMatrix[i].setRotate(rot0);
|
|
fMatrix[i].postScale(sx, sy);
|
|
fMatrix[i].postRotate(rot1);
|
|
}
|
|
}
|
|
virtual void performTest() {
|
|
SkPoint rotation1, scale, rotation2;
|
|
for (int i = 0; i < 10; ++i) {
|
|
(void) SkDecomposeUpper2x2(fMatrix[i], &rotation1, &scale, &rotation2);
|
|
}
|
|
}
|
|
private:
|
|
SkMatrix fMatrix[10];
|
|
SkRandom fRandom;
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
class InvertMapRectMatrixBench : public MatrixBench {
|
|
public:
|
|
InvertMapRectMatrixBench(const char* name, int flags)
|
|
: INHERITED(name)
|
|
, fFlags(flags) {
|
|
fMatrix.reset();
|
|
fIteration = 0;
|
|
if (flags & kScale_Flag) {
|
|
fMatrix.postScale(SkFloatToScalar(1.5f), SkFloatToScalar(2.5f));
|
|
}
|
|
if (flags & kTranslate_Flag) {
|
|
fMatrix.postTranslate(SkFloatToScalar(1.5f), SkFloatToScalar(2.5f));
|
|
}
|
|
if (flags & kRotate_Flag) {
|
|
fMatrix.postRotate(SkFloatToScalar(45.0f));
|
|
}
|
|
if (flags & kPerspective_Flag) {
|
|
fMatrix.setPerspX(SkFloatToScalar(1.5f));
|
|
fMatrix.setPerspY(SkFloatToScalar(2.5f));
|
|
}
|
|
if (0 == (flags & kUncachedTypeMask_Flag)) {
|
|
fMatrix.getType();
|
|
}
|
|
}
|
|
enum Flag {
|
|
kScale_Flag = 0x01,
|
|
kTranslate_Flag = 0x02,
|
|
kRotate_Flag = 0x04,
|
|
kPerspective_Flag = 0x08,
|
|
kUncachedTypeMask_Flag = 0x10,
|
|
};
|
|
protected:
|
|
virtual void performTest() {
|
|
if (fFlags & kUncachedTypeMask_Flag) {
|
|
// This will invalidate the typemask without
|
|
// changing the matrix.
|
|
fMatrix.setPerspX(fMatrix.getPerspX());
|
|
}
|
|
SkMatrix inv;
|
|
bool invertible = fMatrix.invert(&inv);
|
|
SkASSERT(invertible);
|
|
SkRect transformedRect;
|
|
// an arbitrary, small, non-zero rect to transform
|
|
SkRect srcRect = SkRect::MakeWH(SkIntToScalar(10), SkIntToScalar(10));
|
|
if (invertible) {
|
|
inv.mapRect(&transformedRect, srcRect);
|
|
}
|
|
}
|
|
private:
|
|
SkMatrix fMatrix;
|
|
int fFlags;
|
|
unsigned fIteration;
|
|
typedef MatrixBench INHERITED;
|
|
};
|
|
|
|
///////////////////////////////////////////////////////////////////////////////
|
|
|
|
DEF_BENCH( return new EqualsMatrixBench(); )
|
|
DEF_BENCH( return new ScaleMatrixBench(); )
|
|
DEF_BENCH( return new FloatConcatMatrixBench(); )
|
|
DEF_BENCH( return new FloatDoubleConcatMatrixBench(); )
|
|
DEF_BENCH( return new DoubleConcatMatrixBench(); )
|
|
DEF_BENCH( return new GetTypeMatrixBench(); )
|
|
DEF_BENCH( return new DecomposeMatrixBench(); )
|
|
|
|
DEF_BENCH( return new InvertMapRectMatrixBench("invert_maprect_identity", 0); )
|
|
|
|
DEF_BENCH(return new InvertMapRectMatrixBench(
|
|
"invert_maprect_rectstaysrect",
|
|
InvertMapRectMatrixBench::kScale_Flag |
|
|
InvertMapRectMatrixBench::kTranslate_Flag); )
|
|
|
|
DEF_BENCH(return new InvertMapRectMatrixBench(
|
|
"invert_maprect_translate",
|
|
InvertMapRectMatrixBench::kTranslate_Flag); )
|
|
|
|
DEF_BENCH(return new InvertMapRectMatrixBench(
|
|
"invert_maprect_nonpersp",
|
|
InvertMapRectMatrixBench::kScale_Flag |
|
|
InvertMapRectMatrixBench::kRotate_Flag |
|
|
InvertMapRectMatrixBench::kTranslate_Flag); )
|
|
|
|
DEF_BENCH( return new InvertMapRectMatrixBench(
|
|
"invert_maprect_persp",
|
|
InvertMapRectMatrixBench::kPerspective_Flag); )
|
|
|
|
DEF_BENCH( return new InvertMapRectMatrixBench(
|
|
"invert_maprect_typemask_rectstaysrect",
|
|
InvertMapRectMatrixBench::kUncachedTypeMask_Flag |
|
|
InvertMapRectMatrixBench::kScale_Flag |
|
|
InvertMapRectMatrixBench::kTranslate_Flag); )
|
|
|
|
DEF_BENCH( return new InvertMapRectMatrixBench(
|
|
"invert_maprect_typemask_nonpersp",
|
|
InvertMapRectMatrixBench::kUncachedTypeMask_Flag |
|
|
InvertMapRectMatrixBench::kScale_Flag |
|
|
InvertMapRectMatrixBench::kRotate_Flag |
|
|
InvertMapRectMatrixBench::kTranslate_Flag); )
|
|
|
|
DEF_BENCH( return new ScaleTransMixedMatrixBench(); )
|
|
DEF_BENCH( return new ScaleTransDoubleMatrixBench(); )
|