Port SkTextureCompression opts to SkOpts

Pretty vanilla translation.  I cleaned up who calls whom a little.
Used to be utils -> opts -> utils, now it's just utils -> opts.

I may follow up with a pass over the NEON code for readability
and to clean up dead code.

This turns on NEON A8->R11EAC conversion for ARMv8.
Unit tests which now hit the NEON code still pass.
I can't find any related bench.

BUG=skia:4117

Review URL: https://codereview.chromium.org/1273103002
This commit is contained in:
mtklein 2015-08-06 08:17:16 -07:00 committed by Commit bot
parent d0cf9d831f
commit b6394746ff
13 changed files with 304 additions and 412 deletions

View File

@ -92,6 +92,7 @@
'include_dirs': [
'../include/private',
'../src/core',
'../src/utils',
],
'sources': [ '<@(ssse3_sources)' ],
'conditions': [
@ -112,6 +113,7 @@
'include_dirs': [
'../include/private',
'../src/core',
'../src/utils',
],
'sources': [ '<@(sse41_sources)' ],
'conditions': [

View File

@ -7,21 +7,18 @@
'<(skia_src_path)/opts/SkBitmapProcState_opts_none.cpp',
'<(skia_src_path)/opts/SkBlitMask_opts_none.cpp',
'<(skia_src_path)/opts/SkBlitRow_opts_none.cpp',
'<(skia_src_path)/opts/SkTextureCompression_opts_none.cpp',
],
'armv7_sources': [
'<(skia_src_path)/opts/SkBitmapProcState_opts_arm.cpp',
'<(skia_src_path)/opts/SkBlitMask_opts_arm.cpp',
'<(skia_src_path)/opts/SkBlitRow_opts_arm.cpp',
'<(skia_src_path)/opts/SkTextureCompression_opts_arm.cpp',
],
'neon_sources': [
'<(skia_src_path)/opts/SkBitmapProcState_arm_neon.cpp',
'<(skia_src_path)/opts/SkBitmapProcState_matrixProcs_neon.cpp',
'<(skia_src_path)/opts/SkBlitMask_opts_arm_neon.cpp',
'<(skia_src_path)/opts/SkBlitRow_opts_arm_neon.cpp',
'<(skia_src_path)/opts/SkTextureCompression_opts_neon.cpp',
'<(skia_src_path)/opts/SkOpts_neon.cpp',
],
'arm64_sources': [
@ -32,7 +29,6 @@
'<(skia_src_path)/opts/SkBlitMask_opts_arm_neon.cpp',
'<(skia_src_path)/opts/SkBlitRow_opts_arm.cpp',
'<(skia_src_path)/opts/SkBlitRow_opts_arm_neon.cpp',
'<(skia_src_path)/opts/SkTextureCompression_opts_none.cpp',
'<(skia_src_path)/opts/SkOpts_neon.cpp',
],
@ -40,14 +36,12 @@
'<(skia_src_path)/opts/SkBitmapProcState_opts_mips_dsp.cpp',
'<(skia_src_path)/opts/SkBlitMask_opts_none.cpp',
'<(skia_src_path)/opts/SkBlitRow_opts_mips_dsp.cpp',
'<(skia_src_path)/opts/SkTextureCompression_opts_none.cpp',
],
'sse2_sources': [
'<(skia_src_path)/opts/SkBitmapFilter_opts_SSE2.cpp',
'<(skia_src_path)/opts/SkBitmapProcState_opts_SSE2.cpp',
'<(skia_src_path)/opts/SkBlitRow_opts_SSE2.cpp',
'<(skia_src_path)/opts/SkTextureCompression_opts_none.cpp',
'<(skia_src_path)/opts/opts_check_x86.cpp',
'<(skia_src_path)/opts/SkOpts_sse2.cpp',
],

View File

@ -12,6 +12,7 @@
#include "SkBlurImageFilter_opts.h"
#include "SkFloatingPoint_opts.h"
#include "SkMorphologyImageFilter_opts.h"
#include "SkTextureCompressor_opts.h"
#include "SkUtils_opts.h"
#include "SkXfermode_opts.h"
@ -46,6 +47,9 @@ namespace SkOpts {
decltype( erode_x) erode_x = portable::erode_x;
decltype( erode_y) erode_y = portable::erode_y;
decltype(texture_compressor) texture_compressor = portable::texture_compressor;
decltype(fill_block_dimensions) fill_block_dimensions = portable::fill_block_dimensions;
// Each Init_foo() is defined in src/opts/SkOpts_foo.cpp.
void Init_sse2();
void Init_ssse3();

View File

@ -8,6 +8,7 @@
#ifndef SkOpts_DEFINED
#define SkOpts_DEFINED
#include "SkTextureCompressor.h"
#include "SkTypes.h"
#include "SkXfermode.h"
@ -36,6 +37,12 @@ namespace SkOpts {
typedef void (*Morph)(const SkPMColor*, SkPMColor*, int, int, int, int, int);
extern Morph dilate_x, dilate_y, erode_x, erode_y;
typedef bool (*TextureCompressor)(uint8_t* dst, const uint8_t* src,
int width, int height, size_t rowBytes);
extern TextureCompressor (*texture_compressor)(SkColorType, SkTextureCompressor::Format);
extern bool (*fill_block_dimensions)(SkTextureCompressor::Format, int* x, int* y);
}
#endif//SkOpts_DEFINED

View File

@ -11,6 +11,7 @@
#include "SkBlurImageFilter_opts.h"
#include "SkFloatingPoint_opts.h"
#include "SkMorphologyImageFilter_opts.h"
#include "SkTextureCompressor_opts.h"
#include "SkUtils_opts.h"
#include "SkXfermode_opts.h"
@ -29,5 +30,8 @@ namespace SkOpts {
dilate_y = neon::dilate_y;
erode_x = neon::erode_x;
erode_y = neon::erode_y;
texture_compressor = neon::texture_compressor;
fill_block_dimensions = neon::fill_block_dimensions;
}
}

View File

@ -1,21 +0,0 @@
/*
* Copyright 2014
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#ifndef SkTextureCompression_opts_DEFINED
#define SkTextureCompression_opts_DEFINED
#include "SkTextureCompressor.h"
#include "SkImageInfo.h"
SkTextureCompressor::CompressionProc
SkTextureCompressorGetPlatformProc(SkColorType colorType, SkTextureCompressor::Format fmt);
// Returns true if dimX and dimY are set to the block size of the supplied
// compression format according to how the platform can consume them. Returns false otherwise.
bool SkTextureCompressorGetPlatformDims(SkTextureCompressor::Format fmt, int* dimX, int* dimY);
#endif // SkTextureCompression_opts_DEFINED

View File

@ -1,59 +0,0 @@
/*
* Copyright 2014 Google Inc.
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "SkTextureCompression_opts.h"
#include "SkTextureCompression_opts_neon.h"
#include "SkUtilsArm.h"
SkTextureCompressor::CompressionProc
SkTextureCompressorGetPlatformProc(SkColorType colorType, SkTextureCompressor::Format fmt) {
#if SK_ARM_NEON_IS_NONE
return NULL;
#else
#if SK_ARM_NEON_IS_DYNAMIC
if (!sk_cpu_arm_has_neon()) {
return NULL;
}
#endif
switch (colorType) {
case kAlpha_8_SkColorType:
{
switch (fmt) {
case SkTextureCompressor::kR11_EAC_Format:
return CompressA8toR11EAC_NEON;
default:
return NULL;
}
}
break;
default:
return NULL;
}
#endif
}
bool SkTextureCompressorGetPlatformDims(SkTextureCompressor::Format fmt, int* dimX, int* dimY) {
#if SK_ARM_NEON_IS_NONE
return false;
#else
#if SK_ARM_NEON_IS_DYNAMIC
if (!sk_cpu_arm_has_neon()) {
return false;
}
#endif
switch (fmt) {
case SkTextureCompressor::kR11_EAC_Format:
*dimX = 16;
*dimY = 4;
return true;
default:
return false;
}
return false;
#endif
}

View File

@ -1,239 +0,0 @@
/*
* Copyright 2014
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "SkTextureCompressor.h"
#include "SkTextureCompression_opts.h"
#include <arm_neon.h>
// Converts indices in each of the four bits of the register from
// 0, 1, 2, 3, 4, 5, 6, 7
// to
// 3, 2, 1, 0, 4, 5, 6, 7
//
// A more detailed explanation can be found in SkTextureCompressor::convert_indices
static inline uint8x16_t convert_indices(const uint8x16_t &x) {
static const int8x16_t kThree = {
0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03,
0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03,
};
static const int8x16_t kZero = {
0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
};
// Take top three bits
int8x16_t sx = vreinterpretq_s8_u8(x);
// Negate ...
sx = vnegq_s8(sx);
// Add three...
sx = vaddq_s8(sx, kThree);
// Generate negatives mask
const int8x16_t mask = vreinterpretq_s8_u8(vcltq_s8(sx, kZero));
// Absolute value
sx = vabsq_s8(sx);
// Add three to the values that were negative...
return vreinterpretq_u8_s8(vaddq_s8(sx, vandq_s8(mask, kThree)));
}
template<unsigned shift>
static inline uint64x2_t shift_swap(const uint64x2_t &x, const uint64x2_t &mask) {
uint64x2_t t = vandq_u64(mask, veorq_u64(x, vshrq_n_u64(x, shift)));
return veorq_u64(x, veorq_u64(t, vshlq_n_u64(t, shift)));
}
static inline uint64x2_t pack_indices(const uint64x2_t &x) {
// x: 00 a e 00 b f 00 c g 00 d h 00 i m 00 j n 00 k o 00 l p
static const uint64x2_t kMask1 = { 0x3FC0003FC00000ULL, 0x3FC0003FC00000ULL };
uint64x2_t ret = shift_swap<10>(x, kMask1);
// x: b f 00 00 00 a e c g i m 00 00 00 d h j n 00 k o 00 l p
static const uint64x2_t kMask2 = { (0x3FULL << 52), (0x3FULL << 52) };
static const uint64x2_t kMask3 = { (0x3FULL << 28), (0x3FULL << 28) };
const uint64x2_t x1 = vandq_u64(vshlq_n_u64(ret, 52), kMask2);
const uint64x2_t x2 = vandq_u64(vshlq_n_u64(ret, 20), kMask3);
ret = vshrq_n_u64(vorrq_u64(ret, vorrq_u64(x1, x2)), 16);
// x: 00 00 00 00 00 00 00 00 b f l p a e c g i m k o d h j n
static const uint64x2_t kMask4 = { 0xFC0000ULL, 0xFC0000ULL };
ret = shift_swap<6>(ret, kMask4);
#if defined (SK_CPU_BENDIAN)
// x: 00 00 00 00 00 00 00 00 b f l p a e i m c g k o d h j n
static const uint64x2_t kMask5 = { 0x3FULL, 0x3FULL };
ret = shift_swap<36>(ret, kMask5);
// x: 00 00 00 00 00 00 00 00 b f j n a e i m c g k o d h l p
static const uint64x2_t kMask6 = { 0xFFF000000ULL, 0xFFF000000ULL };
ret = shift_swap<12>(ret, kMask6);
#else
// x: 00 00 00 00 00 00 00 00 c g i m d h l p b f j n a e k o
static const uint64x2_t kMask5 = { 0xFC0ULL, 0xFC0ULL };
ret = shift_swap<36>(ret, kMask5);
// x: 00 00 00 00 00 00 00 00 a e i m d h l p b f j n c g k o
static const uint64x2_t kMask6 = { (0xFFFULL << 36), (0xFFFULL << 36) };
static const uint64x2_t kMask7 = { 0xFFFFFFULL, 0xFFFFFFULL };
static const uint64x2_t kMask8 = { 0xFFFULL, 0xFFFULL };
const uint64x2_t y1 = vandq_u64(ret, kMask6);
const uint64x2_t y2 = vshlq_n_u64(vandq_u64(ret, kMask7), 12);
const uint64x2_t y3 = vandq_u64(vshrq_n_u64(ret, 24), kMask8);
ret = vorrq_u64(y1, vorrq_u64(y2, y3));
#endif
// x: 00 00 00 00 00 00 00 00 a e i m b f j n c g k o d h l p
// Set the header
static const uint64x2_t kHeader = { 0x8490000000000000ULL, 0x8490000000000000ULL };
return vorrq_u64(kHeader, ret);
}
// Takes a row of alpha values and places the most significant three bits of each byte into
// the least significant bits of the same byte
static inline uint8x16_t make_index_row(const uint8x16_t &x) {
static const uint8x16_t kTopThreeMask = {
0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0,
0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0,
};
return vshrq_n_u8(vandq_u8(x, kTopThreeMask), 5);
}
// Returns true if all of the bits in x are 0.
static inline bool is_zero(uint8x16_t x) {
// First experiments say that this is way slower than just examining the lanes
// but it might need a little more investigation.
#if 0
// This code path tests the system register for overflow. We trigger
// overflow by adding x to a register with all of its bits set. The
// first instruction sets the bits.
int reg;
asm ("VTST.8 %%q0, %q1, %q1\n"
"VQADD.u8 %q1, %%q0\n"
"VMRS %0, FPSCR\n"
: "=r"(reg) : "w"(vreinterpretq_f32_u8(x)) : "q0", "q1");
// Bit 21 corresponds to the overflow flag.
return reg & (0x1 << 21);
#else
const uint64x2_t cvt = vreinterpretq_u64_u8(x);
const uint64_t l1 = vgetq_lane_u64(cvt, 0);
return (l1 == 0) && (l1 == vgetq_lane_u64(cvt, 1));
#endif
}
#if defined (SK_CPU_BENDIAN)
static inline uint64x2_t fix_endianness(uint64x2_t x) {
return x;
}
#else
static inline uint64x2_t fix_endianness(uint64x2_t x) {
return vreinterpretq_u64_u8(vrev64q_u8(vreinterpretq_u8_u64(x)));
}
#endif
static void compress_r11eac_blocks(uint64_t* dst, const uint8_t* src, size_t rowBytes) {
// Try to avoid switching between vector and non-vector ops...
const uint8_t *const src1 = src;
const uint8_t *const src2 = src + rowBytes;
const uint8_t *const src3 = src + 2*rowBytes;
const uint8_t *const src4 = src + 3*rowBytes;
uint64_t *const dst1 = dst;
uint64_t *const dst2 = dst + 2;
const uint8x16_t alphaRow1 = vld1q_u8(src1);
const uint8x16_t alphaRow2 = vld1q_u8(src2);
const uint8x16_t alphaRow3 = vld1q_u8(src3);
const uint8x16_t alphaRow4 = vld1q_u8(src4);
const uint8x16_t cmp12 = vceqq_u8(alphaRow1, alphaRow2);
const uint8x16_t cmp34 = vceqq_u8(alphaRow3, alphaRow4);
const uint8x16_t cmp13 = vceqq_u8(alphaRow1, alphaRow3);
const uint8x16_t cmp = vandq_u8(vandq_u8(cmp12, cmp34), cmp13);
const uint8x16_t ncmp = vmvnq_u8(cmp);
const uint8x16_t nAlphaRow1 = vmvnq_u8(alphaRow1);
if (is_zero(ncmp)) {
if (is_zero(alphaRow1)) {
static const uint64x2_t kTransparent = { 0x0020000000002000ULL,
0x0020000000002000ULL };
vst1q_u64(dst1, kTransparent);
vst1q_u64(dst2, kTransparent);
return;
} else if (is_zero(nAlphaRow1)) {
vst1q_u64(dst1, vreinterpretq_u64_u8(cmp));
vst1q_u64(dst2, vreinterpretq_u64_u8(cmp));
return;
}
}
const uint8x16_t indexRow1 = convert_indices(make_index_row(alphaRow1));
const uint8x16_t indexRow2 = convert_indices(make_index_row(alphaRow2));
const uint8x16_t indexRow3 = convert_indices(make_index_row(alphaRow3));
const uint8x16_t indexRow4 = convert_indices(make_index_row(alphaRow4));
const uint64x2_t indexRow12 = vreinterpretq_u64_u8(
vorrq_u8(vshlq_n_u8(indexRow1, 3), indexRow2));
const uint64x2_t indexRow34 = vreinterpretq_u64_u8(
vorrq_u8(vshlq_n_u8(indexRow3, 3), indexRow4));
const uint32x4x2_t blockIndices = vtrnq_u32(vreinterpretq_u32_u64(indexRow12),
vreinterpretq_u32_u64(indexRow34));
const uint64x2_t blockIndicesLeft = vreinterpretq_u64_u32(vrev64q_u32(blockIndices.val[0]));
const uint64x2_t blockIndicesRight = vreinterpretq_u64_u32(vrev64q_u32(blockIndices.val[1]));
const uint64x2_t indicesLeft = fix_endianness(pack_indices(blockIndicesLeft));
const uint64x2_t indicesRight = fix_endianness(pack_indices(blockIndicesRight));
const uint64x2_t d1 = vcombine_u64(vget_low_u64(indicesLeft), vget_low_u64(indicesRight));
const uint64x2_t d2 = vcombine_u64(vget_high_u64(indicesLeft), vget_high_u64(indicesRight));
vst1q_u64(dst1, d1);
vst1q_u64(dst2, d2);
}
bool CompressA8toR11EAC_NEON(uint8_t* dst, const uint8_t* src,
int width, int height, size_t rowBytes) {
// Since we're going to operate on 4 blocks at a time, the src width
// must be a multiple of 16. However, the height only needs to be a
// multiple of 4
if (0 == width || 0 == height || (width % 16) != 0 || (height % 4) != 0) {
return SkTextureCompressor::CompressBufferToFormat(
dst, src,
kAlpha_8_SkColorType,
width, height, rowBytes,
SkTextureCompressor::kR11_EAC_Format, false);
}
const int blocksX = width >> 2;
const int blocksY = height >> 2;
SkASSERT((blocksX % 4) == 0);
uint64_t* encPtr = reinterpret_cast<uint64_t*>(dst);
for (int y = 0; y < blocksY; ++y) {
for (int x = 0; x < blocksX; x+=4) {
// Compress it
compress_r11eac_blocks(encPtr, src + 4*x, rowBytes);
encPtr += 4;
}
src += 4 * rowBytes;
}
return true;
}

View File

@ -1,14 +0,0 @@
/*
* Copyright 2014 The Android Open Source Project
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#ifndef SkTextureCompression_opts_neon_h_
#define SkTextureCompression_opts_neon_h_
bool CompressA8toR11EAC_NEON(uint8_t* dst, const uint8_t* src,
int width, int height, size_t rowBytes);
#endif // SkTextureCompression_opts_neon_h_

View File

@ -1,17 +0,0 @@
/*
* Copyright 2014
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "SkTextureCompression_opts.h"
SkTextureCompressor::CompressionProc
SkTextureCompressorGetPlatformProc(SkColorType colorType, SkTextureCompressor::Format fmt) {
return NULL;
}
bool SkTextureCompressorGetPlatformDims(SkTextureCompressor::Format fmt, int* dimX, int* dimY) {
return false;
}

View File

@ -0,0 +1,267 @@
/*
* Copyright 2015 Google Inc.
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#ifndef SkTextureCompressor_opts_DEFINED
#define SkTextureCompressor_opts_DEFINED
#include "SkOpts.h"
namespace SK_OPTS_NS {
#if defined(SK_ARM_HAS_NEON)
// Converts indices in each of the four bits of the register from
// 0, 1, 2, 3, 4, 5, 6, 7
// to
// 3, 2, 1, 0, 4, 5, 6, 7
//
// A more detailed explanation can be found in SkTextureCompressor::convert_indices
static inline uint8x16_t convert_indices(const uint8x16_t &x) {
static const int8x16_t kThree = {
0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03,
0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03, 0x03,
};
static const int8x16_t kZero = {
0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00,
};
// Take top three bits
int8x16_t sx = vreinterpretq_s8_u8(x);
// Negate ...
sx = vnegq_s8(sx);
// Add three...
sx = vaddq_s8(sx, kThree);
// Generate negatives mask
const int8x16_t mask = vreinterpretq_s8_u8(vcltq_s8(sx, kZero));
// Absolute value
sx = vabsq_s8(sx);
// Add three to the values that were negative...
return vreinterpretq_u8_s8(vaddq_s8(sx, vandq_s8(mask, kThree)));
}
template<unsigned shift>
static inline uint64x2_t shift_swap(const uint64x2_t &x, const uint64x2_t &mask) {
uint64x2_t t = vandq_u64(mask, veorq_u64(x, vshrq_n_u64(x, shift)));
return veorq_u64(x, veorq_u64(t, vshlq_n_u64(t, shift)));
}
static inline uint64x2_t pack_indices(const uint64x2_t &x) {
// x: 00 a e 00 b f 00 c g 00 d h 00 i m 00 j n 00 k o 00 l p
static const uint64x2_t kMask1 = { 0x3FC0003FC00000ULL, 0x3FC0003FC00000ULL };
uint64x2_t ret = shift_swap<10>(x, kMask1);
// x: b f 00 00 00 a e c g i m 00 00 00 d h j n 00 k o 00 l p
static const uint64x2_t kMask2 = { (0x3FULL << 52), (0x3FULL << 52) };
static const uint64x2_t kMask3 = { (0x3FULL << 28), (0x3FULL << 28) };
const uint64x2_t x1 = vandq_u64(vshlq_n_u64(ret, 52), kMask2);
const uint64x2_t x2 = vandq_u64(vshlq_n_u64(ret, 20), kMask3);
ret = vshrq_n_u64(vorrq_u64(ret, vorrq_u64(x1, x2)), 16);
// x: 00 00 00 00 00 00 00 00 b f l p a e c g i m k o d h j n
static const uint64x2_t kMask4 = { 0xFC0000ULL, 0xFC0000ULL };
ret = shift_swap<6>(ret, kMask4);
#if defined (SK_CPU_BENDIAN)
// x: 00 00 00 00 00 00 00 00 b f l p a e i m c g k o d h j n
static const uint64x2_t kMask5 = { 0x3FULL, 0x3FULL };
ret = shift_swap<36>(ret, kMask5);
// x: 00 00 00 00 00 00 00 00 b f j n a e i m c g k o d h l p
static const uint64x2_t kMask6 = { 0xFFF000000ULL, 0xFFF000000ULL };
ret = shift_swap<12>(ret, kMask6);
#else
// x: 00 00 00 00 00 00 00 00 c g i m d h l p b f j n a e k o
static const uint64x2_t kMask5 = { 0xFC0ULL, 0xFC0ULL };
ret = shift_swap<36>(ret, kMask5);
// x: 00 00 00 00 00 00 00 00 a e i m d h l p b f j n c g k o
static const uint64x2_t kMask6 = { (0xFFFULL << 36), (0xFFFULL << 36) };
static const uint64x2_t kMask7 = { 0xFFFFFFULL, 0xFFFFFFULL };
static const uint64x2_t kMask8 = { 0xFFFULL, 0xFFFULL };
const uint64x2_t y1 = vandq_u64(ret, kMask6);
const uint64x2_t y2 = vshlq_n_u64(vandq_u64(ret, kMask7), 12);
const uint64x2_t y3 = vandq_u64(vshrq_n_u64(ret, 24), kMask8);
ret = vorrq_u64(y1, vorrq_u64(y2, y3));
#endif
// x: 00 00 00 00 00 00 00 00 a e i m b f j n c g k o d h l p
// Set the header
static const uint64x2_t kHeader = { 0x8490000000000000ULL, 0x8490000000000000ULL };
return vorrq_u64(kHeader, ret);
}
// Takes a row of alpha values and places the most significant three bits of each byte into
// the least significant bits of the same byte
static inline uint8x16_t make_index_row(const uint8x16_t &x) {
static const uint8x16_t kTopThreeMask = {
0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0,
0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0, 0xE0,
};
return vshrq_n_u8(vandq_u8(x, kTopThreeMask), 5);
}
// Returns true if all of the bits in x are 0.
static inline bool is_zero(uint8x16_t x) {
// First experiments say that this is way slower than just examining the lanes
// but it might need a little more investigation.
#if 0
// This code path tests the system register for overflow. We trigger
// overflow by adding x to a register with all of its bits set. The
// first instruction sets the bits.
int reg;
asm ("VTST.8 %%q0, %q1, %q1\n"
"VQADD.u8 %q1, %%q0\n"
"VMRS %0, FPSCR\n"
: "=r"(reg) : "w"(vreinterpretq_f32_u8(x)) : "q0", "q1");
// Bit 21 corresponds to the overflow flag.
return reg & (0x1 << 21);
#else
const uint64x2_t cvt = vreinterpretq_u64_u8(x);
const uint64_t l1 = vgetq_lane_u64(cvt, 0);
return (l1 == 0) && (l1 == vgetq_lane_u64(cvt, 1));
#endif
}
#if defined (SK_CPU_BENDIAN)
static inline uint64x2_t fix_endianness(uint64x2_t x) {
return x;
}
#else
static inline uint64x2_t fix_endianness(uint64x2_t x) {
return vreinterpretq_u64_u8(vrev64q_u8(vreinterpretq_u8_u64(x)));
}
#endif
static void compress_r11eac_blocks(uint64_t* dst, const uint8_t* src, size_t rowBytes) {
// Try to avoid switching between vector and non-vector ops...
const uint8_t *const src1 = src;
const uint8_t *const src2 = src + rowBytes;
const uint8_t *const src3 = src + 2*rowBytes;
const uint8_t *const src4 = src + 3*rowBytes;
uint64_t *const dst1 = dst;
uint64_t *const dst2 = dst + 2;
const uint8x16_t alphaRow1 = vld1q_u8(src1);
const uint8x16_t alphaRow2 = vld1q_u8(src2);
const uint8x16_t alphaRow3 = vld1q_u8(src3);
const uint8x16_t alphaRow4 = vld1q_u8(src4);
const uint8x16_t cmp12 = vceqq_u8(alphaRow1, alphaRow2);
const uint8x16_t cmp34 = vceqq_u8(alphaRow3, alphaRow4);
const uint8x16_t cmp13 = vceqq_u8(alphaRow1, alphaRow3);
const uint8x16_t cmp = vandq_u8(vandq_u8(cmp12, cmp34), cmp13);
const uint8x16_t ncmp = vmvnq_u8(cmp);
const uint8x16_t nAlphaRow1 = vmvnq_u8(alphaRow1);
if (is_zero(ncmp)) {
if (is_zero(alphaRow1)) {
static const uint64x2_t kTransparent = { 0x0020000000002000ULL,
0x0020000000002000ULL };
vst1q_u64(dst1, kTransparent);
vst1q_u64(dst2, kTransparent);
return;
} else if (is_zero(nAlphaRow1)) {
vst1q_u64(dst1, vreinterpretq_u64_u8(cmp));
vst1q_u64(dst2, vreinterpretq_u64_u8(cmp));
return;
}
}
const uint8x16_t indexRow1 = convert_indices(make_index_row(alphaRow1));
const uint8x16_t indexRow2 = convert_indices(make_index_row(alphaRow2));
const uint8x16_t indexRow3 = convert_indices(make_index_row(alphaRow3));
const uint8x16_t indexRow4 = convert_indices(make_index_row(alphaRow4));
const uint64x2_t indexRow12 = vreinterpretq_u64_u8(
vorrq_u8(vshlq_n_u8(indexRow1, 3), indexRow2));
const uint64x2_t indexRow34 = vreinterpretq_u64_u8(
vorrq_u8(vshlq_n_u8(indexRow3, 3), indexRow4));
const uint32x4x2_t blockIndices = vtrnq_u32(vreinterpretq_u32_u64(indexRow12),
vreinterpretq_u32_u64(indexRow34));
const uint64x2_t blockIndicesLeft = vreinterpretq_u64_u32(vrev64q_u32(blockIndices.val[0]));
const uint64x2_t blockIndicesRight = vreinterpretq_u64_u32(vrev64q_u32(blockIndices.val[1]));
const uint64x2_t indicesLeft = fix_endianness(pack_indices(blockIndicesLeft));
const uint64x2_t indicesRight = fix_endianness(pack_indices(blockIndicesRight));
const uint64x2_t d1 = vcombine_u64(vget_low_u64(indicesLeft), vget_low_u64(indicesRight));
const uint64x2_t d2 = vcombine_u64(vget_high_u64(indicesLeft), vget_high_u64(indicesRight));
vst1q_u64(dst1, d1);
vst1q_u64(dst2, d2);
}
static bool compress_a8_r11eac(uint8_t* dst, const uint8_t* src,
int width, int height, size_t rowBytes) {
// Since we're going to operate on 4 blocks at a time, the src width
// must be a multiple of 16. However, the height only needs to be a
// multiple of 4
if (0 == width || 0 == height || (width % 16) != 0 || (height % 4) != 0) {
return false;
}
const int blocksX = width >> 2;
const int blocksY = height >> 2;
SkASSERT((blocksX % 4) == 0);
uint64_t* encPtr = reinterpret_cast<uint64_t*>(dst);
for (int y = 0; y < blocksY; ++y) {
for (int x = 0; x < blocksX; x+=4) {
// Compress it
compress_r11eac_blocks(encPtr, src + 4*x, rowBytes);
encPtr += 4;
}
src += 4 * rowBytes;
}
return true;
}
static SkOpts::TextureCompressor texture_compressor(SkColorType ct,
SkTextureCompressor::Format fmt) {
if (ct == kAlpha_8_SkColorType && fmt == SkTextureCompressor::kR11_EAC_Format) {
return compress_a8_r11eac;
}
return nullptr;
}
static bool fill_block_dimensions(SkTextureCompressor::Format fmt, int* x, int* y) {
if (fmt == SkTextureCompressor::kR11_EAC_Format) {
*x = 16;
*y = 4;
return true;
}
return false;
}
#else
static SkOpts::TextureCompressor texture_compressor(SkColorType, SkTextureCompressor::Format) {
return nullptr;
}
static bool fill_block_dimensions(SkTextureCompressor::Format, int*, int*) {
return false;
}
#endif
} // namespace SK_OPTS_NS
#endif//SkTextureCompressor_opts_DEFINED

View File

@ -14,8 +14,7 @@
#include "SkBitmapProcShader.h"
#include "SkData.h"
#include "SkEndian.h"
#include "SkTextureCompression_opts.h"
#include "SkOpts.h"
#ifndef SK_IGNORE_ETC1_SUPPORT
# include "etc1.h"
@ -40,7 +39,7 @@ void GetBlockDimensions(Format format, int* dimX, int* dimY, bool matchSpec) {
return;
}
if (!matchSpec && SkTextureCompressorGetPlatformDims(format, dimX, dimY)) {
if (!matchSpec && SkOpts::fill_block_dimensions(format, dimX, dimY)) {
return;
}
@ -77,7 +76,7 @@ int GetCompressedDataSize(Format fmt, int width, int height) {
GetBlockDimensions(fmt, &dimX, &dimY, true);
int encodedBlockSize = 0;
switch (fmt) {
// These formats are 64 bits per 4x4 block.
case kLATC_Format:
@ -120,54 +119,26 @@ int GetCompressedDataSize(Format fmt, int width, int height) {
}
bool CompressBufferToFormat(uint8_t* dst, const uint8_t* src, SkColorType srcColorType,
int width, int height, size_t rowBytes, Format format, bool opt) {
CompressionProc proc = NULL;
if (opt) {
proc = SkTextureCompressorGetPlatformProc(srcColorType, format);
int width, int height, size_t rowBytes, Format format) {
SkOpts::TextureCompressor proc = SkOpts::texture_compressor(srcColorType, format);
if (proc && proc(dst, src, width, height, rowBytes)) {
return true;
}
if (NULL == proc) {
switch (srcColorType) {
case kAlpha_8_SkColorType:
{
switch (format) {
case kLATC_Format:
proc = CompressA8ToLATC;
break;
case kR11_EAC_Format:
proc = CompressA8ToR11EAC;
break;
case kASTC_12x12_Format:
proc = CompressA8To12x12ASTC;
break;
default:
// Do nothing...
break;
}
}
switch (srcColorType) {
case kAlpha_8_SkColorType:
if (format == kLATC_Format) { proc = CompressA8ToLATC; }
if (format == kR11_EAC_Format) { proc = CompressA8ToR11EAC; }
if (format == kASTC_12x12_Format) { proc = CompressA8To12x12ASTC; }
break;
case kRGB_565_SkColorType:
{
switch (format) {
case kETC1_Format:
proc = compress_etc1_565;
break;
default:
// Do nothing...
break;
}
}
case kRGB_565_SkColorType:
if (format == kETC1_Format) { proc = compress_etc1_565; }
break;
default:
break;
default:
// Do nothing...
break;
}
}
if (proc) {
return proc(dst, src, width, height, rowBytes);
if (proc && proc(dst, src, width, height, rowBytes)) {
return true;
}
return false;

View File

@ -55,7 +55,7 @@ namespace SkTextureCompressor {
int GetCompressedDataSize(Format fmt, int width, int height);
// Returns an SkData holding a blob of compressed data that corresponds
// to the pixmap. If the pixmap colorType cannot be compressed using the
// to the pixmap. If the pixmap colorType cannot be compressed using the
// associated format, then we return NULL. The caller is responsible for
// calling unref() on the returned data.
SkData* CompressBitmapToFormat(const SkPixmap&, Format format);
@ -64,8 +64,7 @@ namespace SkTextureCompressor {
// large enough to hold width*height pixels. The dst data is expected to
// be large enough to hold the compressed data according to the format.
bool CompressBufferToFormat(uint8_t* dst, const uint8_t* src, SkColorType srcColorType,
int width, int height, size_t rowBytes, Format format,
bool opt = true /* Use optimization if available */);
int width, int height, size_t rowBytes, Format format);
// Decompresses the given src data from the format specified into the
// destination buffer. The width and height of the data passed corresponds
@ -81,12 +80,6 @@ namespace SkTextureCompressor {
bool DecompressBufferFromFormat(uint8_t* dst, int dstRowBytes, const uint8_t* src,
int width, int height, Format format);
// This typedef defines what the nominal aspects of a compression function
// are. The typedef is not meant to be used by clients of the API, but rather
// allows SIMD optimized compression functions to be implemented.
typedef bool (*CompressionProc)(uint8_t* dst, const uint8_t* src,
int width, int height, size_t rowBytes);
// Returns true if there exists a blitter for the specified format.
inline bool ExistsBlitterForFormat(Format format) {
switch (format) {