2013-10-23 11:06:13 +00:00
|
|
|
// Copyright 2010 Google Inc. All Rights Reserved.
|
|
|
|
//
|
|
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
// you may not use this file except in compliance with the License.
|
|
|
|
// You may obtain a copy of the License at
|
|
|
|
//
|
|
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
//
|
|
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
// See the License for the specific language governing permissions and
|
|
|
|
// limitations under the License.
|
|
|
|
//
|
|
|
|
// A (forgetful) hash table to the data seen by the compressor, to
|
|
|
|
// help create backward references to previous data.
|
|
|
|
|
|
|
|
#ifndef BROTLI_ENC_HASH_H_
|
|
|
|
#define BROTLI_ENC_HASH_H_
|
|
|
|
|
|
|
|
#include <stddef.h>
|
|
|
|
#include <stdint.h>
|
|
|
|
#include <string.h>
|
|
|
|
#include <sys/types.h>
|
|
|
|
#include <algorithm>
|
2013-12-12 18:43:05 +00:00
|
|
|
#include <cstdlib>
|
2014-03-20 13:32:35 +00:00
|
|
|
#include <memory>
|
2014-02-17 13:25:36 +00:00
|
|
|
#include <string>
|
2013-10-23 11:06:13 +00:00
|
|
|
|
2015-04-23 13:52:32 +00:00
|
|
|
#include "./dictionary_hash.h"
|
2013-10-23 11:06:13 +00:00
|
|
|
#include "./fast_log.h"
|
|
|
|
#include "./find_match_length.h"
|
|
|
|
#include "./port.h"
|
2014-10-28 12:25:22 +00:00
|
|
|
#include "./prefix.h"
|
2014-03-20 13:32:35 +00:00
|
|
|
#include "./static_dict.h"
|
2015-04-23 13:52:32 +00:00
|
|
|
#include "./transform.h"
|
2013-10-23 11:06:13 +00:00
|
|
|
|
|
|
|
namespace brotli {
|
|
|
|
|
2014-10-28 12:25:22 +00:00
|
|
|
static const int kDistanceCacheIndex[] = {
|
|
|
|
0, 1, 2, 3, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1,
|
|
|
|
};
|
|
|
|
static const int kDistanceCacheOffset[] = {
|
|
|
|
0, 0, 0, 0, -1, 1, -2, 2, -3, 3, -1, 1, -2, 2, -3, 3
|
|
|
|
};
|
|
|
|
|
2013-10-23 11:06:13 +00:00
|
|
|
// kHashMul32 multiplier has these properties:
|
|
|
|
// * The multiplier must be odd. Otherwise we may lose the highest bit.
|
|
|
|
// * No long streaks of 1s or 0s.
|
|
|
|
// * There is no effort to ensure that it is a prime, the oddity is enough
|
|
|
|
// for this use.
|
|
|
|
// * The number has been tuned heuristically against compression benchmarks.
|
|
|
|
static const uint32_t kHashMul32 = 0x1e35a7bd;
|
|
|
|
|
2015-05-11 12:11:07 +00:00
|
|
|
template<int kShiftBits>
|
2014-03-20 13:32:35 +00:00
|
|
|
inline uint32_t Hash(const uint8_t *data) {
|
2015-05-11 12:11:07 +00:00
|
|
|
uint32_t h = BROTLI_UNALIGNED_LOAD32(data) * kHashMul32;
|
|
|
|
// The higher bits contain more mixture from the multiplication,
|
|
|
|
// so we take our results from there.
|
|
|
|
return h >> (32 - kShiftBits);
|
2013-10-23 11:06:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// Usually, we always choose the longest backward reference. This function
|
|
|
|
// allows for the exception of that rule.
|
|
|
|
//
|
|
|
|
// If we choose a backward reference that is further away, it will
|
|
|
|
// usually be coded with more bits. We approximate this by assuming
|
|
|
|
// log2(distance). If the distance can be expressed in terms of the
|
|
|
|
// last four distances, we use some heuristic constants to estimate
|
|
|
|
// the bits cost. For the first up to four literals we use the bit
|
|
|
|
// cost of the literals from the literal cost model, after that we
|
|
|
|
// use the average bit cost of the cost model.
|
|
|
|
//
|
|
|
|
// This function is used to sometimes discard a longer backward reference
|
|
|
|
// when it is not much longer and the bit cost for encoding it is more
|
|
|
|
// than the saved literals.
|
2015-06-12 14:25:41 +00:00
|
|
|
inline double BackwardReferenceScore(int copy_length,
|
2014-03-20 13:32:35 +00:00
|
|
|
int backward_reference_offset) {
|
2015-06-12 14:25:41 +00:00
|
|
|
return 5.4 * copy_length - 1.20 * Log2Floor(backward_reference_offset);
|
2014-03-20 13:32:35 +00:00
|
|
|
}
|
|
|
|
|
2015-06-12 14:25:41 +00:00
|
|
|
inline double BackwardReferenceScoreUsingLastDistance(int copy_length,
|
2014-03-20 13:32:35 +00:00
|
|
|
int distance_short_code) {
|
|
|
|
static const double kDistanceShortCodeBitCost[16] = {
|
|
|
|
-0.6, 0.95, 1.17, 1.27,
|
|
|
|
0.93, 0.93, 0.96, 0.96, 0.99, 0.99,
|
|
|
|
1.05, 1.05, 1.15, 1.15, 1.25, 1.25
|
|
|
|
};
|
2015-06-12 14:25:41 +00:00
|
|
|
return 5.4 * copy_length - kDistanceShortCodeBitCost[distance_short_code];
|
2013-10-23 11:06:13 +00:00
|
|
|
}
|
|
|
|
|
2015-06-12 14:25:41 +00:00
|
|
|
struct BackwardMatch {
|
|
|
|
BackwardMatch() : distance(0), length_and_code(0) {}
|
|
|
|
|
|
|
|
BackwardMatch(int dist, int len)
|
|
|
|
: distance(dist), length_and_code((len << 5)) {}
|
|
|
|
|
|
|
|
BackwardMatch(int dist, int len, int len_code)
|
|
|
|
: distance(dist),
|
|
|
|
length_and_code((len << 5) | (len == len_code ? 0 : len_code)) {}
|
|
|
|
|
|
|
|
int length() const {
|
|
|
|
return length_and_code >> 5;
|
|
|
|
}
|
|
|
|
int length_code() const {
|
|
|
|
int code = length_and_code & 31;
|
|
|
|
return code ? code : length();
|
|
|
|
}
|
|
|
|
|
|
|
|
int distance;
|
|
|
|
int length_and_code;
|
|
|
|
};
|
|
|
|
|
2014-10-28 12:25:22 +00:00
|
|
|
// A (forgetful) hash table to the data seen by the compressor, to
|
|
|
|
// help create backward references to previous data.
|
|
|
|
//
|
|
|
|
// This is a hash map of fixed size (kBucketSize). Starting from the
|
|
|
|
// given index, kBucketSweep buckets are used to store values of a key.
|
2015-05-07 15:30:10 +00:00
|
|
|
template <int kBucketBits, int kBucketSweep, bool kUseDictionary>
|
2014-10-28 12:25:22 +00:00
|
|
|
class HashLongestMatchQuickly {
|
|
|
|
public:
|
|
|
|
HashLongestMatchQuickly() {
|
|
|
|
Reset();
|
|
|
|
}
|
|
|
|
void Reset() {
|
|
|
|
// It is not strictly necessary to fill this buffer here, but
|
|
|
|
// not filling will make the results of the compression stochastic
|
|
|
|
// (but correct). This is because random data would cause the
|
|
|
|
// system to find accidentally good backward references here and there.
|
|
|
|
std::fill(&buckets_[0],
|
|
|
|
&buckets_[sizeof(buckets_) / sizeof(buckets_[0])],
|
|
|
|
0);
|
2015-04-23 13:52:32 +00:00
|
|
|
num_dict_lookups_ = 0;
|
|
|
|
num_dict_matches_ = 0;
|
2014-10-28 12:25:22 +00:00
|
|
|
}
|
|
|
|
// Look at 4 bytes at data.
|
|
|
|
// Compute a hash from these, and store the value somewhere within
|
|
|
|
// [ix .. ix+3].
|
|
|
|
inline void Store(const uint8_t *data, const int ix) {
|
2015-05-11 12:11:07 +00:00
|
|
|
const uint32_t key = Hash<kBucketBits>(data);
|
2014-10-28 12:25:22 +00:00
|
|
|
// Wiggle the value with the bucket sweep range.
|
|
|
|
const uint32_t off = (static_cast<uint32_t>(ix) >> 3) % kBucketSweep;
|
|
|
|
buckets_[key + off] = ix;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Store hashes for a range of data.
|
|
|
|
void StoreHashes(const uint8_t *data, size_t len, int startix, int mask) {
|
|
|
|
for (int p = 0; p < len; ++p) {
|
|
|
|
Store(&data[p & mask], startix + p);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Find a longest backward match of &ring_buffer[cur_ix & ring_buffer_mask]
|
|
|
|
// up to the length of max_length.
|
|
|
|
//
|
|
|
|
// Does not look for matches longer than max_length.
|
|
|
|
// Does not look for matches further away than max_backward.
|
|
|
|
// Writes the best found match length into best_len_out.
|
|
|
|
// Writes the index (&data[index]) of the start of the best match into
|
|
|
|
// best_distance_out.
|
|
|
|
inline bool FindLongestMatch(const uint8_t * __restrict ring_buffer,
|
|
|
|
const size_t ring_buffer_mask,
|
|
|
|
const int* __restrict distance_cache,
|
|
|
|
const uint32_t cur_ix,
|
|
|
|
const uint32_t max_length,
|
|
|
|
const uint32_t max_backward,
|
|
|
|
int * __restrict best_len_out,
|
|
|
|
int * __restrict best_len_code_out,
|
|
|
|
int * __restrict best_distance_out,
|
|
|
|
double* __restrict best_score_out) {
|
|
|
|
const int best_len_in = *best_len_out;
|
|
|
|
const int cur_ix_masked = cur_ix & ring_buffer_mask;
|
|
|
|
int compare_char = ring_buffer[cur_ix_masked + best_len_in];
|
|
|
|
double best_score = *best_score_out;
|
|
|
|
int best_len = best_len_in;
|
|
|
|
int backward = distance_cache[0];
|
|
|
|
size_t prev_ix = cur_ix - backward;
|
|
|
|
bool match_found = false;
|
|
|
|
if (prev_ix < cur_ix) {
|
|
|
|
prev_ix &= ring_buffer_mask;
|
|
|
|
if (compare_char == ring_buffer[prev_ix + best_len]) {
|
|
|
|
int len = FindMatchLengthWithLimit(&ring_buffer[prev_ix],
|
|
|
|
&ring_buffer[cur_ix_masked],
|
|
|
|
max_length);
|
|
|
|
if (len >= 4) {
|
2015-06-12 14:25:41 +00:00
|
|
|
best_score = BackwardReferenceScoreUsingLastDistance(len, 0);
|
2014-10-28 12:25:22 +00:00
|
|
|
best_len = len;
|
|
|
|
*best_len_out = len;
|
|
|
|
*best_len_code_out = len;
|
|
|
|
*best_distance_out = backward;
|
|
|
|
*best_score_out = best_score;
|
|
|
|
compare_char = ring_buffer[cur_ix_masked + best_len];
|
|
|
|
if (kBucketSweep == 1) {
|
|
|
|
return true;
|
|
|
|
} else {
|
|
|
|
match_found = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2015-05-11 12:11:07 +00:00
|
|
|
const uint32_t key = Hash<kBucketBits>(&ring_buffer[cur_ix_masked]);
|
2014-10-28 12:25:22 +00:00
|
|
|
if (kBucketSweep == 1) {
|
|
|
|
// Only one to look for, don't bother to prepare for a loop.
|
|
|
|
prev_ix = buckets_[key];
|
|
|
|
backward = cur_ix - prev_ix;
|
|
|
|
prev_ix &= ring_buffer_mask;
|
|
|
|
if (compare_char != ring_buffer[prev_ix + best_len_in]) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
if (PREDICT_FALSE(backward == 0 || backward > max_backward)) {
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
const int len = FindMatchLengthWithLimit(&ring_buffer[prev_ix],
|
|
|
|
&ring_buffer[cur_ix_masked],
|
|
|
|
max_length);
|
|
|
|
if (len >= 4) {
|
|
|
|
*best_len_out = len;
|
|
|
|
*best_len_code_out = len;
|
|
|
|
*best_distance_out = backward;
|
2015-06-12 14:25:41 +00:00
|
|
|
*best_score_out = BackwardReferenceScore(len, backward);
|
2014-10-28 12:25:22 +00:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
uint32_t *bucket = buckets_ + key;
|
|
|
|
prev_ix = *bucket++;
|
|
|
|
for (int i = 0; i < kBucketSweep; ++i, prev_ix = *bucket++) {
|
|
|
|
const int backward = cur_ix - prev_ix;
|
|
|
|
prev_ix &= ring_buffer_mask;
|
|
|
|
if (compare_char != ring_buffer[prev_ix + best_len]) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (PREDICT_FALSE(backward == 0 || backward > max_backward)) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const int len =
|
|
|
|
FindMatchLengthWithLimit(&ring_buffer[prev_ix],
|
|
|
|
&ring_buffer[cur_ix_masked],
|
|
|
|
max_length);
|
|
|
|
if (len >= 4) {
|
2015-06-12 14:25:41 +00:00
|
|
|
const double score = BackwardReferenceScore(len, backward);
|
2014-10-28 12:25:22 +00:00
|
|
|
if (best_score < score) {
|
|
|
|
best_score = score;
|
|
|
|
best_len = len;
|
|
|
|
*best_len_out = best_len;
|
|
|
|
*best_len_code_out = best_len;
|
|
|
|
*best_distance_out = backward;
|
|
|
|
*best_score_out = score;
|
|
|
|
compare_char = ring_buffer[cur_ix_masked + best_len];
|
|
|
|
match_found = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2015-05-07 15:30:10 +00:00
|
|
|
}
|
|
|
|
if (kUseDictionary && !match_found &&
|
|
|
|
num_dict_matches_ >= (num_dict_lookups_ >> 7)) {
|
|
|
|
++num_dict_lookups_;
|
2015-05-11 12:11:07 +00:00
|
|
|
const uint32_t key = Hash<14>(&ring_buffer[cur_ix_masked]) << 1;
|
2015-05-07 15:30:10 +00:00
|
|
|
const uint16_t v = kStaticDictionaryHash[key];
|
|
|
|
if (v > 0) {
|
|
|
|
const int len = v & 31;
|
|
|
|
const int dist = v >> 5;
|
|
|
|
const int offset = kBrotliDictionaryOffsetsByLength[len] + len * dist;
|
|
|
|
if (len <= max_length) {
|
|
|
|
const int matchlen =
|
|
|
|
FindMatchLengthWithLimit(&ring_buffer[cur_ix_masked],
|
|
|
|
&kBrotliDictionary[offset], len);
|
|
|
|
if (matchlen == len) {
|
|
|
|
const size_t backward = max_backward + dist + 1;
|
2015-06-12 14:25:41 +00:00
|
|
|
const double score = BackwardReferenceScore(len, backward);
|
2015-05-07 15:30:10 +00:00
|
|
|
if (best_score < score) {
|
|
|
|
++num_dict_matches_;
|
|
|
|
best_score = score;
|
|
|
|
best_len = len;
|
|
|
|
*best_len_out = best_len;
|
|
|
|
*best_len_code_out = best_len;
|
|
|
|
*best_distance_out = backward;
|
|
|
|
*best_score_out = best_score;
|
|
|
|
return true;
|
2015-04-23 13:52:32 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2014-10-28 12:25:22 +00:00
|
|
|
}
|
2015-05-07 15:30:10 +00:00
|
|
|
return match_found;
|
2014-10-28 12:25:22 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
static const uint32_t kBucketSize = 1 << kBucketBits;
|
|
|
|
uint32_t buckets_[kBucketSize + kBucketSweep];
|
2015-04-23 13:52:32 +00:00
|
|
|
size_t num_dict_lookups_;
|
|
|
|
size_t num_dict_matches_;
|
2014-10-28 12:25:22 +00:00
|
|
|
};
|
|
|
|
|
2015-06-12 14:25:41 +00:00
|
|
|
// The maximum length for which the zopflification uses distinct distances.
|
|
|
|
static const int kMaxZopfliLen = 325;
|
|
|
|
|
2013-10-23 11:06:13 +00:00
|
|
|
// A (forgetful) hash table to the data seen by the compressor, to
|
|
|
|
// help create backward references to previous data.
|
|
|
|
//
|
|
|
|
// This is a hash map of fixed size (kBucketSize) to a ring buffer of
|
|
|
|
// fixed size (kBlockSize). The ring buffer contains the last kBlockSize
|
|
|
|
// index positions of the given hash key in the compressed data.
|
2014-10-28 12:25:22 +00:00
|
|
|
template <int kBucketBits,
|
|
|
|
int kBlockBits,
|
2015-06-12 14:25:41 +00:00
|
|
|
int kNumLastDistancesToCheck>
|
2013-10-23 11:06:13 +00:00
|
|
|
class HashLongestMatch {
|
|
|
|
public:
|
2014-10-28 12:25:22 +00:00
|
|
|
HashLongestMatch() : static_dict_(NULL) {
|
2013-10-23 11:06:13 +00:00
|
|
|
Reset();
|
|
|
|
}
|
2015-06-12 14:25:41 +00:00
|
|
|
|
2013-10-23 11:06:13 +00:00
|
|
|
void Reset() {
|
|
|
|
std::fill(&num_[0], &num_[sizeof(num_) / sizeof(num_[0])], 0);
|
2015-04-23 13:52:32 +00:00
|
|
|
num_dict_lookups_ = 0;
|
|
|
|
num_dict_matches_ = 0;
|
2013-10-23 11:06:13 +00:00
|
|
|
}
|
2015-06-12 14:25:41 +00:00
|
|
|
|
|
|
|
void CopyTo(HashLongestMatch* target) {
|
|
|
|
bool has_data = false;
|
|
|
|
for (int i = 0; i < kBucketSize; i++) {
|
|
|
|
target->num_[i] = num_[i];
|
|
|
|
has_data = (has_data || num_[i]);
|
|
|
|
}
|
|
|
|
if (has_data) {
|
|
|
|
for (int i = 0; i < kBucketSize; i++) {
|
|
|
|
for (int j = 0; j < kBlockSize; j++) {
|
|
|
|
target->buckets_[i][j] = buckets_[i][j];
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
target->num_dict_lookups_ = num_dict_lookups_;
|
|
|
|
target->num_dict_matches_ = num_dict_matches_;
|
|
|
|
target->static_dict_ = static_dict_;
|
|
|
|
}
|
|
|
|
|
2014-03-20 13:32:35 +00:00
|
|
|
void SetStaticDictionary(const StaticDictionary *dict) {
|
|
|
|
static_dict_ = dict;
|
|
|
|
}
|
2013-10-23 11:06:13 +00:00
|
|
|
|
|
|
|
// Look at 3 bytes at data.
|
|
|
|
// Compute a hash from these, and store the value of ix at that position.
|
|
|
|
inline void Store(const uint8_t *data, const int ix) {
|
2015-05-11 12:11:07 +00:00
|
|
|
const uint32_t key = Hash<kBucketBits>(data);
|
2013-10-23 11:06:13 +00:00
|
|
|
const int minor_ix = num_[key] & kBlockMask;
|
|
|
|
buckets_[key][minor_ix] = ix;
|
|
|
|
++num_[key];
|
|
|
|
}
|
|
|
|
|
|
|
|
// Store hashes for a range of data.
|
|
|
|
void StoreHashes(const uint8_t *data, size_t len, int startix, int mask) {
|
|
|
|
for (int p = 0; p < len; ++p) {
|
|
|
|
Store(&data[p & mask], startix + p);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Find a longest backward match of &data[cur_ix] up to the length of
|
|
|
|
// max_length.
|
|
|
|
//
|
|
|
|
// Does not look for matches longer than max_length.
|
|
|
|
// Does not look for matches further away than max_backward.
|
|
|
|
// Writes the best found match length into best_len_out.
|
|
|
|
// Writes the index (&data[index]) offset from the start of the best match
|
|
|
|
// into best_distance_out.
|
|
|
|
// Write the score of the best match into best_score_out.
|
|
|
|
bool FindLongestMatch(const uint8_t * __restrict data,
|
2013-11-15 18:02:17 +00:00
|
|
|
const size_t ring_buffer_mask,
|
2014-10-28 12:25:22 +00:00
|
|
|
const int* __restrict distance_cache,
|
2013-10-23 11:06:13 +00:00
|
|
|
const uint32_t cur_ix,
|
|
|
|
uint32_t max_length,
|
|
|
|
const uint32_t max_backward,
|
2014-10-28 12:25:22 +00:00
|
|
|
int * __restrict best_len_out,
|
|
|
|
int * __restrict best_len_code_out,
|
|
|
|
int * __restrict best_distance_out,
|
|
|
|
double * __restrict best_score_out) {
|
2014-02-14 14:04:23 +00:00
|
|
|
*best_len_code_out = 0;
|
2013-11-15 18:02:17 +00:00
|
|
|
const size_t cur_ix_masked = cur_ix & ring_buffer_mask;
|
2013-10-23 11:06:13 +00:00
|
|
|
bool match_found = false;
|
|
|
|
// Don't accept a short copy from far away.
|
2014-10-28 12:25:22 +00:00
|
|
|
double best_score = *best_score_out;
|
|
|
|
int best_len = *best_len_out;
|
2013-10-23 11:06:13 +00:00
|
|
|
*best_len_out = 0;
|
|
|
|
// Try last distance first.
|
2014-10-28 12:25:22 +00:00
|
|
|
for (int i = 0; i < kNumLastDistancesToCheck; ++i) {
|
|
|
|
const int idx = kDistanceCacheIndex[i];
|
|
|
|
const int backward = distance_cache[idx] + kDistanceCacheOffset[i];
|
|
|
|
size_t prev_ix = cur_ix - backward;
|
2013-10-23 11:06:13 +00:00
|
|
|
if (prev_ix >= cur_ix) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (PREDICT_FALSE(backward > max_backward)) {
|
|
|
|
continue;
|
|
|
|
}
|
2013-11-15 18:02:17 +00:00
|
|
|
prev_ix &= ring_buffer_mask;
|
2014-10-28 12:25:22 +00:00
|
|
|
|
2014-01-06 15:01:57 +00:00
|
|
|
if (cur_ix_masked + best_len > ring_buffer_mask ||
|
|
|
|
prev_ix + best_len > ring_buffer_mask ||
|
|
|
|
data[cur_ix_masked + best_len] != data[prev_ix + best_len]) {
|
2013-10-23 11:06:13 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const size_t len =
|
2013-11-15 18:02:17 +00:00
|
|
|
FindMatchLengthWithLimit(&data[prev_ix], &data[cur_ix_masked],
|
|
|
|
max_length);
|
2015-05-11 12:11:07 +00:00
|
|
|
if (len >= 3 || (len == 2 && i < 2)) {
|
2013-10-23 11:06:13 +00:00
|
|
|
// Comparing for >= 2 does not change the semantics, but just saves for
|
|
|
|
// a few unnecessary binary logarithms in backward reference score,
|
|
|
|
// since we are not interested in such short matches.
|
2015-06-12 14:25:41 +00:00
|
|
|
double score = BackwardReferenceScoreUsingLastDistance(len, i);
|
2013-10-23 11:06:13 +00:00
|
|
|
if (best_score < score) {
|
|
|
|
best_score = score;
|
|
|
|
best_len = len;
|
|
|
|
*best_len_out = best_len;
|
2013-11-19 22:32:56 +00:00
|
|
|
*best_len_code_out = best_len;
|
2014-10-28 12:25:22 +00:00
|
|
|
*best_distance_out = backward;
|
2013-10-23 11:06:13 +00:00
|
|
|
*best_score_out = best_score;
|
|
|
|
match_found = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2015-05-11 12:11:07 +00:00
|
|
|
const uint32_t key = Hash<kBucketBits>(&data[cur_ix_masked]);
|
2014-03-20 13:32:35 +00:00
|
|
|
const int * __restrict const bucket = &buckets_[key][0];
|
|
|
|
const int down = (num_[key] > kBlockSize) ? (num_[key] - kBlockSize) : 0;
|
2013-10-23 11:06:13 +00:00
|
|
|
for (int i = num_[key] - 1; i >= down; --i) {
|
2013-11-19 22:32:56 +00:00
|
|
|
int prev_ix = bucket[i & kBlockMask];
|
2014-03-20 13:32:35 +00:00
|
|
|
if (prev_ix >= 0) {
|
2013-11-19 22:32:56 +00:00
|
|
|
const size_t backward = cur_ix - prev_ix;
|
|
|
|
if (PREDICT_FALSE(backward > max_backward)) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
prev_ix &= ring_buffer_mask;
|
2014-01-06 15:01:57 +00:00
|
|
|
if (cur_ix_masked + best_len > ring_buffer_mask ||
|
|
|
|
prev_ix + best_len > ring_buffer_mask ||
|
|
|
|
data[cur_ix_masked + best_len] != data[prev_ix + best_len]) {
|
2013-11-19 22:32:56 +00:00
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const size_t len =
|
|
|
|
FindMatchLengthWithLimit(&data[prev_ix], &data[cur_ix_masked],
|
|
|
|
max_length);
|
2015-06-12 14:25:41 +00:00
|
|
|
if (len >= 4) {
|
2013-11-19 22:32:56 +00:00
|
|
|
// Comparing for >= 3 does not change the semantics, but just saves
|
|
|
|
// for a few unnecessary binary logarithms in backward reference
|
|
|
|
// score, since we are not interested in such short matches.
|
2015-06-12 14:25:41 +00:00
|
|
|
double score = BackwardReferenceScore(len, backward);
|
2013-11-19 22:32:56 +00:00
|
|
|
if (best_score < score) {
|
|
|
|
best_score = score;
|
|
|
|
best_len = len;
|
|
|
|
*best_len_out = best_len;
|
|
|
|
*best_len_code_out = best_len;
|
2014-10-28 12:25:22 +00:00
|
|
|
*best_distance_out = backward;
|
2013-11-19 22:32:56 +00:00
|
|
|
*best_score_out = best_score;
|
|
|
|
match_found = true;
|
|
|
|
}
|
2013-10-23 11:06:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2015-04-23 13:52:32 +00:00
|
|
|
if (!match_found && num_dict_matches_ >= (num_dict_lookups_ >> 7)) {
|
2015-05-11 12:11:07 +00:00
|
|
|
uint32_t key = Hash<14>(&data[cur_ix_masked]) << 1;
|
2015-04-23 13:52:32 +00:00
|
|
|
for (int k = 0; k < 2; ++k, ++key) {
|
|
|
|
++num_dict_lookups_;
|
|
|
|
const uint16_t v = kStaticDictionaryHash[key];
|
|
|
|
if (v > 0) {
|
|
|
|
const int len = v & 31;
|
|
|
|
const int dist = v >> 5;
|
|
|
|
const int offset = kBrotliDictionaryOffsetsByLength[len] + len * dist;
|
|
|
|
if (len <= max_length) {
|
|
|
|
const int matchlen =
|
|
|
|
FindMatchLengthWithLimit(&data[cur_ix_masked],
|
|
|
|
&kBrotliDictionary[offset], len);
|
|
|
|
if (matchlen == len) {
|
|
|
|
const size_t backward = max_backward + dist + 1;
|
2015-06-12 14:25:41 +00:00
|
|
|
double score = BackwardReferenceScore(len, backward);
|
2015-04-23 13:52:32 +00:00
|
|
|
if (best_score < score) {
|
|
|
|
++num_dict_matches_;
|
|
|
|
best_score = score;
|
|
|
|
best_len = len;
|
|
|
|
*best_len_out = best_len;
|
|
|
|
*best_len_code_out = best_len;
|
|
|
|
*best_distance_out = backward;
|
|
|
|
*best_score_out = best_score;
|
|
|
|
match_found = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2015-06-12 14:25:41 +00:00
|
|
|
return match_found;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Similar to FindLongestMatch(), but finds all matches.
|
|
|
|
//
|
|
|
|
// Sets *num_matches to the number of matches found, and stores the found
|
|
|
|
// matches in matches[0] to matches[*num_matches - 1].
|
|
|
|
//
|
|
|
|
// If the longest match is longer than kMaxZopfliLen, returns only this
|
|
|
|
// longest match.
|
|
|
|
//
|
|
|
|
// Requires that at least kMaxZopfliLen space is available in matches.
|
|
|
|
void FindAllMatches(const uint8_t* data,
|
|
|
|
const size_t ring_buffer_mask,
|
|
|
|
const uint32_t cur_ix,
|
|
|
|
uint32_t max_length,
|
|
|
|
const uint32_t max_backward,
|
|
|
|
int* num_matches,
|
|
|
|
BackwardMatch* matches) const {
|
|
|
|
BackwardMatch* const orig_matches = matches;
|
|
|
|
const size_t cur_ix_masked = cur_ix & ring_buffer_mask;
|
|
|
|
int best_len = 1;
|
|
|
|
int stop = static_cast<int>(cur_ix) - 64;
|
|
|
|
if (stop < 0) { stop = 0; }
|
|
|
|
for (int i = cur_ix - 1; i > stop && best_len <= 2; --i) {
|
|
|
|
size_t prev_ix = i;
|
|
|
|
const size_t backward = cur_ix - prev_ix;
|
|
|
|
if (PREDICT_FALSE(backward > max_backward)) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
prev_ix &= ring_buffer_mask;
|
|
|
|
if (data[cur_ix_masked] != data[prev_ix] ||
|
|
|
|
data[cur_ix_masked + 1] != data[prev_ix + 1]) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const size_t len =
|
|
|
|
FindMatchLengthWithLimit(&data[prev_ix], &data[cur_ix_masked],
|
|
|
|
max_length);
|
|
|
|
if (len > best_len) {
|
|
|
|
best_len = len;
|
|
|
|
if (len > kMaxZopfliLen) {
|
|
|
|
matches = orig_matches;
|
|
|
|
}
|
|
|
|
*matches++ = BackwardMatch(backward, len);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
const uint32_t key = Hash<kBucketBits>(&data[cur_ix_masked]);
|
|
|
|
const int * __restrict const bucket = &buckets_[key][0];
|
|
|
|
const int down = (num_[key] > kBlockSize) ? (num_[key] - kBlockSize) : 0;
|
|
|
|
for (int i = num_[key] - 1; i >= down; --i) {
|
|
|
|
int prev_ix = bucket[i & kBlockMask];
|
|
|
|
if (prev_ix >= 0) {
|
|
|
|
const size_t backward = cur_ix - prev_ix;
|
|
|
|
if (PREDICT_FALSE(backward > max_backward)) {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
prev_ix &= ring_buffer_mask;
|
|
|
|
if (cur_ix_masked + best_len > ring_buffer_mask ||
|
|
|
|
prev_ix + best_len > ring_buffer_mask ||
|
|
|
|
data[cur_ix_masked + best_len] != data[prev_ix + best_len]) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
const size_t len =
|
|
|
|
FindMatchLengthWithLimit(&data[prev_ix], &data[cur_ix_masked],
|
|
|
|
max_length);
|
|
|
|
if (len > best_len) {
|
|
|
|
best_len = len;
|
|
|
|
if (len > kMaxZopfliLen) {
|
|
|
|
matches = orig_matches;
|
|
|
|
}
|
|
|
|
*matches++ = BackwardMatch(backward, len);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (static_dict_ != NULL) {
|
2014-03-20 13:32:35 +00:00
|
|
|
// We decide based on first 4 bytes how many bytes to test for.
|
2015-04-23 14:20:29 +00:00
|
|
|
uint32_t prefix = BROTLI_UNALIGNED_LOAD32(&data[cur_ix_masked]);
|
2014-03-20 13:32:35 +00:00
|
|
|
int maxlen = static_dict_->GetLength(prefix);
|
|
|
|
for (int len = std::min<size_t>(maxlen, max_length);
|
|
|
|
len > best_len && len >= 4; --len) {
|
|
|
|
std::string snippet((const char *)&data[cur_ix_masked], len);
|
|
|
|
int copy_len_code;
|
|
|
|
int word_id;
|
|
|
|
if (static_dict_->Get(snippet, ©_len_code, &word_id)) {
|
|
|
|
const size_t backward = max_backward + word_id + 1;
|
2015-06-12 14:25:41 +00:00
|
|
|
*matches++ = BackwardMatch(backward, len, copy_len_code);
|
2014-03-20 13:32:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2015-06-12 14:25:41 +00:00
|
|
|
*num_matches += matches - orig_matches;
|
2013-10-23 11:06:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
// Number of hash buckets.
|
|
|
|
static const uint32_t kBucketSize = 1 << kBucketBits;
|
|
|
|
|
|
|
|
// Only kBlockSize newest backward references are kept,
|
|
|
|
// and the older are forgotten.
|
|
|
|
static const uint32_t kBlockSize = 1 << kBlockBits;
|
|
|
|
|
|
|
|
// Mask for accessing entries in a block (in a ringbuffer manner).
|
|
|
|
static const uint32_t kBlockMask = (1 << kBlockBits) - 1;
|
|
|
|
|
|
|
|
// Number of entries in a particular bucket.
|
|
|
|
uint16_t num_[kBucketSize];
|
|
|
|
|
|
|
|
// Buckets containing kBlockSize of backward references.
|
2013-11-19 22:32:56 +00:00
|
|
|
int buckets_[kBucketSize][kBlockSize];
|
2013-10-23 11:06:13 +00:00
|
|
|
|
2015-04-23 13:52:32 +00:00
|
|
|
size_t num_dict_lookups_;
|
|
|
|
size_t num_dict_matches_;
|
|
|
|
|
2014-03-20 13:32:35 +00:00
|
|
|
const StaticDictionary *static_dict_;
|
2013-10-23 11:06:13 +00:00
|
|
|
};
|
|
|
|
|
2014-03-20 13:32:35 +00:00
|
|
|
struct Hashers {
|
2015-05-07 15:30:10 +00:00
|
|
|
// For kBucketSweep == 1, enabling the dictionary lookup makes compression
|
|
|
|
// a little faster (0.5% - 1%) and it compresses 0.15% better on small text
|
|
|
|
// and html inputs.
|
|
|
|
typedef HashLongestMatchQuickly<16, 1, true> H1;
|
|
|
|
typedef HashLongestMatchQuickly<16, 2, false> H2;
|
|
|
|
typedef HashLongestMatchQuickly<16, 4, false> H3;
|
|
|
|
typedef HashLongestMatchQuickly<17, 4, true> H4;
|
2015-06-12 14:25:41 +00:00
|
|
|
typedef HashLongestMatch<14, 4, 4> H5;
|
|
|
|
typedef HashLongestMatch<14, 5, 4> H6;
|
|
|
|
typedef HashLongestMatch<15, 6, 10> H7;
|
|
|
|
typedef HashLongestMatch<15, 7, 10> H8;
|
|
|
|
typedef HashLongestMatch<15, 8, 16> H9;
|
2014-10-28 12:25:22 +00:00
|
|
|
|
|
|
|
void Init(int type) {
|
2014-03-20 13:32:35 +00:00
|
|
|
switch (type) {
|
2014-10-28 12:25:22 +00:00
|
|
|
case 1: hash_h1.reset(new H1); break;
|
|
|
|
case 2: hash_h2.reset(new H2); break;
|
|
|
|
case 3: hash_h3.reset(new H3); break;
|
|
|
|
case 4: hash_h4.reset(new H4); break;
|
|
|
|
case 5: hash_h5.reset(new H5); break;
|
|
|
|
case 6: hash_h6.reset(new H6); break;
|
|
|
|
case 7: hash_h7.reset(new H7); break;
|
|
|
|
case 8: hash_h8.reset(new H8); break;
|
|
|
|
case 9: hash_h9.reset(new H9); break;
|
|
|
|
default: break;
|
2014-03-20 13:32:35 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-06-12 14:25:41 +00:00
|
|
|
// Brotli's built in static transformable dictionary.
|
2014-03-20 13:32:35 +00:00
|
|
|
void SetStaticDictionary(const StaticDictionary *dict) {
|
2015-06-12 14:25:41 +00:00
|
|
|
if (hash_h9.get() != NULL) hash_h9->SetStaticDictionary(dict);
|
2014-03-20 13:32:35 +00:00
|
|
|
}
|
|
|
|
|
2015-06-12 13:43:54 +00:00
|
|
|
template<typename Hasher>
|
|
|
|
void WarmupHash(const size_t size, const uint8_t* dict, Hasher* hasher) {
|
|
|
|
for (size_t i = 0; i < size; i++) {
|
|
|
|
hasher->Store(dict, i);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Custom LZ77 window.
|
|
|
|
void PrependCustomDictionary(
|
|
|
|
int type, const size_t size, const uint8_t* dict) {
|
|
|
|
switch (type) {
|
|
|
|
case 1: WarmupHash(size, dict, hash_h1.get()); break;
|
|
|
|
case 2: WarmupHash(size, dict, hash_h2.get()); break;
|
|
|
|
case 3: WarmupHash(size, dict, hash_h3.get()); break;
|
|
|
|
case 4: WarmupHash(size, dict, hash_h4.get()); break;
|
|
|
|
case 5: WarmupHash(size, dict, hash_h5.get()); break;
|
|
|
|
case 6: WarmupHash(size, dict, hash_h6.get()); break;
|
|
|
|
case 7: WarmupHash(size, dict, hash_h7.get()); break;
|
|
|
|
case 8: WarmupHash(size, dict, hash_h8.get()); break;
|
|
|
|
case 9: WarmupHash(size, dict, hash_h9.get()); break;
|
|
|
|
default: break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-10-28 12:25:22 +00:00
|
|
|
std::unique_ptr<H1> hash_h1;
|
|
|
|
std::unique_ptr<H2> hash_h2;
|
|
|
|
std::unique_ptr<H3> hash_h3;
|
|
|
|
std::unique_ptr<H4> hash_h4;
|
|
|
|
std::unique_ptr<H5> hash_h5;
|
|
|
|
std::unique_ptr<H6> hash_h6;
|
|
|
|
std::unique_ptr<H7> hash_h7;
|
|
|
|
std::unique_ptr<H8> hash_h8;
|
|
|
|
std::unique_ptr<H9> hash_h9;
|
2014-03-20 13:32:35 +00:00
|
|
|
};
|
2013-11-15 18:02:17 +00:00
|
|
|
|
2013-10-23 11:06:13 +00:00
|
|
|
} // namespace brotli
|
|
|
|
|
|
|
|
#endif // BROTLI_ENC_HASH_H_
|