| Index: third_party/libwebp/enc/backward_references.c
|
| diff --git a/third_party/libwebp/enc/backward_references.c b/third_party/libwebp/enc/backward_references.c
|
| index a3c30aa0710e327eebe2234c25d3d6bb5ddba4a7..c39437dd5def4304f4b4383e058920b656473ac3 100644
|
| --- a/third_party/libwebp/enc/backward_references.c
|
| +++ b/third_party/libwebp/enc/backward_references.c
|
| @@ -16,13 +16,12 @@
|
| #include "./backward_references.h"
|
| #include "./histogram.h"
|
| #include "../dsp/lossless.h"
|
| +#include "../dsp/dsp.h"
|
| #include "../utils/color_cache.h"
|
| #include "../utils/utils.h"
|
|
|
| #define VALUES_IN_BYTE 256
|
|
|
| -#define HASH_MULTIPLIER (0xc6a4a7935bd1e995ULL)
|
| -
|
| #define MIN_BLOCK_SIZE 256 // minimum block size for backward references
|
|
|
| #define MAX_ENTROPY (1e30f)
|
| @@ -58,10 +57,28 @@ static int DistanceToPlaneCode(int xsize, int dist) {
|
| return dist + 120;
|
| }
|
|
|
| +// Returns the exact index where array1 and array2 are different if this
|
| +// index is strictly superior to best_len_match. Otherwise, it returns 0.
|
| +// If no two elements are the same, it returns max_limit.
|
| static WEBP_INLINE int FindMatchLength(const uint32_t* const array1,
|
| const uint32_t* const array2,
|
| - const int max_limit) {
|
| - int match_len = 0;
|
| + int best_len_match,
|
| + int max_limit) {
|
| + int match_len;
|
| +
|
| + // Before 'expensive' linear match, check if the two arrays match at the
|
| + // current best length index.
|
| + if (array1[best_len_match] != array2[best_len_match]) return 0;
|
| +
|
| +#if defined(WEBP_USE_SSE2)
|
| + // Check if anything is different up to best_len_match excluded.
|
| + // memcmp seems to be slower on ARM so it is disabled for now.
|
| + if (memcmp(array1, array2, best_len_match * sizeof(*array1))) return 0;
|
| + match_len = best_len_match + 1;
|
| +#else
|
| + match_len = 0;
|
| +#endif
|
| +
|
| while (match_len < max_limit && array1[match_len] == array2[match_len]) {
|
| ++match_len;
|
| }
|
| @@ -178,15 +195,12 @@ int VP8LBackwardRefsCopy(const VP8LBackwardRefs* const src,
|
| // Hash chains
|
|
|
| // initialize as empty
|
| -static void HashChainInit(VP8LHashChain* const p) {
|
| - int i;
|
| +static void HashChainReset(VP8LHashChain* const p) {
|
| assert(p != NULL);
|
| - for (i = 0; i < p->size_; ++i) {
|
| - p->chain_[i] = -1;
|
| - }
|
| - for (i = 0; i < HASH_SIZE; ++i) {
|
| - p->hash_to_first_index_[i] = -1;
|
| - }
|
| + // Set the int32_t arrays to -1.
|
| + memset(p->chain_, 0xff, p->size_ * sizeof(*p->chain_));
|
| + memset(p->hash_to_first_index_, 0xff,
|
| + HASH_SIZE * sizeof(*p->hash_to_first_index_));
|
| }
|
|
|
| int VP8LHashChainInit(VP8LHashChain* const p, int size) {
|
| @@ -196,7 +210,7 @@ int VP8LHashChainInit(VP8LHashChain* const p, int size) {
|
| p->chain_ = (int*)WebPSafeMalloc(size, sizeof(*p->chain_));
|
| if (p->chain_ == NULL) return 0;
|
| p->size_ = size;
|
| - HashChainInit(p);
|
| + HashChainReset(p);
|
| return 1;
|
| }
|
|
|
| @@ -209,209 +223,212 @@ void VP8LHashChainClear(VP8LHashChain* const p) {
|
|
|
| // -----------------------------------------------------------------------------
|
|
|
| -static WEBP_INLINE uint64_t GetPixPairHash64(const uint32_t* const argb) {
|
| - uint64_t key = ((uint64_t)argb[1] << 32) | argb[0];
|
| - key = (key * HASH_MULTIPLIER) >> (64 - HASH_BITS);
|
| +#define HASH_MULTIPLIER_HI (0xc6a4a793U)
|
| +#define HASH_MULTIPLIER_LO (0x5bd1e996U)
|
| +
|
| +static WEBP_INLINE uint32_t GetPixPairHash64(const uint32_t* const argb) {
|
| + uint32_t key;
|
| + key = argb[1] * HASH_MULTIPLIER_HI;
|
| + key += argb[0] * HASH_MULTIPLIER_LO;
|
| + key = key >> (32 - HASH_BITS);
|
| return key;
|
| }
|
|
|
| // Insertion of two pixels at a time.
|
| static void HashChainInsert(VP8LHashChain* const p,
|
| const uint32_t* const argb, int pos) {
|
| - const uint64_t hash_code = GetPixPairHash64(argb);
|
| + const uint32_t hash_code = GetPixPairHash64(argb);
|
| p->chain_[pos] = p->hash_to_first_index_[hash_code];
|
| p->hash_to_first_index_[hash_code] = pos;
|
| }
|
|
|
| -static void GetParamsForHashChainFindCopy(int quality, int xsize,
|
| - int cache_bits, int* window_size,
|
| - int* iter_pos, int* iter_limit) {
|
| - const int iter_mult = (quality < 27) ? 1 : 1 + ((quality - 27) >> 4);
|
| - const int iter_neg = -iter_mult * (quality >> 1);
|
| - // Limit the backward-ref window size for lower qualities.
|
| - const int max_window_size = (quality > 50) ? WINDOW_SIZE
|
| - : (quality > 25) ? (xsize << 8)
|
| +// Returns the maximum number of hash chain lookups to do for a
|
| +// given compression quality. Return value in range [6, 86].
|
| +static int GetMaxItersForQuality(int quality, int low_effort) {
|
| + return (low_effort ? 6 : 8) + (quality * quality) / 128;
|
| +}
|
| +
|
| +static int GetWindowSizeForHashChain(int quality, int xsize) {
|
| + const int max_window_size = (quality > 75) ? WINDOW_SIZE
|
| + : (quality > 50) ? (xsize << 8)
|
| + : (quality > 25) ? (xsize << 6)
|
| : (xsize << 4);
|
| assert(xsize > 0);
|
| - *window_size = (max_window_size > WINDOW_SIZE) ? WINDOW_SIZE
|
| - : max_window_size;
|
| - *iter_pos = 8 + (quality >> 3);
|
| - // For lower entropy images, the rigorous search loop in HashChainFindCopy
|
| - // can be relaxed.
|
| - *iter_limit = (cache_bits > 0) ? iter_neg : iter_neg / 2;
|
| + return (max_window_size > WINDOW_SIZE) ? WINDOW_SIZE : max_window_size;
|
| +}
|
| +
|
| +static WEBP_INLINE int MaxFindCopyLength(int len) {
|
| + return (len < MAX_LENGTH) ? len : MAX_LENGTH;
|
| +}
|
| +
|
| +static void HashChainFindOffset(const VP8LHashChain* const p, int base_position,
|
| + const uint32_t* const argb, int len,
|
| + int window_size, int* const distance_ptr) {
|
| + const uint32_t* const argb_start = argb + base_position;
|
| + const int min_pos =
|
| + (base_position > window_size) ? base_position - window_size : 0;
|
| + int pos;
|
| + assert(len <= MAX_LENGTH);
|
| + for (pos = p->hash_to_first_index_[GetPixPairHash64(argb_start)];
|
| + pos >= min_pos;
|
| + pos = p->chain_[pos]) {
|
| + const int curr_length =
|
| + FindMatchLength(argb + pos, argb_start, len - 1, len);
|
| + if (curr_length == len) break;
|
| + }
|
| + *distance_ptr = base_position - pos;
|
| }
|
|
|
| static int HashChainFindCopy(const VP8LHashChain* const p,
|
| - int base_position, int xsize_signed,
|
| + int base_position,
|
| const uint32_t* const argb, int max_len,
|
| - int window_size, int iter_pos, int iter_limit,
|
| + int window_size, int iter_max,
|
| int* const distance_ptr,
|
| int* const length_ptr) {
|
| const uint32_t* const argb_start = argb + base_position;
|
| - uint64_t best_val = 0;
|
| - uint32_t best_length = 1;
|
| - uint32_t best_distance = 0;
|
| - const uint32_t xsize = (uint32_t)xsize_signed;
|
| + int iter = iter_max;
|
| + int best_length = 0;
|
| + int best_distance = 0;
|
| const int min_pos =
|
| (base_position > window_size) ? base_position - window_size : 0;
|
| int pos;
|
| - assert(xsize > 0);
|
| - if (max_len > MAX_LENGTH) {
|
| - max_len = MAX_LENGTH;
|
| + int length_max = 256;
|
| + if (max_len < length_max) {
|
| + length_max = max_len;
|
| }
|
| for (pos = p->hash_to_first_index_[GetPixPairHash64(argb_start)];
|
| pos >= min_pos;
|
| pos = p->chain_[pos]) {
|
| - uint64_t val;
|
| - uint32_t curr_length;
|
| - uint32_t distance;
|
| - const uint32_t* const ptr1 = (argb + pos + best_length - 1);
|
| - const uint32_t* const ptr2 = (argb_start + best_length - 1);
|
| -
|
| - if (iter_pos < 0) {
|
| - if (iter_pos < iter_limit || best_val >= 0xff0000) {
|
| - break;
|
| - }
|
| + int curr_length;
|
| + int distance;
|
| + if (--iter < 0) {
|
| + break;
|
| }
|
| - --iter_pos;
|
| -
|
| - // Before 'expensive' linear match, check if the two arrays match at the
|
| - // current best length index and also for the succeeding elements.
|
| - if (ptr1[0] != ptr2[0] || ptr1[1] != ptr2[1]) continue;
|
| -
|
| - curr_length = FindMatchLength(argb + pos, argb_start, max_len);
|
| - if (curr_length < best_length) continue;
|
| -
|
| - distance = (uint32_t)(base_position - pos);
|
| - val = curr_length << 16;
|
| - // Favoring 2d locality here gives savings for certain images.
|
| - if (distance < 9 * xsize) {
|
| - const uint32_t y = distance / xsize;
|
| - uint32_t x = distance % xsize;
|
| - if (x > (xsize >> 1)) {
|
| - x = xsize - x;
|
| - }
|
| - if (x <= 7) {
|
| - val += 9 * 9 + 9 * 9;
|
| - val -= y * y + x * x;
|
| - }
|
| - }
|
| - if (best_val < val) {
|
| - best_val = val;
|
| +
|
| + curr_length = FindMatchLength(argb + pos, argb_start, best_length, max_len);
|
| + if (best_length < curr_length) {
|
| + distance = base_position - pos;
|
| best_length = curr_length;
|
| best_distance = distance;
|
| - if (curr_length >= (uint32_t)max_len) {
|
| - break;
|
| - }
|
| - if ((best_distance == 1 || distance == xsize) &&
|
| - best_length >= 128) {
|
| + if (curr_length >= length_max) {
|
| break;
|
| }
|
| }
|
| }
|
| - *distance_ptr = (int)best_distance;
|
| + *distance_ptr = best_distance;
|
| *length_ptr = best_length;
|
| return (best_length >= MIN_LENGTH);
|
| }
|
|
|
| -static WEBP_INLINE void PushBackCopy(VP8LBackwardRefs* const refs, int length) {
|
| - while (length >= MAX_LENGTH) {
|
| - BackwardRefsCursorAdd(refs, PixOrCopyCreateCopy(1, MAX_LENGTH));
|
| - length -= MAX_LENGTH;
|
| - }
|
| - if (length > 0) {
|
| - BackwardRefsCursorAdd(refs, PixOrCopyCreateCopy(1, length));
|
| +static WEBP_INLINE void AddSingleLiteral(uint32_t pixel, int use_color_cache,
|
| + VP8LColorCache* const hashers,
|
| + VP8LBackwardRefs* const refs) {
|
| + PixOrCopy v;
|
| + if (use_color_cache) {
|
| + const uint32_t key = VP8LColorCacheGetIndex(hashers, pixel);
|
| + if (VP8LColorCacheLookup(hashers, key) == pixel) {
|
| + v = PixOrCopyCreateCacheIdx(key);
|
| + } else {
|
| + v = PixOrCopyCreateLiteral(pixel);
|
| + VP8LColorCacheSet(hashers, key, pixel);
|
| + }
|
| + } else {
|
| + v = PixOrCopyCreateLiteral(pixel);
|
| }
|
| + BackwardRefsCursorAdd(refs, v);
|
| }
|
|
|
| static int BackwardReferencesRle(int xsize, int ysize,
|
| const uint32_t* const argb,
|
| - VP8LBackwardRefs* const refs) {
|
| + int cache_bits, VP8LBackwardRefs* const refs) {
|
| const int pix_count = xsize * ysize;
|
| - int match_len = 0;
|
| - int i;
|
| + int i, k;
|
| + const int use_color_cache = (cache_bits > 0);
|
| + VP8LColorCache hashers;
|
| +
|
| + if (use_color_cache && !VP8LColorCacheInit(&hashers, cache_bits)) {
|
| + return 0;
|
| + }
|
| ClearBackwardRefs(refs);
|
| - PushBackCopy(refs, match_len); // i=0 case
|
| - BackwardRefsCursorAdd(refs, PixOrCopyCreateLiteral(argb[0]));
|
| - for (i = 1; i < pix_count; ++i) {
|
| - if (argb[i] == argb[i - 1]) {
|
| - ++match_len;
|
| + // Add first pixel as literal.
|
| + AddSingleLiteral(argb[0], use_color_cache, &hashers, refs);
|
| + i = 1;
|
| + while (i < pix_count) {
|
| + const int max_len = MaxFindCopyLength(pix_count - i);
|
| + const int kMinLength = 4;
|
| + const int rle_len = FindMatchLength(argb + i, argb + i - 1, 0, max_len);
|
| + const int prev_row_len = (i < xsize) ? 0 :
|
| + FindMatchLength(argb + i, argb + i - xsize, 0, max_len);
|
| + if (rle_len >= prev_row_len && rle_len >= kMinLength) {
|
| + BackwardRefsCursorAdd(refs, PixOrCopyCreateCopy(1, rle_len));
|
| + // We don't need to update the color cache here since it is always the
|
| + // same pixel being copied, and that does not change the color cache
|
| + // state.
|
| + i += rle_len;
|
| + } else if (prev_row_len >= kMinLength) {
|
| + BackwardRefsCursorAdd(refs, PixOrCopyCreateCopy(xsize, prev_row_len));
|
| + if (use_color_cache) {
|
| + for (k = 0; k < prev_row_len; ++k) {
|
| + VP8LColorCacheInsert(&hashers, argb[i + k]);
|
| + }
|
| + }
|
| + i += prev_row_len;
|
| } else {
|
| - PushBackCopy(refs, match_len);
|
| - match_len = 0;
|
| - BackwardRefsCursorAdd(refs, PixOrCopyCreateLiteral(argb[i]));
|
| + AddSingleLiteral(argb[i], use_color_cache, &hashers, refs);
|
| + i++;
|
| }
|
| }
|
| - PushBackCopy(refs, match_len);
|
| + if (use_color_cache) VP8LColorCacheClear(&hashers);
|
| return !refs->error_;
|
| }
|
|
|
| -static int BackwardReferencesHashChain(int xsize, int ysize,
|
| - const uint32_t* const argb,
|
| - int cache_bits, int quality,
|
| - VP8LHashChain* const hash_chain,
|
| - VP8LBackwardRefs* const refs) {
|
| +static int BackwardReferencesLz77(int xsize, int ysize,
|
| + const uint32_t* const argb, int cache_bits,
|
| + int quality, int low_effort,
|
| + VP8LHashChain* const hash_chain,
|
| + VP8LBackwardRefs* const refs) {
|
| int i;
|
| int ok = 0;
|
| int cc_init = 0;
|
| const int use_color_cache = (cache_bits > 0);
|
| const int pix_count = xsize * ysize;
|
| VP8LColorCache hashers;
|
| - int window_size = WINDOW_SIZE;
|
| - int iter_pos = 1;
|
| - int iter_limit = -1;
|
| + int iter_max = GetMaxItersForQuality(quality, low_effort);
|
| + const int window_size = GetWindowSizeForHashChain(quality, xsize);
|
| + int min_matches = 32;
|
|
|
| if (use_color_cache) {
|
| cc_init = VP8LColorCacheInit(&hashers, cache_bits);
|
| if (!cc_init) goto Error;
|
| }
|
| -
|
| ClearBackwardRefs(refs);
|
| - GetParamsForHashChainFindCopy(quality, xsize, cache_bits,
|
| - &window_size, &iter_pos, &iter_limit);
|
| - HashChainInit(hash_chain);
|
| - for (i = 0; i < pix_count; ) {
|
| + HashChainReset(hash_chain);
|
| + for (i = 0; i < pix_count - 2; ) {
|
| // Alternative#1: Code the pixels starting at 'i' using backward reference.
|
| int offset = 0;
|
| int len = 0;
|
| - if (i < pix_count - 1) { // FindCopy(i,..) reads pixels at [i] and [i + 1].
|
| - int max_len = pix_count - i;
|
| - HashChainFindCopy(hash_chain, i, xsize, argb, max_len,
|
| - window_size, iter_pos, iter_limit,
|
| - &offset, &len);
|
| - }
|
| - if (len >= MIN_LENGTH) {
|
| - // Alternative#2: Insert the pixel at 'i' as literal, and code the
|
| - // pixels starting at 'i + 1' using backward reference.
|
| + const int max_len = MaxFindCopyLength(pix_count - i);
|
| + HashChainFindCopy(hash_chain, i, argb, max_len, window_size,
|
| + iter_max, &offset, &len);
|
| + if (len > MIN_LENGTH || (len == MIN_LENGTH && offset <= 512)) {
|
| int offset2 = 0;
|
| int len2 = 0;
|
| int k;
|
| + min_matches = 8;
|
| HashChainInsert(hash_chain, &argb[i], i);
|
| - if (i < pix_count - 2) { // FindCopy(i+1,..) reads [i + 1] and [i + 2].
|
| - int max_len = pix_count - (i + 1);
|
| - HashChainFindCopy(hash_chain, i + 1, xsize, argb, max_len,
|
| - window_size, iter_pos, iter_limit,
|
| - &offset2, &len2);
|
| + if ((len < (max_len >> 2)) && !low_effort) {
|
| + // Evaluate Alternative#2: Insert the pixel at 'i' as literal, and code
|
| + // the pixels starting at 'i + 1' using backward reference.
|
| + HashChainFindCopy(hash_chain, i + 1, argb, max_len - 1,
|
| + window_size, iter_max, &offset2,
|
| + &len2);
|
| if (len2 > len + 1) {
|
| - const uint32_t pixel = argb[i];
|
| - // Alternative#2 is a better match. So push pixel at 'i' as literal.
|
| - PixOrCopy v;
|
| - if (use_color_cache && VP8LColorCacheContains(&hashers, pixel)) {
|
| - const int ix = VP8LColorCacheGetIndex(&hashers, pixel);
|
| - v = PixOrCopyCreateCacheIdx(ix);
|
| - } else {
|
| - if (use_color_cache) VP8LColorCacheInsert(&hashers, pixel);
|
| - v = PixOrCopyCreateLiteral(pixel);
|
| - }
|
| - BackwardRefsCursorAdd(refs, v);
|
| + AddSingleLiteral(argb[i], use_color_cache, &hashers, refs);
|
| i++; // Backward reference to be done for next pixel.
|
| len = len2;
|
| offset = offset2;
|
| }
|
| }
|
| - if (len >= MAX_LENGTH) {
|
| - len = MAX_LENGTH - 1;
|
| - }
|
| BackwardRefsCursorAdd(refs, PixOrCopyCreateCopy(offset, len));
|
| if (use_color_cache) {
|
| for (k = 0; k < len; ++k) {
|
| @@ -419,33 +436,36 @@ static int BackwardReferencesHashChain(int xsize, int ysize,
|
| }
|
| }
|
| // Add to the hash_chain (but cannot add the last pixel).
|
| - {
|
| + if (offset >= 3 && offset != xsize) {
|
| const int last = (len < pix_count - 1 - i) ? len : pix_count - 1 - i;
|
| - for (k = 1; k < last; ++k) {
|
| + for (k = 2; k < last - 8; k += 2) {
|
| + HashChainInsert(hash_chain, &argb[i + k], i + k);
|
| + }
|
| + for (; k < last; ++k) {
|
| HashChainInsert(hash_chain, &argb[i + k], i + k);
|
| }
|
| }
|
| i += len;
|
| } else {
|
| - const uint32_t pixel = argb[i];
|
| - PixOrCopy v;
|
| - if (use_color_cache && VP8LColorCacheContains(&hashers, pixel)) {
|
| - // push pixel as a PixOrCopyCreateCacheIdx pixel
|
| - const int ix = VP8LColorCacheGetIndex(&hashers, pixel);
|
| - v = PixOrCopyCreateCacheIdx(ix);
|
| - } else {
|
| - if (use_color_cache) VP8LColorCacheInsert(&hashers, pixel);
|
| - v = PixOrCopyCreateLiteral(pixel);
|
| - }
|
| - BackwardRefsCursorAdd(refs, v);
|
| - if (i + 1 < pix_count) {
|
| + AddSingleLiteral(argb[i], use_color_cache, &hashers, refs);
|
| + HashChainInsert(hash_chain, &argb[i], i);
|
| + ++i;
|
| + --min_matches;
|
| + if (min_matches <= 0) {
|
| + AddSingleLiteral(argb[i], use_color_cache, &hashers, refs);
|
| HashChainInsert(hash_chain, &argb[i], i);
|
| + ++i;
|
| }
|
| - ++i;
|
| }
|
| }
|
| + while (i < pix_count) {
|
| + // Handle the last pixel(s).
|
| + AddSingleLiteral(argb[i], use_color_cache, &hashers, refs);
|
| + ++i;
|
| + }
|
| +
|
| ok = !refs->error_;
|
| -Error:
|
| + Error:
|
| if (cc_init) VP8LColorCacheClear(&hashers);
|
| return ok;
|
| }
|
| @@ -455,15 +475,14 @@ Error:
|
| typedef struct {
|
| double alpha_[VALUES_IN_BYTE];
|
| double red_[VALUES_IN_BYTE];
|
| - double literal_[PIX_OR_COPY_CODES_MAX];
|
| double blue_[VALUES_IN_BYTE];
|
| double distance_[NUM_DISTANCE_CODES];
|
| + double* literal_;
|
| } CostModel;
|
|
|
| static int BackwardReferencesTraceBackwards(
|
| - int xsize, int ysize, int recursive_cost_model,
|
| - const uint32_t* const argb, int quality, int cache_bits,
|
| - VP8LHashChain* const hash_chain,
|
| + int xsize, int ysize, const uint32_t* const argb, int quality,
|
| + int cache_bits, VP8LHashChain* const hash_chain,
|
| VP8LBackwardRefs* const refs);
|
|
|
| static void ConvertPopulationCountTableToBitEstimates(
|
| @@ -487,28 +506,10 @@ static void ConvertPopulationCountTableToBitEstimates(
|
| }
|
| }
|
|
|
| -static int CostModelBuild(CostModel* const m, int xsize, int ysize,
|
| - int recursion_level, const uint32_t* const argb,
|
| - int quality, int cache_bits,
|
| - VP8LHashChain* const hash_chain,
|
| +static int CostModelBuild(CostModel* const m, int cache_bits,
|
| VP8LBackwardRefs* const refs) {
|
| int ok = 0;
|
| - VP8LHistogram* histo = NULL;
|
| -
|
| - ClearBackwardRefs(refs);
|
| - if (recursion_level > 0) {
|
| - if (!BackwardReferencesTraceBackwards(xsize, ysize, recursion_level - 1,
|
| - argb, quality, cache_bits, hash_chain,
|
| - refs)) {
|
| - goto Error;
|
| - }
|
| - } else {
|
| - if (!BackwardReferencesHashChain(xsize, ysize, argb, cache_bits, quality,
|
| - hash_chain, refs)) {
|
| - goto Error;
|
| - }
|
| - }
|
| - histo = VP8LAllocateHistogram(cache_bits);
|
| + VP8LHistogram* const histo = VP8LAllocateHistogram(cache_bits);
|
| if (histo == NULL) goto Error;
|
|
|
| VP8LHistogramCreate(histo, refs, cache_bits);
|
| @@ -557,10 +558,35 @@ static WEBP_INLINE double GetDistanceCost(const CostModel* const m,
|
| return m->distance_[code] + extra_bits;
|
| }
|
|
|
| +static void AddSingleLiteralWithCostModel(
|
| + const uint32_t* const argb, VP8LHashChain* const hash_chain,
|
| + VP8LColorCache* const hashers, const CostModel* const cost_model, int idx,
|
| + int is_last, int use_color_cache, double prev_cost, float* const cost,
|
| + uint16_t* const dist_array) {
|
| + double cost_val = prev_cost;
|
| + const uint32_t color = argb[0];
|
| + if (!is_last) {
|
| + HashChainInsert(hash_chain, argb, idx);
|
| + }
|
| + if (use_color_cache && VP8LColorCacheContains(hashers, color)) {
|
| + const double mul0 = 0.68;
|
| + const int ix = VP8LColorCacheGetIndex(hashers, color);
|
| + cost_val += GetCacheCost(cost_model, ix) * mul0;
|
| + } else {
|
| + const double mul1 = 0.82;
|
| + if (use_color_cache) VP8LColorCacheInsert(hashers, color);
|
| + cost_val += GetLiteralCost(cost_model, color) * mul1;
|
| + }
|
| + if (cost[idx] > cost_val) {
|
| + cost[idx] = (float)cost_val;
|
| + dist_array[idx] = 1; // only one is inserted.
|
| + }
|
| +}
|
| +
|
| static int BackwardReferencesHashChainDistanceOnly(
|
| - int xsize, int ysize, int recursive_cost_model, const uint32_t* const argb,
|
| + int xsize, int ysize, const uint32_t* const argb,
|
| int quality, int cache_bits, VP8LHashChain* const hash_chain,
|
| - VP8LBackwardRefs* const refs, uint32_t* const dist_array) {
|
| + VP8LBackwardRefs* const refs, uint16_t* const dist_array) {
|
| int i;
|
| int ok = 0;
|
| int cc_init = 0;
|
| @@ -568,24 +594,27 @@ static int BackwardReferencesHashChainDistanceOnly(
|
| const int use_color_cache = (cache_bits > 0);
|
| float* const cost =
|
| (float*)WebPSafeMalloc(pix_count, sizeof(*cost));
|
| - CostModel* cost_model = (CostModel*)WebPSafeMalloc(1ULL, sizeof(*cost_model));
|
| + const size_t literal_array_size = sizeof(double) *
|
| + (NUM_LITERAL_CODES + NUM_LENGTH_CODES +
|
| + ((cache_bits > 0) ? (1 << cache_bits) : 0));
|
| + const size_t cost_model_size = sizeof(CostModel) + literal_array_size;
|
| + CostModel* const cost_model =
|
| + (CostModel*)WebPSafeMalloc(1ULL, cost_model_size);
|
| VP8LColorCache hashers;
|
| - const double mul0 = (recursive_cost_model != 0) ? 1.0 : 0.68;
|
| - const double mul1 = (recursive_cost_model != 0) ? 1.0 : 0.82;
|
| - const int min_distance_code = 2; // TODO(vikasa): tune as function of quality
|
| - int window_size = WINDOW_SIZE;
|
| - int iter_pos = 1;
|
| - int iter_limit = -1;
|
| + const int skip_length = 32 + quality;
|
| + const int skip_min_distance_code = 2;
|
| + int iter_max = GetMaxItersForQuality(quality, 0);
|
| + const int window_size = GetWindowSizeForHashChain(quality, xsize);
|
|
|
| if (cost == NULL || cost_model == NULL) goto Error;
|
|
|
| + cost_model->literal_ = (double*)(cost_model + 1);
|
| if (use_color_cache) {
|
| cc_init = VP8LColorCacheInit(&hashers, cache_bits);
|
| if (!cc_init) goto Error;
|
| }
|
|
|
| - if (!CostModelBuild(cost_model, xsize, ysize, recursive_cost_model, argb,
|
| - quality, cache_bits, hash_chain, refs)) {
|
| + if (!CostModelBuild(cost_model, cache_bits, refs)) {
|
| goto Error;
|
| }
|
|
|
| @@ -594,85 +623,80 @@ static int BackwardReferencesHashChainDistanceOnly(
|
| // We loop one pixel at a time, but store all currently best points to
|
| // non-processed locations from this point.
|
| dist_array[0] = 0;
|
| - GetParamsForHashChainFindCopy(quality, xsize, cache_bits,
|
| - &window_size, &iter_pos, &iter_limit);
|
| - HashChainInit(hash_chain);
|
| - for (i = 0; i < pix_count; ++i) {
|
| - double prev_cost = 0.0;
|
| - int shortmax;
|
| - if (i > 0) {
|
| - prev_cost = cost[i - 1];
|
| - }
|
| - for (shortmax = 0; shortmax < 2; ++shortmax) {
|
| - int offset = 0;
|
| - int len = 0;
|
| - if (i < pix_count - 1) { // FindCopy reads pixels at [i] and [i + 1].
|
| - int max_len = shortmax ? 2 : pix_count - i;
|
| - HashChainFindCopy(hash_chain, i, xsize, argb, max_len,
|
| - window_size, iter_pos, iter_limit,
|
| - &offset, &len);
|
| + HashChainReset(hash_chain);
|
| + // Add first pixel as literal.
|
| + AddSingleLiteralWithCostModel(argb + 0, hash_chain, &hashers, cost_model, 0,
|
| + 0, use_color_cache, 0.0, cost, dist_array);
|
| + for (i = 1; i < pix_count - 1; ++i) {
|
| + int offset = 0;
|
| + int len = 0;
|
| + double prev_cost = cost[i - 1];
|
| + const int max_len = MaxFindCopyLength(pix_count - i);
|
| + HashChainFindCopy(hash_chain, i, argb, max_len, window_size,
|
| + iter_max, &offset, &len);
|
| + if (len >= MIN_LENGTH) {
|
| + const int code = DistanceToPlaneCode(xsize, offset);
|
| + const double distance_cost =
|
| + prev_cost + GetDistanceCost(cost_model, code);
|
| + int k;
|
| + for (k = 1; k < len; ++k) {
|
| + const double cost_val = distance_cost + GetLengthCost(cost_model, k);
|
| + if (cost[i + k] > cost_val) {
|
| + cost[i + k] = (float)cost_val;
|
| + dist_array[i + k] = k + 1;
|
| + }
|
| }
|
| - if (len >= MIN_LENGTH) {
|
| - const int code = DistanceToPlaneCode(xsize, offset);
|
| - const double distance_cost =
|
| - prev_cost + GetDistanceCost(cost_model, code);
|
| - int k;
|
| - for (k = 1; k < len; ++k) {
|
| - const double cost_val = distance_cost + GetLengthCost(cost_model, k);
|
| - if (cost[i + k] > cost_val) {
|
| - cost[i + k] = (float)cost_val;
|
| - dist_array[i + k] = k + 1;
|
| + // This if is for speedup only. It roughly doubles the speed, and
|
| + // makes compression worse by .1 %.
|
| + if (len >= skip_length && code <= skip_min_distance_code) {
|
| + // Long copy for short distances, let's skip the middle
|
| + // lookups for better copies.
|
| + // 1) insert the hashes.
|
| + if (use_color_cache) {
|
| + for (k = 0; k < len; ++k) {
|
| + VP8LColorCacheInsert(&hashers, argb[i + k]);
|
| }
|
| }
|
| - // This if is for speedup only. It roughly doubles the speed, and
|
| - // makes compression worse by .1 %.
|
| - if (len >= 128 && code <= min_distance_code) {
|
| - // Long copy for short distances, let's skip the middle
|
| - // lookups for better copies.
|
| - // 1) insert the hashes.
|
| - if (use_color_cache) {
|
| - for (k = 0; k < len; ++k) {
|
| - VP8LColorCacheInsert(&hashers, argb[i + k]);
|
| - }
|
| - }
|
| - // 2) Add to the hash_chain (but cannot add the last pixel)
|
| - {
|
| - const int last = (len + i < pix_count - 1) ? len + i
|
| - : pix_count - 1;
|
| - for (k = i; k < last; ++k) {
|
| - HashChainInsert(hash_chain, &argb[k], k);
|
| - }
|
| + // 2) Add to the hash_chain (but cannot add the last pixel)
|
| + {
|
| + const int last = (len + i < pix_count - 1) ? len + i
|
| + : pix_count - 1;
|
| + for (k = i; k < last; ++k) {
|
| + HashChainInsert(hash_chain, &argb[k], k);
|
| }
|
| - // 3) jump.
|
| - i += len - 1; // for loop does ++i, thus -1 here.
|
| - goto next_symbol;
|
| }
|
| + // 3) jump.
|
| + i += len - 1; // for loop does ++i, thus -1 here.
|
| + goto next_symbol;
|
| }
|
| - }
|
| - if (i < pix_count - 1) {
|
| - HashChainInsert(hash_chain, &argb[i], i);
|
| - }
|
| - {
|
| - // inserting a literal pixel
|
| - double cost_val = prev_cost;
|
| - if (use_color_cache && VP8LColorCacheContains(&hashers, argb[i])) {
|
| - const int ix = VP8LColorCacheGetIndex(&hashers, argb[i]);
|
| - cost_val += GetCacheCost(cost_model, ix) * mul0;
|
| - } else {
|
| - if (use_color_cache) VP8LColorCacheInsert(&hashers, argb[i]);
|
| - cost_val += GetLiteralCost(cost_model, argb[i]) * mul1;
|
| - }
|
| - if (cost[i] > cost_val) {
|
| - cost[i] = (float)cost_val;
|
| - dist_array[i] = 1; // only one is inserted.
|
| + if (len != MIN_LENGTH) {
|
| + int code_min_length;
|
| + double cost_total;
|
| + HashChainFindOffset(hash_chain, i, argb, MIN_LENGTH, window_size,
|
| + &offset);
|
| + code_min_length = DistanceToPlaneCode(xsize, offset);
|
| + cost_total = prev_cost +
|
| + GetDistanceCost(cost_model, code_min_length) +
|
| + GetLengthCost(cost_model, 1);
|
| + if (cost[i + 1] > cost_total) {
|
| + cost[i + 1] = (float)cost_total;
|
| + dist_array[i + 1] = 2;
|
| + }
|
| }
|
| }
|
| + AddSingleLiteralWithCostModel(argb + i, hash_chain, &hashers, cost_model, i,
|
| + 0, use_color_cache, prev_cost, cost,
|
| + dist_array);
|
| next_symbol: ;
|
| }
|
| - // Last pixel still to do, it can only be a single step if not reached
|
| - // through cheaper means already.
|
| + // Handle the last pixel.
|
| + if (i == (pix_count - 1)) {
|
| + AddSingleLiteralWithCostModel(argb + i, hash_chain, &hashers, cost_model, i,
|
| + 1, use_color_cache, cost[pix_count - 2], cost,
|
| + dist_array);
|
| + }
|
| ok = !refs->error_;
|
| -Error:
|
| + Error:
|
| if (cc_init) VP8LColorCacheClear(&hashers);
|
| WebPSafeFree(cost_model);
|
| WebPSafeFree(cost);
|
| @@ -682,12 +706,12 @@ Error:
|
| // We pack the path at the end of *dist_array and return
|
| // a pointer to this part of the array. Example:
|
| // dist_array = [1x2xx3x2] => packed [1x2x1232], chosen_path = [1232]
|
| -static void TraceBackwards(uint32_t* const dist_array,
|
| +static void TraceBackwards(uint16_t* const dist_array,
|
| int dist_array_size,
|
| - uint32_t** const chosen_path,
|
| + uint16_t** const chosen_path,
|
| int* const chosen_path_size) {
|
| - uint32_t* path = dist_array + dist_array_size;
|
| - uint32_t* cur = dist_array + dist_array_size - 1;
|
| + uint16_t* path = dist_array + dist_array_size;
|
| + uint16_t* cur = dist_array + dist_array_size - 1;
|
| while (cur >= dist_array) {
|
| const int k = *cur;
|
| --path;
|
| @@ -701,20 +725,16 @@ static void TraceBackwards(uint32_t* const dist_array,
|
| static int BackwardReferencesHashChainFollowChosenPath(
|
| int xsize, int ysize, const uint32_t* const argb,
|
| int quality, int cache_bits,
|
| - const uint32_t* const chosen_path, int chosen_path_size,
|
| + const uint16_t* const chosen_path, int chosen_path_size,
|
| VP8LHashChain* const hash_chain,
|
| VP8LBackwardRefs* const refs) {
|
| const int pix_count = xsize * ysize;
|
| const int use_color_cache = (cache_bits > 0);
|
| - int size = 0;
|
| - int i = 0;
|
| - int k;
|
| int ix;
|
| + int i = 0;
|
| int ok = 0;
|
| int cc_init = 0;
|
| - int window_size = WINDOW_SIZE;
|
| - int iter_pos = 1;
|
| - int iter_limit = -1;
|
| + const int window_size = GetWindowSizeForHashChain(quality, xsize);
|
| VP8LColorCache hashers;
|
|
|
| if (use_color_cache) {
|
| @@ -723,18 +743,13 @@ static int BackwardReferencesHashChainFollowChosenPath(
|
| }
|
|
|
| ClearBackwardRefs(refs);
|
| - GetParamsForHashChainFindCopy(quality, xsize, cache_bits,
|
| - &window_size, &iter_pos, &iter_limit);
|
| - HashChainInit(hash_chain);
|
| - for (ix = 0; ix < chosen_path_size; ++ix, ++size) {
|
| + HashChainReset(hash_chain);
|
| + for (ix = 0; ix < chosen_path_size; ++ix) {
|
| int offset = 0;
|
| - int len = 0;
|
| - int max_len = chosen_path[ix];
|
| - if (max_len != 1) {
|
| - HashChainFindCopy(hash_chain, i, xsize, argb, max_len,
|
| - window_size, iter_pos, iter_limit,
|
| - &offset, &len);
|
| - assert(len == max_len);
|
| + const int len = chosen_path[ix];
|
| + if (len != 1) {
|
| + int k;
|
| + HashChainFindOffset(hash_chain, i, argb, len, window_size, &offset);
|
| BackwardRefsCursorAdd(refs, PixOrCopyCreateCopy(offset, len));
|
| if (use_color_cache) {
|
| for (k = 0; k < len; ++k) {
|
| @@ -766,29 +781,28 @@ static int BackwardReferencesHashChainFollowChosenPath(
|
| }
|
| }
|
| ok = !refs->error_;
|
| -Error:
|
| + Error:
|
| if (cc_init) VP8LColorCacheClear(&hashers);
|
| return ok;
|
| }
|
|
|
| // Returns 1 on success.
|
| static int BackwardReferencesTraceBackwards(int xsize, int ysize,
|
| - int recursive_cost_model,
|
| const uint32_t* const argb,
|
| int quality, int cache_bits,
|
| VP8LHashChain* const hash_chain,
|
| VP8LBackwardRefs* const refs) {
|
| int ok = 0;
|
| const int dist_array_size = xsize * ysize;
|
| - uint32_t* chosen_path = NULL;
|
| + uint16_t* chosen_path = NULL;
|
| int chosen_path_size = 0;
|
| - uint32_t* dist_array =
|
| - (uint32_t*)WebPSafeMalloc(dist_array_size, sizeof(*dist_array));
|
| + uint16_t* dist_array =
|
| + (uint16_t*)WebPSafeMalloc(dist_array_size, sizeof(*dist_array));
|
|
|
| if (dist_array == NULL) goto Error;
|
|
|
| if (!BackwardReferencesHashChainDistanceOnly(
|
| - xsize, ysize, recursive_cost_model, argb, quality, cache_bits, hash_chain,
|
| + xsize, ysize, argb, quality, cache_bits, hash_chain,
|
| refs, dist_array)) {
|
| goto Error;
|
| }
|
| @@ -817,72 +831,10 @@ static void BackwardReferences2DLocality(int xsize,
|
| }
|
| }
|
|
|
| -VP8LBackwardRefs* VP8LGetBackwardReferences(
|
| - int width, int height, const uint32_t* const argb, int quality,
|
| - int cache_bits, int use_2d_locality, VP8LHashChain* const hash_chain,
|
| - VP8LBackwardRefs refs_array[2]) {
|
| - int lz77_is_useful;
|
| - const int num_pix = width * height;
|
| - VP8LBackwardRefs* best = NULL;
|
| - VP8LBackwardRefs* const refs_lz77 = &refs_array[0];
|
| - VP8LBackwardRefs* const refs_rle = &refs_array[1];
|
| -
|
| - if (!BackwardReferencesHashChain(width, height, argb, cache_bits, quality,
|
| - hash_chain, refs_lz77)) {
|
| - return NULL;
|
| - }
|
| - if (!BackwardReferencesRle(width, height, argb, refs_rle)) {
|
| - return NULL;
|
| - }
|
| -
|
| - {
|
| - double bit_cost_lz77, bit_cost_rle;
|
| - VP8LHistogram* const histo = VP8LAllocateHistogram(cache_bits);
|
| - if (histo == NULL) return NULL;
|
| - // Evaluate LZ77 coding.
|
| - VP8LHistogramCreate(histo, refs_lz77, cache_bits);
|
| - bit_cost_lz77 = VP8LHistogramEstimateBits(histo);
|
| - // Evaluate RLE coding.
|
| - VP8LHistogramCreate(histo, refs_rle, cache_bits);
|
| - bit_cost_rle = VP8LHistogramEstimateBits(histo);
|
| - // Decide if LZ77 is useful.
|
| - lz77_is_useful = (bit_cost_lz77 < bit_cost_rle);
|
| - VP8LFreeHistogram(histo);
|
| - }
|
| -
|
| - // Choose appropriate backward reference.
|
| - if (lz77_is_useful) {
|
| - // TraceBackwards is costly. Don't execute it at lower quality.
|
| - const int try_lz77_trace_backwards = (quality >= 25);
|
| - best = refs_lz77; // default guess: lz77 is better
|
| - if (try_lz77_trace_backwards) {
|
| - // Set recursion level for large images using a color cache.
|
| - const int recursion_level =
|
| - (num_pix < 320 * 200) && (cache_bits > 0) ? 1 : 0;
|
| - VP8LBackwardRefs* const refs_trace = &refs_array[1];
|
| - ClearBackwardRefs(refs_trace);
|
| - if (BackwardReferencesTraceBackwards(width, height, recursion_level, argb,
|
| - quality, cache_bits, hash_chain,
|
| - refs_trace)) {
|
| - best = refs_trace;
|
| - }
|
| - }
|
| - } else {
|
| - best = refs_rle;
|
| - }
|
| -
|
| - if (use_2d_locality) BackwardReferences2DLocality(width, best);
|
| -
|
| - return best;
|
| -}
|
| -
|
| // Returns entropy for the given cache bits.
|
| -static double ComputeCacheEntropy(const uint32_t* const argb,
|
| - int xsize, int ysize,
|
| +static double ComputeCacheEntropy(const uint32_t* argb,
|
| const VP8LBackwardRefs* const refs,
|
| int cache_bits) {
|
| - int pixel_index = 0;
|
| - uint32_t k;
|
| const int use_color_cache = (cache_bits > 0);
|
| int cc_init = 0;
|
| double entropy = MAX_ENTROPY;
|
| @@ -896,33 +848,40 @@ static double ComputeCacheEntropy(const uint32_t* const argb,
|
| cc_init = VP8LColorCacheInit(&hashers, cache_bits);
|
| if (!cc_init) goto Error;
|
| }
|
| -
|
| - while (VP8LRefsCursorOk(&c)) {
|
| - const PixOrCopy* const v = c.cur_pos;
|
| - if (PixOrCopyIsLiteral(v)) {
|
| - if (use_color_cache &&
|
| - VP8LColorCacheContains(&hashers, argb[pixel_index])) {
|
| - // push pixel as a cache index
|
| - const int ix = VP8LColorCacheGetIndex(&hashers, argb[pixel_index]);
|
| - const PixOrCopy token = PixOrCopyCreateCacheIdx(ix);
|
| - VP8LHistogramAddSinglePixOrCopy(histo, &token);
|
| - } else {
|
| - VP8LHistogramAddSinglePixOrCopy(histo, v);
|
| - }
|
| - } else {
|
| - VP8LHistogramAddSinglePixOrCopy(histo, v);
|
| + if (!use_color_cache) {
|
| + while (VP8LRefsCursorOk(&c)) {
|
| + VP8LHistogramAddSinglePixOrCopy(histo, c.cur_pos);
|
| + VP8LRefsCursorNext(&c);
|
| }
|
| - if (use_color_cache) {
|
| - for (k = 0; k < PixOrCopyLength(v); ++k) {
|
| - VP8LColorCacheInsert(&hashers, argb[pixel_index + k]);
|
| + } else {
|
| + while (VP8LRefsCursorOk(&c)) {
|
| + const PixOrCopy* const v = c.cur_pos;
|
| + if (PixOrCopyIsLiteral(v)) {
|
| + const uint32_t pix = *argb++;
|
| + const uint32_t key = VP8LColorCacheGetIndex(&hashers, pix);
|
| + if (VP8LColorCacheLookup(&hashers, key) == pix) {
|
| + ++histo->literal_[NUM_LITERAL_CODES + NUM_LENGTH_CODES + key];
|
| + } else {
|
| + VP8LColorCacheSet(&hashers, key, pix);
|
| + ++histo->blue_[pix & 0xff];
|
| + ++histo->literal_[(pix >> 8) & 0xff];
|
| + ++histo->red_[(pix >> 16) & 0xff];
|
| + ++histo->alpha_[pix >> 24];
|
| + }
|
| + } else {
|
| + int len = PixOrCopyLength(v);
|
| + int code, extra_bits;
|
| + VP8LPrefixEncodeBits(len, &code, &extra_bits);
|
| + ++histo->literal_[NUM_LITERAL_CODES + code];
|
| + VP8LPrefixEncodeBits(PixOrCopyDistance(v), &code, &extra_bits);
|
| + ++histo->distance_[code];
|
| + do {
|
| + VP8LColorCacheInsert(&hashers, *argb++);
|
| + } while(--len != 0);
|
| }
|
| + VP8LRefsCursorNext(&c);
|
| }
|
| - pixel_index += PixOrCopyLength(v);
|
| - VP8LRefsCursorNext(&c);
|
| }
|
| - assert(pixel_index == xsize * ysize);
|
| - (void)xsize; // xsize is not used in non-debug compilations otherwise.
|
| - (void)ysize; // ysize is not used in non-debug compilations otherwise.
|
| entropy = VP8LHistogramEstimateBits(histo) +
|
| kSmallPenaltyForLargeCache * cache_bits;
|
| Error:
|
| @@ -931,45 +890,204 @@ static double ComputeCacheEntropy(const uint32_t* const argb,
|
| return entropy;
|
| }
|
|
|
| -// *best_cache_bits will contain how many bits are to be used for a color cache.
|
| +// Evaluate optimal cache bits for the local color cache.
|
| +// The input *best_cache_bits sets the maximum cache bits to use (passing 0
|
| +// implies disabling the local color cache). The local color cache is also
|
| +// disabled for the lower (<= 25) quality.
|
| // Returns 0 in case of memory error.
|
| -int VP8LCalculateEstimateForCacheSize(const uint32_t* const argb,
|
| - int xsize, int ysize, int quality,
|
| - VP8LHashChain* const hash_chain,
|
| - VP8LBackwardRefs* const refs,
|
| - int* const best_cache_bits) {
|
| +static int CalculateBestCacheSize(const uint32_t* const argb,
|
| + int xsize, int ysize, int quality,
|
| + VP8LHashChain* const hash_chain,
|
| + VP8LBackwardRefs* const refs,
|
| + int* const lz77_computed,
|
| + int* const best_cache_bits) {
|
| int eval_low = 1;
|
| int eval_high = 1;
|
| double entropy_low = MAX_ENTROPY;
|
| double entropy_high = MAX_ENTROPY;
|
| + const double cost_mul = 5e-4;
|
| int cache_bits_low = 0;
|
| - int cache_bits_high = MAX_COLOR_CACHE_BITS;
|
| + int cache_bits_high = (quality <= 25) ? 0 : *best_cache_bits;
|
|
|
| - if (!BackwardReferencesHashChain(xsize, ysize, argb, 0, quality, hash_chain,
|
| - refs)) {
|
| + assert(cache_bits_high <= MAX_COLOR_CACHE_BITS);
|
| +
|
| + *lz77_computed = 0;
|
| + if (cache_bits_high == 0) {
|
| + *best_cache_bits = 0;
|
| + // Local color cache is disabled.
|
| + return 1;
|
| + }
|
| + if (!BackwardReferencesLz77(xsize, ysize, argb, cache_bits_low, quality, 0,
|
| + hash_chain, refs)) {
|
| return 0;
|
| }
|
| // Do a binary search to find the optimal entropy for cache_bits.
|
| - while (cache_bits_high - cache_bits_low > 1) {
|
| + while (eval_low || eval_high) {
|
| if (eval_low) {
|
| - entropy_low =
|
| - ComputeCacheEntropy(argb, xsize, ysize, refs, cache_bits_low);
|
| + entropy_low = ComputeCacheEntropy(argb, refs, cache_bits_low);
|
| + entropy_low += entropy_low * cache_bits_low * cost_mul;
|
| eval_low = 0;
|
| }
|
| if (eval_high) {
|
| - entropy_high =
|
| - ComputeCacheEntropy(argb, xsize, ysize, refs, cache_bits_high);
|
| + entropy_high = ComputeCacheEntropy(argb, refs, cache_bits_high);
|
| + entropy_high += entropy_high * cache_bits_high * cost_mul;
|
| eval_high = 0;
|
| }
|
| if (entropy_high < entropy_low) {
|
| + const int prev_cache_bits_low = cache_bits_low;
|
| *best_cache_bits = cache_bits_high;
|
| cache_bits_low = (cache_bits_low + cache_bits_high) / 2;
|
| - eval_low = 1;
|
| + if (cache_bits_low != prev_cache_bits_low) eval_low = 1;
|
| } else {
|
| *best_cache_bits = cache_bits_low;
|
| cache_bits_high = (cache_bits_low + cache_bits_high) / 2;
|
| - eval_high = 1;
|
| + if (cache_bits_high != cache_bits_low) eval_high = 1;
|
| }
|
| }
|
| + *lz77_computed = 1;
|
| return 1;
|
| }
|
| +
|
| +// Update (in-place) backward references for specified cache_bits.
|
| +static int BackwardRefsWithLocalCache(const uint32_t* const argb,
|
| + int cache_bits,
|
| + VP8LBackwardRefs* const refs) {
|
| + int pixel_index = 0;
|
| + VP8LColorCache hashers;
|
| + VP8LRefsCursor c = VP8LRefsCursorInit(refs);
|
| + if (!VP8LColorCacheInit(&hashers, cache_bits)) return 0;
|
| +
|
| + while (VP8LRefsCursorOk(&c)) {
|
| + PixOrCopy* const v = c.cur_pos;
|
| + if (PixOrCopyIsLiteral(v)) {
|
| + const uint32_t argb_literal = v->argb_or_distance;
|
| + if (VP8LColorCacheContains(&hashers, argb_literal)) {
|
| + const int ix = VP8LColorCacheGetIndex(&hashers, argb_literal);
|
| + *v = PixOrCopyCreateCacheIdx(ix);
|
| + } else {
|
| + VP8LColorCacheInsert(&hashers, argb_literal);
|
| + }
|
| + ++pixel_index;
|
| + } else {
|
| + // refs was created without local cache, so it can not have cache indexes.
|
| + int k;
|
| + assert(PixOrCopyIsCopy(v));
|
| + for (k = 0; k < v->len; ++k) {
|
| + VP8LColorCacheInsert(&hashers, argb[pixel_index++]);
|
| + }
|
| + }
|
| + VP8LRefsCursorNext(&c);
|
| + }
|
| + VP8LColorCacheClear(&hashers);
|
| + return 1;
|
| +}
|
| +
|
| +static VP8LBackwardRefs* GetBackwardReferencesLowEffort(
|
| + int width, int height, const uint32_t* const argb, int quality,
|
| + int* const cache_bits, VP8LHashChain* const hash_chain,
|
| + VP8LBackwardRefs refs_array[2]) {
|
| + VP8LBackwardRefs* refs_lz77 = &refs_array[0];
|
| + *cache_bits = 0;
|
| + if (!BackwardReferencesLz77(width, height, argb, 0, quality,
|
| + 1 /* Low effort. */, hash_chain, refs_lz77)) {
|
| + return NULL;
|
| + }
|
| + BackwardReferences2DLocality(width, refs_lz77);
|
| + return refs_lz77;
|
| +}
|
| +
|
| +static VP8LBackwardRefs* GetBackwardReferences(
|
| + int width, int height, const uint32_t* const argb, int quality,
|
| + int* const cache_bits, VP8LHashChain* const hash_chain,
|
| + VP8LBackwardRefs refs_array[2]) {
|
| + int lz77_is_useful;
|
| + int lz77_computed;
|
| + double bit_cost_lz77, bit_cost_rle;
|
| + VP8LBackwardRefs* best = NULL;
|
| + VP8LBackwardRefs* refs_lz77 = &refs_array[0];
|
| + VP8LBackwardRefs* refs_rle = &refs_array[1];
|
| + VP8LHistogram* histo = NULL;
|
| +
|
| + if (!CalculateBestCacheSize(argb, width, height, quality, hash_chain,
|
| + refs_lz77, &lz77_computed, cache_bits)) {
|
| + goto Error;
|
| + }
|
| +
|
| + if (lz77_computed) {
|
| + // Transform refs_lz77 for the optimized cache_bits.
|
| + if (*cache_bits > 0) {
|
| + if (!BackwardRefsWithLocalCache(argb, *cache_bits, refs_lz77)) {
|
| + goto Error;
|
| + }
|
| + }
|
| + } else {
|
| + if (!BackwardReferencesLz77(width, height, argb, *cache_bits, quality,
|
| + 0 /* Low effort. */, hash_chain, refs_lz77)) {
|
| + goto Error;
|
| + }
|
| + }
|
| +
|
| + if (!BackwardReferencesRle(width, height, argb, *cache_bits, refs_rle)) {
|
| + goto Error;
|
| + }
|
| +
|
| + histo = VP8LAllocateHistogram(*cache_bits);
|
| + if (histo == NULL) goto Error;
|
| +
|
| + {
|
| + // Evaluate LZ77 coding.
|
| + VP8LHistogramCreate(histo, refs_lz77, *cache_bits);
|
| + bit_cost_lz77 = VP8LHistogramEstimateBits(histo);
|
| + // Evaluate RLE coding.
|
| + VP8LHistogramCreate(histo, refs_rle, *cache_bits);
|
| + bit_cost_rle = VP8LHistogramEstimateBits(histo);
|
| + // Decide if LZ77 is useful.
|
| + lz77_is_useful = (bit_cost_lz77 < bit_cost_rle);
|
| + }
|
| +
|
| + // Choose appropriate backward reference.
|
| + if (lz77_is_useful) {
|
| + // TraceBackwards is costly. Don't execute it at lower quality.
|
| + const int try_lz77_trace_backwards = (quality >= 25);
|
| + best = refs_lz77; // default guess: lz77 is better
|
| + if (try_lz77_trace_backwards) {
|
| + VP8LBackwardRefs* const refs_trace = refs_rle;
|
| + if (!VP8LBackwardRefsCopy(refs_lz77, refs_trace)) {
|
| + best = NULL;
|
| + goto Error;
|
| + }
|
| + if (BackwardReferencesTraceBackwards(width, height, argb, quality,
|
| + *cache_bits, hash_chain,
|
| + refs_trace)) {
|
| + double bit_cost_trace;
|
| + // Evaluate LZ77 coding.
|
| + VP8LHistogramCreate(histo, refs_trace, *cache_bits);
|
| + bit_cost_trace = VP8LHistogramEstimateBits(histo);
|
| + if (bit_cost_trace < bit_cost_lz77) {
|
| + best = refs_trace;
|
| + }
|
| + }
|
| + }
|
| + } else {
|
| + best = refs_rle;
|
| + }
|
| +
|
| + BackwardReferences2DLocality(width, best);
|
| +
|
| + Error:
|
| + VP8LFreeHistogram(histo);
|
| + return best;
|
| +}
|
| +
|
| +VP8LBackwardRefs* VP8LGetBackwardReferences(
|
| + int width, int height, const uint32_t* const argb, int quality,
|
| + int low_effort, int* const cache_bits, VP8LHashChain* const hash_chain,
|
| + VP8LBackwardRefs refs_array[2]) {
|
| + if (low_effort) {
|
| + return GetBackwardReferencesLowEffort(width, height, argb, quality,
|
| + cache_bits, hash_chain, refs_array);
|
| + } else {
|
| + return GetBackwardReferences(width, height, argb, quality, cache_bits,
|
| + hash_chain, refs_array);
|
| + }
|
| +}
|
|
|