Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(1743)

Unified Diff: third_party/brotli/enc/bit_cost.h

Issue 1956893002: Added brotli enc/ and tools/ directories. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Updated to most recent build tools Created 4 years, 7 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « third_party/brotli/enc/backward_references.cc ('k') | third_party/brotli/enc/block_splitter.h » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: third_party/brotli/enc/bit_cost.h
diff --git a/third_party/brotli/enc/bit_cost.h b/third_party/brotli/enc/bit_cost.h
new file mode 100644
index 0000000000000000000000000000000000000000..4652006864c581c6589b92e50932251fca5a9124
--- /dev/null
+++ b/third_party/brotli/enc/bit_cost.h
@@ -0,0 +1,161 @@
+/* Copyright 2013 Google Inc. All Rights Reserved.
+
+ Distributed under MIT license.
+ See file LICENSE for detail or copy at https://opensource.org/licenses/MIT
+*/
+
+// Functions to estimate the bit cost of Huffman trees.
+
+#ifndef BROTLI_ENC_BIT_COST_H_
+#define BROTLI_ENC_BIT_COST_H_
+
+#include "./entropy_encode.h"
+#include "./fast_log.h"
+#include "./types.h"
+
+namespace brotli {
+
+static inline double ShannonEntropy(const uint32_t *population, size_t size,
+ size_t *total) {
+ size_t sum = 0;
+ double retval = 0;
+ const uint32_t *population_end = population + size;
+ size_t p;
+ if (size & 1) {
+ goto odd_number_of_elements_left;
+ }
+ while (population < population_end) {
+ p = *population++;
+ sum += p;
+ retval -= static_cast<double>(p) * FastLog2(p);
+ odd_number_of_elements_left:
+ p = *population++;
+ sum += p;
+ retval -= static_cast<double>(p) * FastLog2(p);
+ }
+ if (sum) retval += static_cast<double>(sum) * FastLog2(sum);
+ *total = sum;
+ return retval;
+}
+
+static inline double BitsEntropy(const uint32_t *population, size_t size) {
+ size_t sum;
+ double retval = ShannonEntropy(population, size, &sum);
+ if (retval < sum) {
+ // At least one bit per literal is needed.
+ retval = static_cast<double>(sum);
+ }
+ return retval;
+}
+
+template<int kSize>
+double PopulationCost(const Histogram<kSize>& histogram) {
+ static const double kOneSymbolHistogramCost = 12;
+ static const double kTwoSymbolHistogramCost = 20;
+ static const double kThreeSymbolHistogramCost = 28;
+ static const double kFourSymbolHistogramCost = 37;
+ if (histogram.total_count_ == 0) {
+ return kOneSymbolHistogramCost;
+ }
+ int count = 0;
+ int s[5];
+ for (int i = 0; i < kSize; ++i) {
+ if (histogram.data_[i] > 0) {
+ s[count] = i;
+ ++count;
+ if (count > 4) break;
+ }
+ }
+ if (count == 1) {
+ return kOneSymbolHistogramCost;
+ }
+ if (count == 2) {
+ return (kTwoSymbolHistogramCost +
+ static_cast<double>(histogram.total_count_));
+ }
+ if (count == 3) {
+ const uint32_t histo0 = histogram.data_[s[0]];
+ const uint32_t histo1 = histogram.data_[s[1]];
+ const uint32_t histo2 = histogram.data_[s[2]];
+ const uint32_t histomax = std::max(histo0, std::max(histo1, histo2));
+ return (kThreeSymbolHistogramCost +
+ 2 * (histo0 + histo1 + histo2) - histomax);
+ }
+ if (count == 4) {
+ uint32_t histo[4];
+ for (int i = 0; i < 4; ++i) {
+ histo[i] = histogram.data_[s[i]];
+ }
+ // Sort
+ for (int i = 0; i < 4; ++i) {
+ for (int j = i + 1; j < 4; ++j) {
+ if (histo[j] > histo[i]) {
+ std::swap(histo[j], histo[i]);
+ }
+ }
+ }
+ const uint32_t h23 = histo[2] + histo[3];
+ const uint32_t histomax = std::max(h23, histo[0]);
+ return (kFourSymbolHistogramCost +
+ 3 * h23 + 2 * (histo[0] + histo[1]) - histomax);
+ }
+
+ // In this loop we compute the entropy of the histogram and simultaneously
+ // build a simplified histogram of the code length codes where we use the
+ // zero repeat code 17, but we don't use the non-zero repeat code 16.
+ double bits = 0;
+ size_t max_depth = 1;
+ uint32_t depth_histo[kCodeLengthCodes] = { 0 };
+ const double log2total = FastLog2(histogram.total_count_);
+ for (size_t i = 0; i < kSize;) {
+ if (histogram.data_[i] > 0) {
+ // Compute -log2(P(symbol)) = -log2(count(symbol)/total_count) =
+ // = log2(total_count) - log2(count(symbol))
+ double log2p = log2total - FastLog2(histogram.data_[i]);
+ // Approximate the bit depth by round(-log2(P(symbol)))
+ size_t depth = static_cast<size_t>(log2p + 0.5);
+ bits += histogram.data_[i] * log2p;
+ if (depth > 15) {
+ depth = 15;
+ }
+ if (depth > max_depth) {
+ max_depth = depth;
+ }
+ ++depth_histo[depth];
+ ++i;
+ } else {
+ // Compute the run length of zeros and add the appropriate number of 0 and
+ // 17 code length codes to the code length code histogram.
+ uint32_t reps = 1;
+ for (size_t k = i + 1; k < kSize && histogram.data_[k] == 0; ++k) {
+ ++reps;
+ }
+ i += reps;
+ if (i == kSize) {
+ // Don't add any cost for the last zero run, since these are encoded
+ // only implicitly.
+ break;
+ }
+ if (reps < 3) {
+ depth_histo[0] += reps;
+ } else {
+ reps -= 2;
+ while (reps > 0) {
+ ++depth_histo[17];
+ // Add the 3 extra bits for the 17 code length code.
+ bits += 3;
+ reps >>= 3;
+ }
+ }
+ }
+ }
+ // Add the estimated encoding cost of the code length code histogram.
+ bits += static_cast<double>(18 + 2 * max_depth);
+ // Add the entropy of the code length code histogram.
+ bits += BitsEntropy(depth_histo, kCodeLengthCodes);
+ return bits;
+}
+
+} // namespace brotli
+
+#endif // BROTLI_ENC_BIT_COST_H_
« no previous file with comments | « third_party/brotli/enc/backward_references.cc ('k') | third_party/brotli/enc/block_splitter.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698