Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(215)

Side by Side Diff: third_party/brotli/enc/brotli_bit_stream.cc

Issue 2537133002: Update brotli to v1.0.0-snapshot. (Closed)
Patch Set: Fixed typo Created 4 years ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
« no previous file with comments | « third_party/brotli/enc/brotli_bit_stream.c ('k') | third_party/brotli/enc/cluster.h » ('j') | no next file with comments »
Toggle Intra-line Diffs ('i') | Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
OLDNEW
(Empty)
1 /* Copyright 2014 Google Inc. All Rights Reserved.
2
3 Distributed under MIT license.
4 See file LICENSE for detail or copy at https://opensource.org/licenses/MIT
5 */
6
7 // Brotli bit stream functions to support the low level format. There are no
8 // compression algorithms here, just the right ordering of bits to match the
9 // specs.
10
11 #include "./brotli_bit_stream.h"
12
13 #include <algorithm>
14 #include <cstdlib> /* free, malloc */
15 #include <cstring>
16 #include <limits>
17 #include <vector>
18
19 #include "./bit_cost.h"
20 #include "./context.h"
21 #include "./entropy_encode.h"
22 #include "./entropy_encode_static.h"
23 #include "./fast_log.h"
24 #include "./prefix.h"
25 #include "./write_bits.h"
26
27 namespace brotli {
28
29 namespace {
30
31 static const size_t kMaxHuffmanTreeSize = 2 * kNumCommandPrefixes + 1;
32 // Context map alphabet has 256 context id symbols plus max 16 rle symbols.
33 static const size_t kContextMapAlphabetSize = 256 + 16;
34 // Block type alphabet has 256 block id symbols plus 2 special symbols.
35 static const size_t kBlockTypeAlphabetSize = 256 + 2;
36
37 // nibblesbits represents the 2 bits to encode MNIBBLES (0-3)
38 // REQUIRES: length > 0
39 // REQUIRES: length <= (1 << 24)
40 void EncodeMlen(size_t length, uint64_t* bits,
41 size_t* numbits, uint64_t* nibblesbits) {
42 assert(length > 0);
43 assert(length <= (1 << 24));
44 length--; // MLEN - 1 is encoded
45 size_t lg = length == 0 ? 1 : Log2FloorNonZero(
46 static_cast<uint32_t>(length)) + 1;
47 assert(lg <= 24);
48 size_t mnibbles = (lg < 16 ? 16 : (lg + 3)) / 4;
49 *nibblesbits = mnibbles - 4;
50 *numbits = mnibbles * 4;
51 *bits = length;
52 }
53
54 static inline void StoreCommandExtra(
55 const Command& cmd, size_t* storage_ix, uint8_t* storage) {
56 uint32_t copylen_code = cmd.copy_len_code();
57 uint16_t inscode = GetInsertLengthCode(cmd.insert_len_);
58 uint16_t copycode = GetCopyLengthCode(copylen_code);
59 uint32_t insnumextra = GetInsertExtra(inscode);
60 uint64_t insextraval = cmd.insert_len_ - GetInsertBase(inscode);
61 uint64_t copyextraval = copylen_code - GetCopyBase(copycode);
62 uint64_t bits = (copyextraval << insnumextra) | insextraval;
63 WriteBits(insnumextra + GetCopyExtra(copycode), bits, storage_ix, storage);
64 }
65
66 } // namespace
67
68 void StoreVarLenUint8(size_t n, size_t* storage_ix, uint8_t* storage) {
69 if (n == 0) {
70 WriteBits(1, 0, storage_ix, storage);
71 } else {
72 WriteBits(1, 1, storage_ix, storage);
73 size_t nbits = Log2FloorNonZero(n);
74 WriteBits(3, nbits, storage_ix, storage);
75 WriteBits(nbits, n - (1 << nbits), storage_ix, storage);
76 }
77 }
78
79 void StoreCompressedMetaBlockHeader(bool final_block,
80 size_t length,
81 size_t* storage_ix,
82 uint8_t* storage) {
83 // Write ISLAST bit.
84 WriteBits(1, final_block, storage_ix, storage);
85 // Write ISEMPTY bit.
86 if (final_block) {
87 WriteBits(1, 0, storage_ix, storage);
88 }
89
90 uint64_t lenbits;
91 size_t nlenbits;
92 uint64_t nibblesbits;
93 EncodeMlen(length, &lenbits, &nlenbits, &nibblesbits);
94 WriteBits(2, nibblesbits, storage_ix, storage);
95 WriteBits(nlenbits, lenbits, storage_ix, storage);
96
97 if (!final_block) {
98 // Write ISUNCOMPRESSED bit.
99 WriteBits(1, 0, storage_ix, storage);
100 }
101 }
102
103 void StoreUncompressedMetaBlockHeader(size_t length,
104 size_t* storage_ix,
105 uint8_t* storage) {
106 // Write ISLAST bit. Uncompressed block cannot be the last one, so set to 0.
107 WriteBits(1, 0, storage_ix, storage);
108 uint64_t lenbits;
109 size_t nlenbits;
110 uint64_t nibblesbits;
111 EncodeMlen(length, &lenbits, &nlenbits, &nibblesbits);
112 WriteBits(2, nibblesbits, storage_ix, storage);
113 WriteBits(nlenbits, lenbits, storage_ix, storage);
114 // Write ISUNCOMPRESSED bit.
115 WriteBits(1, 1, storage_ix, storage);
116 }
117
118 void StoreHuffmanTreeOfHuffmanTreeToBitMask(
119 const int num_codes,
120 const uint8_t *code_length_bitdepth,
121 size_t *storage_ix,
122 uint8_t *storage) {
123 static const uint8_t kStorageOrder[kCodeLengthCodes] = {
124 1, 2, 3, 4, 0, 5, 17, 6, 16, 7, 8, 9, 10, 11, 12, 13, 14, 15
125 };
126 // The bit lengths of the Huffman code over the code length alphabet
127 // are compressed with the following static Huffman code:
128 // Symbol Code
129 // ------ ----
130 // 0 00
131 // 1 1110
132 // 2 110
133 // 3 01
134 // 4 10
135 // 5 1111
136 static const uint8_t kHuffmanBitLengthHuffmanCodeSymbols[6] = {
137 0, 7, 3, 2, 1, 15
138 };
139 static const uint8_t kHuffmanBitLengthHuffmanCodeBitLengths[6] = {
140 2, 4, 3, 2, 2, 4
141 };
142
143 // Throw away trailing zeros:
144 size_t codes_to_store = kCodeLengthCodes;
145 if (num_codes > 1) {
146 for (; codes_to_store > 0; --codes_to_store) {
147 if (code_length_bitdepth[kStorageOrder[codes_to_store - 1]] != 0) {
148 break;
149 }
150 }
151 }
152 size_t skip_some = 0; // skips none.
153 if (code_length_bitdepth[kStorageOrder[0]] == 0 &&
154 code_length_bitdepth[kStorageOrder[1]] == 0) {
155 skip_some = 2; // skips two.
156 if (code_length_bitdepth[kStorageOrder[2]] == 0) {
157 skip_some = 3; // skips three.
158 }
159 }
160 WriteBits(2, skip_some, storage_ix, storage);
161 for (size_t i = skip_some; i < codes_to_store; ++i) {
162 size_t l = code_length_bitdepth[kStorageOrder[i]];
163 WriteBits(kHuffmanBitLengthHuffmanCodeBitLengths[l],
164 kHuffmanBitLengthHuffmanCodeSymbols[l], storage_ix, storage);
165 }
166 }
167
168 static void StoreHuffmanTreeToBitMask(
169 const size_t huffman_tree_size,
170 const uint8_t* huffman_tree,
171 const uint8_t* huffman_tree_extra_bits,
172 const uint8_t* code_length_bitdepth,
173 const uint16_t* code_length_bitdepth_symbols,
174 size_t * __restrict storage_ix,
175 uint8_t * __restrict storage) {
176 for (size_t i = 0; i < huffman_tree_size; ++i) {
177 size_t ix = huffman_tree[i];
178 WriteBits(code_length_bitdepth[ix], code_length_bitdepth_symbols[ix],
179 storage_ix, storage);
180 // Extra bits
181 switch (ix) {
182 case 16:
183 WriteBits(2, huffman_tree_extra_bits[i], storage_ix, storage);
184 break;
185 case 17:
186 WriteBits(3, huffman_tree_extra_bits[i], storage_ix, storage);
187 break;
188 }
189 }
190 }
191
192 static void StoreSimpleHuffmanTree(const uint8_t* depths,
193 size_t symbols[4],
194 size_t num_symbols,
195 size_t max_bits,
196 size_t *storage_ix, uint8_t *storage) {
197 // value of 1 indicates a simple Huffman code
198 WriteBits(2, 1, storage_ix, storage);
199 WriteBits(2, num_symbols - 1, storage_ix, storage); // NSYM - 1
200
201 // Sort
202 for (size_t i = 0; i < num_symbols; i++) {
203 for (size_t j = i + 1; j < num_symbols; j++) {
204 if (depths[symbols[j]] < depths[symbols[i]]) {
205 std::swap(symbols[j], symbols[i]);
206 }
207 }
208 }
209
210 if (num_symbols == 2) {
211 WriteBits(max_bits, symbols[0], storage_ix, storage);
212 WriteBits(max_bits, symbols[1], storage_ix, storage);
213 } else if (num_symbols == 3) {
214 WriteBits(max_bits, symbols[0], storage_ix, storage);
215 WriteBits(max_bits, symbols[1], storage_ix, storage);
216 WriteBits(max_bits, symbols[2], storage_ix, storage);
217 } else {
218 WriteBits(max_bits, symbols[0], storage_ix, storage);
219 WriteBits(max_bits, symbols[1], storage_ix, storage);
220 WriteBits(max_bits, symbols[2], storage_ix, storage);
221 WriteBits(max_bits, symbols[3], storage_ix, storage);
222 // tree-select
223 WriteBits(1, depths[symbols[0]] == 1 ? 1 : 0, storage_ix, storage);
224 }
225 }
226
227 // num = alphabet size
228 // depths = symbol depths
229 void StoreHuffmanTree(const uint8_t* depths, size_t num,
230 HuffmanTree* tree,
231 size_t *storage_ix, uint8_t *storage) {
232 // Write the Huffman tree into the brotli-representation.
233 // The command alphabet is the largest, so this allocation will fit all
234 // alphabets.
235 assert(num <= kNumCommandPrefixes);
236 uint8_t huffman_tree[kNumCommandPrefixes];
237 uint8_t huffman_tree_extra_bits[kNumCommandPrefixes];
238 size_t huffman_tree_size = 0;
239 WriteHuffmanTree(depths, num, &huffman_tree_size, huffman_tree,
240 huffman_tree_extra_bits);
241
242 // Calculate the statistics of the Huffman tree in brotli-representation.
243 uint32_t huffman_tree_histogram[kCodeLengthCodes] = { 0 };
244 for (size_t i = 0; i < huffman_tree_size; ++i) {
245 ++huffman_tree_histogram[huffman_tree[i]];
246 }
247
248 int num_codes = 0;
249 int code = 0;
250 for (int i = 0; i < kCodeLengthCodes; ++i) {
251 if (huffman_tree_histogram[i]) {
252 if (num_codes == 0) {
253 code = i;
254 num_codes = 1;
255 } else if (num_codes == 1) {
256 num_codes = 2;
257 break;
258 }
259 }
260 }
261
262 // Calculate another Huffman tree to use for compressing both the
263 // earlier Huffman tree with.
264 uint8_t code_length_bitdepth[kCodeLengthCodes] = { 0 };
265 uint16_t code_length_bitdepth_symbols[kCodeLengthCodes] = { 0 };
266 CreateHuffmanTree(&huffman_tree_histogram[0], kCodeLengthCodes,
267 5, tree, &code_length_bitdepth[0]);
268 ConvertBitDepthsToSymbols(code_length_bitdepth, kCodeLengthCodes,
269 &code_length_bitdepth_symbols[0]);
270
271 // Now, we have all the data, let's start storing it
272 StoreHuffmanTreeOfHuffmanTreeToBitMask(num_codes, code_length_bitdepth,
273 storage_ix, storage);
274
275 if (num_codes == 1) {
276 code_length_bitdepth[code] = 0;
277 }
278
279 // Store the real huffman tree now.
280 StoreHuffmanTreeToBitMask(huffman_tree_size,
281 huffman_tree,
282 huffman_tree_extra_bits,
283 &code_length_bitdepth[0],
284 code_length_bitdepth_symbols,
285 storage_ix, storage);
286 }
287
288 void BuildAndStoreHuffmanTree(const uint32_t *histogram,
289 const size_t length,
290 HuffmanTree* tree,
291 uint8_t* depth,
292 uint16_t* bits,
293 size_t* storage_ix,
294 uint8_t* storage) {
295 size_t count = 0;
296 size_t s4[4] = { 0 };
297 for (size_t i = 0; i < length; i++) {
298 if (histogram[i]) {
299 if (count < 4) {
300 s4[count] = i;
301 } else if (count > 4) {
302 break;
303 }
304 count++;
305 }
306 }
307
308 size_t max_bits_counter = length - 1;
309 size_t max_bits = 0;
310 while (max_bits_counter) {
311 max_bits_counter >>= 1;
312 ++max_bits;
313 }
314
315 if (count <= 1) {
316 WriteBits(4, 1, storage_ix, storage);
317 WriteBits(max_bits, s4[0], storage_ix, storage);
318 return;
319 }
320
321 CreateHuffmanTree(histogram, length, 15, tree, depth);
322 ConvertBitDepthsToSymbols(depth, length, bits);
323
324 if (count <= 4) {
325 StoreSimpleHuffmanTree(depth, s4, count, max_bits, storage_ix, storage);
326 } else {
327 StoreHuffmanTree(depth, length, tree, storage_ix, storage);
328 }
329 }
330
331 static inline bool SortHuffmanTree(const HuffmanTree& v0,
332 const HuffmanTree& v1) {
333 return v0.total_count_ < v1.total_count_;
334 }
335
336 void BuildAndStoreHuffmanTreeFast(const uint32_t *histogram,
337 const size_t histogram_total,
338 const size_t max_bits,
339 uint8_t* depth,
340 uint16_t* bits,
341 size_t* storage_ix,
342 uint8_t* storage) {
343 size_t count = 0;
344 size_t symbols[4] = { 0 };
345 size_t length = 0;
346 size_t total = histogram_total;
347 while (total != 0) {
348 if (histogram[length]) {
349 if (count < 4) {
350 symbols[count] = length;
351 }
352 ++count;
353 total -= histogram[length];
354 }
355 ++length;
356 }
357
358 if (count <= 1) {
359 WriteBits(4, 1, storage_ix, storage);
360 WriteBits(max_bits, symbols[0], storage_ix, storage);
361 return;
362 }
363
364 const size_t max_tree_size = 2 * length + 1;
365 HuffmanTree* const tree =
366 static_cast<HuffmanTree*>(malloc(max_tree_size * sizeof(HuffmanTree)));
367 for (uint32_t count_limit = 1; ; count_limit *= 2) {
368 HuffmanTree* node = tree;
369 for (size_t i = length; i != 0;) {
370 --i;
371 if (histogram[i]) {
372 if (PREDICT_TRUE(histogram[i] >= count_limit)) {
373 *node = HuffmanTree(histogram[i], -1, static_cast<int16_t>(i));
374 } else {
375 *node = HuffmanTree(count_limit, -1, static_cast<int16_t>(i));
376 }
377 ++node;
378 }
379 }
380 const int n = static_cast<int>(node - tree);
381 std::sort(tree, node, SortHuffmanTree);
382 // The nodes are:
383 // [0, n): the sorted leaf nodes that we start with.
384 // [n]: we add a sentinel here.
385 // [n + 1, 2n): new parent nodes are added here, starting from
386 // (n+1). These are naturally in ascending order.
387 // [2n]: we add a sentinel at the end as well.
388 // There will be (2n+1) elements at the end.
389 const HuffmanTree sentinel(std::numeric_limits<int>::max(), -1, -1);
390 *node++ = sentinel;
391 *node++ = sentinel;
392
393 int i = 0; // Points to the next leaf node.
394 int j = n + 1; // Points to the next non-leaf node.
395 for (int k = n - 1; k > 0; --k) {
396 int left, right;
397 if (tree[i].total_count_ <= tree[j].total_count_) {
398 left = i;
399 ++i;
400 } else {
401 left = j;
402 ++j;
403 }
404 if (tree[i].total_count_ <= tree[j].total_count_) {
405 right = i;
406 ++i;
407 } else {
408 right = j;
409 ++j;
410 }
411 // The sentinel node becomes the parent node.
412 node[-1].total_count_ =
413 tree[left].total_count_ + tree[right].total_count_;
414 node[-1].index_left_ = static_cast<int16_t>(left);
415 node[-1].index_right_or_value_ = static_cast<int16_t>(right);
416 // Add back the last sentinel node.
417 *node++ = sentinel;
418 }
419 SetDepth(tree[2 * n - 1], &tree[0], depth, 0);
420 // We need to pack the Huffman tree in 14 bits.
421 // If this was not successful, add fake entities to the lowest values
422 // and retry.
423 if (PREDICT_TRUE(*std::max_element(&depth[0], &depth[length]) <= 14)) {
424 break;
425 }
426 }
427 free(tree);
428 ConvertBitDepthsToSymbols(depth, length, bits);
429 if (count <= 4) {
430 // value of 1 indicates a simple Huffman code
431 WriteBits(2, 1, storage_ix, storage);
432 WriteBits(2, count - 1, storage_ix, storage); // NSYM - 1
433
434 // Sort
435 for (size_t i = 0; i < count; i++) {
436 for (size_t j = i + 1; j < count; j++) {
437 if (depth[symbols[j]] < depth[symbols[i]]) {
438 std::swap(symbols[j], symbols[i]);
439 }
440 }
441 }
442
443 if (count == 2) {
444 WriteBits(max_bits, symbols[0], storage_ix, storage);
445 WriteBits(max_bits, symbols[1], storage_ix, storage);
446 } else if (count == 3) {
447 WriteBits(max_bits, symbols[0], storage_ix, storage);
448 WriteBits(max_bits, symbols[1], storage_ix, storage);
449 WriteBits(max_bits, symbols[2], storage_ix, storage);
450 } else {
451 WriteBits(max_bits, symbols[0], storage_ix, storage);
452 WriteBits(max_bits, symbols[1], storage_ix, storage);
453 WriteBits(max_bits, symbols[2], storage_ix, storage);
454 WriteBits(max_bits, symbols[3], storage_ix, storage);
455 // tree-select
456 WriteBits(1, depth[symbols[0]] == 1 ? 1 : 0, storage_ix, storage);
457 }
458 } else {
459 // Complex Huffman Tree
460 StoreStaticCodeLengthCode(storage_ix, storage);
461
462 // Actual rle coding.
463 uint8_t previous_value = 8;
464 for (size_t i = 0; i < length;) {
465 const uint8_t value = depth[i];
466 size_t reps = 1;
467 for (size_t k = i + 1; k < length && depth[k] == value; ++k) {
468 ++reps;
469 }
470 i += reps;
471 if (value == 0) {
472 WriteBits(kZeroRepsDepth[reps], kZeroRepsBits[reps],
473 storage_ix, storage);
474 } else {
475 if (previous_value != value) {
476 WriteBits(kCodeLengthDepth[value], kCodeLengthBits[value],
477 storage_ix, storage);
478 --reps;
479 }
480 if (reps < 3) {
481 while (reps != 0) {
482 reps--;
483 WriteBits(kCodeLengthDepth[value], kCodeLengthBits[value],
484 storage_ix, storage);
485 }
486 } else {
487 reps -= 3;
488 WriteBits(kNonZeroRepsDepth[reps], kNonZeroRepsBits[reps],
489 storage_ix, storage);
490 }
491 previous_value = value;
492 }
493 }
494 }
495 }
496
497 static size_t IndexOf(const uint8_t* v, size_t v_size, uint8_t value) {
498 size_t i = 0;
499 for (; i < v_size; ++i) {
500 if (v[i] == value) return i;
501 }
502 return i;
503 }
504
505 static void MoveToFront(uint8_t* v, size_t index) {
506 uint8_t value = v[index];
507 for (size_t i = index; i != 0; --i) {
508 v[i] = v[i - 1];
509 }
510 v[0] = value;
511 }
512
513 static void MoveToFrontTransform(const uint32_t* __restrict v_in,
514 const size_t v_size,
515 uint32_t* v_out) {
516 if (v_size == 0) {
517 return;
518 }
519 uint32_t max_value = *std::max_element(v_in, v_in + v_size);
520 assert(max_value < 256u);
521 uint8_t mtf[256];
522 size_t mtf_size = max_value + 1;
523 for (uint32_t i = 0; i <= max_value; ++i) {
524 mtf[i] = static_cast<uint8_t>(i);
525 }
526 for (size_t i = 0; i < v_size; ++i) {
527 size_t index = IndexOf(mtf, mtf_size, static_cast<uint8_t>(v_in[i]));
528 assert(index < mtf_size);
529 v_out[i] = static_cast<uint32_t>(index);
530 MoveToFront(mtf, index);
531 }
532 }
533
534 // Finds runs of zeros in v[0..in_size) and replaces them with a prefix code of
535 // the run length plus extra bits (lower 9 bits is the prefix code and the rest
536 // are the extra bits). Non-zero values in v[] are shifted by
537 // *max_length_prefix. Will not create prefix codes bigger than the initial
538 // value of *max_run_length_prefix. The prefix code of run length L is simply
539 // Log2Floor(L) and the number of extra bits is the same as the prefix code.
540 static void RunLengthCodeZeros(const size_t in_size,
541 uint32_t* __restrict v,
542 size_t* __restrict out_size,
543 uint32_t* __restrict max_run_length_prefix) {
544 uint32_t max_reps = 0;
545 for (size_t i = 0; i < in_size;) {
546 for (; i < in_size && v[i] != 0; ++i) ;
547 uint32_t reps = 0;
548 for (; i < in_size && v[i] == 0; ++i) {
549 ++reps;
550 }
551 max_reps = std::max(reps, max_reps);
552 }
553 uint32_t max_prefix = max_reps > 0 ? Log2FloorNonZero(max_reps) : 0;
554 max_prefix = std::min(max_prefix, *max_run_length_prefix);
555 *max_run_length_prefix = max_prefix;
556 *out_size = 0;
557 for (size_t i = 0; i < in_size;) {
558 assert(*out_size <= i);
559 if (v[i] != 0) {
560 v[*out_size] = v[i] + *max_run_length_prefix;
561 ++i;
562 ++(*out_size);
563 } else {
564 uint32_t reps = 1;
565 for (size_t k = i + 1; k < in_size && v[k] == 0; ++k) {
566 ++reps;
567 }
568 i += reps;
569 while (reps != 0) {
570 if (reps < (2u << max_prefix)) {
571 uint32_t run_length_prefix = Log2FloorNonZero(reps);
572 const uint32_t extra_bits = reps - (1u << run_length_prefix);
573 v[*out_size] = run_length_prefix + (extra_bits << 9);
574 ++(*out_size);
575 break;
576 } else {
577 const uint32_t extra_bits = (1u << max_prefix) - 1u;
578 v[*out_size] = max_prefix + (extra_bits << 9);
579 reps -= (2u << max_prefix) - 1u;
580 ++(*out_size);
581 }
582 }
583 }
584 }
585 }
586
587 void EncodeContextMap(const std::vector<uint32_t>& context_map,
588 size_t num_clusters,
589 HuffmanTree* tree,
590 size_t* storage_ix, uint8_t* storage) {
591 StoreVarLenUint8(num_clusters - 1, storage_ix, storage);
592
593 if (num_clusters == 1) {
594 return;
595 }
596
597 uint32_t* rle_symbols = new uint32_t[context_map.size()];
598 MoveToFrontTransform(&context_map[0], context_map.size(), rle_symbols);
599 uint32_t max_run_length_prefix = 6;
600 size_t num_rle_symbols = 0;
601 RunLengthCodeZeros(context_map.size(), rle_symbols,
602 &num_rle_symbols, &max_run_length_prefix);
603 uint32_t histogram[kContextMapAlphabetSize];
604 memset(histogram, 0, sizeof(histogram));
605 static const int kSymbolBits = 9;
606 static const uint32_t kSymbolMask = (1u << kSymbolBits) - 1u;
607 for (size_t i = 0; i < num_rle_symbols; ++i) {
608 ++histogram[rle_symbols[i] & kSymbolMask];
609 }
610 bool use_rle = max_run_length_prefix > 0;
611 WriteBits(1, use_rle, storage_ix, storage);
612 if (use_rle) {
613 WriteBits(4, max_run_length_prefix - 1, storage_ix, storage);
614 }
615 uint8_t depths[kContextMapAlphabetSize];
616 uint16_t bits[kContextMapAlphabetSize];
617 memset(depths, 0, sizeof(depths));
618 memset(bits, 0, sizeof(bits));
619 BuildAndStoreHuffmanTree(histogram, num_clusters + max_run_length_prefix,
620 tree, depths, bits, storage_ix, storage);
621 for (size_t i = 0; i < num_rle_symbols; ++i) {
622 const uint32_t rle_symbol = rle_symbols[i] & kSymbolMask;
623 const uint32_t extra_bits_val = rle_symbols[i] >> kSymbolBits;
624 WriteBits(depths[rle_symbol], bits[rle_symbol], storage_ix, storage);
625 if (rle_symbol > 0 && rle_symbol <= max_run_length_prefix) {
626 WriteBits(rle_symbol, extra_bits_val, storage_ix, storage);
627 }
628 }
629 WriteBits(1, 1, storage_ix, storage); // use move-to-front
630 delete[] rle_symbols;
631 }
632
633 void StoreBlockSwitch(const BlockSplitCode& code,
634 const size_t block_ix,
635 size_t* storage_ix,
636 uint8_t* storage) {
637 if (block_ix > 0) {
638 size_t typecode = code.type_code[block_ix];
639 WriteBits(code.type_depths[typecode], code.type_bits[typecode],
640 storage_ix, storage);
641 }
642 size_t lencode = code.length_prefix[block_ix];
643 WriteBits(code.length_depths[lencode], code.length_bits[lencode],
644 storage_ix, storage);
645 WriteBits(code.length_nextra[block_ix], code.length_extra[block_ix],
646 storage_ix, storage);
647 }
648
649 static void BuildAndStoreBlockSplitCode(const std::vector<uint8_t>& types,
650 const std::vector<uint32_t>& lengths,
651 const size_t num_types,
652 HuffmanTree* tree,
653 BlockSplitCode* code,
654 size_t* storage_ix,
655 uint8_t* storage) {
656 const size_t num_blocks = types.size();
657 uint32_t type_histo[kBlockTypeAlphabetSize];
658 uint32_t length_histo[kNumBlockLenPrefixes];
659 memset(type_histo, 0, (num_types + 2) * sizeof(type_histo[0]));
660 memset(length_histo, 0, sizeof(length_histo));
661 size_t last_type = 1;
662 size_t second_last_type = 0;
663 code->type_code.resize(num_blocks);
664 code->length_prefix.resize(num_blocks);
665 code->length_nextra.resize(num_blocks);
666 code->length_extra.resize(num_blocks);
667 code->type_depths.resize(num_types + 2);
668 code->type_bits.resize(num_types + 2);
669 memset(code->length_depths, 0, sizeof(code->length_depths));
670 memset(code->length_bits, 0, sizeof(code->length_bits));
671 for (size_t i = 0; i < num_blocks; ++i) {
672 size_t type = types[i];
673 size_t type_code = (type == last_type + 1 ? 1 :
674 type == second_last_type ? 0 :
675 type + 2);
676 second_last_type = last_type;
677 last_type = type;
678 code->type_code[i] = static_cast<uint32_t>(type_code);
679 if (i != 0) ++type_histo[type_code];
680 GetBlockLengthPrefixCode(lengths[i],
681 &code->length_prefix[i],
682 &code->length_nextra[i],
683 &code->length_extra[i]);
684 ++length_histo[code->length_prefix[i]];
685 }
686 StoreVarLenUint8(num_types - 1, storage_ix, storage);
687 if (num_types > 1) {
688 BuildAndStoreHuffmanTree(&type_histo[0], num_types + 2, tree,
689 &code->type_depths[0], &code->type_bits[0],
690 storage_ix, storage);
691 BuildAndStoreHuffmanTree(&length_histo[0], kNumBlockLenPrefixes, tree,
692 &code->length_depths[0], &code->length_bits[0],
693 storage_ix, storage);
694 StoreBlockSwitch(*code, 0, storage_ix, storage);
695 }
696 }
697
698 void StoreTrivialContextMap(size_t num_types,
699 size_t context_bits,
700 HuffmanTree* tree,
701 size_t* storage_ix,
702 uint8_t* storage) {
703 StoreVarLenUint8(num_types - 1, storage_ix, storage);
704 if (num_types > 1) {
705 size_t repeat_code = context_bits - 1u;
706 size_t repeat_bits = (1u << repeat_code) - 1u;
707 size_t alphabet_size = num_types + repeat_code;
708 uint32_t histogram[kContextMapAlphabetSize];
709 uint8_t depths[kContextMapAlphabetSize];
710 uint16_t bits[kContextMapAlphabetSize];
711 memset(histogram, 0, alphabet_size * sizeof(histogram[0]));
712 memset(depths, 0, alphabet_size * sizeof(depths[0]));
713 memset(bits, 0, alphabet_size * sizeof(bits[0]));
714 // Write RLEMAX.
715 WriteBits(1, 1, storage_ix, storage);
716 WriteBits(4, repeat_code - 1, storage_ix, storage);
717 histogram[repeat_code] = static_cast<uint32_t>(num_types);
718 histogram[0] = 1;
719 for (size_t i = context_bits; i < alphabet_size; ++i) {
720 histogram[i] = 1;
721 }
722 BuildAndStoreHuffmanTree(&histogram[0], alphabet_size, tree,
723 &depths[0], &bits[0],
724 storage_ix, storage);
725 for (size_t i = 0; i < num_types; ++i) {
726 size_t code = (i == 0 ? 0 : i + context_bits - 1);
727 WriteBits(depths[code], bits[code], storage_ix, storage);
728 WriteBits(depths[repeat_code], bits[repeat_code], storage_ix, storage);
729 WriteBits(repeat_code, repeat_bits, storage_ix, storage);
730 }
731 // Write IMTF (inverse-move-to-front) bit.
732 WriteBits(1, 1, storage_ix, storage);
733 }
734 }
735
736 // Manages the encoding of one block category (literal, command or distance).
737 class BlockEncoder {
738 public:
739 BlockEncoder(size_t alphabet_size,
740 size_t num_block_types,
741 const std::vector<uint8_t>& block_types,
742 const std::vector<uint32_t>& block_lengths)
743 : alphabet_size_(alphabet_size),
744 num_block_types_(num_block_types),
745 block_types_(block_types),
746 block_lengths_(block_lengths),
747 block_ix_(0),
748 block_len_(block_lengths.empty() ? 0 : block_lengths[0]),
749 entropy_ix_(0) {}
750
751 // Creates entropy codes of block lengths and block types and stores them
752 // to the bit stream.
753 void BuildAndStoreBlockSwitchEntropyCodes(HuffmanTree* tree,
754 size_t* storage_ix,
755 uint8_t* storage) {
756 BuildAndStoreBlockSplitCode(
757 block_types_, block_lengths_, num_block_types_,
758 tree, &block_split_code_, storage_ix, storage);
759 }
760
761 // Creates entropy codes for all block types and stores them to the bit
762 // stream.
763 template<int kSize>
764 void BuildAndStoreEntropyCodes(
765 const std::vector<Histogram<kSize> >& histograms,
766 HuffmanTree* tree,
767 size_t* storage_ix, uint8_t* storage) {
768 depths_.resize(histograms.size() * alphabet_size_);
769 bits_.resize(histograms.size() * alphabet_size_);
770 for (size_t i = 0; i < histograms.size(); ++i) {
771 size_t ix = i * alphabet_size_;
772 BuildAndStoreHuffmanTree(&histograms[i].data_[0], alphabet_size_,
773 tree,
774 &depths_[ix], &bits_[ix],
775 storage_ix, storage);
776 }
777 }
778
779 // Stores the next symbol with the entropy code of the current block type.
780 // Updates the block type and block length at block boundaries.
781 void StoreSymbol(size_t symbol, size_t* storage_ix, uint8_t* storage) {
782 if (block_len_ == 0) {
783 ++block_ix_;
784 block_len_ = block_lengths_[block_ix_];
785 entropy_ix_ = block_types_[block_ix_] * alphabet_size_;
786 StoreBlockSwitch(block_split_code_, block_ix_, storage_ix, storage);
787 }
788 --block_len_;
789 size_t ix = entropy_ix_ + symbol;
790 WriteBits(depths_[ix], bits_[ix], storage_ix, storage);
791 }
792
793 // Stores the next symbol with the entropy code of the current block type and
794 // context value.
795 // Updates the block type and block length at block boundaries.
796 template<int kContextBits>
797 void StoreSymbolWithContext(size_t symbol, size_t context,
798 const std::vector<uint32_t>& context_map,
799 size_t* storage_ix, uint8_t* storage) {
800 if (block_len_ == 0) {
801 ++block_ix_;
802 block_len_ = block_lengths_[block_ix_];
803 size_t block_type = block_types_[block_ix_];
804 entropy_ix_ = block_type << kContextBits;
805 StoreBlockSwitch(block_split_code_, block_ix_, storage_ix, storage);
806 }
807 --block_len_;
808 size_t histo_ix = context_map[entropy_ix_ + context];
809 size_t ix = histo_ix * alphabet_size_ + symbol;
810 WriteBits(depths_[ix], bits_[ix], storage_ix, storage);
811 }
812
813 private:
814 const size_t alphabet_size_;
815 const size_t num_block_types_;
816 const std::vector<uint8_t>& block_types_;
817 const std::vector<uint32_t>& block_lengths_;
818 BlockSplitCode block_split_code_;
819 size_t block_ix_;
820 size_t block_len_;
821 size_t entropy_ix_;
822 std::vector<uint8_t> depths_;
823 std::vector<uint16_t> bits_;
824 };
825
826 static void JumpToByteBoundary(size_t* storage_ix, uint8_t* storage) {
827 *storage_ix = (*storage_ix + 7u) & ~7u;
828 storage[*storage_ix >> 3] = 0;
829 }
830
831 void StoreMetaBlock(const uint8_t* input,
832 size_t start_pos,
833 size_t length,
834 size_t mask,
835 uint8_t prev_byte,
836 uint8_t prev_byte2,
837 bool is_last,
838 uint32_t num_direct_distance_codes,
839 uint32_t distance_postfix_bits,
840 ContextType literal_context_mode,
841 const brotli::Command *commands,
842 size_t n_commands,
843 const MetaBlockSplit& mb,
844 size_t *storage_ix,
845 uint8_t *storage) {
846 StoreCompressedMetaBlockHeader(is_last, length, storage_ix, storage);
847
848 size_t num_distance_codes =
849 kNumDistanceShortCodes + num_direct_distance_codes +
850 (48u << distance_postfix_bits);
851
852 HuffmanTree* tree = static_cast<HuffmanTree*>(
853 malloc(kMaxHuffmanTreeSize * sizeof(HuffmanTree)));
854 BlockEncoder literal_enc(256,
855 mb.literal_split.num_types,
856 mb.literal_split.types,
857 mb.literal_split.lengths);
858 BlockEncoder command_enc(kNumCommandPrefixes,
859 mb.command_split.num_types,
860 mb.command_split.types,
861 mb.command_split.lengths);
862 BlockEncoder distance_enc(num_distance_codes,
863 mb.distance_split.num_types,
864 mb.distance_split.types,
865 mb.distance_split.lengths);
866
867 literal_enc.BuildAndStoreBlockSwitchEntropyCodes(tree, storage_ix, storage);
868 command_enc.BuildAndStoreBlockSwitchEntropyCodes(tree, storage_ix, storage);
869 distance_enc.BuildAndStoreBlockSwitchEntropyCodes(tree, storage_ix, storage);
870
871 WriteBits(2, distance_postfix_bits, storage_ix, storage);
872 WriteBits(4, num_direct_distance_codes >> distance_postfix_bits,
873 storage_ix, storage);
874 for (size_t i = 0; i < mb.literal_split.num_types; ++i) {
875 WriteBits(2, literal_context_mode, storage_ix, storage);
876 }
877
878 size_t num_literal_histograms = mb.literal_histograms.size();
879 if (mb.literal_context_map.empty()) {
880 StoreTrivialContextMap(num_literal_histograms, kLiteralContextBits, tree,
881 storage_ix, storage);
882 } else {
883 EncodeContextMap(mb.literal_context_map, num_literal_histograms, tree,
884 storage_ix, storage);
885 }
886
887 size_t num_dist_histograms = mb.distance_histograms.size();
888 if (mb.distance_context_map.empty()) {
889 StoreTrivialContextMap(num_dist_histograms, kDistanceContextBits, tree,
890 storage_ix, storage);
891 } else {
892 EncodeContextMap(mb.distance_context_map, num_dist_histograms, tree,
893 storage_ix, storage);
894 }
895
896 literal_enc.BuildAndStoreEntropyCodes(mb.literal_histograms, tree,
897 storage_ix, storage);
898 command_enc.BuildAndStoreEntropyCodes(mb.command_histograms, tree,
899 storage_ix, storage);
900 distance_enc.BuildAndStoreEntropyCodes(mb.distance_histograms, tree,
901 storage_ix, storage);
902 free(tree);
903
904 size_t pos = start_pos;
905 for (size_t i = 0; i < n_commands; ++i) {
906 const Command cmd = commands[i];
907 size_t cmd_code = cmd.cmd_prefix_;
908 command_enc.StoreSymbol(cmd_code, storage_ix, storage);
909 StoreCommandExtra(cmd, storage_ix, storage);
910 if (mb.literal_context_map.empty()) {
911 for (size_t j = cmd.insert_len_; j != 0; --j) {
912 literal_enc.StoreSymbol(input[pos & mask], storage_ix, storage);
913 ++pos;
914 }
915 } else {
916 for (size_t j = cmd.insert_len_; j != 0; --j) {
917 size_t context = Context(prev_byte, prev_byte2, literal_context_mode);
918 uint8_t literal = input[pos & mask];
919 literal_enc.StoreSymbolWithContext<kLiteralContextBits>(
920 literal, context, mb.literal_context_map, storage_ix, storage);
921 prev_byte2 = prev_byte;
922 prev_byte = literal;
923 ++pos;
924 }
925 }
926 pos += cmd.copy_len();
927 if (cmd.copy_len()) {
928 prev_byte2 = input[(pos - 2) & mask];
929 prev_byte = input[(pos - 1) & mask];
930 if (cmd.cmd_prefix_ >= 128) {
931 size_t dist_code = cmd.dist_prefix_;
932 uint32_t distnumextra = cmd.dist_extra_ >> 24;
933 uint64_t distextra = cmd.dist_extra_ & 0xffffff;
934 if (mb.distance_context_map.empty()) {
935 distance_enc.StoreSymbol(dist_code, storage_ix, storage);
936 } else {
937 size_t context = cmd.DistanceContext();
938 distance_enc.StoreSymbolWithContext<kDistanceContextBits>(
939 dist_code, context, mb.distance_context_map, storage_ix, storage);
940 }
941 brotli::WriteBits(distnumextra, distextra, storage_ix, storage);
942 }
943 }
944 }
945 if (is_last) {
946 JumpToByteBoundary(storage_ix, storage);
947 }
948 }
949
950 static void BuildHistograms(const uint8_t* input,
951 size_t start_pos,
952 size_t mask,
953 const brotli::Command *commands,
954 size_t n_commands,
955 HistogramLiteral* lit_histo,
956 HistogramCommand* cmd_histo,
957 HistogramDistance* dist_histo) {
958 size_t pos = start_pos;
959 for (size_t i = 0; i < n_commands; ++i) {
960 const Command cmd = commands[i];
961 cmd_histo->Add(cmd.cmd_prefix_);
962 for (size_t j = cmd.insert_len_; j != 0; --j) {
963 lit_histo->Add(input[pos & mask]);
964 ++pos;
965 }
966 pos += cmd.copy_len();
967 if (cmd.copy_len() && cmd.cmd_prefix_ >= 128) {
968 dist_histo->Add(cmd.dist_prefix_);
969 }
970 }
971 }
972
973 static void StoreDataWithHuffmanCodes(const uint8_t* input,
974 size_t start_pos,
975 size_t mask,
976 const brotli::Command *commands,
977 size_t n_commands,
978 const uint8_t* lit_depth,
979 const uint16_t* lit_bits,
980 const uint8_t* cmd_depth,
981 const uint16_t* cmd_bits,
982 const uint8_t* dist_depth,
983 const uint16_t* dist_bits,
984 size_t* storage_ix,
985 uint8_t* storage) {
986 size_t pos = start_pos;
987 for (size_t i = 0; i < n_commands; ++i) {
988 const Command cmd = commands[i];
989 const size_t cmd_code = cmd.cmd_prefix_;
990 WriteBits(cmd_depth[cmd_code], cmd_bits[cmd_code], storage_ix, storage);
991 StoreCommandExtra(cmd, storage_ix, storage);
992 for (size_t j = cmd.insert_len_; j != 0; --j) {
993 const uint8_t literal = input[pos & mask];
994 WriteBits(lit_depth[literal], lit_bits[literal], storage_ix, storage);
995 ++pos;
996 }
997 pos += cmd.copy_len();
998 if (cmd.copy_len() && cmd.cmd_prefix_ >= 128) {
999 const size_t dist_code = cmd.dist_prefix_;
1000 const uint32_t distnumextra = cmd.dist_extra_ >> 24;
1001 const uint32_t distextra = cmd.dist_extra_ & 0xffffff;
1002 WriteBits(dist_depth[dist_code], dist_bits[dist_code],
1003 storage_ix, storage);
1004 WriteBits(distnumextra, distextra, storage_ix, storage);
1005 }
1006 }
1007 }
1008
1009 void StoreMetaBlockTrivial(const uint8_t* input,
1010 size_t start_pos,
1011 size_t length,
1012 size_t mask,
1013 bool is_last,
1014 const brotli::Command *commands,
1015 size_t n_commands,
1016 size_t *storage_ix,
1017 uint8_t *storage) {
1018 StoreCompressedMetaBlockHeader(is_last, length, storage_ix, storage);
1019
1020 HistogramLiteral lit_histo;
1021 HistogramCommand cmd_histo;
1022 HistogramDistance dist_histo;
1023
1024 BuildHistograms(input, start_pos, mask, commands, n_commands,
1025 &lit_histo, &cmd_histo, &dist_histo);
1026
1027 WriteBits(13, 0, storage_ix, storage);
1028
1029 std::vector<uint8_t> lit_depth(256);
1030 std::vector<uint16_t> lit_bits(256);
1031 std::vector<uint8_t> cmd_depth(kNumCommandPrefixes);
1032 std::vector<uint16_t> cmd_bits(kNumCommandPrefixes);
1033 std::vector<uint8_t> dist_depth(64);
1034 std::vector<uint16_t> dist_bits(64);
1035
1036 HuffmanTree* tree = static_cast<HuffmanTree*>(
1037 malloc(kMaxHuffmanTreeSize * sizeof(HuffmanTree)));
1038 BuildAndStoreHuffmanTree(&lit_histo.data_[0], 256, tree,
1039 &lit_depth[0], &lit_bits[0],
1040 storage_ix, storage);
1041 BuildAndStoreHuffmanTree(&cmd_histo.data_[0], kNumCommandPrefixes, tree,
1042 &cmd_depth[0], &cmd_bits[0],
1043 storage_ix, storage);
1044 BuildAndStoreHuffmanTree(&dist_histo.data_[0], 64, tree,
1045 &dist_depth[0], &dist_bits[0],
1046 storage_ix, storage);
1047 free(tree);
1048 StoreDataWithHuffmanCodes(input, start_pos, mask, commands,
1049 n_commands, &lit_depth[0], &lit_bits[0],
1050 &cmd_depth[0], &cmd_bits[0],
1051 &dist_depth[0], &dist_bits[0],
1052 storage_ix, storage);
1053 if (is_last) {
1054 JumpToByteBoundary(storage_ix, storage);
1055 }
1056 }
1057
1058 void StoreMetaBlockFast(const uint8_t* input,
1059 size_t start_pos,
1060 size_t length,
1061 size_t mask,
1062 bool is_last,
1063 const brotli::Command *commands,
1064 size_t n_commands,
1065 size_t *storage_ix,
1066 uint8_t *storage) {
1067 StoreCompressedMetaBlockHeader(is_last, length, storage_ix, storage);
1068
1069 WriteBits(13, 0, storage_ix, storage);
1070
1071 if (n_commands <= 128) {
1072 uint32_t histogram[256] = { 0 };
1073 size_t pos = start_pos;
1074 size_t num_literals = 0;
1075 for (size_t i = 0; i < n_commands; ++i) {
1076 const Command cmd = commands[i];
1077 for (size_t j = cmd.insert_len_; j != 0; --j) {
1078 ++histogram[input[pos & mask]];
1079 ++pos;
1080 }
1081 num_literals += cmd.insert_len_;
1082 pos += cmd.copy_len();
1083 }
1084 uint8_t lit_depth[256] = { 0 };
1085 uint16_t lit_bits[256] = { 0 };
1086 BuildAndStoreHuffmanTreeFast(histogram, num_literals,
1087 /* max_bits = */ 8,
1088 lit_depth, lit_bits,
1089 storage_ix, storage);
1090 StoreStaticCommandHuffmanTree(storage_ix, storage);
1091 StoreStaticDistanceHuffmanTree(storage_ix, storage);
1092 StoreDataWithHuffmanCodes(input, start_pos, mask, commands,
1093 n_commands, &lit_depth[0], &lit_bits[0],
1094 kStaticCommandCodeDepth,
1095 kStaticCommandCodeBits,
1096 kStaticDistanceCodeDepth,
1097 kStaticDistanceCodeBits,
1098 storage_ix, storage);
1099 } else {
1100 HistogramLiteral lit_histo;
1101 HistogramCommand cmd_histo;
1102 HistogramDistance dist_histo;
1103 BuildHistograms(input, start_pos, mask, commands, n_commands,
1104 &lit_histo, &cmd_histo, &dist_histo);
1105 std::vector<uint8_t> lit_depth(256);
1106 std::vector<uint16_t> lit_bits(256);
1107 std::vector<uint8_t> cmd_depth(kNumCommandPrefixes);
1108 std::vector<uint16_t> cmd_bits(kNumCommandPrefixes);
1109 std::vector<uint8_t> dist_depth(64);
1110 std::vector<uint16_t> dist_bits(64);
1111 BuildAndStoreHuffmanTreeFast(&lit_histo.data_[0], lit_histo.total_count_,
1112 /* max_bits = */ 8,
1113 &lit_depth[0], &lit_bits[0],
1114 storage_ix, storage);
1115 BuildAndStoreHuffmanTreeFast(&cmd_histo.data_[0], cmd_histo.total_count_,
1116 /* max_bits = */ 10,
1117 &cmd_depth[0], &cmd_bits[0],
1118 storage_ix, storage);
1119 BuildAndStoreHuffmanTreeFast(&dist_histo.data_[0], dist_histo.total_count_,
1120 /* max_bits = */ 6,
1121 &dist_depth[0], &dist_bits[0],
1122 storage_ix, storage);
1123 StoreDataWithHuffmanCodes(input, start_pos, mask, commands,
1124 n_commands, &lit_depth[0], &lit_bits[0],
1125 &cmd_depth[0], &cmd_bits[0],
1126 &dist_depth[0], &dist_bits[0],
1127 storage_ix, storage);
1128 }
1129
1130 if (is_last) {
1131 JumpToByteBoundary(storage_ix, storage);
1132 }
1133 }
1134
1135 // This is for storing uncompressed blocks (simple raw storage of
1136 // bytes-as-bytes).
1137 void StoreUncompressedMetaBlock(bool final_block,
1138 const uint8_t * __restrict input,
1139 size_t position, size_t mask,
1140 size_t len,
1141 size_t * __restrict storage_ix,
1142 uint8_t * __restrict storage) {
1143 StoreUncompressedMetaBlockHeader(len, storage_ix, storage);
1144 JumpToByteBoundary(storage_ix, storage);
1145
1146 size_t masked_pos = position & mask;
1147 if (masked_pos + len > mask + 1) {
1148 size_t len1 = mask + 1 - masked_pos;
1149 memcpy(&storage[*storage_ix >> 3], &input[masked_pos], len1);
1150 *storage_ix += len1 << 3;
1151 len -= len1;
1152 masked_pos = 0;
1153 }
1154 memcpy(&storage[*storage_ix >> 3], &input[masked_pos], len);
1155 *storage_ix += len << 3;
1156
1157 // We need to clear the next 4 bytes to continue to be
1158 // compatible with WriteBits.
1159 brotli::WriteBitsPrepareStorage(*storage_ix, storage);
1160
1161 // Since the uncompressed block itself may not be the final block, add an
1162 // empty one after this.
1163 if (final_block) {
1164 brotli::WriteBits(1, 1, storage_ix, storage); // islast
1165 brotli::WriteBits(1, 1, storage_ix, storage); // isempty
1166 JumpToByteBoundary(storage_ix, storage);
1167 }
1168 }
1169
1170 void StoreSyncMetaBlock(size_t * __restrict storage_ix,
1171 uint8_t * __restrict storage) {
1172 // Empty metadata meta-block bit pattern:
1173 // 1 bit: is_last (0)
1174 // 2 bits: num nibbles (3)
1175 // 1 bit: reserved (0)
1176 // 2 bits: metadata length bytes (0)
1177 WriteBits(6, 6, storage_ix, storage);
1178 JumpToByteBoundary(storage_ix, storage);
1179 }
1180
1181 } // namespace brotli
OLDNEW
« no previous file with comments | « third_party/brotli/enc/brotli_bit_stream.c ('k') | third_party/brotli/enc/cluster.h » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698