| /* NOLINT(build/header_guard) */ | 
 | /* Copyright 2013 Google Inc. All Rights Reserved. | 
 |  | 
 |    Distributed under MIT license. | 
 |    See file LICENSE for detail or copy at https://opensource.org/licenses/MIT | 
 | */ | 
 |  | 
 | /* template parameters: EXPORT_FN, FN */ | 
 |  | 
 | static BROTLI_NOINLINE void EXPORT_FN(CreateBackwardReferences)( | 
 |     size_t num_bytes, size_t position, | 
 |     const uint8_t* ringbuffer, size_t ringbuffer_mask, | 
 |     ContextLut literal_context_lut, const BrotliEncoderParams* params, | 
 |     Hasher* hasher, int* dist_cache, size_t* last_insert_len, | 
 |     Command* commands, size_t* num_commands, size_t* num_literals) { | 
 |   HASHER()* privat = &hasher->privat.FN(_); | 
 |   /* Set maximum distance, see section 9.1. of the spec. */ | 
 |   const size_t max_backward_limit = BROTLI_MAX_BACKWARD_LIMIT(params->lgwin); | 
 |   const size_t position_offset = params->stream_offset; | 
 |  | 
 |   const Command* const orig_commands = commands; | 
 |   size_t insert_length = *last_insert_len; | 
 |   const size_t pos_end = position + num_bytes; | 
 |   const size_t store_end = num_bytes >= FN(StoreLookahead)() ? | 
 |       position + num_bytes - FN(StoreLookahead)() + 1 : position; | 
 |  | 
 |   /* For speed up heuristics for random data. */ | 
 |   const size_t random_heuristics_window_size = | 
 |       LiteralSpreeLengthForSparseSearch(params); | 
 |   size_t apply_random_heuristics = position + random_heuristics_window_size; | 
 |   const size_t gap = params->dictionary.compound.total_size; | 
 |  | 
 |   /* Minimum score to accept a backward reference. */ | 
 |   const score_t kMinScore = BROTLI_SCORE_BASE + 100; | 
 |  | 
 |   FN(PrepareDistanceCache)(privat, dist_cache); | 
 |  | 
 |   while (position + FN(HashTypeLength)() < pos_end) { | 
 |     size_t max_length = pos_end - position; | 
 |     size_t max_distance = BROTLI_MIN(size_t, position, max_backward_limit); | 
 |     size_t dictionary_start = BROTLI_MIN(size_t, | 
 |         position + position_offset, max_backward_limit); | 
 |     HasherSearchResult sr; | 
 |     int dict_id = 0; | 
 |     uint8_t p1 = 0; | 
 |     uint8_t p2 = 0; | 
 |     if (params->dictionary.contextual.context_based) { | 
 |       p1 = position >= 1 ? | 
 |           ringbuffer[(size_t)(position - 1) & ringbuffer_mask] : 0; | 
 |       p2 = position >= 2 ? | 
 |           ringbuffer[(size_t)(position - 2) & ringbuffer_mask] : 0; | 
 |       dict_id = params->dictionary.contextual.context_map[ | 
 |           BROTLI_CONTEXT(p1, p2, literal_context_lut)]; | 
 |     } | 
 |     sr.len = 0; | 
 |     sr.len_code_delta = 0; | 
 |     sr.distance = 0; | 
 |     sr.score = kMinScore; | 
 |     FN(FindLongestMatch)(privat, params->dictionary.contextual.dict[dict_id], | 
 |         ringbuffer, ringbuffer_mask, dist_cache, position, max_length, | 
 |         max_distance, dictionary_start + gap, params->dist.max_distance, &sr); | 
 |     if (ENABLE_COMPOUND_DICTIONARY) { | 
 |       LookupCompoundDictionaryMatch(¶ms->dictionary.compound, ringbuffer, | 
 |           ringbuffer_mask, dist_cache, position, max_length, | 
 |           dictionary_start, params->dist.max_distance, &sr); | 
 |     } | 
 |     if (sr.score > kMinScore) { | 
 |       /* Found a match. Let's look for something even better ahead. */ | 
 |       int delayed_backward_references_in_row = 0; | 
 |       --max_length; | 
 |       for (;; --max_length) { | 
 |         const score_t cost_diff_lazy = 175; | 
 |         HasherSearchResult sr2; | 
 |         sr2.len = params->quality < MIN_QUALITY_FOR_EXTENSIVE_REFERENCE_SEARCH ? | 
 |             BROTLI_MIN(size_t, sr.len - 1, max_length) : 0; | 
 |         sr2.len_code_delta = 0; | 
 |         sr2.distance = 0; | 
 |         sr2.score = kMinScore; | 
 |         max_distance = BROTLI_MIN(size_t, position + 1, max_backward_limit); | 
 |         dictionary_start = BROTLI_MIN(size_t, | 
 |             position + 1 + position_offset, max_backward_limit); | 
 |         if (params->dictionary.contextual.context_based) { | 
 |           p2 = p1; | 
 |           p1 = ringbuffer[position & ringbuffer_mask]; | 
 |           dict_id = params->dictionary.contextual.context_map[ | 
 |               BROTLI_CONTEXT(p1, p2, literal_context_lut)]; | 
 |         } | 
 |         FN(FindLongestMatch)(privat, | 
 |             params->dictionary.contextual.dict[dict_id], | 
 |             ringbuffer, ringbuffer_mask, dist_cache, position + 1, max_length, | 
 |             max_distance, dictionary_start + gap, params->dist.max_distance, | 
 |             &sr2); | 
 |         if (ENABLE_COMPOUND_DICTIONARY) { | 
 |           LookupCompoundDictionaryMatch( | 
 |               ¶ms->dictionary.compound, ringbuffer, | 
 |               ringbuffer_mask, dist_cache, position + 1, max_length, | 
 |               dictionary_start, params->dist.max_distance, &sr2); | 
 |         } | 
 |         if (sr2.score >= sr.score + cost_diff_lazy) { | 
 |           /* Ok, let's just write one byte for now and start a match from the | 
 |              next byte. */ | 
 |           ++position; | 
 |           ++insert_length; | 
 |           sr = sr2; | 
 |           if (++delayed_backward_references_in_row < 4 && | 
 |               position + FN(HashTypeLength)() < pos_end) { | 
 |             continue; | 
 |           } | 
 |         } | 
 |         break; | 
 |       } | 
 |       apply_random_heuristics = | 
 |           position + 2 * sr.len + random_heuristics_window_size; | 
 |       dictionary_start = BROTLI_MIN(size_t, | 
 |           position + position_offset, max_backward_limit); | 
 |       { | 
 |         /* The first 16 codes are special short-codes, | 
 |            and the minimum offset is 1. */ | 
 |         size_t distance_code = ComputeDistanceCode( | 
 |             sr.distance, dictionary_start + gap, dist_cache); | 
 |         if ((sr.distance <= (dictionary_start + gap)) && distance_code > 0) { | 
 |           dist_cache[3] = dist_cache[2]; | 
 |           dist_cache[2] = dist_cache[1]; | 
 |           dist_cache[1] = dist_cache[0]; | 
 |           dist_cache[0] = (int)sr.distance; | 
 |           FN(PrepareDistanceCache)(privat, dist_cache); | 
 |         } | 
 |         InitCommand(commands++, ¶ms->dist, insert_length, | 
 |             sr.len, sr.len_code_delta, distance_code); | 
 |       } | 
 |       *num_literals += insert_length; | 
 |       insert_length = 0; | 
 |       /* Put the hash keys into the table, if there are enough bytes left. | 
 |          Depending on the hasher implementation, it can push all positions | 
 |          in the given range or only a subset of them. | 
 |          Avoid hash poisoning with RLE data. */ | 
 |       { | 
 |         size_t range_start = position + 2; | 
 |         size_t range_end = BROTLI_MIN(size_t, position + sr.len, store_end); | 
 |         if (sr.distance < (sr.len >> 2)) { | 
 |           range_start = BROTLI_MIN(size_t, range_end, BROTLI_MAX(size_t, | 
 |               range_start, position + sr.len - (sr.distance << 2))); | 
 |         } | 
 |         FN(StoreRange)(privat, ringbuffer, ringbuffer_mask, range_start, | 
 |                        range_end); | 
 |       } | 
 |       position += sr.len; | 
 |     } else { | 
 |       ++insert_length; | 
 |       ++position; | 
 |       /* If we have not seen matches for a long time, we can skip some | 
 |          match lookups. Unsuccessful match lookups are very very expensive | 
 |          and this kind of a heuristic speeds up compression quite | 
 |          a lot. */ | 
 |       if (position > apply_random_heuristics) { | 
 |         /* Going through uncompressible data, jump. */ | 
 |         if (position > | 
 |             apply_random_heuristics + 4 * random_heuristics_window_size) { | 
 |           /* It is quite a long time since we saw a copy, so we assume | 
 |              that this data is not compressible, and store hashes less | 
 |              often. Hashes of non compressible data are less likely to | 
 |              turn out to be useful in the future, too, so we store less of | 
 |              them to not to flood out the hash table of good compressible | 
 |              data. */ | 
 |           const size_t kMargin = | 
 |               BROTLI_MAX(size_t, FN(StoreLookahead)() - 1, 4); | 
 |           size_t pos_jump = | 
 |               BROTLI_MIN(size_t, position + 16, pos_end - kMargin); | 
 |           for (; position < pos_jump; position += 4) { | 
 |             FN(Store)(privat, ringbuffer, ringbuffer_mask, position); | 
 |             insert_length += 4; | 
 |           } | 
 |         } else { | 
 |           const size_t kMargin = | 
 |               BROTLI_MAX(size_t, FN(StoreLookahead)() - 1, 2); | 
 |           size_t pos_jump = | 
 |               BROTLI_MIN(size_t, position + 8, pos_end - kMargin); | 
 |           for (; position < pos_jump; position += 2) { | 
 |             FN(Store)(privat, ringbuffer, ringbuffer_mask, position); | 
 |             insert_length += 2; | 
 |           } | 
 |         } | 
 |       } | 
 |     } | 
 |   } | 
 |   insert_length += pos_end - position; | 
 |   *last_insert_len = insert_length; | 
 |   *num_commands += (size_t)(commands - orig_commands); | 
 | } |