+
+ /* Calculate the cost to reach the next position by coding a literal. */
+ optimum[1].queue = c->queue;
+ optimum[1].cost = lzx_literal_cost(c->cur_window[c->match_window_pos - 1],
+ &c->costs);
+ optimum[1].prev.link = 0;
+
+ /* Calculate the cost to reach any position up to and including that
+ * reached by the longest match.
+ *
+ * Note: We consider only the lowest-offset match that reaches each
+ * position.
+ *
+ * Note: Some of the cost calculation stays the same for each offset,
+ * regardless of how many lengths it gets used for. Therefore, to
+ * improve performance, we hand-code the cost calculation instead of
+ * calling lzx_match_cost() to do a from-scratch cost evaluation at each
+ * length. */
+ for (unsigned i = 0, len = 2; i < num_matches; i++) {
+ u32 offset;
+ struct lzx_lru_queue queue;
+ u32 position_cost;
+ unsigned position_slot;
+ unsigned num_extra_bits;
+
+ offset = matches[i].offset;
+ queue = c->queue;
+ position_cost = 0;
+
+ position_slot = lzx_get_position_slot(offset, &queue);
+ num_extra_bits = lzx_get_num_extra_bits(position_slot);
+ if (num_extra_bits >= 3) {
+ position_cost += num_extra_bits - 3;
+ position_cost += c->costs.aligned[(offset + LZX_OFFSET_OFFSET) & 7];
+ } else {
+ position_cost += num_extra_bits;
+ }
+
+ do {
+ unsigned len_header;
+ unsigned main_symbol;
+ u32 cost;
+
+ cost = position_cost;
+
+ len_header = min(len - LZX_MIN_MATCH_LEN, LZX_NUM_PRIMARY_LENS);
+ main_symbol = ((position_slot << 3) | len_header) + LZX_NUM_CHARS;
+ cost += c->costs.main[main_symbol];
+ if (len_header == LZX_NUM_PRIMARY_LENS)
+ cost += c->costs.len[len - LZX_MIN_MATCH_LEN - LZX_NUM_PRIMARY_LENS];
+
+ optimum[len].queue = queue;
+ optimum[len].prev.link = 0;
+ optimum[len].prev.match_offset = offset;
+ optimum[len].cost = cost;
+ } while (++len <= matches[i].len);
+ }
+ end_pos = longest_len;
+
+ if (longest_rep_len >= LZX_MIN_MATCH_LEN) {
+ u32 cost;
+
+ while (end_pos < longest_rep_len)
+ optimum[++end_pos].cost = MC_INFINITE_COST;
+
+ cost = lzx_repmatch_cost(longest_rep_len, longest_rep_slot,
+ &c->costs);
+ if (cost <= optimum[longest_rep_len].cost) {
+ optimum[longest_rep_len].queue = c->queue;
+ swap(optimum[longest_rep_len].queue.R[0],
+ optimum[longest_rep_len].queue.R[longest_rep_slot]);
+ optimum[longest_rep_len].prev.link = 0;
+ optimum[longest_rep_len].prev.match_offset =
+ optimum[longest_rep_len].queue.R[0];
+ optimum[longest_rep_len].cost = cost;
+ }
+ }
+
+ /* Step forward, calculating the estimated minimum cost to reach each
+ * position. The algorithm may find multiple paths to reach each
+ * position; only the lowest-cost path is saved.
+ *
+ * The progress of the parse is tracked in the @optimum array, which for
+ * each position contains the minimum cost to reach that position, the
+ * index of the start of the match/literal taken to reach that position
+ * through the minimum-cost path, the offset of the match taken (not
+ * relevant for literals), and the adaptive state that will exist at
+ * that position after the minimum-cost path is taken. The @cur_pos
+ * variable stores the position at which the algorithm is currently
+ * considering coding choices, and the @end_pos variable stores the
+ * greatest position at which the costs of coding choices have been
+ * saved.
+ *
+ * The loop terminates when any one of the following conditions occurs:
+ *
+ * 1. A match with length greater than or equal to @nice_match_length is
+ * found. When this occurs, the algorithm chooses this match
+ * unconditionally, and consequently the near-optimal match/literal
+ * sequence up to and including that match is fully determined and it
+ * can begin returning the match/literal list.
+ *
+ * 2. @cur_pos reaches a position not overlapped by a preceding match.
+ * In such cases, the near-optimal match/literal sequence up to
+ * @cur_pos is fully determined and it can begin returning the
+ * match/literal list.
+ *
+ * 3. Failing either of the above in a degenerate case, the loop
+ * terminates when space in the @optimum array is exhausted.
+ * This terminates the algorithm and forces it to start returning
+ * matches/literals even though they may not be globally optimal.
+ *
+ * Upon loop termination, a nonempty list of matches/literals will have
+ * been produced and stored in the @optimum array. These
+ * matches/literals are linked in reverse order, so the last thing this
+ * function does is reverse this list and return the first
+ * match/literal, leaving the rest to be returned immediately by
+ * subsequent calls to this function.
+ */
+ cur_pos = 0;
+ for (;;) {
+ u32 cost;
+
+ /* Advance to next position. */
+ cur_pos++;
+
+ /* Check termination conditions (2) and (3) noted above. */
+ if (cur_pos == end_pos || cur_pos == LZX_OPTIM_ARRAY_LENGTH)
+ return lzx_match_chooser_reverse_list(c, cur_pos);
+
+ /* Search for matches at repeat offsets. Again, as a heuristic
+ * we only keep the longest one. */
+ longest_rep_len = LZX_MIN_MATCH_LEN - 1;
+ unsigned limit = min(LZX_MAX_MATCH_LEN,
+ c->match_window_end - c->match_window_pos);
+ for (int i = 0; i < LZX_NUM_RECENT_OFFSETS; i++) {
+ u32 offset = optimum[cur_pos].queue.R[i];
+ const u8 *strptr = &c->cur_window[c->match_window_pos];
+ const u8 *matchptr = strptr - offset;
+ unsigned len = 0;
+ while (len < limit && strptr[len] == matchptr[len])
+ len++;
+ if (len > longest_rep_len) {
+ longest_rep_len = len;
+ longest_rep_slot = i;
+ }
+ }
+
+ /* If we found a long match at a repeat offset, choose it
+ * immediately. */
+ if (longest_rep_len >= c->params.nice_match_length) {
+ /* Build the list of matches to return and get
+ * the first one. */
+ match = lzx_match_chooser_reverse_list(c, cur_pos);
+
+ /* Append the long match to the end of the list. */
+ optimum[cur_pos].next.match_offset =
+ optimum[cur_pos].queue.R[longest_rep_slot];
+ optimum[cur_pos].next.link = cur_pos + longest_rep_len;
+ c->optimum_end_idx = cur_pos + longest_rep_len;
+
+ /* Skip over the remaining bytes of the long match. */
+ lzx_skip_bytes(c, longest_rep_len);
+
+ /* Return first match in the list. */
+ return match;
+ }
+
+ /* Find other matches. */
+ num_matches = lzx_get_matches(c, &matches);
+
+ /* If there's a long match, choose it immediately. */
+ if (num_matches) {
+ longest_len = matches[num_matches - 1].len;
+ if (longest_len >= c->params.nice_match_length) {
+ /* Build the list of matches to return and get
+ * the first one. */
+ match = lzx_match_chooser_reverse_list(c, cur_pos);
+
+ /* Append the long match to the end of the list. */
+ optimum[cur_pos].next.match_offset =
+ matches[num_matches - 1].offset;
+ optimum[cur_pos].next.link = cur_pos + longest_len;
+ c->optimum_end_idx = cur_pos + longest_len;
+
+ /* Skip over the remaining bytes of the long match. */
+ lzx_skip_bytes(c, longest_len - 1);
+
+ /* Return first match in the list. */
+ return match;
+ }
+ } else {
+ longest_len = 1;
+ }
+
+ /* If we are reaching any positions for the first time, we need
+ * to initialize their costs to infinity. */
+ while (end_pos < cur_pos + longest_len)
+ optimum[++end_pos].cost = MC_INFINITE_COST;
+
+ /* Consider coding a literal. */
+ cost = optimum[cur_pos].cost +
+ lzx_literal_cost(c->cur_window[c->match_window_pos - 1],
+ &c->costs);
+ if (cost < optimum[cur_pos + 1].cost) {
+ optimum[cur_pos + 1].queue = optimum[cur_pos].queue;
+ optimum[cur_pos + 1].cost = cost;
+ optimum[cur_pos + 1].prev.link = cur_pos;
+ }
+
+ /* Consider coding a match.
+ *
+ * The hard-coded cost calculation is done for the same reason
+ * stated in the comment for the similar loop earlier.
+ * Actually, it is *this* one that has the biggest effect on
+ * performance; overall LZX compression is > 10% faster with
+ * this code compared to calling lzx_match_cost() with each
+ * length. */
+ for (unsigned i = 0, len = 2; i < num_matches; i++) {
+ u32 offset;
+ u32 position_cost;
+ unsigned position_slot;
+ unsigned num_extra_bits;
+
+ offset = matches[i].offset;
+ position_cost = optimum[cur_pos].cost;
+
+ /* Yet another optimization: instead of calling
+ * lzx_get_position_slot(), hand-inline the search of
+ * the repeat offset queue. Then we can omit the
+ * extra_bits calculation for repeat offset matches, and
+ * also only compute the updated queue if we actually do
+ * find a new lowest cost path. */
+ for (position_slot = 0; position_slot < LZX_NUM_RECENT_OFFSETS; position_slot++)
+ if (offset == optimum[cur_pos].queue.R[position_slot])
+ goto have_position_cost;
+
+ position_slot = lzx_get_position_slot_raw(offset + LZX_OFFSET_OFFSET);
+
+ num_extra_bits = lzx_get_num_extra_bits(position_slot);
+ if (num_extra_bits >= 3) {
+ position_cost += num_extra_bits - 3;
+ position_cost += c->costs.aligned[
+ (offset + LZX_OFFSET_OFFSET) & 7];
+ } else {
+ position_cost += num_extra_bits;
+ }
+
+ have_position_cost:
+
+ do {
+ unsigned len_header;
+ unsigned main_symbol;
+ u32 cost;
+
+ cost = position_cost;
+
+ len_header = min(len - LZX_MIN_MATCH_LEN,
+ LZX_NUM_PRIMARY_LENS);
+ main_symbol = ((position_slot << 3) | len_header) +
+ LZX_NUM_CHARS;
+ cost += c->costs.main[main_symbol];
+ if (len_header == LZX_NUM_PRIMARY_LENS) {
+ cost += c->costs.len[len -
+ LZX_MIN_MATCH_LEN -
+ LZX_NUM_PRIMARY_LENS];
+ }
+ if (cost < optimum[cur_pos + len].cost) {
+ if (position_slot < LZX_NUM_RECENT_OFFSETS) {
+ optimum[cur_pos + len].queue = optimum[cur_pos].queue;
+ swap(optimum[cur_pos + len].queue.R[0],
+ optimum[cur_pos + len].queue.R[position_slot]);
+ } else {
+ optimum[cur_pos + len].queue.R[0] = offset;
+ optimum[cur_pos + len].queue.R[1] = optimum[cur_pos].queue.R[0];
+ optimum[cur_pos + len].queue.R[2] = optimum[cur_pos].queue.R[1];
+ }
+ optimum[cur_pos + len].prev.link = cur_pos;
+ optimum[cur_pos + len].prev.match_offset = offset;
+ optimum[cur_pos + len].cost = cost;
+ }
+ } while (++len <= matches[i].len);
+ }
+
+ /* Consider coding a repeat offset match.
+ *
+ * As a heuristic, we only consider the longest length of the
+ * longest repeat offset match. This does not, however,
+ * necessarily mean that we will never consider any other repeat
+ * offsets, because above we detect repeat offset matches that
+ * were found by the regular match-finder. Therefore, this
+ * special handling of the longest repeat-offset match is only
+ * helpful for coding a repeat offset match that was *not* found
+ * by the match-finder, e.g. due to being obscured by a less
+ * distant match that is at least as long.
+ *
+ * Note: an alternative, used in LZMA, is to consider every
+ * length of every repeat offset match. This is a more thorough
+ * search, and it makes it unnecessary to detect repeat offset
+ * matches that were found by the regular match-finder. But by
+ * my tests, for LZX the LZMA method slows down the compressor
+ * by ~10% and doesn't actually help the compression ratio too
+ * much.
+ *
+ * Also tested a compromise approach: consider every 3rd length
+ * of the longest repeat offset match. Still didn't seem quite
+ * worth it, though.
+ */
+ if (longest_rep_len >= LZX_MIN_MATCH_LEN) {
+
+ while (end_pos < cur_pos + longest_rep_len)
+ optimum[++end_pos].cost = MC_INFINITE_COST;
+
+ cost = optimum[cur_pos].cost +
+ lzx_repmatch_cost(longest_rep_len, longest_rep_slot,
+ &c->costs);
+ if (cost <= optimum[cur_pos + longest_rep_len].cost) {
+ optimum[cur_pos + longest_rep_len].queue =
+ optimum[cur_pos].queue;
+ swap(optimum[cur_pos + longest_rep_len].queue.R[0],
+ optimum[cur_pos + longest_rep_len].queue.R[longest_rep_slot]);
+ optimum[cur_pos + longest_rep_len].prev.link =
+ cur_pos;
+ optimum[cur_pos + longest_rep_len].prev.match_offset =
+ optimum[cur_pos + longest_rep_len].queue.R[0];
+ optimum[cur_pos + longest_rep_len].cost =
+ cost;
+ }
+ }
+ }
+}
+
+static struct lz_match
+lzx_choose_lazy_item(struct lzx_compressor *c)
+{
+ const struct lz_match *matches;
+ struct lz_match cur_match;
+ struct lz_match next_match;
+ u32 num_matches;
+
+ if (c->prev_match.len) {
+ cur_match = c->prev_match;
+ c->prev_match.len = 0;
+ } else {
+ num_matches = lzx_get_matches(c, &matches);
+ if (num_matches == 0 ||
+ (matches[num_matches - 1].len <= 3 &&
+ (matches[num_matches - 1].len <= 2 ||
+ matches[num_matches - 1].offset > 4096)))
+ {
+ return (struct lz_match) { };
+ }
+
+ cur_match = matches[num_matches - 1];
+ }
+
+ if (cur_match.len >= c->params.nice_match_length) {
+ lzx_skip_bytes(c, cur_match.len - 1);
+ return cur_match;
+ }
+
+ num_matches = lzx_get_matches(c, &matches);
+ if (num_matches == 0 ||
+ (matches[num_matches - 1].len <= 3 &&
+ (matches[num_matches - 1].len <= 2 ||
+ matches[num_matches - 1].offset > 4096)))
+ {
+ lzx_skip_bytes(c, cur_match.len - 2);
+ return cur_match;
+ }
+
+ next_match = matches[num_matches - 1];
+
+ if (next_match.len <= cur_match.len) {
+ lzx_skip_bytes(c, cur_match.len - 2);
+ return cur_match;
+ } else {
+ c->prev_match = next_match;
+ return (struct lz_match) { };
+ }
+}
+
+/*
+ * Return the next match or literal to use, delegating to the currently selected
+ * match-choosing algorithm.
+ *
+ * If the length of the returned 'struct lz_match' is less than
+ * LZX_MIN_MATCH_LEN, then it is really a literal.
+ */
+static inline struct lz_match
+lzx_choose_item(struct lzx_compressor *c)
+{
+ return (*c->params.choose_item_func)(c);