static inline void
check_that_powers_fit_in_bitfield(void)
{
- BUILD_BUG_ON(LZMS_NUM_DELTA_POWER_SYMS > (1 << (31 - DELTA_SOURCE_POWER_SHIFT)));
+ STATIC_ASSERT(LZMS_NUM_DELTA_POWER_SYMS <= (1 << (31 - DELTA_SOURCE_POWER_SHIFT)));
}
/* A stripped-down version of the adaptive state in LZMS which excludes the
*
* Note: this adaptive state structure also does not include the
* probability entries or current Huffman codewords. Those aren't
- * maintained per-position and are only updated occassionally.
+ * maintained per-position and are only updated occasionally.
*/
struct lzms_adaptive_state state;
} _aligned_attribute(64);
unsigned lz_rep_states[LZMS_NUM_LZ_REP_DECISIONS];
unsigned delta_state;
unsigned delta_rep_states[LZMS_NUM_DELTA_REP_DECISIONS];
- struct lzms_probability_entry main_probs[LZMS_NUM_MAIN_PROBS];
- struct lzms_probability_entry match_probs[LZMS_NUM_MATCH_PROBS];
- struct lzms_probability_entry lz_probs[LZMS_NUM_LZ_PROBS];
- struct lzms_probability_entry lz_rep_probs[LZMS_NUM_LZ_REP_DECISIONS]
- [LZMS_NUM_LZ_REP_PROBS];
- struct lzms_probability_entry delta_probs[LZMS_NUM_DELTA_PROBS];
- struct lzms_probability_entry delta_rep_probs[LZMS_NUM_DELTA_REP_DECISIONS]
- [LZMS_NUM_DELTA_REP_PROBS];
+ struct lzms_probabilites probs;
/* Huffman codes */
do {
if (likely(rc->next >= rc->begin)) {
if (rc->next != rc->end) {
- put_unaligned_u16_le(rc->cache +
+ put_unaligned_le16(rc->cache +
(u16)(rc->lower_bound >> 32),
rc->next++);
}
lzms_encode_main_bit(struct lzms_compressor *c, int bit)
{
lzms_encode_bit(bit, &c->main_state, LZMS_NUM_MAIN_PROBS,
- c->main_probs, &c->rc);
+ c->probs.main, &c->rc);
}
static void
lzms_encode_match_bit(struct lzms_compressor *c, int bit)
{
lzms_encode_bit(bit, &c->match_state, LZMS_NUM_MATCH_PROBS,
- c->match_probs, &c->rc);
+ c->probs.match, &c->rc);
}
static void
lzms_encode_lz_bit(struct lzms_compressor *c, int bit)
{
lzms_encode_bit(bit, &c->lz_state, LZMS_NUM_LZ_PROBS,
- c->lz_probs, &c->rc);
+ c->probs.lz, &c->rc);
}
static void
lzms_encode_lz_rep_bit(struct lzms_compressor *c, int bit, int idx)
{
lzms_encode_bit(bit, &c->lz_rep_states[idx], LZMS_NUM_LZ_REP_PROBS,
- c->lz_rep_probs[idx], &c->rc);
+ c->probs.lz_rep[idx], &c->rc);
}
static void
lzms_encode_delta_bit(struct lzms_compressor *c, int bit)
{
lzms_encode_bit(bit, &c->delta_state, LZMS_NUM_DELTA_PROBS,
- c->delta_probs, &c->rc);
+ c->probs.delta, &c->rc);
}
static void
lzms_encode_delta_rep_bit(struct lzms_compressor *c, int bit, int idx)
{
lzms_encode_bit(bit, &c->delta_rep_states[idx], LZMS_NUM_DELTA_REP_PROBS,
- c->delta_rep_probs[idx], &c->rc);
+ c->probs.delta_rep[idx], &c->rc);
}
/******************************************************************************
/* Write a coding unit, unless it would underflow the buffer. */
if (os->next != os->begin)
- put_unaligned_u16_le(os->bitbuf >> os->bitcount, --os->next);
+ put_unaligned_le16(os->bitbuf >> os->bitcount, --os->next);
/* Optimization for call sites that never write more than 16
* bits at once. */
return false;
if (os->bitcount != 0)
- put_unaligned_u16_le(os->bitbuf << (16 - os->bitcount), --os->next);
+ put_unaligned_le16(os->bitbuf << (16 - os->bitcount), --os->next);
return true;
}
check_cost_shift(void)
{
/* lzms_bit_costs is hard-coded to the current COST_SHIFT. */
- BUILD_BUG_ON(COST_SHIFT != 6);
+ STATIC_ASSERT(COST_SHIFT == 6);
}
#if 0
static inline u32
lzms_literal_cost(struct lzms_compressor *c, unsigned main_state, unsigned literal)
{
- return lzms_bit_0_cost(main_state, c->main_probs) +
+ return lzms_bit_0_cost(main_state, c->probs.main) +
((u32)c->literal_lens[literal] << COST_SHIFT);
}
lzms_init_delta_matchfinder(struct lzms_compressor *c)
{
/* Set all entries to use an invalid power, which will never match. */
- BUILD_BUG_ON(NUM_POWERS_TO_CONSIDER >= (1 << (32 - DELTA_SOURCE_POWER_SHIFT)));
+ STATIC_ASSERT(NUM_POWERS_TO_CONSIDER < (1 << (32 - DELTA_SOURCE_POWER_SHIFT)));
memset(c->delta_hash_table, 0xFF, sizeof(c->delta_hash_table));
/* Initialize the next hash code for each power. We can just use zeroes
* include in the hash code computation the span and the low-order bits
* of the current position. */
- BUILD_BUG_ON(NBYTES_HASHED_FOR_DELTA != 3);
+ STATIC_ASSERT(NBYTES_HASHED_FOR_DELTA == 3);
u8 d0 = *(p + 0) - *(p + 0 - span);
u8 d1 = *(p + 1) - *(p + 1 - span);
u8 d2 = *(p + 2) - *(p + 2 - span);
c->delta_hash_table[hash] =
(power << DELTA_SOURCE_POWER_SHIFT) | pos;
c->next_delta_hashes[power] = next_hash;
- prefetch(&c->delta_hash_table[next_hash]);
+ prefetchw(&c->delta_hash_table[next_hash]);
}
} while (in_next++, pos++, --count);
}
u32 base_cost = cur_node->cost +
lzms_bit_1_cost(cur_node->state.main_state,
- c->main_probs) +
+ c->probs.main) +
lzms_bit_0_cost(cur_node->state.match_state,
- c->match_probs) +
+ c->probs.match) +
lzms_bit_1_cost(cur_node->state.lz_state,
- c->lz_probs);
+ c->probs.lz);
for (int i = 0; i < rep_idx; i++)
base_cost += lzms_bit_1_cost(cur_node->state.lz_rep_states[i],
- c->lz_rep_probs[i]);
+ c->probs.lz_rep[i]);
if (rep_idx < LZMS_NUM_LZ_REP_DECISIONS)
base_cost += lzms_bit_0_cost(cur_node->state.lz_rep_states[rep_idx],
- c->lz_rep_probs[rep_idx]);
+ c->probs.lz_rep[rep_idx]);
u32 len = 2;
do {
main_state = ((main_state << 1) | 0) % LZMS_NUM_MAIN_PROBS;
/* add LZ-rep0 cost */
- cost += lzms_bit_1_cost(main_state, c->main_probs) +
- lzms_bit_0_cost(match_state, c->match_probs) +
- lzms_bit_1_cost(lz_state, c->lz_probs) +
- lzms_bit_0_cost(lz_rep0_state, c->lz_rep_probs[0]) +
+ cost += lzms_bit_1_cost(main_state, c->probs.main) +
+ lzms_bit_0_cost(match_state, c->probs.match) +
+ lzms_bit_1_cost(lz_state, c->probs.lz) +
+ lzms_bit_0_cost(lz_rep0_state, c->probs.lz_rep[0]) +
lzms_fast_length_cost(c, rep0_len);
const u32 total_len = rep_len + 1 + rep0_len;
u32 base_cost = cur_node->cost +
lzms_bit_1_cost(cur_node->state.main_state,
- c->main_probs) +
+ c->probs.main) +
lzms_bit_1_cost(cur_node->state.match_state,
- c->match_probs) +
+ c->probs.match) +
lzms_bit_1_cost(cur_node->state.delta_state,
- c->delta_probs);
+ c->probs.delta);
for (int i = 0; i < rep_idx; i++)
base_cost += lzms_bit_1_cost(cur_node->state.delta_rep_states[i],
- c->delta_rep_probs[i]);
+ c->probs.delta_rep[i]);
if (rep_idx < LZMS_NUM_DELTA_REP_DECISIONS)
base_cost += lzms_bit_0_cost(cur_node->state.delta_rep_states[rep_idx],
- c->delta_rep_probs[rep_idx]);
+ c->probs.delta_rep[rep_idx]);
u32 len = 2;
do {
u32 base_cost = cur_node->cost +
lzms_bit_1_cost(cur_node->state.main_state,
- c->main_probs) +
+ c->probs.main) +
lzms_bit_0_cost(cur_node->state.match_state,
- c->match_probs) +
+ c->probs.match) +
lzms_bit_0_cost(cur_node->state.lz_state,
- c->lz_probs);
+ c->probs.lz);
if (c->try_lzmatch_lit_lzrep0 &&
likely(in_end - (in_next + c->matches[0].length) >= 3))
main_state = ((main_state << 1) | 0) % LZMS_NUM_MAIN_PROBS;
/* add LZ-rep0 cost */
- cost += lzms_bit_1_cost(main_state, c->main_probs) +
- lzms_bit_0_cost(match_state, c->match_probs) +
- lzms_bit_1_cost(lz_state, c->lz_probs) +
+ cost += lzms_bit_1_cost(main_state, c->probs.main) +
+ lzms_bit_0_cost(match_state, c->probs.match) +
+ lzms_bit_1_cost(lz_state, c->probs.lz) +
lzms_bit_0_cost(cur_node->state.lz_rep_states[0],
- c->lz_rep_probs[0]) +
+ c->probs.lz_rep[0]) +
lzms_fast_length_cost(c, rep0_len);
const u32 total_len = len + 1 + rep0_len;
const u32 pos = in_next - c->in_buffer;
/* Consider each possible power (log2 of span) */
- BUILD_BUG_ON(NUM_POWERS_TO_CONSIDER > LZMS_NUM_DELTA_POWER_SYMS);
+ STATIC_ASSERT(NUM_POWERS_TO_CONSIDER <= LZMS_NUM_DELTA_POWER_SYMS);
for (u32 power = 0; power < NUM_POWERS_TO_CONSIDER; power++) {
const u32 span = (u32)1 << power;
c->delta_hash_table[hash] = (power << DELTA_SOURCE_POWER_SHIFT) | pos;
c->next_delta_hashes[power] = next_hash;
- prefetch(&c->delta_hash_table[next_hash]);
+ prefetchw(&c->delta_hash_table[next_hash]);
if (power != cur_match >> DELTA_SOURCE_POWER_SHIFT)
continue;
/* Check the first 3 bytes before entering the
* extension loop. */
- BUILD_BUG_ON(NBYTES_HASHED_FOR_DELTA != 3);
+ STATIC_ASSERT(NBYTES_HASHED_FOR_DELTA == 3);
if (((u8)(*(in_next + 0) - *(in_next + 0 - span)) !=
(u8)(*(matchptr + 0) - *(matchptr + 0 - span))) ||
((u8)(*(in_next + 1) - *(in_next + 1 - span)) !=
u32 base_cost = cur_node->cost +
lzms_bit_1_cost(cur_node->state.main_state,
- c->main_probs) +
+ c->probs.main) +
lzms_bit_1_cost(cur_node->state.match_state,
- c->match_probs) +
+ c->probs.match) +
lzms_bit_0_cost(cur_node->state.delta_state,
- c->delta_probs) +
+ c->probs.delta) +
lzms_delta_source_cost(c, power, raw_offset);
u32 l = NBYTES_HASHED_FOR_DELTA;
/* Add cost of LZ-rep0 */
const u32 cost = cur_and_lit_cost +
- lzms_bit_1_cost(main_state, c->main_probs) +
+ lzms_bit_1_cost(main_state, c->probs.main) +
lzms_bit_0_cost(cur_node->state.match_state,
- c->match_probs) +
+ c->probs.match) +
lzms_bit_1_cost(cur_node->state.lz_state,
- c->lz_probs) +
+ c->probs.lz) +
lzms_bit_0_cost(cur_node->state.lz_rep_states[0],
- c->lz_rep_probs[0]) +
+ c->probs.lz_rep[0]) +
lzms_fast_length_cost(c, rep0_len);
const u32 total_len = 1 + rep0_len;
for (int i = 0; i < LZMS_NUM_DELTA_REP_DECISIONS; i++)
c->delta_rep_states[i] = 0;
- lzms_init_probability_entries(c->main_probs, LZMS_NUM_MAIN_PROBS);
- lzms_init_probability_entries(c->match_probs, LZMS_NUM_MATCH_PROBS);
- lzms_init_probability_entries(c->lz_probs, LZMS_NUM_LZ_PROBS);
- for (int i = 0; i < LZMS_NUM_LZ_REP_DECISIONS; i++)
- lzms_init_probability_entries(c->lz_rep_probs[i], LZMS_NUM_LZ_REP_PROBS);
- lzms_init_probability_entries(c->delta_probs, LZMS_NUM_DELTA_PROBS);
- for (int i = 0; i < LZMS_NUM_DELTA_REP_DECISIONS; i++)
- lzms_init_probability_entries(c->delta_rep_probs[i], LZMS_NUM_DELTA_REP_PROBS);
+ lzms_init_probabilities(&c->probs);
}
static void
}
static size_t
-lzms_compress(const void *in, size_t in_nbytes,
- void *out, size_t out_nbytes_avail, void *_c)
+lzms_compress(const void *restrict in, size_t in_nbytes,
+ void *restrict out, size_t out_nbytes_avail, void *restrict _c)
{
struct lzms_compressor *c = _c;
size_t result;