ref: d3a1ebad5226e2510ca65ef02e148967511df7b3
dir: /src/env.h/
/* * Copyright © 2018, VideoLAN and dav1d authors * Copyright © 2018, Two Orioles, LLC * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, this * list of conditions and the following disclaimer. * * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #ifndef __DAV1D_SRC_ENV_H__ #define __DAV1D_SRC_ENV_H__ #include <assert.h> #include <stddef.h> #include <stdint.h> #include <stdlib.h> #include "src/levels.h" #include "src/ref_mvs.h" #include "src/tables.h" typedef struct BlockContext { uint8_t ALIGN(mode[32], 8); uint8_t ALIGN(lcoef[32], 8); uint8_t ALIGN(ccoef[2][32], 8); uint8_t ALIGN(seg_pred[32], 8); uint8_t ALIGN(skip[32], 8); uint8_t ALIGN(skip_mode[32], 8); uint8_t ALIGN(intra[32], 8); uint8_t ALIGN(comp_type[32], 8); int8_t ALIGN(ref[2][32], 8); // -1 means intra uint8_t ALIGN(filter[2][32], 8); // 3 means unset int8_t ALIGN(tx_intra[32], 8); int8_t ALIGN(tx[32], 8); uint8_t ALIGN(tx_lpf_y[32], 8); uint8_t ALIGN(tx_lpf_uv[32], 8); uint8_t ALIGN(partition[16], 8); uint8_t ALIGN(uvmode[32], 8); uint8_t ALIGN(pal_sz[32], 8); } BlockContext; static inline int get_intra_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { if (have_left) { if (have_top) { const int ctx = l->intra[yb4] + a->intra[xb4]; return ctx + (ctx == 2); } else return l->intra[yb4] * 2; } else { return have_top ? a->intra[xb4] * 2 : 0; } } static inline int get_tx_ctx(const BlockContext *const a, const BlockContext *const l, const TxfmInfo *const max_tx, const int yb4, const int xb4) { return (l->tx_intra[yb4] >= max_tx->lh) + (a->tx_intra[xb4] >= max_tx->lw); } static inline int get_partition_ctx(const BlockContext *const a, const BlockContext *const l, const enum BlockLevel bl, const int yb8, const int xb8) { return ((a->partition[xb8] >> (4 - bl)) & 1) + (((l->partition[yb8] >> (4 - bl)) & 1) << 1); } static inline unsigned cdf_element_prob(const uint16_t *const cdf, const int e) { assert(e > 0); return cdf[e - 1] - cdf[e]; } static inline unsigned gather_left_partition_prob(const uint16_t *const in, const enum BlockLevel bl) { unsigned out = 0; out += cdf_element_prob(in, PARTITION_H); if (bl != BL_128X128) out += cdf_element_prob(in, PARTITION_H4); // Exploit the fact that cdfs for PARTITION_SPLIT, PARTITION_T_TOP_SPLIT, // PARTITION_T_BOTTOM_SPLIT and PARTITION_T_LEFT_SPLIT are neighbors. out += in[PARTITION_SPLIT - 1] - in[PARTITION_T_LEFT_SPLIT]; return out; } static inline unsigned gather_top_partition_prob(const uint16_t *const in, const enum BlockLevel bl) { unsigned out = 0; if (bl != BL_128X128) out += cdf_element_prob(in, PARTITION_V4); // Exploit the fact that cdfs for PARTITION_T_LEFT_SPLIT and PARTITION_T_RIGHT_SPLIT, // and PARTITION_V, PARTITION_SPLIT and PARTITION_T_TOP_SPLIT are neighbors. out += in[PARTITION_T_LEFT_SPLIT - 1] - in[PARTITION_T_RIGHT_SPLIT]; out += in[PARTITION_V - 1] - in[PARTITION_T_TOP_SPLIT]; return out; } static inline enum TxfmTypeSet get_ext_txtp_set(const enum RectTxfmSize tx, const int inter, const Dav1dFrameHeader *const hdr, const int seg_id) { if (!hdr->segmentation.qidx[seg_id]) { if (hdr->segmentation.lossless[seg_id]) { assert(tx == (int) TX_4X4); return TXTP_SET_LOSSLESS; } else { return TXTP_SET_DCT; } } const TxfmInfo *const t_dim = &dav1d_txfm_dimensions[tx]; if (t_dim->max >= TX_64X64) return TXTP_SET_DCT; if (t_dim->max == TX_32X32) return inter ? TXTP_SET_DCT_ID : TXTP_SET_DCT; if (hdr->reduced_txtp_set) return inter ? TXTP_SET_DCT_ID : TXTP_SET_DT4_ID; const enum TxfmSize txsqsz = t_dim->min; if (inter) return txsqsz == TX_16X16 ? TXTP_SET_DT9_ID_1D : TXTP_SET_ALL; else return txsqsz == TX_16X16 ? TXTP_SET_DT4_ID : TXTP_SET_DT4_ID_1D; } static inline enum TxfmType get_uv_intra_txtp(const enum IntraPredMode uv_mode, const enum RectTxfmSize tx, const Dav1dFrameHeader *const hdr, const int seg_id) { if (hdr->segmentation.lossless[seg_id]) { assert(tx == (int) TX_4X4); return WHT_WHT; } const TxfmInfo *const t_dim = &dav1d_txfm_dimensions[tx]; return t_dim->max == TX_32X32 ? DCT_DCT : dav1d_txtp_from_uvmode[uv_mode]; } static inline enum TxfmType get_uv_inter_txtp(const TxfmInfo *const uvt_dim, const enum TxfmType ytxtp, const Dav1dFrameHeader *const hdr, const int seg_id) { if (hdr->segmentation.lossless[seg_id]) { assert(uvt_dim->max == TX_4X4); return WHT_WHT; } if (uvt_dim->max == TX_32X32) return ytxtp == IDTX ? IDTX : DCT_DCT; if (uvt_dim->min == TX_16X16 && ((1 << ytxtp) & ((1 << H_FLIPADST) | (1 << V_FLIPADST) | (1 << H_ADST) | (1 << V_ADST)))) { return DCT_DCT; } return ytxtp; } static inline int get_filter_ctx(const BlockContext *const a, const BlockContext *const l, const int comp, const int dir, const int ref, const int yb4, const int xb4) { const int a_filter = (a->ref[0][xb4] == ref || a->ref[1][xb4] == ref) ? a->filter[dir][xb4] : DAV1D_N_SWITCHABLE_FILTERS; const int l_filter = (l->ref[0][yb4] == ref || l->ref[1][yb4] == ref) ? l->filter[dir][yb4] : DAV1D_N_SWITCHABLE_FILTERS; if (a_filter == l_filter) { return comp * 4 + a_filter; } else if (a_filter == DAV1D_N_SWITCHABLE_FILTERS) { return comp * 4 + l_filter; } else if (l_filter == DAV1D_N_SWITCHABLE_FILTERS) { return comp * 4 + a_filter; } else { return comp * 4 + DAV1D_N_SWITCHABLE_FILTERS; } } static inline int get_comp_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { if (have_top) { if (have_left) { if (a->comp_type[xb4]) { if (l->comp_type[yb4]) { return 4; } else { // 4U means intra (-1) or bwd (>= 4) return 2 + ((unsigned)l->ref[0][yb4] >= 4U); } } else if (l->comp_type[yb4]) { // 4U means intra (-1) or bwd (>= 4) return 2 + ((unsigned)a->ref[0][xb4] >= 4U); } else { return (l->ref[0][yb4] >= 4) ^ (a->ref[0][xb4] >= 4); } } else { return a->comp_type[xb4] ? 3 : a->ref[0][xb4] >= 4; } } else if (have_left) { return l->comp_type[yb4] ? 3 : l->ref[0][yb4] >= 4; } else { return 1; } } static inline int get_comp_dir_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { #define has_uni_comp(edge, off) \ ((edge->ref[0][off] < 4) == (edge->ref[1][off] < 4)) if (have_top && have_left) { const int a_intra = a->intra[xb4], l_intra = l->intra[yb4]; if (a_intra && l_intra) return 2; if (a_intra || l_intra) { const BlockContext *const edge = a_intra ? l : a; const int off = a_intra ? yb4 : xb4; if (edge->comp_type[off] == COMP_INTER_NONE) return 2; return 1 + 2 * has_uni_comp(edge, off); } const int a_comp = a->comp_type[xb4] != COMP_INTER_NONE; const int l_comp = l->comp_type[yb4] != COMP_INTER_NONE; const int a_ref0 = a->ref[0][xb4], l_ref0 = l->ref[0][yb4]; if (!a_comp && !l_comp) { return 1 + 2 * ((a_ref0 >= 4) == (l_ref0 >= 4)); } else if (!a_comp || !l_comp) { const BlockContext *const edge = a_comp ? a : l; const int off = a_comp ? xb4 : yb4; if (!has_uni_comp(edge, off)) return 1; return 3 + ((a_ref0 >= 4) == (l_ref0 >= 4)); } else { const int a_uni = has_uni_comp(a, xb4), l_uni = has_uni_comp(l, yb4); if (!a_uni && !l_uni) return 0; if (!a_uni || !l_uni) return 2; return 3 + ((a_ref0 == 4) == (l_ref0 == 4)); } } else if (have_top || have_left) { const BlockContext *const edge = have_left ? l : a; const int off = have_left ? yb4 : xb4; if (edge->intra[off]) return 2; if (edge->comp_type[off] == COMP_INTER_NONE) return 2; return 4 * has_uni_comp(edge, off); } else { return 2; } } static inline int get_poc_diff(const int order_hint_n_bits, const int poc0, const int poc1) { if (!order_hint_n_bits) return 0; const int mask = 1 << (order_hint_n_bits - 1); const int diff = poc0 - poc1; return (diff & (mask - 1)) - (diff & mask); } static inline int get_jnt_comp_ctx(const int order_hint_n_bits, const unsigned poc, const unsigned ref0poc, const unsigned ref1poc, const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4) { const unsigned d0 = abs(get_poc_diff(order_hint_n_bits, ref0poc, poc)); const unsigned d1 = abs(get_poc_diff(order_hint_n_bits, poc, ref1poc)); const int offset = d0 == d1; const int a_ctx = a->comp_type[xb4] >= COMP_INTER_AVG || a->ref[0][xb4] == 6; const int l_ctx = l->comp_type[yb4] >= COMP_INTER_AVG || l->ref[0][yb4] == 6; return 3 * offset + a_ctx + l_ctx; } static inline int get_mask_comp_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4) { const int a_ctx = a->comp_type[xb4] >= COMP_INTER_SEG ? 1 : a->ref[0][xb4] == 6 ? 3 : 0; const int l_ctx = l->comp_type[yb4] >= COMP_INTER_SEG ? 1 : l->ref[0][yb4] == 6 ? 3 : 0; return imin(a_ctx + l_ctx, 5); } #define av1_get_ref_2_ctx av1_get_bwd_ref_ctx #define av1_get_ref_3_ctx av1_get_fwd_ref_ctx #define av1_get_ref_4_ctx av1_get_fwd_ref_1_ctx #define av1_get_ref_5_ctx av1_get_fwd_ref_2_ctx #define av1_get_ref_6_ctx av1_get_bwd_ref_1_ctx #define av1_get_uni_p_ctx av1_get_ref_ctx #define av1_get_uni_p2_ctx av1_get_fwd_ref_2_ctx static inline int av1_get_ref_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, int have_top, int have_left) { int cnt[2] = { 0 }; if (have_top && !a->intra[xb4]) { cnt[a->ref[0][xb4] >= 4]++; if (a->comp_type[xb4]) cnt[a->ref[1][xb4] >= 4]++; } if (have_left && !l->intra[yb4]) { cnt[l->ref[0][yb4] >= 4]++; if (l->comp_type[yb4]) cnt[l->ref[1][yb4] >= 4]++; } return cnt[0] == cnt[1] ? 1 : cnt[0] < cnt[1] ? 0 : 2; } static inline int av1_get_fwd_ref_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { int cnt[4] = { 0 }; if (have_top && !a->intra[xb4]) { if (a->ref[0][xb4] < 4) cnt[a->ref[0][xb4]]++; if (a->comp_type[xb4] && a->ref[1][xb4] < 4) cnt[a->ref[1][xb4]]++; } if (have_left && !l->intra[yb4]) { if (l->ref[0][yb4] < 4) cnt[l->ref[0][yb4]]++; if (l->comp_type[yb4] && l->ref[1][yb4] < 4) cnt[l->ref[1][yb4]]++; } cnt[0] += cnt[1]; cnt[2] += cnt[3]; return cnt[0] == cnt[2] ? 1 : cnt[0] < cnt[2] ? 0 : 2; } static inline int av1_get_fwd_ref_1_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { int cnt[2] = { 0 }; if (have_top && !a->intra[xb4]) { if (a->ref[0][xb4] < 2) cnt[a->ref[0][xb4]]++; if (a->comp_type[xb4] && a->ref[1][xb4] < 2) cnt[a->ref[1][xb4]]++; } if (have_left && !l->intra[yb4]) { if (l->ref[0][yb4] < 2) cnt[l->ref[0][yb4]]++; if (l->comp_type[yb4] && l->ref[1][yb4] < 2) cnt[l->ref[1][yb4]]++; } return cnt[0] == cnt[1] ? 1 : cnt[0] < cnt[1] ? 0 : 2; } static inline int av1_get_fwd_ref_2_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { int cnt[2] = { 0 }; if (have_top && !a->intra[xb4]) { if ((a->ref[0][xb4] ^ 2U) < 2) cnt[a->ref[0][xb4] - 2]++; if (a->comp_type[xb4] && (a->ref[1][xb4] ^ 2U) < 2) cnt[a->ref[1][xb4] - 2]++; } if (have_left && !l->intra[yb4]) { if ((l->ref[0][yb4] ^ 2U) < 2) cnt[l->ref[0][yb4] - 2]++; if (l->comp_type[yb4] && (l->ref[1][yb4] ^ 2U) < 2) cnt[l->ref[1][yb4] - 2]++; } return cnt[0] == cnt[1] ? 1 : cnt[0] < cnt[1] ? 0 : 2; } static inline int av1_get_bwd_ref_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { int cnt[3] = { 0 }; if (have_top && !a->intra[xb4]) { if (a->ref[0][xb4] >= 4) cnt[a->ref[0][xb4] - 4]++; if (a->comp_type[xb4] && a->ref[1][xb4] >= 4) cnt[a->ref[1][xb4] - 4]++; } if (have_left && !l->intra[yb4]) { if (l->ref[0][yb4] >= 4) cnt[l->ref[0][yb4] - 4]++; if (l->comp_type[yb4] && l->ref[1][yb4] >= 4) cnt[l->ref[1][yb4] - 4]++; } cnt[1] += cnt[0]; return cnt[2] == cnt[1] ? 1 : cnt[1] < cnt[2] ? 0 : 2; } static inline int av1_get_bwd_ref_1_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { int cnt[3] = { 0 }; if (have_top && !a->intra[xb4]) { if (a->ref[0][xb4] >= 4) cnt[a->ref[0][xb4] - 4]++; if (a->comp_type[xb4] && a->ref[1][xb4] >= 4) cnt[a->ref[1][xb4] - 4]++; } if (have_left && !l->intra[yb4]) { if (l->ref[0][yb4] >= 4) cnt[l->ref[0][yb4] - 4]++; if (l->comp_type[yb4] && l->ref[1][yb4] >= 4) cnt[l->ref[1][yb4] - 4]++; } return cnt[0] == cnt[1] ? 1 : cnt[0] < cnt[1] ? 0 : 2; } static inline int av1_get_uni_p1_ctx(const BlockContext *const a, const BlockContext *const l, const int yb4, const int xb4, const int have_top, const int have_left) { int cnt[3] = { 0 }; if (have_top && !a->intra[xb4]) { if (a->ref[0][xb4] - 1U < 3) cnt[a->ref[0][xb4] - 1]++; if (a->comp_type[xb4] && a->ref[1][xb4] - 1U < 3) cnt[a->ref[1][xb4] - 1]++; } if (have_left && !l->intra[yb4]) { if (l->ref[0][yb4] - 1U < 3) cnt[l->ref[0][yb4] - 1]++; if (l->comp_type[yb4] && l->ref[1][yb4] - 1U < 3) cnt[l->ref[1][yb4] - 1]++; } cnt[1] += cnt[2]; return cnt[0] == cnt[1] ? 1 : cnt[0] < cnt[1] ? 0 : 2; } static inline int get_drl_context(const candidate_mv *const ref_mv_stack, const int ref_idx) { if (ref_mv_stack[ref_idx].weight >= 640 && ref_mv_stack[ref_idx + 1].weight >= 640) return 0; if (ref_mv_stack[ref_idx].weight >= 640 && ref_mv_stack[ref_idx + 1].weight < 640) return 1; if (ref_mv_stack[ref_idx].weight < 640 && ref_mv_stack[ref_idx + 1].weight < 640) return 2; return 0; } static inline unsigned get_cur_frame_segid(const int by, const int bx, const int have_top, const int have_left, int *const seg_ctx, const uint8_t *cur_seg_map, const ptrdiff_t stride) { cur_seg_map += bx + by * stride; if (have_left && have_top) { const int l = cur_seg_map[-1]; const int a = cur_seg_map[-stride]; const int al = cur_seg_map[-(stride + 1)]; if (l == a && al == l) *seg_ctx = 2; else if (l == a || al == l || a == al) *seg_ctx = 1; else *seg_ctx = 0; return a == al ? a : l; } else { *seg_ctx = 0; return have_left ? cur_seg_map[-1] : have_top ? cur_seg_map[-stride] : 0; } } static inline int get_coef_skip_ctx(const TxfmInfo *const t_dim, const enum BlockSize bs, const uint8_t *const a, const uint8_t *const l, const int chroma, const enum Dav1dPixelLayout layout) { const uint8_t *const b_dim = dav1d_block_dimensions[bs]; if (chroma) { const int ss_ver = layout == DAV1D_PIXEL_LAYOUT_I420; const int ss_hor = layout != DAV1D_PIXEL_LAYOUT_I444; const int not_one_blk = b_dim[2] - (!!b_dim[2] && ss_hor) > t_dim->lw || b_dim[3] - (!!b_dim[3] && ss_ver) > t_dim->lh; int ca, cl; #define MERGE_CTX(dir, type, mask) \ c##dir = !!((*(const type *) dir) & mask); \ break switch (t_dim->lw) { case TX_4X4: MERGE_CTX(a, uint8_t, 0x3F); case TX_8X8: MERGE_CTX(a, uint16_t, 0x3F3F); case TX_16X16: MERGE_CTX(a, uint32_t, 0x3F3F3F3FU); case TX_32X32: MERGE_CTX(a, uint64_t, 0x3F3F3F3F3F3F3F3FULL); default: abort(); } switch (t_dim->lh) { case TX_4X4: MERGE_CTX(l, uint8_t, 0x3F); case TX_8X8: MERGE_CTX(l, uint16_t, 0x3F3F); case TX_16X16: MERGE_CTX(l, uint32_t, 0x3F3F3F3FU); case TX_32X32: MERGE_CTX(l, uint64_t, 0x3F3F3F3F3F3F3F3FULL); default: abort(); } #undef MERGE_CTX return 7 + not_one_blk * 3 + ca + cl; } else if (b_dim[2] == t_dim->lw && b_dim[3] == t_dim->lh) { return 0; } else { static const uint8_t skip_contexts[5][5] = { { 1, 2, 2, 2, 3 }, { 1, 4, 4, 4, 5 }, { 1, 4, 4, 4, 5 }, { 1, 4, 4, 4, 5 }, { 1, 4, 4, 4, 6 } }; uint64_t la, ll; #define MERGE_CTX(dir, type, tx) do { \ l##dir = *(const type *) dir; \ if (tx == TX_64X64) \ l##dir |= *(const type *) &dir[sizeof(type)]; \ if (tx >= TX_32X32) l##dir |= l##dir >> 32; \ if (tx >= TX_16X16) l##dir |= l##dir >> 16; \ if (tx >= TX_8X8) l##dir |= l##dir >> 8; \ l##dir &= 0x3F; \ } while (0); \ break switch (t_dim->lw) { case TX_4X4: MERGE_CTX(a, uint8_t, TX_4X4); case TX_8X8: MERGE_CTX(a, uint16_t, TX_8X8); case TX_16X16: MERGE_CTX(a, uint32_t, TX_16X16); case TX_32X32: MERGE_CTX(a, uint64_t, TX_32X32); case TX_64X64: MERGE_CTX(a, uint64_t, TX_64X64); } switch (t_dim->lh) { case TX_4X4: MERGE_CTX(l, uint8_t, TX_4X4); case TX_8X8: MERGE_CTX(l, uint16_t, TX_8X8); case TX_16X16: MERGE_CTX(l, uint32_t, TX_16X16); case TX_32X32: MERGE_CTX(l, uint64_t, TX_32X32); case TX_64X64: MERGE_CTX(l, uint64_t, TX_64X64); } #undef MERGE_CTX const int max = imin(la | ll, 4); const int min = imin(imin(la, ll), 4); return skip_contexts[min][max]; } } static inline int get_coef_nz_ctx(uint8_t *const levels, const int scan_idx, const int rc, const int is_eob, const enum RectTxfmSize tx, const enum TxClass tx_class) { const TxfmInfo *const t_dim = &dav1d_txfm_dimensions[tx]; if (is_eob) { if (scan_idx == 0) return 0; const int eighth_sz = imin(t_dim->w, 8) * imin(t_dim->h, 8) * 2; if (scan_idx <= eighth_sz) return 1; const int quart_sz = eighth_sz * 2; if (scan_idx <= quart_sz) return 2; return 3; } const int x = rc >> (2 + imin(t_dim->lh, 3)); const int y = rc & (4 * imin(t_dim->h, 8) - 1); const ptrdiff_t stride = 4 * (imin(t_dim->h, 8) + 1); static const uint8_t offsets[3][5][2 /* x, y */] = { [TX_CLASS_2D] = { { 0, 1 }, { 1, 0 }, { 2, 0 }, { 0, 2 }, { 1, 1 } }, [TX_CLASS_V] = { { 0, 1 }, { 1, 0 }, { 0, 2 }, { 0, 3 }, { 0, 4 } }, [TX_CLASS_H] = { { 0, 1 }, { 1, 0 }, { 2, 0 }, { 3, 0 }, { 4, 0 } } }; const uint8_t (*const off)[2] = offsets[tx_class]; int mag = 0; for (int i = 0; i < 5; i++) mag += imin(levels[(x + off[i][0]) * stride + (y + off[i][1])], 3); const int ctx = imin((mag + 1) >> 1, 4); if (tx_class == TX_CLASS_2D) { return !rc ? 0 : dav1d_nz_map_ctx_offset[tx][imin(y, 4)][imin(x, 4)] + ctx; } else { return 26 + imin((tx_class == TX_CLASS_V) ? y : x, 2) * 5 + ctx; } } static inline int get_dc_sign_ctx(const TxfmInfo *const t_dim, const uint8_t *const a, const uint8_t *const l) { uint64_t sa, sl; #define MERGE_CTX(dir, type, tx, mask) do { \ s##dir = ((*(const type *) dir) >> 6) & mask; \ if (tx == TX_64X64) \ s##dir += ((*(const type *) &dir[sizeof(type)]) >> 6) & mask; \ if (tx >= TX_32X32) s##dir += s##dir >> 32; \ if (tx >= TX_16X16) s##dir += s##dir >> 16; \ if (tx >= TX_8X8) s##dir += s##dir >> 8; \ } while (0); \ break switch (t_dim->lw) { case TX_4X4: MERGE_CTX(a, uint8_t, TX_4X4, 0x03); case TX_8X8: MERGE_CTX(a, uint16_t, TX_8X8, 0x0303); case TX_16X16: MERGE_CTX(a, uint32_t, TX_16X16, 0x03030303U); case TX_32X32: MERGE_CTX(a, uint64_t, TX_32X32, 0x0303030303030303ULL); case TX_64X64: MERGE_CTX(a, uint64_t, TX_64X64, 0x0303030303030303ULL); } switch (t_dim->lh) { case TX_4X4: MERGE_CTX(l, uint8_t, TX_4X4, 0x03); case TX_8X8: MERGE_CTX(l, uint16_t, TX_8X8, 0x0303); case TX_16X16: MERGE_CTX(l, uint32_t, TX_16X16, 0x03030303U); case TX_32X32: MERGE_CTX(l, uint64_t, TX_32X32, 0x0303030303030303ULL); case TX_64X64: MERGE_CTX(l, uint64_t, TX_64X64, 0x0303030303030303ULL); } #undef MERGE_CTX const int s = ((int) ((sa + sl) & 0xFF)) - (t_dim->w + t_dim->h); return s < 0 ? 1 : s > 0 ? 2 : 0; } static inline int get_br_ctx(const uint8_t *const levels, const int rc, const enum RectTxfmSize tx, const enum TxClass tx_class) { const TxfmInfo *const t_dim = &dav1d_txfm_dimensions[tx]; const int x = rc >> (imin(t_dim->lh, 3) + 2); const int y = rc & (4 * imin(t_dim->h, 8) - 1); const int stride = 4 * (imin(t_dim->h, 8) + 1); int mag = 0; static const uint8_t offsets_from_txclass[3][3][2] = { [TX_CLASS_2D] = { { 0, 1 }, { 1, 0 }, { 1, 1 } }, [TX_CLASS_H] = { { 0, 1 }, { 1, 0 }, { 0, 2 } }, [TX_CLASS_V] = { { 0, 1 }, { 1, 0 }, { 2, 0 } } }; const uint8_t (*const offsets)[2] = offsets_from_txclass[tx_class]; for (int i = 0; i < 3; i++) mag += levels[(x + offsets[i][1]) * stride + y + offsets[i][0]]; mag = imin((mag + 1) >> 1, 6); if (rc == 0) return mag; switch (tx_class) { case TX_CLASS_2D: if (y < 2 && x < 2) return mag + 7; break; case TX_CLASS_H: if (x == 0) return mag + 7; break; case TX_CLASS_V: if (y == 0) return mag + 7; break; } return mag + 14; } static inline mv get_gmv_2d(const Dav1dWarpedMotionParams *const gmv, const int bx4, const int by4, const int bw4, const int bh4, const Dav1dFrameHeader *const hdr) { switch (gmv->type) { case DAV1D_WM_TYPE_ROT_ZOOM: assert(gmv->matrix[5] == gmv->matrix[2]); assert(gmv->matrix[4] == -gmv->matrix[3]); // fall-through default: case DAV1D_WM_TYPE_AFFINE: { const int x = bx4 * 4 + bw4 * 2 - 1; const int y = by4 * 4 + bh4 * 2 - 1; const int xc = (gmv->matrix[2] - (1 << 16)) * x + gmv->matrix[3] * y + gmv->matrix[0]; const int yc = (gmv->matrix[5] - (1 << 16)) * y + gmv->matrix[4] * x + gmv->matrix[1]; const int shift = 16 - (3 - !hdr->hp); const int round = (1 << shift) >> 1; return (mv) { .y = apply_sign(((abs(yc) + round) >> shift) << !hdr->hp, yc), .x = apply_sign(((abs(xc) + round) >> shift) << !hdr->hp, xc), }; } case DAV1D_WM_TYPE_TRANSLATION: return (mv) { .y = gmv->matrix[0] >> 13, .x = gmv->matrix[1] >> 13, }; case DAV1D_WM_TYPE_IDENTITY: return (mv) { .x = 0, .y = 0 }; } } #endif /* __DAV1D_SRC_ENV_H__ */