1*b2055c35SXin Li // Copyright 2011 Google Inc. All Rights Reserved.
2*b2055c35SXin Li //
3*b2055c35SXin Li // Use of this source code is governed by a BSD-style license
4*b2055c35SXin Li // that can be found in the COPYING file in the root of the source
5*b2055c35SXin Li // tree. An additional intellectual property rights grant can be found
6*b2055c35SXin Li // in the file PATENTS. All contributing project authors may
7*b2055c35SXin Li // be found in the AUTHORS file in the root of the source tree.
8*b2055c35SXin Li // -----------------------------------------------------------------------------
9*b2055c35SXin Li //
10*b2055c35SXin Li // Quantization
11*b2055c35SXin Li //
12*b2055c35SXin Li // Author: Skal ([email protected])
13*b2055c35SXin Li
14*b2055c35SXin Li #include <assert.h>
15*b2055c35SXin Li #include <math.h>
16*b2055c35SXin Li #include <stdlib.h> // for abs()
17*b2055c35SXin Li
18*b2055c35SXin Li #include "src/dsp/quant.h"
19*b2055c35SXin Li #include "src/enc/vp8i_enc.h"
20*b2055c35SXin Li #include "src/enc/cost_enc.h"
21*b2055c35SXin Li
22*b2055c35SXin Li #define DO_TRELLIS_I4 1
23*b2055c35SXin Li #define DO_TRELLIS_I16 1 // not a huge gain, but ok at low bitrate.
24*b2055c35SXin Li #define DO_TRELLIS_UV 0 // disable trellis for UV. Risky. Not worth.
25*b2055c35SXin Li #define USE_TDISTO 1
26*b2055c35SXin Li
27*b2055c35SXin Li #define MID_ALPHA 64 // neutral value for susceptibility
28*b2055c35SXin Li #define MIN_ALPHA 30 // lowest usable value for susceptibility
29*b2055c35SXin Li #define MAX_ALPHA 100 // higher meaningful value for susceptibility
30*b2055c35SXin Li
31*b2055c35SXin Li #define SNS_TO_DQ 0.9 // Scaling constant between the sns value and the QP
32*b2055c35SXin Li // power-law modulation. Must be strictly less than 1.
33*b2055c35SXin Li
34*b2055c35SXin Li // number of non-zero coeffs below which we consider the block very flat
35*b2055c35SXin Li // (and apply a penalty to complex predictions)
36*b2055c35SXin Li #define FLATNESS_LIMIT_I16 0 // I16 mode (special case)
37*b2055c35SXin Li #define FLATNESS_LIMIT_I4 3 // I4 mode
38*b2055c35SXin Li #define FLATNESS_LIMIT_UV 2 // UV mode
39*b2055c35SXin Li #define FLATNESS_PENALTY 140 // roughly ~1bit per block
40*b2055c35SXin Li
41*b2055c35SXin Li #define MULT_8B(a, b) (((a) * (b) + 128) >> 8)
42*b2055c35SXin Li
43*b2055c35SXin Li #define RD_DISTO_MULT 256 // distortion multiplier (equivalent of lambda)
44*b2055c35SXin Li
45*b2055c35SXin Li // #define DEBUG_BLOCK
46*b2055c35SXin Li
47*b2055c35SXin Li //------------------------------------------------------------------------------
48*b2055c35SXin Li
49*b2055c35SXin Li #if defined(DEBUG_BLOCK)
50*b2055c35SXin Li
51*b2055c35SXin Li #include <stdio.h>
52*b2055c35SXin Li #include <stdlib.h>
53*b2055c35SXin Li
PrintBlockInfo(const VP8EncIterator * const it,const VP8ModeScore * const rd)54*b2055c35SXin Li static void PrintBlockInfo(const VP8EncIterator* const it,
55*b2055c35SXin Li const VP8ModeScore* const rd) {
56*b2055c35SXin Li int i, j;
57*b2055c35SXin Li const int is_i16 = (it->mb_->type_ == 1);
58*b2055c35SXin Li const uint8_t* const y_in = it->yuv_in_ + Y_OFF_ENC;
59*b2055c35SXin Li const uint8_t* const y_out = it->yuv_out_ + Y_OFF_ENC;
60*b2055c35SXin Li const uint8_t* const uv_in = it->yuv_in_ + U_OFF_ENC;
61*b2055c35SXin Li const uint8_t* const uv_out = it->yuv_out_ + U_OFF_ENC;
62*b2055c35SXin Li printf("SOURCE / OUTPUT / ABS DELTA\n");
63*b2055c35SXin Li for (j = 0; j < 16; ++j) {
64*b2055c35SXin Li for (i = 0; i < 16; ++i) printf("%3d ", y_in[i + j * BPS]);
65*b2055c35SXin Li printf(" ");
66*b2055c35SXin Li for (i = 0; i < 16; ++i) printf("%3d ", y_out[i + j * BPS]);
67*b2055c35SXin Li printf(" ");
68*b2055c35SXin Li for (i = 0; i < 16; ++i) {
69*b2055c35SXin Li printf("%1d ", abs(y_in[i + j * BPS] - y_out[i + j * BPS]));
70*b2055c35SXin Li }
71*b2055c35SXin Li printf("\n");
72*b2055c35SXin Li }
73*b2055c35SXin Li printf("\n"); // newline before the U/V block
74*b2055c35SXin Li for (j = 0; j < 8; ++j) {
75*b2055c35SXin Li for (i = 0; i < 8; ++i) printf("%3d ", uv_in[i + j * BPS]);
76*b2055c35SXin Li printf(" ");
77*b2055c35SXin Li for (i = 8; i < 16; ++i) printf("%3d ", uv_in[i + j * BPS]);
78*b2055c35SXin Li printf(" ");
79*b2055c35SXin Li for (i = 0; i < 8; ++i) printf("%3d ", uv_out[i + j * BPS]);
80*b2055c35SXin Li printf(" ");
81*b2055c35SXin Li for (i = 8; i < 16; ++i) printf("%3d ", uv_out[i + j * BPS]);
82*b2055c35SXin Li printf(" ");
83*b2055c35SXin Li for (i = 0; i < 8; ++i) {
84*b2055c35SXin Li printf("%1d ", abs(uv_out[i + j * BPS] - uv_in[i + j * BPS]));
85*b2055c35SXin Li }
86*b2055c35SXin Li printf(" ");
87*b2055c35SXin Li for (i = 8; i < 16; ++i) {
88*b2055c35SXin Li printf("%1d ", abs(uv_out[i + j * BPS] - uv_in[i + j * BPS]));
89*b2055c35SXin Li }
90*b2055c35SXin Li printf("\n");
91*b2055c35SXin Li }
92*b2055c35SXin Li printf("\nD:%d SD:%d R:%d H:%d nz:0x%x score:%d\n",
93*b2055c35SXin Li (int)rd->D, (int)rd->SD, (int)rd->R, (int)rd->H, (int)rd->nz,
94*b2055c35SXin Li (int)rd->score);
95*b2055c35SXin Li if (is_i16) {
96*b2055c35SXin Li printf("Mode: %d\n", rd->mode_i16);
97*b2055c35SXin Li printf("y_dc_levels:");
98*b2055c35SXin Li for (i = 0; i < 16; ++i) printf("%3d ", rd->y_dc_levels[i]);
99*b2055c35SXin Li printf("\n");
100*b2055c35SXin Li } else {
101*b2055c35SXin Li printf("Modes[16]: ");
102*b2055c35SXin Li for (i = 0; i < 16; ++i) printf("%d ", rd->modes_i4[i]);
103*b2055c35SXin Li printf("\n");
104*b2055c35SXin Li }
105*b2055c35SXin Li printf("y_ac_levels:\n");
106*b2055c35SXin Li for (j = 0; j < 16; ++j) {
107*b2055c35SXin Li for (i = is_i16 ? 1 : 0; i < 16; ++i) {
108*b2055c35SXin Li printf("%4d ", rd->y_ac_levels[j][i]);
109*b2055c35SXin Li }
110*b2055c35SXin Li printf("\n");
111*b2055c35SXin Li }
112*b2055c35SXin Li printf("\n");
113*b2055c35SXin Li printf("uv_levels (mode=%d):\n", rd->mode_uv);
114*b2055c35SXin Li for (j = 0; j < 8; ++j) {
115*b2055c35SXin Li for (i = 0; i < 16; ++i) {
116*b2055c35SXin Li printf("%4d ", rd->uv_levels[j][i]);
117*b2055c35SXin Li }
118*b2055c35SXin Li printf("\n");
119*b2055c35SXin Li }
120*b2055c35SXin Li }
121*b2055c35SXin Li
122*b2055c35SXin Li #endif // DEBUG_BLOCK
123*b2055c35SXin Li
124*b2055c35SXin Li //------------------------------------------------------------------------------
125*b2055c35SXin Li
clip(int v,int m,int M)126*b2055c35SXin Li static WEBP_INLINE int clip(int v, int m, int M) {
127*b2055c35SXin Li return v < m ? m : v > M ? M : v;
128*b2055c35SXin Li }
129*b2055c35SXin Li
130*b2055c35SXin Li static const uint8_t kZigzag[16] = {
131*b2055c35SXin Li 0, 1, 4, 8, 5, 2, 3, 6, 9, 12, 13, 10, 7, 11, 14, 15
132*b2055c35SXin Li };
133*b2055c35SXin Li
134*b2055c35SXin Li static const uint8_t kDcTable[128] = {
135*b2055c35SXin Li 4, 5, 6, 7, 8, 9, 10, 10,
136*b2055c35SXin Li 11, 12, 13, 14, 15, 16, 17, 17,
137*b2055c35SXin Li 18, 19, 20, 20, 21, 21, 22, 22,
138*b2055c35SXin Li 23, 23, 24, 25, 25, 26, 27, 28,
139*b2055c35SXin Li 29, 30, 31, 32, 33, 34, 35, 36,
140*b2055c35SXin Li 37, 37, 38, 39, 40, 41, 42, 43,
141*b2055c35SXin Li 44, 45, 46, 46, 47, 48, 49, 50,
142*b2055c35SXin Li 51, 52, 53, 54, 55, 56, 57, 58,
143*b2055c35SXin Li 59, 60, 61, 62, 63, 64, 65, 66,
144*b2055c35SXin Li 67, 68, 69, 70, 71, 72, 73, 74,
145*b2055c35SXin Li 75, 76, 76, 77, 78, 79, 80, 81,
146*b2055c35SXin Li 82, 83, 84, 85, 86, 87, 88, 89,
147*b2055c35SXin Li 91, 93, 95, 96, 98, 100, 101, 102,
148*b2055c35SXin Li 104, 106, 108, 110, 112, 114, 116, 118,
149*b2055c35SXin Li 122, 124, 126, 128, 130, 132, 134, 136,
150*b2055c35SXin Li 138, 140, 143, 145, 148, 151, 154, 157
151*b2055c35SXin Li };
152*b2055c35SXin Li
153*b2055c35SXin Li static const uint16_t kAcTable[128] = {
154*b2055c35SXin Li 4, 5, 6, 7, 8, 9, 10, 11,
155*b2055c35SXin Li 12, 13, 14, 15, 16, 17, 18, 19,
156*b2055c35SXin Li 20, 21, 22, 23, 24, 25, 26, 27,
157*b2055c35SXin Li 28, 29, 30, 31, 32, 33, 34, 35,
158*b2055c35SXin Li 36, 37, 38, 39, 40, 41, 42, 43,
159*b2055c35SXin Li 44, 45, 46, 47, 48, 49, 50, 51,
160*b2055c35SXin Li 52, 53, 54, 55, 56, 57, 58, 60,
161*b2055c35SXin Li 62, 64, 66, 68, 70, 72, 74, 76,
162*b2055c35SXin Li 78, 80, 82, 84, 86, 88, 90, 92,
163*b2055c35SXin Li 94, 96, 98, 100, 102, 104, 106, 108,
164*b2055c35SXin Li 110, 112, 114, 116, 119, 122, 125, 128,
165*b2055c35SXin Li 131, 134, 137, 140, 143, 146, 149, 152,
166*b2055c35SXin Li 155, 158, 161, 164, 167, 170, 173, 177,
167*b2055c35SXin Li 181, 185, 189, 193, 197, 201, 205, 209,
168*b2055c35SXin Li 213, 217, 221, 225, 229, 234, 239, 245,
169*b2055c35SXin Li 249, 254, 259, 264, 269, 274, 279, 284
170*b2055c35SXin Li };
171*b2055c35SXin Li
172*b2055c35SXin Li static const uint16_t kAcTable2[128] = {
173*b2055c35SXin Li 8, 8, 9, 10, 12, 13, 15, 17,
174*b2055c35SXin Li 18, 20, 21, 23, 24, 26, 27, 29,
175*b2055c35SXin Li 31, 32, 34, 35, 37, 38, 40, 41,
176*b2055c35SXin Li 43, 44, 46, 48, 49, 51, 52, 54,
177*b2055c35SXin Li 55, 57, 58, 60, 62, 63, 65, 66,
178*b2055c35SXin Li 68, 69, 71, 72, 74, 75, 77, 79,
179*b2055c35SXin Li 80, 82, 83, 85, 86, 88, 89, 93,
180*b2055c35SXin Li 96, 99, 102, 105, 108, 111, 114, 117,
181*b2055c35SXin Li 120, 124, 127, 130, 133, 136, 139, 142,
182*b2055c35SXin Li 145, 148, 151, 155, 158, 161, 164, 167,
183*b2055c35SXin Li 170, 173, 176, 179, 184, 189, 193, 198,
184*b2055c35SXin Li 203, 207, 212, 217, 221, 226, 230, 235,
185*b2055c35SXin Li 240, 244, 249, 254, 258, 263, 268, 274,
186*b2055c35SXin Li 280, 286, 292, 299, 305, 311, 317, 323,
187*b2055c35SXin Li 330, 336, 342, 348, 354, 362, 370, 379,
188*b2055c35SXin Li 385, 393, 401, 409, 416, 424, 432, 440
189*b2055c35SXin Li };
190*b2055c35SXin Li
191*b2055c35SXin Li static const uint8_t kBiasMatrices[3][2] = { // [luma-ac,luma-dc,chroma][dc,ac]
192*b2055c35SXin Li { 96, 110 }, { 96, 108 }, { 110, 115 }
193*b2055c35SXin Li };
194*b2055c35SXin Li
195*b2055c35SXin Li // Sharpening by (slightly) raising the hi-frequency coeffs.
196*b2055c35SXin Li // Hack-ish but helpful for mid-bitrate range. Use with care.
197*b2055c35SXin Li #define SHARPEN_BITS 11 // number of descaling bits for sharpening bias
198*b2055c35SXin Li static const uint8_t kFreqSharpening[16] = {
199*b2055c35SXin Li 0, 30, 60, 90,
200*b2055c35SXin Li 30, 60, 90, 90,
201*b2055c35SXin Li 60, 90, 90, 90,
202*b2055c35SXin Li 90, 90, 90, 90
203*b2055c35SXin Li };
204*b2055c35SXin Li
205*b2055c35SXin Li //------------------------------------------------------------------------------
206*b2055c35SXin Li // Initialize quantization parameters in VP8Matrix
207*b2055c35SXin Li
208*b2055c35SXin Li // Returns the average quantizer
ExpandMatrix(VP8Matrix * const m,int type)209*b2055c35SXin Li static int ExpandMatrix(VP8Matrix* const m, int type) {
210*b2055c35SXin Li int i, sum;
211*b2055c35SXin Li for (i = 0; i < 2; ++i) {
212*b2055c35SXin Li const int is_ac_coeff = (i > 0);
213*b2055c35SXin Li const int bias = kBiasMatrices[type][is_ac_coeff];
214*b2055c35SXin Li m->iq_[i] = (1 << QFIX) / m->q_[i];
215*b2055c35SXin Li m->bias_[i] = BIAS(bias);
216*b2055c35SXin Li // zthresh_ is the exact value such that QUANTDIV(coeff, iQ, B) is:
217*b2055c35SXin Li // * zero if coeff <= zthresh
218*b2055c35SXin Li // * non-zero if coeff > zthresh
219*b2055c35SXin Li m->zthresh_[i] = ((1 << QFIX) - 1 - m->bias_[i]) / m->iq_[i];
220*b2055c35SXin Li }
221*b2055c35SXin Li for (i = 2; i < 16; ++i) {
222*b2055c35SXin Li m->q_[i] = m->q_[1];
223*b2055c35SXin Li m->iq_[i] = m->iq_[1];
224*b2055c35SXin Li m->bias_[i] = m->bias_[1];
225*b2055c35SXin Li m->zthresh_[i] = m->zthresh_[1];
226*b2055c35SXin Li }
227*b2055c35SXin Li for (sum = 0, i = 0; i < 16; ++i) {
228*b2055c35SXin Li if (type == 0) { // we only use sharpening for AC luma coeffs
229*b2055c35SXin Li m->sharpen_[i] = (kFreqSharpening[i] * m->q_[i]) >> SHARPEN_BITS;
230*b2055c35SXin Li } else {
231*b2055c35SXin Li m->sharpen_[i] = 0;
232*b2055c35SXin Li }
233*b2055c35SXin Li sum += m->q_[i];
234*b2055c35SXin Li }
235*b2055c35SXin Li return (sum + 8) >> 4;
236*b2055c35SXin Li }
237*b2055c35SXin Li
CheckLambdaValue(int * const v)238*b2055c35SXin Li static void CheckLambdaValue(int* const v) { if (*v < 1) *v = 1; }
239*b2055c35SXin Li
SetupMatrices(VP8Encoder * enc)240*b2055c35SXin Li static void SetupMatrices(VP8Encoder* enc) {
241*b2055c35SXin Li int i;
242*b2055c35SXin Li const int tlambda_scale =
243*b2055c35SXin Li (enc->method_ >= 4) ? enc->config_->sns_strength
244*b2055c35SXin Li : 0;
245*b2055c35SXin Li const int num_segments = enc->segment_hdr_.num_segments_;
246*b2055c35SXin Li for (i = 0; i < num_segments; ++i) {
247*b2055c35SXin Li VP8SegmentInfo* const m = &enc->dqm_[i];
248*b2055c35SXin Li const int q = m->quant_;
249*b2055c35SXin Li int q_i4, q_i16, q_uv;
250*b2055c35SXin Li m->y1_.q_[0] = kDcTable[clip(q + enc->dq_y1_dc_, 0, 127)];
251*b2055c35SXin Li m->y1_.q_[1] = kAcTable[clip(q, 0, 127)];
252*b2055c35SXin Li
253*b2055c35SXin Li m->y2_.q_[0] = kDcTable[ clip(q + enc->dq_y2_dc_, 0, 127)] * 2;
254*b2055c35SXin Li m->y2_.q_[1] = kAcTable2[clip(q + enc->dq_y2_ac_, 0, 127)];
255*b2055c35SXin Li
256*b2055c35SXin Li m->uv_.q_[0] = kDcTable[clip(q + enc->dq_uv_dc_, 0, 117)];
257*b2055c35SXin Li m->uv_.q_[1] = kAcTable[clip(q + enc->dq_uv_ac_, 0, 127)];
258*b2055c35SXin Li
259*b2055c35SXin Li q_i4 = ExpandMatrix(&m->y1_, 0);
260*b2055c35SXin Li q_i16 = ExpandMatrix(&m->y2_, 1);
261*b2055c35SXin Li q_uv = ExpandMatrix(&m->uv_, 2);
262*b2055c35SXin Li
263*b2055c35SXin Li m->lambda_i4_ = (3 * q_i4 * q_i4) >> 7;
264*b2055c35SXin Li m->lambda_i16_ = (3 * q_i16 * q_i16);
265*b2055c35SXin Li m->lambda_uv_ = (3 * q_uv * q_uv) >> 6;
266*b2055c35SXin Li m->lambda_mode_ = (1 * q_i4 * q_i4) >> 7;
267*b2055c35SXin Li m->lambda_trellis_i4_ = (7 * q_i4 * q_i4) >> 3;
268*b2055c35SXin Li m->lambda_trellis_i16_ = (q_i16 * q_i16) >> 2;
269*b2055c35SXin Li m->lambda_trellis_uv_ = (q_uv * q_uv) << 1;
270*b2055c35SXin Li m->tlambda_ = (tlambda_scale * q_i4) >> 5;
271*b2055c35SXin Li
272*b2055c35SXin Li // none of these constants should be < 1
273*b2055c35SXin Li CheckLambdaValue(&m->lambda_i4_);
274*b2055c35SXin Li CheckLambdaValue(&m->lambda_i16_);
275*b2055c35SXin Li CheckLambdaValue(&m->lambda_uv_);
276*b2055c35SXin Li CheckLambdaValue(&m->lambda_mode_);
277*b2055c35SXin Li CheckLambdaValue(&m->lambda_trellis_i4_);
278*b2055c35SXin Li CheckLambdaValue(&m->lambda_trellis_i16_);
279*b2055c35SXin Li CheckLambdaValue(&m->lambda_trellis_uv_);
280*b2055c35SXin Li CheckLambdaValue(&m->tlambda_);
281*b2055c35SXin Li
282*b2055c35SXin Li m->min_disto_ = 20 * m->y1_.q_[0]; // quantization-aware min disto
283*b2055c35SXin Li m->max_edge_ = 0;
284*b2055c35SXin Li
285*b2055c35SXin Li m->i4_penalty_ = 1000 * q_i4 * q_i4;
286*b2055c35SXin Li }
287*b2055c35SXin Li }
288*b2055c35SXin Li
289*b2055c35SXin Li //------------------------------------------------------------------------------
290*b2055c35SXin Li // Initialize filtering parameters
291*b2055c35SXin Li
292*b2055c35SXin Li // Very small filter-strength values have close to no visual effect. So we can
293*b2055c35SXin Li // save a little decoding-CPU by turning filtering off for these.
294*b2055c35SXin Li #define FSTRENGTH_CUTOFF 2
295*b2055c35SXin Li
SetupFilterStrength(VP8Encoder * const enc)296*b2055c35SXin Li static void SetupFilterStrength(VP8Encoder* const enc) {
297*b2055c35SXin Li int i;
298*b2055c35SXin Li // level0 is in [0..500]. Using '-f 50' as filter_strength is mid-filtering.
299*b2055c35SXin Li const int level0 = 5 * enc->config_->filter_strength;
300*b2055c35SXin Li for (i = 0; i < NUM_MB_SEGMENTS; ++i) {
301*b2055c35SXin Li VP8SegmentInfo* const m = &enc->dqm_[i];
302*b2055c35SXin Li // We focus on the quantization of AC coeffs.
303*b2055c35SXin Li const int qstep = kAcTable[clip(m->quant_, 0, 127)] >> 2;
304*b2055c35SXin Li const int base_strength =
305*b2055c35SXin Li VP8FilterStrengthFromDelta(enc->filter_hdr_.sharpness_, qstep);
306*b2055c35SXin Li // Segments with lower complexity ('beta') will be less filtered.
307*b2055c35SXin Li const int f = base_strength * level0 / (256 + m->beta_);
308*b2055c35SXin Li m->fstrength_ = (f < FSTRENGTH_CUTOFF) ? 0 : (f > 63) ? 63 : f;
309*b2055c35SXin Li }
310*b2055c35SXin Li // We record the initial strength (mainly for the case of 1-segment only).
311*b2055c35SXin Li enc->filter_hdr_.level_ = enc->dqm_[0].fstrength_;
312*b2055c35SXin Li enc->filter_hdr_.simple_ = (enc->config_->filter_type == 0);
313*b2055c35SXin Li enc->filter_hdr_.sharpness_ = enc->config_->filter_sharpness;
314*b2055c35SXin Li }
315*b2055c35SXin Li
316*b2055c35SXin Li //------------------------------------------------------------------------------
317*b2055c35SXin Li
318*b2055c35SXin Li // Note: if you change the values below, remember that the max range
319*b2055c35SXin Li // allowed by the syntax for DQ_UV is [-16,16].
320*b2055c35SXin Li #define MAX_DQ_UV (6)
321*b2055c35SXin Li #define MIN_DQ_UV (-4)
322*b2055c35SXin Li
323*b2055c35SXin Li // We want to emulate jpeg-like behaviour where the expected "good" quality
324*b2055c35SXin Li // is around q=75. Internally, our "good" middle is around c=50. So we
325*b2055c35SXin Li // map accordingly using linear piece-wise function
QualityToCompression(double c)326*b2055c35SXin Li static double QualityToCompression(double c) {
327*b2055c35SXin Li const double linear_c = (c < 0.75) ? c * (2. / 3.) : 2. * c - 1.;
328*b2055c35SXin Li // The file size roughly scales as pow(quantizer, 3.). Actually, the
329*b2055c35SXin Li // exponent is somewhere between 2.8 and 3.2, but we're mostly interested
330*b2055c35SXin Li // in the mid-quant range. So we scale the compressibility inversely to
331*b2055c35SXin Li // this power-law: quant ~= compression ^ 1/3. This law holds well for
332*b2055c35SXin Li // low quant. Finer modeling for high-quant would make use of kAcTable[]
333*b2055c35SXin Li // more explicitly.
334*b2055c35SXin Li const double v = pow(linear_c, 1 / 3.);
335*b2055c35SXin Li return v;
336*b2055c35SXin Li }
337*b2055c35SXin Li
QualityToJPEGCompression(double c,double alpha)338*b2055c35SXin Li static double QualityToJPEGCompression(double c, double alpha) {
339*b2055c35SXin Li // We map the complexity 'alpha' and quality setting 'c' to a compression
340*b2055c35SXin Li // exponent empirically matched to the compression curve of libjpeg6b.
341*b2055c35SXin Li // On average, the WebP output size will be roughly similar to that of a
342*b2055c35SXin Li // JPEG file compressed with same quality factor.
343*b2055c35SXin Li const double amin = 0.30;
344*b2055c35SXin Li const double amax = 0.85;
345*b2055c35SXin Li const double exp_min = 0.4;
346*b2055c35SXin Li const double exp_max = 0.9;
347*b2055c35SXin Li const double slope = (exp_min - exp_max) / (amax - amin);
348*b2055c35SXin Li // Linearly interpolate 'expn' from exp_min to exp_max
349*b2055c35SXin Li // in the [amin, amax] range.
350*b2055c35SXin Li const double expn = (alpha > amax) ? exp_min
351*b2055c35SXin Li : (alpha < amin) ? exp_max
352*b2055c35SXin Li : exp_max + slope * (alpha - amin);
353*b2055c35SXin Li const double v = pow(c, expn);
354*b2055c35SXin Li return v;
355*b2055c35SXin Li }
356*b2055c35SXin Li
SegmentsAreEquivalent(const VP8SegmentInfo * const S1,const VP8SegmentInfo * const S2)357*b2055c35SXin Li static int SegmentsAreEquivalent(const VP8SegmentInfo* const S1,
358*b2055c35SXin Li const VP8SegmentInfo* const S2) {
359*b2055c35SXin Li return (S1->quant_ == S2->quant_) && (S1->fstrength_ == S2->fstrength_);
360*b2055c35SXin Li }
361*b2055c35SXin Li
SimplifySegments(VP8Encoder * const enc)362*b2055c35SXin Li static void SimplifySegments(VP8Encoder* const enc) {
363*b2055c35SXin Li int map[NUM_MB_SEGMENTS] = { 0, 1, 2, 3 };
364*b2055c35SXin Li // 'num_segments_' is previously validated and <= NUM_MB_SEGMENTS, but an
365*b2055c35SXin Li // explicit check is needed to avoid a spurious warning about 'i' exceeding
366*b2055c35SXin Li // array bounds of 'dqm_' with some compilers (noticed with gcc-4.9).
367*b2055c35SXin Li const int num_segments = (enc->segment_hdr_.num_segments_ < NUM_MB_SEGMENTS)
368*b2055c35SXin Li ? enc->segment_hdr_.num_segments_
369*b2055c35SXin Li : NUM_MB_SEGMENTS;
370*b2055c35SXin Li int num_final_segments = 1;
371*b2055c35SXin Li int s1, s2;
372*b2055c35SXin Li for (s1 = 1; s1 < num_segments; ++s1) { // find similar segments
373*b2055c35SXin Li const VP8SegmentInfo* const S1 = &enc->dqm_[s1];
374*b2055c35SXin Li int found = 0;
375*b2055c35SXin Li // check if we already have similar segment
376*b2055c35SXin Li for (s2 = 0; s2 < num_final_segments; ++s2) {
377*b2055c35SXin Li const VP8SegmentInfo* const S2 = &enc->dqm_[s2];
378*b2055c35SXin Li if (SegmentsAreEquivalent(S1, S2)) {
379*b2055c35SXin Li found = 1;
380*b2055c35SXin Li break;
381*b2055c35SXin Li }
382*b2055c35SXin Li }
383*b2055c35SXin Li map[s1] = s2;
384*b2055c35SXin Li if (!found) {
385*b2055c35SXin Li if (num_final_segments != s1) {
386*b2055c35SXin Li enc->dqm_[num_final_segments] = enc->dqm_[s1];
387*b2055c35SXin Li }
388*b2055c35SXin Li ++num_final_segments;
389*b2055c35SXin Li }
390*b2055c35SXin Li }
391*b2055c35SXin Li if (num_final_segments < num_segments) { // Remap
392*b2055c35SXin Li int i = enc->mb_w_ * enc->mb_h_;
393*b2055c35SXin Li while (i-- > 0) enc->mb_info_[i].segment_ = map[enc->mb_info_[i].segment_];
394*b2055c35SXin Li enc->segment_hdr_.num_segments_ = num_final_segments;
395*b2055c35SXin Li // Replicate the trailing segment infos (it's mostly cosmetics)
396*b2055c35SXin Li for (i = num_final_segments; i < num_segments; ++i) {
397*b2055c35SXin Li enc->dqm_[i] = enc->dqm_[num_final_segments - 1];
398*b2055c35SXin Li }
399*b2055c35SXin Li }
400*b2055c35SXin Li }
401*b2055c35SXin Li
VP8SetSegmentParams(VP8Encoder * const enc,float quality)402*b2055c35SXin Li void VP8SetSegmentParams(VP8Encoder* const enc, float quality) {
403*b2055c35SXin Li int i;
404*b2055c35SXin Li int dq_uv_ac, dq_uv_dc;
405*b2055c35SXin Li const int num_segments = enc->segment_hdr_.num_segments_;
406*b2055c35SXin Li const double amp = SNS_TO_DQ * enc->config_->sns_strength / 100. / 128.;
407*b2055c35SXin Li const double Q = quality / 100.;
408*b2055c35SXin Li const double c_base = enc->config_->emulate_jpeg_size ?
409*b2055c35SXin Li QualityToJPEGCompression(Q, enc->alpha_ / 255.) :
410*b2055c35SXin Li QualityToCompression(Q);
411*b2055c35SXin Li for (i = 0; i < num_segments; ++i) {
412*b2055c35SXin Li // We modulate the base coefficient to accommodate for the quantization
413*b2055c35SXin Li // susceptibility and allow denser segments to be quantized more.
414*b2055c35SXin Li const double expn = 1. - amp * enc->dqm_[i].alpha_;
415*b2055c35SXin Li const double c = pow(c_base, expn);
416*b2055c35SXin Li const int q = (int)(127. * (1. - c));
417*b2055c35SXin Li assert(expn > 0.);
418*b2055c35SXin Li enc->dqm_[i].quant_ = clip(q, 0, 127);
419*b2055c35SXin Li }
420*b2055c35SXin Li
421*b2055c35SXin Li // purely indicative in the bitstream (except for the 1-segment case)
422*b2055c35SXin Li enc->base_quant_ = enc->dqm_[0].quant_;
423*b2055c35SXin Li
424*b2055c35SXin Li // fill-in values for the unused segments (required by the syntax)
425*b2055c35SXin Li for (i = num_segments; i < NUM_MB_SEGMENTS; ++i) {
426*b2055c35SXin Li enc->dqm_[i].quant_ = enc->base_quant_;
427*b2055c35SXin Li }
428*b2055c35SXin Li
429*b2055c35SXin Li // uv_alpha_ is normally spread around ~60. The useful range is
430*b2055c35SXin Li // typically ~30 (quite bad) to ~100 (ok to decimate UV more).
431*b2055c35SXin Li // We map it to the safe maximal range of MAX/MIN_DQ_UV for dq_uv.
432*b2055c35SXin Li dq_uv_ac = (enc->uv_alpha_ - MID_ALPHA) * (MAX_DQ_UV - MIN_DQ_UV)
433*b2055c35SXin Li / (MAX_ALPHA - MIN_ALPHA);
434*b2055c35SXin Li // we rescale by the user-defined strength of adaptation
435*b2055c35SXin Li dq_uv_ac = dq_uv_ac * enc->config_->sns_strength / 100;
436*b2055c35SXin Li // and make it safe.
437*b2055c35SXin Li dq_uv_ac = clip(dq_uv_ac, MIN_DQ_UV, MAX_DQ_UV);
438*b2055c35SXin Li // We also boost the dc-uv-quant a little, based on sns-strength, since
439*b2055c35SXin Li // U/V channels are quite more reactive to high quants (flat DC-blocks
440*b2055c35SXin Li // tend to appear, and are unpleasant).
441*b2055c35SXin Li dq_uv_dc = -4 * enc->config_->sns_strength / 100;
442*b2055c35SXin Li dq_uv_dc = clip(dq_uv_dc, -15, 15); // 4bit-signed max allowed
443*b2055c35SXin Li
444*b2055c35SXin Li enc->dq_y1_dc_ = 0; // TODO(skal): dq-lum
445*b2055c35SXin Li enc->dq_y2_dc_ = 0;
446*b2055c35SXin Li enc->dq_y2_ac_ = 0;
447*b2055c35SXin Li enc->dq_uv_dc_ = dq_uv_dc;
448*b2055c35SXin Li enc->dq_uv_ac_ = dq_uv_ac;
449*b2055c35SXin Li
450*b2055c35SXin Li SetupFilterStrength(enc); // initialize segments' filtering, eventually
451*b2055c35SXin Li
452*b2055c35SXin Li if (num_segments > 1) SimplifySegments(enc);
453*b2055c35SXin Li
454*b2055c35SXin Li SetupMatrices(enc); // finalize quantization matrices
455*b2055c35SXin Li }
456*b2055c35SXin Li
457*b2055c35SXin Li //------------------------------------------------------------------------------
458*b2055c35SXin Li // Form the predictions in cache
459*b2055c35SXin Li
460*b2055c35SXin Li // Must be ordered using {DC_PRED, TM_PRED, V_PRED, H_PRED} as index
461*b2055c35SXin Li const uint16_t VP8I16ModeOffsets[4] = { I16DC16, I16TM16, I16VE16, I16HE16 };
462*b2055c35SXin Li const uint16_t VP8UVModeOffsets[4] = { C8DC8, C8TM8, C8VE8, C8HE8 };
463*b2055c35SXin Li
464*b2055c35SXin Li // Must be indexed using {B_DC_PRED -> B_HU_PRED} as index
465*b2055c35SXin Li const uint16_t VP8I4ModeOffsets[NUM_BMODES] = {
466*b2055c35SXin Li I4DC4, I4TM4, I4VE4, I4HE4, I4RD4, I4VR4, I4LD4, I4VL4, I4HD4, I4HU4
467*b2055c35SXin Li };
468*b2055c35SXin Li
VP8MakeLuma16Preds(const VP8EncIterator * const it)469*b2055c35SXin Li void VP8MakeLuma16Preds(const VP8EncIterator* const it) {
470*b2055c35SXin Li const uint8_t* const left = it->x_ ? it->y_left_ : NULL;
471*b2055c35SXin Li const uint8_t* const top = it->y_ ? it->y_top_ : NULL;
472*b2055c35SXin Li VP8EncPredLuma16(it->yuv_p_, left, top);
473*b2055c35SXin Li }
474*b2055c35SXin Li
VP8MakeChroma8Preds(const VP8EncIterator * const it)475*b2055c35SXin Li void VP8MakeChroma8Preds(const VP8EncIterator* const it) {
476*b2055c35SXin Li const uint8_t* const left = it->x_ ? it->u_left_ : NULL;
477*b2055c35SXin Li const uint8_t* const top = it->y_ ? it->uv_top_ : NULL;
478*b2055c35SXin Li VP8EncPredChroma8(it->yuv_p_, left, top);
479*b2055c35SXin Li }
480*b2055c35SXin Li
VP8MakeIntra4Preds(const VP8EncIterator * const it)481*b2055c35SXin Li void VP8MakeIntra4Preds(const VP8EncIterator* const it) {
482*b2055c35SXin Li VP8EncPredLuma4(it->yuv_p_, it->i4_top_);
483*b2055c35SXin Li }
484*b2055c35SXin Li
485*b2055c35SXin Li //------------------------------------------------------------------------------
486*b2055c35SXin Li // Quantize
487*b2055c35SXin Li
488*b2055c35SXin Li // Layout:
489*b2055c35SXin Li // +----+----+
490*b2055c35SXin Li // |YYYY|UUVV| 0
491*b2055c35SXin Li // |YYYY|UUVV| 4
492*b2055c35SXin Li // |YYYY|....| 8
493*b2055c35SXin Li // |YYYY|....| 12
494*b2055c35SXin Li // +----+----+
495*b2055c35SXin Li
496*b2055c35SXin Li const uint16_t VP8Scan[16] = { // Luma
497*b2055c35SXin Li 0 + 0 * BPS, 4 + 0 * BPS, 8 + 0 * BPS, 12 + 0 * BPS,
498*b2055c35SXin Li 0 + 4 * BPS, 4 + 4 * BPS, 8 + 4 * BPS, 12 + 4 * BPS,
499*b2055c35SXin Li 0 + 8 * BPS, 4 + 8 * BPS, 8 + 8 * BPS, 12 + 8 * BPS,
500*b2055c35SXin Li 0 + 12 * BPS, 4 + 12 * BPS, 8 + 12 * BPS, 12 + 12 * BPS,
501*b2055c35SXin Li };
502*b2055c35SXin Li
503*b2055c35SXin Li static const uint16_t VP8ScanUV[4 + 4] = {
504*b2055c35SXin Li 0 + 0 * BPS, 4 + 0 * BPS, 0 + 4 * BPS, 4 + 4 * BPS, // U
505*b2055c35SXin Li 8 + 0 * BPS, 12 + 0 * BPS, 8 + 4 * BPS, 12 + 4 * BPS // V
506*b2055c35SXin Li };
507*b2055c35SXin Li
508*b2055c35SXin Li //------------------------------------------------------------------------------
509*b2055c35SXin Li // Distortion measurement
510*b2055c35SXin Li
511*b2055c35SXin Li static const uint16_t kWeightY[16] = {
512*b2055c35SXin Li 38, 32, 20, 9, 32, 28, 17, 7, 20, 17, 10, 4, 9, 7, 4, 2
513*b2055c35SXin Li };
514*b2055c35SXin Li
515*b2055c35SXin Li static const uint16_t kWeightTrellis[16] = {
516*b2055c35SXin Li #if USE_TDISTO == 0
517*b2055c35SXin Li 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16, 16
518*b2055c35SXin Li #else
519*b2055c35SXin Li 30, 27, 19, 11,
520*b2055c35SXin Li 27, 24, 17, 10,
521*b2055c35SXin Li 19, 17, 12, 8,
522*b2055c35SXin Li 11, 10, 8, 6
523*b2055c35SXin Li #endif
524*b2055c35SXin Li };
525*b2055c35SXin Li
526*b2055c35SXin Li // Init/Copy the common fields in score.
InitScore(VP8ModeScore * const rd)527*b2055c35SXin Li static void InitScore(VP8ModeScore* const rd) {
528*b2055c35SXin Li rd->D = 0;
529*b2055c35SXin Li rd->SD = 0;
530*b2055c35SXin Li rd->R = 0;
531*b2055c35SXin Li rd->H = 0;
532*b2055c35SXin Li rd->nz = 0;
533*b2055c35SXin Li rd->score = MAX_COST;
534*b2055c35SXin Li }
535*b2055c35SXin Li
CopyScore(VP8ModeScore * WEBP_RESTRICT const dst,const VP8ModeScore * WEBP_RESTRICT const src)536*b2055c35SXin Li static void CopyScore(VP8ModeScore* WEBP_RESTRICT const dst,
537*b2055c35SXin Li const VP8ModeScore* WEBP_RESTRICT const src) {
538*b2055c35SXin Li dst->D = src->D;
539*b2055c35SXin Li dst->SD = src->SD;
540*b2055c35SXin Li dst->R = src->R;
541*b2055c35SXin Li dst->H = src->H;
542*b2055c35SXin Li dst->nz = src->nz; // note that nz is not accumulated, but just copied.
543*b2055c35SXin Li dst->score = src->score;
544*b2055c35SXin Li }
545*b2055c35SXin Li
AddScore(VP8ModeScore * WEBP_RESTRICT const dst,const VP8ModeScore * WEBP_RESTRICT const src)546*b2055c35SXin Li static void AddScore(VP8ModeScore* WEBP_RESTRICT const dst,
547*b2055c35SXin Li const VP8ModeScore* WEBP_RESTRICT const src) {
548*b2055c35SXin Li dst->D += src->D;
549*b2055c35SXin Li dst->SD += src->SD;
550*b2055c35SXin Li dst->R += src->R;
551*b2055c35SXin Li dst->H += src->H;
552*b2055c35SXin Li dst->nz |= src->nz; // here, new nz bits are accumulated.
553*b2055c35SXin Li dst->score += src->score;
554*b2055c35SXin Li }
555*b2055c35SXin Li
556*b2055c35SXin Li //------------------------------------------------------------------------------
557*b2055c35SXin Li // Performs trellis-optimized quantization.
558*b2055c35SXin Li
559*b2055c35SXin Li // Trellis node
560*b2055c35SXin Li typedef struct {
561*b2055c35SXin Li int8_t prev; // best previous node
562*b2055c35SXin Li int8_t sign; // sign of coeff_i
563*b2055c35SXin Li int16_t level; // level
564*b2055c35SXin Li } Node;
565*b2055c35SXin Li
566*b2055c35SXin Li // Score state
567*b2055c35SXin Li typedef struct {
568*b2055c35SXin Li score_t score; // partial RD score
569*b2055c35SXin Li const uint16_t* costs; // shortcut to cost tables
570*b2055c35SXin Li } ScoreState;
571*b2055c35SXin Li
572*b2055c35SXin Li // If a coefficient was quantized to a value Q (using a neutral bias),
573*b2055c35SXin Li // we test all alternate possibilities between [Q-MIN_DELTA, Q+MAX_DELTA]
574*b2055c35SXin Li // We don't test negative values though.
575*b2055c35SXin Li #define MIN_DELTA 0 // how much lower level to try
576*b2055c35SXin Li #define MAX_DELTA 1 // how much higher
577*b2055c35SXin Li #define NUM_NODES (MIN_DELTA + 1 + MAX_DELTA)
578*b2055c35SXin Li #define NODE(n, l) (nodes[(n)][(l) + MIN_DELTA])
579*b2055c35SXin Li #define SCORE_STATE(n, l) (score_states[n][(l) + MIN_DELTA])
580*b2055c35SXin Li
SetRDScore(int lambda,VP8ModeScore * const rd)581*b2055c35SXin Li static WEBP_INLINE void SetRDScore(int lambda, VP8ModeScore* const rd) {
582*b2055c35SXin Li rd->score = (rd->R + rd->H) * lambda + RD_DISTO_MULT * (rd->D + rd->SD);
583*b2055c35SXin Li }
584*b2055c35SXin Li
RDScoreTrellis(int lambda,score_t rate,score_t distortion)585*b2055c35SXin Li static WEBP_INLINE score_t RDScoreTrellis(int lambda, score_t rate,
586*b2055c35SXin Li score_t distortion) {
587*b2055c35SXin Li return rate * lambda + RD_DISTO_MULT * distortion;
588*b2055c35SXin Li }
589*b2055c35SXin Li
590*b2055c35SXin Li // Coefficient type.
591*b2055c35SXin Li enum { TYPE_I16_AC = 0, TYPE_I16_DC = 1, TYPE_CHROMA_A = 2, TYPE_I4_AC = 3 };
592*b2055c35SXin Li
TrellisQuantizeBlock(const VP8Encoder * WEBP_RESTRICT const enc,int16_t in[16],int16_t out[16],int ctx0,int coeff_type,const VP8Matrix * WEBP_RESTRICT const mtx,int lambda)593*b2055c35SXin Li static int TrellisQuantizeBlock(const VP8Encoder* WEBP_RESTRICT const enc,
594*b2055c35SXin Li int16_t in[16], int16_t out[16],
595*b2055c35SXin Li int ctx0, int coeff_type,
596*b2055c35SXin Li const VP8Matrix* WEBP_RESTRICT const mtx,
597*b2055c35SXin Li int lambda) {
598*b2055c35SXin Li const ProbaArray* const probas = enc->proba_.coeffs_[coeff_type];
599*b2055c35SXin Li CostArrayPtr const costs =
600*b2055c35SXin Li (CostArrayPtr)enc->proba_.remapped_costs_[coeff_type];
601*b2055c35SXin Li const int first = (coeff_type == TYPE_I16_AC) ? 1 : 0;
602*b2055c35SXin Li Node nodes[16][NUM_NODES];
603*b2055c35SXin Li ScoreState score_states[2][NUM_NODES];
604*b2055c35SXin Li ScoreState* ss_cur = &SCORE_STATE(0, MIN_DELTA);
605*b2055c35SXin Li ScoreState* ss_prev = &SCORE_STATE(1, MIN_DELTA);
606*b2055c35SXin Li int best_path[3] = {-1, -1, -1}; // store best-last/best-level/best-previous
607*b2055c35SXin Li score_t best_score;
608*b2055c35SXin Li int n, m, p, last;
609*b2055c35SXin Li
610*b2055c35SXin Li {
611*b2055c35SXin Li score_t cost;
612*b2055c35SXin Li const int thresh = mtx->q_[1] * mtx->q_[1] / 4;
613*b2055c35SXin Li const int last_proba = probas[VP8EncBands[first]][ctx0][0];
614*b2055c35SXin Li
615*b2055c35SXin Li // compute the position of the last interesting coefficient
616*b2055c35SXin Li last = first - 1;
617*b2055c35SXin Li for (n = 15; n >= first; --n) {
618*b2055c35SXin Li const int j = kZigzag[n];
619*b2055c35SXin Li const int err = in[j] * in[j];
620*b2055c35SXin Li if (err > thresh) {
621*b2055c35SXin Li last = n;
622*b2055c35SXin Li break;
623*b2055c35SXin Li }
624*b2055c35SXin Li }
625*b2055c35SXin Li // we don't need to go inspect up to n = 16 coeffs. We can just go up
626*b2055c35SXin Li // to last + 1 (inclusive) without losing much.
627*b2055c35SXin Li if (last < 15) ++last;
628*b2055c35SXin Li
629*b2055c35SXin Li // compute 'skip' score. This is the max score one can do.
630*b2055c35SXin Li cost = VP8BitCost(0, last_proba);
631*b2055c35SXin Li best_score = RDScoreTrellis(lambda, cost, 0);
632*b2055c35SXin Li
633*b2055c35SXin Li // initialize source node.
634*b2055c35SXin Li for (m = -MIN_DELTA; m <= MAX_DELTA; ++m) {
635*b2055c35SXin Li const score_t rate = (ctx0 == 0) ? VP8BitCost(1, last_proba) : 0;
636*b2055c35SXin Li ss_cur[m].score = RDScoreTrellis(lambda, rate, 0);
637*b2055c35SXin Li ss_cur[m].costs = costs[first][ctx0];
638*b2055c35SXin Li }
639*b2055c35SXin Li }
640*b2055c35SXin Li
641*b2055c35SXin Li // traverse trellis.
642*b2055c35SXin Li for (n = first; n <= last; ++n) {
643*b2055c35SXin Li const int j = kZigzag[n];
644*b2055c35SXin Li const uint32_t Q = mtx->q_[j];
645*b2055c35SXin Li const uint32_t iQ = mtx->iq_[j];
646*b2055c35SXin Li const uint32_t B = BIAS(0x00); // neutral bias
647*b2055c35SXin Li // note: it's important to take sign of the _original_ coeff,
648*b2055c35SXin Li // so we don't have to consider level < 0 afterward.
649*b2055c35SXin Li const int sign = (in[j] < 0);
650*b2055c35SXin Li const uint32_t coeff0 = (sign ? -in[j] : in[j]) + mtx->sharpen_[j];
651*b2055c35SXin Li int level0 = QUANTDIV(coeff0, iQ, B);
652*b2055c35SXin Li int thresh_level = QUANTDIV(coeff0, iQ, BIAS(0x80));
653*b2055c35SXin Li if (thresh_level > MAX_LEVEL) thresh_level = MAX_LEVEL;
654*b2055c35SXin Li if (level0 > MAX_LEVEL) level0 = MAX_LEVEL;
655*b2055c35SXin Li
656*b2055c35SXin Li { // Swap current and previous score states
657*b2055c35SXin Li ScoreState* const tmp = ss_cur;
658*b2055c35SXin Li ss_cur = ss_prev;
659*b2055c35SXin Li ss_prev = tmp;
660*b2055c35SXin Li }
661*b2055c35SXin Li
662*b2055c35SXin Li // test all alternate level values around level0.
663*b2055c35SXin Li for (m = -MIN_DELTA; m <= MAX_DELTA; ++m) {
664*b2055c35SXin Li Node* const cur = &NODE(n, m);
665*b2055c35SXin Li const int level = level0 + m;
666*b2055c35SXin Li const int ctx = (level > 2) ? 2 : level;
667*b2055c35SXin Li const int band = VP8EncBands[n + 1];
668*b2055c35SXin Li score_t base_score;
669*b2055c35SXin Li score_t best_cur_score;
670*b2055c35SXin Li int best_prev;
671*b2055c35SXin Li score_t cost, score;
672*b2055c35SXin Li
673*b2055c35SXin Li ss_cur[m].costs = costs[n + 1][ctx];
674*b2055c35SXin Li if (level < 0 || level > thresh_level) {
675*b2055c35SXin Li ss_cur[m].score = MAX_COST;
676*b2055c35SXin Li // Node is dead.
677*b2055c35SXin Li continue;
678*b2055c35SXin Li }
679*b2055c35SXin Li
680*b2055c35SXin Li {
681*b2055c35SXin Li // Compute delta_error = how much coding this level will
682*b2055c35SXin Li // subtract to max_error as distortion.
683*b2055c35SXin Li // Here, distortion = sum of (|coeff_i| - level_i * Q_i)^2
684*b2055c35SXin Li const int new_error = coeff0 - level * Q;
685*b2055c35SXin Li const int delta_error =
686*b2055c35SXin Li kWeightTrellis[j] * (new_error * new_error - coeff0 * coeff0);
687*b2055c35SXin Li base_score = RDScoreTrellis(lambda, 0, delta_error);
688*b2055c35SXin Li }
689*b2055c35SXin Li
690*b2055c35SXin Li // Inspect all possible non-dead predecessors. Retain only the best one.
691*b2055c35SXin Li // The base_score is added to all scores so it is only added for the final
692*b2055c35SXin Li // value after the loop.
693*b2055c35SXin Li cost = VP8LevelCost(ss_prev[-MIN_DELTA].costs, level);
694*b2055c35SXin Li best_cur_score =
695*b2055c35SXin Li ss_prev[-MIN_DELTA].score + RDScoreTrellis(lambda, cost, 0);
696*b2055c35SXin Li best_prev = -MIN_DELTA;
697*b2055c35SXin Li for (p = -MIN_DELTA + 1; p <= MAX_DELTA; ++p) {
698*b2055c35SXin Li // Dead nodes (with ss_prev[p].score >= MAX_COST) are automatically
699*b2055c35SXin Li // eliminated since their score can't be better than the current best.
700*b2055c35SXin Li cost = VP8LevelCost(ss_prev[p].costs, level);
701*b2055c35SXin Li // Examine node assuming it's a non-terminal one.
702*b2055c35SXin Li score = ss_prev[p].score + RDScoreTrellis(lambda, cost, 0);
703*b2055c35SXin Li if (score < best_cur_score) {
704*b2055c35SXin Li best_cur_score = score;
705*b2055c35SXin Li best_prev = p;
706*b2055c35SXin Li }
707*b2055c35SXin Li }
708*b2055c35SXin Li best_cur_score += base_score;
709*b2055c35SXin Li // Store best finding in current node.
710*b2055c35SXin Li cur->sign = sign;
711*b2055c35SXin Li cur->level = level;
712*b2055c35SXin Li cur->prev = best_prev;
713*b2055c35SXin Li ss_cur[m].score = best_cur_score;
714*b2055c35SXin Li
715*b2055c35SXin Li // Now, record best terminal node (and thus best entry in the graph).
716*b2055c35SXin Li if (level != 0 && best_cur_score < best_score) {
717*b2055c35SXin Li const score_t last_pos_cost =
718*b2055c35SXin Li (n < 15) ? VP8BitCost(0, probas[band][ctx][0]) : 0;
719*b2055c35SXin Li const score_t last_pos_score = RDScoreTrellis(lambda, last_pos_cost, 0);
720*b2055c35SXin Li score = best_cur_score + last_pos_score;
721*b2055c35SXin Li if (score < best_score) {
722*b2055c35SXin Li best_score = score;
723*b2055c35SXin Li best_path[0] = n; // best eob position
724*b2055c35SXin Li best_path[1] = m; // best node index
725*b2055c35SXin Li best_path[2] = best_prev; // best predecessor
726*b2055c35SXin Li }
727*b2055c35SXin Li }
728*b2055c35SXin Li }
729*b2055c35SXin Li }
730*b2055c35SXin Li
731*b2055c35SXin Li // Fresh start
732*b2055c35SXin Li // Beware! We must preserve in[0]/out[0] value for TYPE_I16_AC case.
733*b2055c35SXin Li if (coeff_type == TYPE_I16_AC) {
734*b2055c35SXin Li memset(in + 1, 0, 15 * sizeof(*in));
735*b2055c35SXin Li memset(out + 1, 0, 15 * sizeof(*out));
736*b2055c35SXin Li } else {
737*b2055c35SXin Li memset(in, 0, 16 * sizeof(*in));
738*b2055c35SXin Li memset(out, 0, 16 * sizeof(*out));
739*b2055c35SXin Li }
740*b2055c35SXin Li if (best_path[0] == -1) {
741*b2055c35SXin Li return 0; // skip!
742*b2055c35SXin Li }
743*b2055c35SXin Li
744*b2055c35SXin Li {
745*b2055c35SXin Li // Unwind the best path.
746*b2055c35SXin Li // Note: best-prev on terminal node is not necessarily equal to the
747*b2055c35SXin Li // best_prev for non-terminal. So we patch best_path[2] in.
748*b2055c35SXin Li int nz = 0;
749*b2055c35SXin Li int best_node = best_path[1];
750*b2055c35SXin Li n = best_path[0];
751*b2055c35SXin Li NODE(n, best_node).prev = best_path[2]; // force best-prev for terminal
752*b2055c35SXin Li
753*b2055c35SXin Li for (; n >= first; --n) {
754*b2055c35SXin Li const Node* const node = &NODE(n, best_node);
755*b2055c35SXin Li const int j = kZigzag[n];
756*b2055c35SXin Li out[n] = node->sign ? -node->level : node->level;
757*b2055c35SXin Li nz |= node->level;
758*b2055c35SXin Li in[j] = out[n] * mtx->q_[j];
759*b2055c35SXin Li best_node = node->prev;
760*b2055c35SXin Li }
761*b2055c35SXin Li return (nz != 0);
762*b2055c35SXin Li }
763*b2055c35SXin Li }
764*b2055c35SXin Li
765*b2055c35SXin Li #undef NODE
766*b2055c35SXin Li
767*b2055c35SXin Li //------------------------------------------------------------------------------
768*b2055c35SXin Li // Performs: difference, transform, quantize, back-transform, add
769*b2055c35SXin Li // all at once. Output is the reconstructed block in *yuv_out, and the
770*b2055c35SXin Li // quantized levels in *levels.
771*b2055c35SXin Li
ReconstructIntra16(VP8EncIterator * WEBP_RESTRICT const it,VP8ModeScore * WEBP_RESTRICT const rd,uint8_t * WEBP_RESTRICT const yuv_out,int mode)772*b2055c35SXin Li static int ReconstructIntra16(VP8EncIterator* WEBP_RESTRICT const it,
773*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd,
774*b2055c35SXin Li uint8_t* WEBP_RESTRICT const yuv_out,
775*b2055c35SXin Li int mode) {
776*b2055c35SXin Li const VP8Encoder* const enc = it->enc_;
777*b2055c35SXin Li const uint8_t* const ref = it->yuv_p_ + VP8I16ModeOffsets[mode];
778*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + Y_OFF_ENC;
779*b2055c35SXin Li const VP8SegmentInfo* const dqm = &enc->dqm_[it->mb_->segment_];
780*b2055c35SXin Li int nz = 0;
781*b2055c35SXin Li int n;
782*b2055c35SXin Li int16_t tmp[16][16], dc_tmp[16];
783*b2055c35SXin Li
784*b2055c35SXin Li for (n = 0; n < 16; n += 2) {
785*b2055c35SXin Li VP8FTransform2(src + VP8Scan[n], ref + VP8Scan[n], tmp[n]);
786*b2055c35SXin Li }
787*b2055c35SXin Li VP8FTransformWHT(tmp[0], dc_tmp);
788*b2055c35SXin Li nz |= VP8EncQuantizeBlockWHT(dc_tmp, rd->y_dc_levels, &dqm->y2_) << 24;
789*b2055c35SXin Li
790*b2055c35SXin Li if (DO_TRELLIS_I16 && it->do_trellis_) {
791*b2055c35SXin Li int x, y;
792*b2055c35SXin Li VP8IteratorNzToBytes(it);
793*b2055c35SXin Li for (y = 0, n = 0; y < 4; ++y) {
794*b2055c35SXin Li for (x = 0; x < 4; ++x, ++n) {
795*b2055c35SXin Li const int ctx = it->top_nz_[x] + it->left_nz_[y];
796*b2055c35SXin Li const int non_zero = TrellisQuantizeBlock(
797*b2055c35SXin Li enc, tmp[n], rd->y_ac_levels[n], ctx, TYPE_I16_AC, &dqm->y1_,
798*b2055c35SXin Li dqm->lambda_trellis_i16_);
799*b2055c35SXin Li it->top_nz_[x] = it->left_nz_[y] = non_zero;
800*b2055c35SXin Li rd->y_ac_levels[n][0] = 0;
801*b2055c35SXin Li nz |= non_zero << n;
802*b2055c35SXin Li }
803*b2055c35SXin Li }
804*b2055c35SXin Li } else {
805*b2055c35SXin Li for (n = 0; n < 16; n += 2) {
806*b2055c35SXin Li // Zero-out the first coeff, so that: a) nz is correct below, and
807*b2055c35SXin Li // b) finding 'last' non-zero coeffs in SetResidualCoeffs() is simplified.
808*b2055c35SXin Li tmp[n][0] = tmp[n + 1][0] = 0;
809*b2055c35SXin Li nz |= VP8EncQuantize2Blocks(tmp[n], rd->y_ac_levels[n], &dqm->y1_) << n;
810*b2055c35SXin Li assert(rd->y_ac_levels[n + 0][0] == 0);
811*b2055c35SXin Li assert(rd->y_ac_levels[n + 1][0] == 0);
812*b2055c35SXin Li }
813*b2055c35SXin Li }
814*b2055c35SXin Li
815*b2055c35SXin Li // Transform back
816*b2055c35SXin Li VP8TransformWHT(dc_tmp, tmp[0]);
817*b2055c35SXin Li for (n = 0; n < 16; n += 2) {
818*b2055c35SXin Li VP8ITransform(ref + VP8Scan[n], tmp[n], yuv_out + VP8Scan[n], 1);
819*b2055c35SXin Li }
820*b2055c35SXin Li
821*b2055c35SXin Li return nz;
822*b2055c35SXin Li }
823*b2055c35SXin Li
ReconstructIntra4(VP8EncIterator * WEBP_RESTRICT const it,int16_t levels[16],const uint8_t * WEBP_RESTRICT const src,uint8_t * WEBP_RESTRICT const yuv_out,int mode)824*b2055c35SXin Li static int ReconstructIntra4(VP8EncIterator* WEBP_RESTRICT const it,
825*b2055c35SXin Li int16_t levels[16],
826*b2055c35SXin Li const uint8_t* WEBP_RESTRICT const src,
827*b2055c35SXin Li uint8_t* WEBP_RESTRICT const yuv_out,
828*b2055c35SXin Li int mode) {
829*b2055c35SXin Li const VP8Encoder* const enc = it->enc_;
830*b2055c35SXin Li const uint8_t* const ref = it->yuv_p_ + VP8I4ModeOffsets[mode];
831*b2055c35SXin Li const VP8SegmentInfo* const dqm = &enc->dqm_[it->mb_->segment_];
832*b2055c35SXin Li int nz = 0;
833*b2055c35SXin Li int16_t tmp[16];
834*b2055c35SXin Li
835*b2055c35SXin Li VP8FTransform(src, ref, tmp);
836*b2055c35SXin Li if (DO_TRELLIS_I4 && it->do_trellis_) {
837*b2055c35SXin Li const int x = it->i4_ & 3, y = it->i4_ >> 2;
838*b2055c35SXin Li const int ctx = it->top_nz_[x] + it->left_nz_[y];
839*b2055c35SXin Li nz = TrellisQuantizeBlock(enc, tmp, levels, ctx, TYPE_I4_AC, &dqm->y1_,
840*b2055c35SXin Li dqm->lambda_trellis_i4_);
841*b2055c35SXin Li } else {
842*b2055c35SXin Li nz = VP8EncQuantizeBlock(tmp, levels, &dqm->y1_);
843*b2055c35SXin Li }
844*b2055c35SXin Li VP8ITransform(ref, tmp, yuv_out, 0);
845*b2055c35SXin Li return nz;
846*b2055c35SXin Li }
847*b2055c35SXin Li
848*b2055c35SXin Li //------------------------------------------------------------------------------
849*b2055c35SXin Li // DC-error diffusion
850*b2055c35SXin Li
851*b2055c35SXin Li // Diffusion weights. We under-correct a bit (15/16th of the error is actually
852*b2055c35SXin Li // diffused) to avoid 'rainbow' chessboard pattern of blocks at q~=0.
853*b2055c35SXin Li #define C1 7 // fraction of error sent to the 4x4 block below
854*b2055c35SXin Li #define C2 8 // fraction of error sent to the 4x4 block on the right
855*b2055c35SXin Li #define DSHIFT 4
856*b2055c35SXin Li #define DSCALE 1 // storage descaling, needed to make the error fit int8_t
857*b2055c35SXin Li
858*b2055c35SXin Li // Quantize as usual, but also compute and return the quantization error.
859*b2055c35SXin Li // Error is already divided by DSHIFT.
QuantizeSingle(int16_t * WEBP_RESTRICT const v,const VP8Matrix * WEBP_RESTRICT const mtx)860*b2055c35SXin Li static int QuantizeSingle(int16_t* WEBP_RESTRICT const v,
861*b2055c35SXin Li const VP8Matrix* WEBP_RESTRICT const mtx) {
862*b2055c35SXin Li int V = *v;
863*b2055c35SXin Li const int sign = (V < 0);
864*b2055c35SXin Li if (sign) V = -V;
865*b2055c35SXin Li if (V > (int)mtx->zthresh_[0]) {
866*b2055c35SXin Li const int qV = QUANTDIV(V, mtx->iq_[0], mtx->bias_[0]) * mtx->q_[0];
867*b2055c35SXin Li const int err = (V - qV);
868*b2055c35SXin Li *v = sign ? -qV : qV;
869*b2055c35SXin Li return (sign ? -err : err) >> DSCALE;
870*b2055c35SXin Li }
871*b2055c35SXin Li *v = 0;
872*b2055c35SXin Li return (sign ? -V : V) >> DSCALE;
873*b2055c35SXin Li }
874*b2055c35SXin Li
CorrectDCValues(const VP8EncIterator * WEBP_RESTRICT const it,const VP8Matrix * WEBP_RESTRICT const mtx,int16_t tmp[][16],VP8ModeScore * WEBP_RESTRICT const rd)875*b2055c35SXin Li static void CorrectDCValues(const VP8EncIterator* WEBP_RESTRICT const it,
876*b2055c35SXin Li const VP8Matrix* WEBP_RESTRICT const mtx,
877*b2055c35SXin Li int16_t tmp[][16],
878*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd) {
879*b2055c35SXin Li // | top[0] | top[1]
880*b2055c35SXin Li // --------+--------+---------
881*b2055c35SXin Li // left[0] | tmp[0] tmp[1] <-> err0 err1
882*b2055c35SXin Li // left[1] | tmp[2] tmp[3] err2 err3
883*b2055c35SXin Li //
884*b2055c35SXin Li // Final errors {err1,err2,err3} are preserved and later restored
885*b2055c35SXin Li // as top[]/left[] on the next block.
886*b2055c35SXin Li int ch;
887*b2055c35SXin Li for (ch = 0; ch <= 1; ++ch) {
888*b2055c35SXin Li const int8_t* const top = it->top_derr_[it->x_][ch];
889*b2055c35SXin Li const int8_t* const left = it->left_derr_[ch];
890*b2055c35SXin Li int16_t (* const c)[16] = &tmp[ch * 4];
891*b2055c35SXin Li int err0, err1, err2, err3;
892*b2055c35SXin Li c[0][0] += (C1 * top[0] + C2 * left[0]) >> (DSHIFT - DSCALE);
893*b2055c35SXin Li err0 = QuantizeSingle(&c[0][0], mtx);
894*b2055c35SXin Li c[1][0] += (C1 * top[1] + C2 * err0) >> (DSHIFT - DSCALE);
895*b2055c35SXin Li err1 = QuantizeSingle(&c[1][0], mtx);
896*b2055c35SXin Li c[2][0] += (C1 * err0 + C2 * left[1]) >> (DSHIFT - DSCALE);
897*b2055c35SXin Li err2 = QuantizeSingle(&c[2][0], mtx);
898*b2055c35SXin Li c[3][0] += (C1 * err1 + C2 * err2) >> (DSHIFT - DSCALE);
899*b2055c35SXin Li err3 = QuantizeSingle(&c[3][0], mtx);
900*b2055c35SXin Li // error 'err' is bounded by mtx->q_[0] which is 132 at max. Hence
901*b2055c35SXin Li // err >> DSCALE will fit in an int8_t type if DSCALE>=1.
902*b2055c35SXin Li assert(abs(err1) <= 127 && abs(err2) <= 127 && abs(err3) <= 127);
903*b2055c35SXin Li rd->derr[ch][0] = (int8_t)err1;
904*b2055c35SXin Li rd->derr[ch][1] = (int8_t)err2;
905*b2055c35SXin Li rd->derr[ch][2] = (int8_t)err3;
906*b2055c35SXin Li }
907*b2055c35SXin Li }
908*b2055c35SXin Li
StoreDiffusionErrors(VP8EncIterator * WEBP_RESTRICT const it,const VP8ModeScore * WEBP_RESTRICT const rd)909*b2055c35SXin Li static void StoreDiffusionErrors(VP8EncIterator* WEBP_RESTRICT const it,
910*b2055c35SXin Li const VP8ModeScore* WEBP_RESTRICT const rd) {
911*b2055c35SXin Li int ch;
912*b2055c35SXin Li for (ch = 0; ch <= 1; ++ch) {
913*b2055c35SXin Li int8_t* const top = it->top_derr_[it->x_][ch];
914*b2055c35SXin Li int8_t* const left = it->left_derr_[ch];
915*b2055c35SXin Li left[0] = rd->derr[ch][0]; // restore err1
916*b2055c35SXin Li left[1] = 3 * rd->derr[ch][2] >> 2; // ... 3/4th of err3
917*b2055c35SXin Li top[0] = rd->derr[ch][1]; // ... err2
918*b2055c35SXin Li top[1] = rd->derr[ch][2] - left[1]; // ... 1/4th of err3.
919*b2055c35SXin Li }
920*b2055c35SXin Li }
921*b2055c35SXin Li
922*b2055c35SXin Li #undef C1
923*b2055c35SXin Li #undef C2
924*b2055c35SXin Li #undef DSHIFT
925*b2055c35SXin Li #undef DSCALE
926*b2055c35SXin Li
927*b2055c35SXin Li //------------------------------------------------------------------------------
928*b2055c35SXin Li
ReconstructUV(VP8EncIterator * WEBP_RESTRICT const it,VP8ModeScore * WEBP_RESTRICT const rd,uint8_t * WEBP_RESTRICT const yuv_out,int mode)929*b2055c35SXin Li static int ReconstructUV(VP8EncIterator* WEBP_RESTRICT const it,
930*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd,
931*b2055c35SXin Li uint8_t* WEBP_RESTRICT const yuv_out, int mode) {
932*b2055c35SXin Li const VP8Encoder* const enc = it->enc_;
933*b2055c35SXin Li const uint8_t* const ref = it->yuv_p_ + VP8UVModeOffsets[mode];
934*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + U_OFF_ENC;
935*b2055c35SXin Li const VP8SegmentInfo* const dqm = &enc->dqm_[it->mb_->segment_];
936*b2055c35SXin Li int nz = 0;
937*b2055c35SXin Li int n;
938*b2055c35SXin Li int16_t tmp[8][16];
939*b2055c35SXin Li
940*b2055c35SXin Li for (n = 0; n < 8; n += 2) {
941*b2055c35SXin Li VP8FTransform2(src + VP8ScanUV[n], ref + VP8ScanUV[n], tmp[n]);
942*b2055c35SXin Li }
943*b2055c35SXin Li if (it->top_derr_ != NULL) CorrectDCValues(it, &dqm->uv_, tmp, rd);
944*b2055c35SXin Li
945*b2055c35SXin Li if (DO_TRELLIS_UV && it->do_trellis_) {
946*b2055c35SXin Li int ch, x, y;
947*b2055c35SXin Li for (ch = 0, n = 0; ch <= 2; ch += 2) {
948*b2055c35SXin Li for (y = 0; y < 2; ++y) {
949*b2055c35SXin Li for (x = 0; x < 2; ++x, ++n) {
950*b2055c35SXin Li const int ctx = it->top_nz_[4 + ch + x] + it->left_nz_[4 + ch + y];
951*b2055c35SXin Li const int non_zero = TrellisQuantizeBlock(
952*b2055c35SXin Li enc, tmp[n], rd->uv_levels[n], ctx, TYPE_CHROMA_A, &dqm->uv_,
953*b2055c35SXin Li dqm->lambda_trellis_uv_);
954*b2055c35SXin Li it->top_nz_[4 + ch + x] = it->left_nz_[4 + ch + y] = non_zero;
955*b2055c35SXin Li nz |= non_zero << n;
956*b2055c35SXin Li }
957*b2055c35SXin Li }
958*b2055c35SXin Li }
959*b2055c35SXin Li } else {
960*b2055c35SXin Li for (n = 0; n < 8; n += 2) {
961*b2055c35SXin Li nz |= VP8EncQuantize2Blocks(tmp[n], rd->uv_levels[n], &dqm->uv_) << n;
962*b2055c35SXin Li }
963*b2055c35SXin Li }
964*b2055c35SXin Li
965*b2055c35SXin Li for (n = 0; n < 8; n += 2) {
966*b2055c35SXin Li VP8ITransform(ref + VP8ScanUV[n], tmp[n], yuv_out + VP8ScanUV[n], 1);
967*b2055c35SXin Li }
968*b2055c35SXin Li return (nz << 16);
969*b2055c35SXin Li }
970*b2055c35SXin Li
971*b2055c35SXin Li //------------------------------------------------------------------------------
972*b2055c35SXin Li // RD-opt decision. Reconstruct each modes, evalue distortion and bit-cost.
973*b2055c35SXin Li // Pick the mode is lower RD-cost = Rate + lambda * Distortion.
974*b2055c35SXin Li
StoreMaxDelta(VP8SegmentInfo * const dqm,const int16_t DCs[16])975*b2055c35SXin Li static void StoreMaxDelta(VP8SegmentInfo* const dqm, const int16_t DCs[16]) {
976*b2055c35SXin Li // We look at the first three AC coefficients to determine what is the average
977*b2055c35SXin Li // delta between each sub-4x4 block.
978*b2055c35SXin Li const int v0 = abs(DCs[1]);
979*b2055c35SXin Li const int v1 = abs(DCs[2]);
980*b2055c35SXin Li const int v2 = abs(DCs[4]);
981*b2055c35SXin Li int max_v = (v1 > v0) ? v1 : v0;
982*b2055c35SXin Li max_v = (v2 > max_v) ? v2 : max_v;
983*b2055c35SXin Li if (max_v > dqm->max_edge_) dqm->max_edge_ = max_v;
984*b2055c35SXin Li }
985*b2055c35SXin Li
SwapModeScore(VP8ModeScore ** a,VP8ModeScore ** b)986*b2055c35SXin Li static void SwapModeScore(VP8ModeScore** a, VP8ModeScore** b) {
987*b2055c35SXin Li VP8ModeScore* const tmp = *a;
988*b2055c35SXin Li *a = *b;
989*b2055c35SXin Li *b = tmp;
990*b2055c35SXin Li }
991*b2055c35SXin Li
SwapPtr(uint8_t ** a,uint8_t ** b)992*b2055c35SXin Li static void SwapPtr(uint8_t** a, uint8_t** b) {
993*b2055c35SXin Li uint8_t* const tmp = *a;
994*b2055c35SXin Li *a = *b;
995*b2055c35SXin Li *b = tmp;
996*b2055c35SXin Li }
997*b2055c35SXin Li
SwapOut(VP8EncIterator * const it)998*b2055c35SXin Li static void SwapOut(VP8EncIterator* const it) {
999*b2055c35SXin Li SwapPtr(&it->yuv_out_, &it->yuv_out2_);
1000*b2055c35SXin Li }
1001*b2055c35SXin Li
PickBestIntra16(VP8EncIterator * WEBP_RESTRICT const it,VP8ModeScore * WEBP_RESTRICT rd)1002*b2055c35SXin Li static void PickBestIntra16(VP8EncIterator* WEBP_RESTRICT const it,
1003*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT rd) {
1004*b2055c35SXin Li const int kNumBlocks = 16;
1005*b2055c35SXin Li VP8SegmentInfo* const dqm = &it->enc_->dqm_[it->mb_->segment_];
1006*b2055c35SXin Li const int lambda = dqm->lambda_i16_;
1007*b2055c35SXin Li const int tlambda = dqm->tlambda_;
1008*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + Y_OFF_ENC;
1009*b2055c35SXin Li VP8ModeScore rd_tmp;
1010*b2055c35SXin Li VP8ModeScore* rd_cur = &rd_tmp;
1011*b2055c35SXin Li VP8ModeScore* rd_best = rd;
1012*b2055c35SXin Li int mode;
1013*b2055c35SXin Li int is_flat = IsFlatSource16(it->yuv_in_ + Y_OFF_ENC);
1014*b2055c35SXin Li
1015*b2055c35SXin Li rd->mode_i16 = -1;
1016*b2055c35SXin Li for (mode = 0; mode < NUM_PRED_MODES; ++mode) {
1017*b2055c35SXin Li uint8_t* const tmp_dst = it->yuv_out2_ + Y_OFF_ENC; // scratch buffer
1018*b2055c35SXin Li rd_cur->mode_i16 = mode;
1019*b2055c35SXin Li
1020*b2055c35SXin Li // Reconstruct
1021*b2055c35SXin Li rd_cur->nz = ReconstructIntra16(it, rd_cur, tmp_dst, mode);
1022*b2055c35SXin Li
1023*b2055c35SXin Li // Measure RD-score
1024*b2055c35SXin Li rd_cur->D = VP8SSE16x16(src, tmp_dst);
1025*b2055c35SXin Li rd_cur->SD =
1026*b2055c35SXin Li tlambda ? MULT_8B(tlambda, VP8TDisto16x16(src, tmp_dst, kWeightY)) : 0;
1027*b2055c35SXin Li rd_cur->H = VP8FixedCostsI16[mode];
1028*b2055c35SXin Li rd_cur->R = VP8GetCostLuma16(it, rd_cur);
1029*b2055c35SXin Li if (is_flat) {
1030*b2055c35SXin Li // refine the first impression (which was in pixel space)
1031*b2055c35SXin Li is_flat = IsFlat(rd_cur->y_ac_levels[0], kNumBlocks, FLATNESS_LIMIT_I16);
1032*b2055c35SXin Li if (is_flat) {
1033*b2055c35SXin Li // Block is very flat. We put emphasis on the distortion being very low!
1034*b2055c35SXin Li rd_cur->D *= 2;
1035*b2055c35SXin Li rd_cur->SD *= 2;
1036*b2055c35SXin Li }
1037*b2055c35SXin Li }
1038*b2055c35SXin Li
1039*b2055c35SXin Li // Since we always examine Intra16 first, we can overwrite *rd directly.
1040*b2055c35SXin Li SetRDScore(lambda, rd_cur);
1041*b2055c35SXin Li if (mode == 0 || rd_cur->score < rd_best->score) {
1042*b2055c35SXin Li SwapModeScore(&rd_cur, &rd_best);
1043*b2055c35SXin Li SwapOut(it);
1044*b2055c35SXin Li }
1045*b2055c35SXin Li }
1046*b2055c35SXin Li if (rd_best != rd) {
1047*b2055c35SXin Li memcpy(rd, rd_best, sizeof(*rd));
1048*b2055c35SXin Li }
1049*b2055c35SXin Li SetRDScore(dqm->lambda_mode_, rd); // finalize score for mode decision.
1050*b2055c35SXin Li VP8SetIntra16Mode(it, rd->mode_i16);
1051*b2055c35SXin Li
1052*b2055c35SXin Li // we have a blocky macroblock (only DCs are non-zero) with fairly high
1053*b2055c35SXin Li // distortion, record max delta so we can later adjust the minimal filtering
1054*b2055c35SXin Li // strength needed to smooth these blocks out.
1055*b2055c35SXin Li if ((rd->nz & 0x100ffff) == 0x1000000 && rd->D > dqm->min_disto_) {
1056*b2055c35SXin Li StoreMaxDelta(dqm, rd->y_dc_levels);
1057*b2055c35SXin Li }
1058*b2055c35SXin Li }
1059*b2055c35SXin Li
1060*b2055c35SXin Li //------------------------------------------------------------------------------
1061*b2055c35SXin Li
1062*b2055c35SXin Li // return the cost array corresponding to the surrounding prediction modes.
GetCostModeI4(VP8EncIterator * WEBP_RESTRICT const it,const uint8_t modes[16])1063*b2055c35SXin Li static const uint16_t* GetCostModeI4(VP8EncIterator* WEBP_RESTRICT const it,
1064*b2055c35SXin Li const uint8_t modes[16]) {
1065*b2055c35SXin Li const int preds_w = it->enc_->preds_w_;
1066*b2055c35SXin Li const int x = (it->i4_ & 3), y = it->i4_ >> 2;
1067*b2055c35SXin Li const int left = (x == 0) ? it->preds_[y * preds_w - 1] : modes[it->i4_ - 1];
1068*b2055c35SXin Li const int top = (y == 0) ? it->preds_[-preds_w + x] : modes[it->i4_ - 4];
1069*b2055c35SXin Li return VP8FixedCostsI4[top][left];
1070*b2055c35SXin Li }
1071*b2055c35SXin Li
PickBestIntra4(VP8EncIterator * WEBP_RESTRICT const it,VP8ModeScore * WEBP_RESTRICT const rd)1072*b2055c35SXin Li static int PickBestIntra4(VP8EncIterator* WEBP_RESTRICT const it,
1073*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd) {
1074*b2055c35SXin Li const VP8Encoder* const enc = it->enc_;
1075*b2055c35SXin Li const VP8SegmentInfo* const dqm = &enc->dqm_[it->mb_->segment_];
1076*b2055c35SXin Li const int lambda = dqm->lambda_i4_;
1077*b2055c35SXin Li const int tlambda = dqm->tlambda_;
1078*b2055c35SXin Li const uint8_t* const src0 = it->yuv_in_ + Y_OFF_ENC;
1079*b2055c35SXin Li uint8_t* const best_blocks = it->yuv_out2_ + Y_OFF_ENC;
1080*b2055c35SXin Li int total_header_bits = 0;
1081*b2055c35SXin Li VP8ModeScore rd_best;
1082*b2055c35SXin Li
1083*b2055c35SXin Li if (enc->max_i4_header_bits_ == 0) {
1084*b2055c35SXin Li return 0;
1085*b2055c35SXin Li }
1086*b2055c35SXin Li
1087*b2055c35SXin Li InitScore(&rd_best);
1088*b2055c35SXin Li rd_best.H = 211; // '211' is the value of VP8BitCost(0, 145)
1089*b2055c35SXin Li SetRDScore(dqm->lambda_mode_, &rd_best);
1090*b2055c35SXin Li VP8IteratorStartI4(it);
1091*b2055c35SXin Li do {
1092*b2055c35SXin Li const int kNumBlocks = 1;
1093*b2055c35SXin Li VP8ModeScore rd_i4;
1094*b2055c35SXin Li int mode;
1095*b2055c35SXin Li int best_mode = -1;
1096*b2055c35SXin Li const uint8_t* const src = src0 + VP8Scan[it->i4_];
1097*b2055c35SXin Li const uint16_t* const mode_costs = GetCostModeI4(it, rd->modes_i4);
1098*b2055c35SXin Li uint8_t* best_block = best_blocks + VP8Scan[it->i4_];
1099*b2055c35SXin Li uint8_t* tmp_dst = it->yuv_p_ + I4TMP; // scratch buffer.
1100*b2055c35SXin Li
1101*b2055c35SXin Li InitScore(&rd_i4);
1102*b2055c35SXin Li VP8MakeIntra4Preds(it);
1103*b2055c35SXin Li for (mode = 0; mode < NUM_BMODES; ++mode) {
1104*b2055c35SXin Li VP8ModeScore rd_tmp;
1105*b2055c35SXin Li int16_t tmp_levels[16];
1106*b2055c35SXin Li
1107*b2055c35SXin Li // Reconstruct
1108*b2055c35SXin Li rd_tmp.nz =
1109*b2055c35SXin Li ReconstructIntra4(it, tmp_levels, src, tmp_dst, mode) << it->i4_;
1110*b2055c35SXin Li
1111*b2055c35SXin Li // Compute RD-score
1112*b2055c35SXin Li rd_tmp.D = VP8SSE4x4(src, tmp_dst);
1113*b2055c35SXin Li rd_tmp.SD =
1114*b2055c35SXin Li tlambda ? MULT_8B(tlambda, VP8TDisto4x4(src, tmp_dst, kWeightY))
1115*b2055c35SXin Li : 0;
1116*b2055c35SXin Li rd_tmp.H = mode_costs[mode];
1117*b2055c35SXin Li
1118*b2055c35SXin Li // Add flatness penalty, to avoid flat area to be mispredicted
1119*b2055c35SXin Li // by a complex mode.
1120*b2055c35SXin Li if (mode > 0 && IsFlat(tmp_levels, kNumBlocks, FLATNESS_LIMIT_I4)) {
1121*b2055c35SXin Li rd_tmp.R = FLATNESS_PENALTY * kNumBlocks;
1122*b2055c35SXin Li } else {
1123*b2055c35SXin Li rd_tmp.R = 0;
1124*b2055c35SXin Li }
1125*b2055c35SXin Li
1126*b2055c35SXin Li // early-out check
1127*b2055c35SXin Li SetRDScore(lambda, &rd_tmp);
1128*b2055c35SXin Li if (best_mode >= 0 && rd_tmp.score >= rd_i4.score) continue;
1129*b2055c35SXin Li
1130*b2055c35SXin Li // finish computing score
1131*b2055c35SXin Li rd_tmp.R += VP8GetCostLuma4(it, tmp_levels);
1132*b2055c35SXin Li SetRDScore(lambda, &rd_tmp);
1133*b2055c35SXin Li
1134*b2055c35SXin Li if (best_mode < 0 || rd_tmp.score < rd_i4.score) {
1135*b2055c35SXin Li CopyScore(&rd_i4, &rd_tmp);
1136*b2055c35SXin Li best_mode = mode;
1137*b2055c35SXin Li SwapPtr(&tmp_dst, &best_block);
1138*b2055c35SXin Li memcpy(rd_best.y_ac_levels[it->i4_], tmp_levels,
1139*b2055c35SXin Li sizeof(rd_best.y_ac_levels[it->i4_]));
1140*b2055c35SXin Li }
1141*b2055c35SXin Li }
1142*b2055c35SXin Li SetRDScore(dqm->lambda_mode_, &rd_i4);
1143*b2055c35SXin Li AddScore(&rd_best, &rd_i4);
1144*b2055c35SXin Li if (rd_best.score >= rd->score) {
1145*b2055c35SXin Li return 0;
1146*b2055c35SXin Li }
1147*b2055c35SXin Li total_header_bits += (int)rd_i4.H; // <- equal to mode_costs[best_mode];
1148*b2055c35SXin Li if (total_header_bits > enc->max_i4_header_bits_) {
1149*b2055c35SXin Li return 0;
1150*b2055c35SXin Li }
1151*b2055c35SXin Li // Copy selected samples if not in the right place already.
1152*b2055c35SXin Li if (best_block != best_blocks + VP8Scan[it->i4_]) {
1153*b2055c35SXin Li VP8Copy4x4(best_block, best_blocks + VP8Scan[it->i4_]);
1154*b2055c35SXin Li }
1155*b2055c35SXin Li rd->modes_i4[it->i4_] = best_mode;
1156*b2055c35SXin Li it->top_nz_[it->i4_ & 3] = it->left_nz_[it->i4_ >> 2] = (rd_i4.nz ? 1 : 0);
1157*b2055c35SXin Li } while (VP8IteratorRotateI4(it, best_blocks));
1158*b2055c35SXin Li
1159*b2055c35SXin Li // finalize state
1160*b2055c35SXin Li CopyScore(rd, &rd_best);
1161*b2055c35SXin Li VP8SetIntra4Mode(it, rd->modes_i4);
1162*b2055c35SXin Li SwapOut(it);
1163*b2055c35SXin Li memcpy(rd->y_ac_levels, rd_best.y_ac_levels, sizeof(rd->y_ac_levels));
1164*b2055c35SXin Li return 1; // select intra4x4 over intra16x16
1165*b2055c35SXin Li }
1166*b2055c35SXin Li
1167*b2055c35SXin Li //------------------------------------------------------------------------------
1168*b2055c35SXin Li
PickBestUV(VP8EncIterator * WEBP_RESTRICT const it,VP8ModeScore * WEBP_RESTRICT const rd)1169*b2055c35SXin Li static void PickBestUV(VP8EncIterator* WEBP_RESTRICT const it,
1170*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd) {
1171*b2055c35SXin Li const int kNumBlocks = 8;
1172*b2055c35SXin Li const VP8SegmentInfo* const dqm = &it->enc_->dqm_[it->mb_->segment_];
1173*b2055c35SXin Li const int lambda = dqm->lambda_uv_;
1174*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + U_OFF_ENC;
1175*b2055c35SXin Li uint8_t* tmp_dst = it->yuv_out2_ + U_OFF_ENC; // scratch buffer
1176*b2055c35SXin Li uint8_t* dst0 = it->yuv_out_ + U_OFF_ENC;
1177*b2055c35SXin Li uint8_t* dst = dst0;
1178*b2055c35SXin Li VP8ModeScore rd_best;
1179*b2055c35SXin Li int mode;
1180*b2055c35SXin Li
1181*b2055c35SXin Li rd->mode_uv = -1;
1182*b2055c35SXin Li InitScore(&rd_best);
1183*b2055c35SXin Li for (mode = 0; mode < NUM_PRED_MODES; ++mode) {
1184*b2055c35SXin Li VP8ModeScore rd_uv;
1185*b2055c35SXin Li
1186*b2055c35SXin Li // Reconstruct
1187*b2055c35SXin Li rd_uv.nz = ReconstructUV(it, &rd_uv, tmp_dst, mode);
1188*b2055c35SXin Li
1189*b2055c35SXin Li // Compute RD-score
1190*b2055c35SXin Li rd_uv.D = VP8SSE16x8(src, tmp_dst);
1191*b2055c35SXin Li rd_uv.SD = 0; // not calling TDisto here: it tends to flatten areas.
1192*b2055c35SXin Li rd_uv.H = VP8FixedCostsUV[mode];
1193*b2055c35SXin Li rd_uv.R = VP8GetCostUV(it, &rd_uv);
1194*b2055c35SXin Li if (mode > 0 && IsFlat(rd_uv.uv_levels[0], kNumBlocks, FLATNESS_LIMIT_UV)) {
1195*b2055c35SXin Li rd_uv.R += FLATNESS_PENALTY * kNumBlocks;
1196*b2055c35SXin Li }
1197*b2055c35SXin Li
1198*b2055c35SXin Li SetRDScore(lambda, &rd_uv);
1199*b2055c35SXin Li if (mode == 0 || rd_uv.score < rd_best.score) {
1200*b2055c35SXin Li CopyScore(&rd_best, &rd_uv);
1201*b2055c35SXin Li rd->mode_uv = mode;
1202*b2055c35SXin Li memcpy(rd->uv_levels, rd_uv.uv_levels, sizeof(rd->uv_levels));
1203*b2055c35SXin Li if (it->top_derr_ != NULL) {
1204*b2055c35SXin Li memcpy(rd->derr, rd_uv.derr, sizeof(rd_uv.derr));
1205*b2055c35SXin Li }
1206*b2055c35SXin Li SwapPtr(&dst, &tmp_dst);
1207*b2055c35SXin Li }
1208*b2055c35SXin Li }
1209*b2055c35SXin Li VP8SetIntraUVMode(it, rd->mode_uv);
1210*b2055c35SXin Li AddScore(rd, &rd_best);
1211*b2055c35SXin Li if (dst != dst0) { // copy 16x8 block if needed
1212*b2055c35SXin Li VP8Copy16x8(dst, dst0);
1213*b2055c35SXin Li }
1214*b2055c35SXin Li if (it->top_derr_ != NULL) { // store diffusion errors for next block
1215*b2055c35SXin Li StoreDiffusionErrors(it, rd);
1216*b2055c35SXin Li }
1217*b2055c35SXin Li }
1218*b2055c35SXin Li
1219*b2055c35SXin Li //------------------------------------------------------------------------------
1220*b2055c35SXin Li // Final reconstruction and quantization.
1221*b2055c35SXin Li
SimpleQuantize(VP8EncIterator * WEBP_RESTRICT const it,VP8ModeScore * WEBP_RESTRICT const rd)1222*b2055c35SXin Li static void SimpleQuantize(VP8EncIterator* WEBP_RESTRICT const it,
1223*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd) {
1224*b2055c35SXin Li const VP8Encoder* const enc = it->enc_;
1225*b2055c35SXin Li const int is_i16 = (it->mb_->type_ == 1);
1226*b2055c35SXin Li int nz = 0;
1227*b2055c35SXin Li
1228*b2055c35SXin Li if (is_i16) {
1229*b2055c35SXin Li nz = ReconstructIntra16(it, rd, it->yuv_out_ + Y_OFF_ENC, it->preds_[0]);
1230*b2055c35SXin Li } else {
1231*b2055c35SXin Li VP8IteratorStartI4(it);
1232*b2055c35SXin Li do {
1233*b2055c35SXin Li const int mode =
1234*b2055c35SXin Li it->preds_[(it->i4_ & 3) + (it->i4_ >> 2) * enc->preds_w_];
1235*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + Y_OFF_ENC + VP8Scan[it->i4_];
1236*b2055c35SXin Li uint8_t* const dst = it->yuv_out_ + Y_OFF_ENC + VP8Scan[it->i4_];
1237*b2055c35SXin Li VP8MakeIntra4Preds(it);
1238*b2055c35SXin Li nz |= ReconstructIntra4(it, rd->y_ac_levels[it->i4_],
1239*b2055c35SXin Li src, dst, mode) << it->i4_;
1240*b2055c35SXin Li } while (VP8IteratorRotateI4(it, it->yuv_out_ + Y_OFF_ENC));
1241*b2055c35SXin Li }
1242*b2055c35SXin Li
1243*b2055c35SXin Li nz |= ReconstructUV(it, rd, it->yuv_out_ + U_OFF_ENC, it->mb_->uv_mode_);
1244*b2055c35SXin Li rd->nz = nz;
1245*b2055c35SXin Li }
1246*b2055c35SXin Li
1247*b2055c35SXin Li // Refine intra16/intra4 sub-modes based on distortion only (not rate).
RefineUsingDistortion(VP8EncIterator * WEBP_RESTRICT const it,int try_both_modes,int refine_uv_mode,VP8ModeScore * WEBP_RESTRICT const rd)1248*b2055c35SXin Li static void RefineUsingDistortion(VP8EncIterator* WEBP_RESTRICT const it,
1249*b2055c35SXin Li int try_both_modes, int refine_uv_mode,
1250*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd) {
1251*b2055c35SXin Li score_t best_score = MAX_COST;
1252*b2055c35SXin Li int nz = 0;
1253*b2055c35SXin Li int mode;
1254*b2055c35SXin Li int is_i16 = try_both_modes || (it->mb_->type_ == 1);
1255*b2055c35SXin Li
1256*b2055c35SXin Li const VP8SegmentInfo* const dqm = &it->enc_->dqm_[it->mb_->segment_];
1257*b2055c35SXin Li // Some empiric constants, of approximate order of magnitude.
1258*b2055c35SXin Li const int lambda_d_i16 = 106;
1259*b2055c35SXin Li const int lambda_d_i4 = 11;
1260*b2055c35SXin Li const int lambda_d_uv = 120;
1261*b2055c35SXin Li score_t score_i4 = dqm->i4_penalty_;
1262*b2055c35SXin Li score_t i4_bit_sum = 0;
1263*b2055c35SXin Li const score_t bit_limit = try_both_modes ? it->enc_->mb_header_limit_
1264*b2055c35SXin Li : MAX_COST; // no early-out allowed
1265*b2055c35SXin Li
1266*b2055c35SXin Li if (is_i16) { // First, evaluate Intra16 distortion
1267*b2055c35SXin Li int best_mode = -1;
1268*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + Y_OFF_ENC;
1269*b2055c35SXin Li for (mode = 0; mode < NUM_PRED_MODES; ++mode) {
1270*b2055c35SXin Li const uint8_t* const ref = it->yuv_p_ + VP8I16ModeOffsets[mode];
1271*b2055c35SXin Li const score_t score = (score_t)VP8SSE16x16(src, ref) * RD_DISTO_MULT
1272*b2055c35SXin Li + VP8FixedCostsI16[mode] * lambda_d_i16;
1273*b2055c35SXin Li if (mode > 0 && VP8FixedCostsI16[mode] > bit_limit) {
1274*b2055c35SXin Li continue;
1275*b2055c35SXin Li }
1276*b2055c35SXin Li
1277*b2055c35SXin Li if (score < best_score) {
1278*b2055c35SXin Li best_mode = mode;
1279*b2055c35SXin Li best_score = score;
1280*b2055c35SXin Li }
1281*b2055c35SXin Li }
1282*b2055c35SXin Li if (it->x_ == 0 || it->y_ == 0) {
1283*b2055c35SXin Li // avoid starting a checkerboard resonance from the border. See bug #432.
1284*b2055c35SXin Li if (IsFlatSource16(src)) {
1285*b2055c35SXin Li best_mode = (it->x_ == 0) ? 0 : 2;
1286*b2055c35SXin Li try_both_modes = 0; // stick to i16
1287*b2055c35SXin Li }
1288*b2055c35SXin Li }
1289*b2055c35SXin Li VP8SetIntra16Mode(it, best_mode);
1290*b2055c35SXin Li // we'll reconstruct later, if i16 mode actually gets selected
1291*b2055c35SXin Li }
1292*b2055c35SXin Li
1293*b2055c35SXin Li // Next, evaluate Intra4
1294*b2055c35SXin Li if (try_both_modes || !is_i16) {
1295*b2055c35SXin Li // We don't evaluate the rate here, but just account for it through a
1296*b2055c35SXin Li // constant penalty (i4 mode usually needs more bits compared to i16).
1297*b2055c35SXin Li is_i16 = 0;
1298*b2055c35SXin Li VP8IteratorStartI4(it);
1299*b2055c35SXin Li do {
1300*b2055c35SXin Li int best_i4_mode = -1;
1301*b2055c35SXin Li score_t best_i4_score = MAX_COST;
1302*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + Y_OFF_ENC + VP8Scan[it->i4_];
1303*b2055c35SXin Li const uint16_t* const mode_costs = GetCostModeI4(it, rd->modes_i4);
1304*b2055c35SXin Li
1305*b2055c35SXin Li VP8MakeIntra4Preds(it);
1306*b2055c35SXin Li for (mode = 0; mode < NUM_BMODES; ++mode) {
1307*b2055c35SXin Li const uint8_t* const ref = it->yuv_p_ + VP8I4ModeOffsets[mode];
1308*b2055c35SXin Li const score_t score = VP8SSE4x4(src, ref) * RD_DISTO_MULT
1309*b2055c35SXin Li + mode_costs[mode] * lambda_d_i4;
1310*b2055c35SXin Li if (score < best_i4_score) {
1311*b2055c35SXin Li best_i4_mode = mode;
1312*b2055c35SXin Li best_i4_score = score;
1313*b2055c35SXin Li }
1314*b2055c35SXin Li }
1315*b2055c35SXin Li i4_bit_sum += mode_costs[best_i4_mode];
1316*b2055c35SXin Li rd->modes_i4[it->i4_] = best_i4_mode;
1317*b2055c35SXin Li score_i4 += best_i4_score;
1318*b2055c35SXin Li if (score_i4 >= best_score || i4_bit_sum > bit_limit) {
1319*b2055c35SXin Li // Intra4 won't be better than Intra16. Bail out and pick Intra16.
1320*b2055c35SXin Li is_i16 = 1;
1321*b2055c35SXin Li break;
1322*b2055c35SXin Li } else { // reconstruct partial block inside yuv_out2_ buffer
1323*b2055c35SXin Li uint8_t* const tmp_dst = it->yuv_out2_ + Y_OFF_ENC + VP8Scan[it->i4_];
1324*b2055c35SXin Li nz |= ReconstructIntra4(it, rd->y_ac_levels[it->i4_],
1325*b2055c35SXin Li src, tmp_dst, best_i4_mode) << it->i4_;
1326*b2055c35SXin Li }
1327*b2055c35SXin Li } while (VP8IteratorRotateI4(it, it->yuv_out2_ + Y_OFF_ENC));
1328*b2055c35SXin Li }
1329*b2055c35SXin Li
1330*b2055c35SXin Li // Final reconstruction, depending on which mode is selected.
1331*b2055c35SXin Li if (!is_i16) {
1332*b2055c35SXin Li VP8SetIntra4Mode(it, rd->modes_i4);
1333*b2055c35SXin Li SwapOut(it);
1334*b2055c35SXin Li best_score = score_i4;
1335*b2055c35SXin Li } else {
1336*b2055c35SXin Li nz = ReconstructIntra16(it, rd, it->yuv_out_ + Y_OFF_ENC, it->preds_[0]);
1337*b2055c35SXin Li }
1338*b2055c35SXin Li
1339*b2055c35SXin Li // ... and UV!
1340*b2055c35SXin Li if (refine_uv_mode) {
1341*b2055c35SXin Li int best_mode = -1;
1342*b2055c35SXin Li score_t best_uv_score = MAX_COST;
1343*b2055c35SXin Li const uint8_t* const src = it->yuv_in_ + U_OFF_ENC;
1344*b2055c35SXin Li for (mode = 0; mode < NUM_PRED_MODES; ++mode) {
1345*b2055c35SXin Li const uint8_t* const ref = it->yuv_p_ + VP8UVModeOffsets[mode];
1346*b2055c35SXin Li const score_t score = VP8SSE16x8(src, ref) * RD_DISTO_MULT
1347*b2055c35SXin Li + VP8FixedCostsUV[mode] * lambda_d_uv;
1348*b2055c35SXin Li if (score < best_uv_score) {
1349*b2055c35SXin Li best_mode = mode;
1350*b2055c35SXin Li best_uv_score = score;
1351*b2055c35SXin Li }
1352*b2055c35SXin Li }
1353*b2055c35SXin Li VP8SetIntraUVMode(it, best_mode);
1354*b2055c35SXin Li }
1355*b2055c35SXin Li nz |= ReconstructUV(it, rd, it->yuv_out_ + U_OFF_ENC, it->mb_->uv_mode_);
1356*b2055c35SXin Li
1357*b2055c35SXin Li rd->nz = nz;
1358*b2055c35SXin Li rd->score = best_score;
1359*b2055c35SXin Li }
1360*b2055c35SXin Li
1361*b2055c35SXin Li //------------------------------------------------------------------------------
1362*b2055c35SXin Li // Entry point
1363*b2055c35SXin Li
VP8Decimate(VP8EncIterator * WEBP_RESTRICT const it,VP8ModeScore * WEBP_RESTRICT const rd,VP8RDLevel rd_opt)1364*b2055c35SXin Li int VP8Decimate(VP8EncIterator* WEBP_RESTRICT const it,
1365*b2055c35SXin Li VP8ModeScore* WEBP_RESTRICT const rd,
1366*b2055c35SXin Li VP8RDLevel rd_opt) {
1367*b2055c35SXin Li int is_skipped;
1368*b2055c35SXin Li const int method = it->enc_->method_;
1369*b2055c35SXin Li
1370*b2055c35SXin Li InitScore(rd);
1371*b2055c35SXin Li
1372*b2055c35SXin Li // We can perform predictions for Luma16x16 and Chroma8x8 already.
1373*b2055c35SXin Li // Luma4x4 predictions needs to be done as-we-go.
1374*b2055c35SXin Li VP8MakeLuma16Preds(it);
1375*b2055c35SXin Li VP8MakeChroma8Preds(it);
1376*b2055c35SXin Li
1377*b2055c35SXin Li if (rd_opt > RD_OPT_NONE) {
1378*b2055c35SXin Li it->do_trellis_ = (rd_opt >= RD_OPT_TRELLIS_ALL);
1379*b2055c35SXin Li PickBestIntra16(it, rd);
1380*b2055c35SXin Li if (method >= 2) {
1381*b2055c35SXin Li PickBestIntra4(it, rd);
1382*b2055c35SXin Li }
1383*b2055c35SXin Li PickBestUV(it, rd);
1384*b2055c35SXin Li if (rd_opt == RD_OPT_TRELLIS) { // finish off with trellis-optim now
1385*b2055c35SXin Li it->do_trellis_ = 1;
1386*b2055c35SXin Li SimpleQuantize(it, rd);
1387*b2055c35SXin Li }
1388*b2055c35SXin Li } else {
1389*b2055c35SXin Li // At this point we have heuristically decided intra16 / intra4.
1390*b2055c35SXin Li // For method >= 2, pick the best intra4/intra16 based on SSE (~tad slower).
1391*b2055c35SXin Li // For method <= 1, we don't re-examine the decision but just go ahead with
1392*b2055c35SXin Li // quantization/reconstruction.
1393*b2055c35SXin Li RefineUsingDistortion(it, (method >= 2), (method >= 1), rd);
1394*b2055c35SXin Li }
1395*b2055c35SXin Li is_skipped = (rd->nz == 0);
1396*b2055c35SXin Li VP8SetSkip(it, is_skipped);
1397*b2055c35SXin Li return is_skipped;
1398*b2055c35SXin Li }
1399