Use a faster way to compute 255-val: Instead of creating a vector of
[libav.git] / libavcodec / h264.c
CommitLineData
0da71265
MN
1/*
2 * H.26L/H.264/AVC/JVT/14496-10/... encoder/decoder
3 * Copyright (c) 2003 Michael Niedermayer <michaelni@gmx.at>
4 *
b78e7197
DB
5 * This file is part of FFmpeg.
6 *
7 * FFmpeg is free software; you can redistribute it and/or
0da71265
MN
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
b78e7197 10 * version 2.1 of the License, or (at your option) any later version.
0da71265 11 *
b78e7197 12 * FFmpeg is distributed in the hope that it will be useful,
0da71265
MN
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
b78e7197 18 * License along with FFmpeg; if not, write to the Free Software
5509bffa 19 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
0da71265
MN
20 *
21 */
115329f1 22
0da71265
MN
23/**
24 * @file h264.c
25 * H.264 / AVC / MPEG4 part10 codec.
26 * @author Michael Niedermayer <michaelni@gmx.at>
27 */
28
0da71265
MN
29#include "dsputil.h"
30#include "avcodec.h"
31#include "mpegvideo.h"
26b4fe82 32#include "h264.h"
0da71265 33#include "h264data.h"
26b4fe82 34#include "h264_parser.h"
0da71265
MN
35#include "golomb.h"
36
e5017ab8
LA
37#include "cabac.h"
38
2848ce84 39//#undef NDEBUG
0da71265
MN
40#include <assert.h>
41
0da71265
MN
42static VLC coeff_token_vlc[4];
43static VLC chroma_dc_coeff_token_vlc;
44
45static VLC total_zeros_vlc[15];
46static VLC chroma_dc_total_zeros_vlc[3];
47
48static VLC run_vlc[6];
49static VLC run7_vlc;
50
8b82a956
MN
51static void svq3_luma_dc_dequant_idct_c(DCTELEM *block, int qp);
52static void svq3_add_idct_c(uint8_t *dst, DCTELEM *block, int stride, int qp, int dc);
6ba71fc4 53static void filter_mb( H264Context *h, int mb_x, int mb_y, uint8_t *img_y, uint8_t *img_cb, uint8_t *img_cr, unsigned int linesize, unsigned int uvlinesize);
3e20143e 54static void filter_mb_fast( H264Context *h, int mb_x, int mb_y, uint8_t *img_y, uint8_t *img_cb, uint8_t *img_cr, unsigned int linesize, unsigned int uvlinesize);
8b82a956 55
849f1035 56static av_always_inline uint32_t pack16to32(int a, int b){
377ec888
MN
57#ifdef WORDS_BIGENDIAN
58 return (b&0xFFFF) + (a<<16);
59#else
60 return (a&0xFFFF) + (b<<16);
61#endif
62}
63
acd8d10f
PI
64const uint8_t ff_rem6[52]={
650, 1, 2, 3, 4, 5, 0, 1, 2, 3, 4, 5, 0, 1, 2, 3, 4, 5, 0, 1, 2, 3, 4, 5, 0, 1, 2, 3, 4, 5, 0, 1, 2, 3, 4, 5, 0, 1, 2, 3, 4, 5, 0, 1, 2, 3, 4, 5, 0, 1, 2, 3,
66};
67
68const uint8_t ff_div6[52]={
690, 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 3, 3, 3, 3, 3, 3, 4, 4, 4, 4, 4, 4, 5, 5, 5, 5, 5, 5, 6, 6, 6, 6, 6, 6, 7, 7, 7, 7, 7, 7, 8, 8, 8, 8,
70};
71
72
0da71265
MN
73/**
74 * fill a rectangle.
5175b937
LLL
75 * @param h height of the rectangle, should be a constant
76 * @param w width of the rectangle, should be a constant
0da71265
MN
77 * @param size the size of val (1 or 4), should be a constant
78 */
849f1035 79static av_always_inline void fill_rectangle(void *vp, int w, int h, int stride, uint32_t val, int size){
af6e2fed 80 uint8_t *p= (uint8_t*)vp;
0da71265 81 assert(size==1 || size==4);
67a82086 82 assert(w<=4);
115329f1 83
0da71265
MN
84 w *= size;
85 stride *= size;
115329f1 86
4733abcb 87 assert((((long)vp)&(FFMIN(w, STRIDE_ALIGN)-1)) == 0);
d9c780a8 88 assert((stride&(w-1))==0);
67a82086
LM
89 if(w==2){
90 const uint16_t v= size==4 ? val : val*0x0101;
91 *(uint16_t*)(p + 0*stride)= v;
92 if(h==1) return;
93 *(uint16_t*)(p + 1*stride)= v;
94 if(h==2) return;
0da71265 95 *(uint16_t*)(p + 2*stride)=
67a82086
LM
96 *(uint16_t*)(p + 3*stride)= v;
97 }else if(w==4){
98 const uint32_t v= size==4 ? val : val*0x01010101;
99 *(uint32_t*)(p + 0*stride)= v;
100 if(h==1) return;
101 *(uint32_t*)(p + 1*stride)= v;
102 if(h==2) return;
0da71265 103 *(uint32_t*)(p + 2*stride)=
67a82086
LM
104 *(uint32_t*)(p + 3*stride)= v;
105 }else if(w==8){
106 //gcc can't optimize 64bit math on x86_32
107#if defined(ARCH_X86_64) || (defined(MP_WORDSIZE) && MP_WORDSIZE >= 64)
108 const uint64_t v= val*0x0100000001ULL;
109 *(uint64_t*)(p + 0*stride)= v;
110 if(h==1) return;
111 *(uint64_t*)(p + 1*stride)= v;
112 if(h==2) return;
0da71265 113 *(uint64_t*)(p + 2*stride)=
67a82086
LM
114 *(uint64_t*)(p + 3*stride)= v;
115 }else if(w==16){
116 const uint64_t v= val*0x0100000001ULL;
0da71265
MN
117 *(uint64_t*)(p + 0+0*stride)=
118 *(uint64_t*)(p + 8+0*stride)=
119 *(uint64_t*)(p + 0+1*stride)=
67a82086
LM
120 *(uint64_t*)(p + 8+1*stride)= v;
121 if(h==2) return;
0da71265
MN
122 *(uint64_t*)(p + 0+2*stride)=
123 *(uint64_t*)(p + 8+2*stride)=
124 *(uint64_t*)(p + 0+3*stride)=
67a82086
LM
125 *(uint64_t*)(p + 8+3*stride)= v;
126#else
127 *(uint32_t*)(p + 0+0*stride)=
128 *(uint32_t*)(p + 4+0*stride)= val;
129 if(h==1) return;
130 *(uint32_t*)(p + 0+1*stride)=
131 *(uint32_t*)(p + 4+1*stride)= val;
132 if(h==2) return;
133 *(uint32_t*)(p + 0+2*stride)=
134 *(uint32_t*)(p + 4+2*stride)=
135 *(uint32_t*)(p + 0+3*stride)=
136 *(uint32_t*)(p + 4+3*stride)= val;
137 }else if(w==16){
138 *(uint32_t*)(p + 0+0*stride)=
139 *(uint32_t*)(p + 4+0*stride)=
140 *(uint32_t*)(p + 8+0*stride)=
141 *(uint32_t*)(p +12+0*stride)=
142 *(uint32_t*)(p + 0+1*stride)=
143 *(uint32_t*)(p + 4+1*stride)=
144 *(uint32_t*)(p + 8+1*stride)=
145 *(uint32_t*)(p +12+1*stride)= val;
146 if(h==2) return;
147 *(uint32_t*)(p + 0+2*stride)=
148 *(uint32_t*)(p + 4+2*stride)=
149 *(uint32_t*)(p + 8+2*stride)=
150 *(uint32_t*)(p +12+2*stride)=
151 *(uint32_t*)(p + 0+3*stride)=
152 *(uint32_t*)(p + 4+3*stride)=
153 *(uint32_t*)(p + 8+3*stride)=
154 *(uint32_t*)(p +12+3*stride)= val;
155#endif
0da71265
MN
156 }else
157 assert(0);
67a82086 158 assert(h==4);
0da71265
MN
159}
160
70abb407 161static void fill_caches(H264Context *h, int mb_type, int for_deblock){
0da71265 162 MpegEncContext * const s = &h->s;
7bc9090a 163 const int mb_xy= s->mb_x + s->mb_y*s->mb_stride;
0da71265
MN
164 int topleft_xy, top_xy, topright_xy, left_xy[2];
165 int topleft_type, top_type, topright_type, left_type[2];
6867a90b 166 int left_block[8];
0da71265
MN
167 int i;
168
717b1733
LM
169 //FIXME deblocking could skip the intra and nnz parts.
170 if(for_deblock && (h->slice_num == 1 || h->slice_table[mb_xy] == h->slice_table[mb_xy-s->mb_stride]) && !FRAME_MBAFF)
e2e5894a
LM
171 return;
172
115329f1
DB
173 //wow what a mess, why didn't they simplify the interlacing&intra stuff, i can't imagine that these complex rules are worth it
174
6867a90b
LLL
175 top_xy = mb_xy - s->mb_stride;
176 topleft_xy = top_xy - 1;
177 topright_xy= top_xy + 1;
178 left_xy[1] = left_xy[0] = mb_xy-1;
179 left_block[0]= 0;
180 left_block[1]= 1;
181 left_block[2]= 2;
182 left_block[3]= 3;
183 left_block[4]= 7;
184 left_block[5]= 10;
185 left_block[6]= 8;
186 left_block[7]= 11;
5d18eaad 187 if(FRAME_MBAFF){
6867a90b
LLL
188 const int pair_xy = s->mb_x + (s->mb_y & ~1)*s->mb_stride;
189 const int top_pair_xy = pair_xy - s->mb_stride;
190 const int topleft_pair_xy = top_pair_xy - 1;
191 const int topright_pair_xy = top_pair_xy + 1;
192 const int topleft_mb_frame_flag = !IS_INTERLACED(s->current_picture.mb_type[topleft_pair_xy]);
193 const int top_mb_frame_flag = !IS_INTERLACED(s->current_picture.mb_type[top_pair_xy]);
194 const int topright_mb_frame_flag = !IS_INTERLACED(s->current_picture.mb_type[topright_pair_xy]);
195 const int left_mb_frame_flag = !IS_INTERLACED(s->current_picture.mb_type[pair_xy-1]);
196 const int curr_mb_frame_flag = !IS_INTERLACED(mb_type);
197 const int bottom = (s->mb_y & 1);
a9c9a240 198 tprintf(s->avctx, "fill_caches: curr_mb_frame_flag:%d, left_mb_frame_flag:%d, topleft_mb_frame_flag:%d, top_mb_frame_flag:%d, topright_mb_frame_flag:%d\n", curr_mb_frame_flag, left_mb_frame_flag, topleft_mb_frame_flag, top_mb_frame_flag, topright_mb_frame_flag);
6867a90b
LLL
199 if (bottom
200 ? !curr_mb_frame_flag // bottom macroblock
201 : (!curr_mb_frame_flag && !top_mb_frame_flag) // top macroblock
202 ) {
203 top_xy -= s->mb_stride;
204 }
205 if (bottom
206 ? !curr_mb_frame_flag // bottom macroblock
207 : (!curr_mb_frame_flag && !topleft_mb_frame_flag) // top macroblock
208 ) {
209 topleft_xy -= s->mb_stride;
210 }
211 if (bottom
212 ? !curr_mb_frame_flag // bottom macroblock
213 : (!curr_mb_frame_flag && !topright_mb_frame_flag) // top macroblock
214 ) {
215 topright_xy -= s->mb_stride;
216 }
217 if (left_mb_frame_flag != curr_mb_frame_flag) {
218 left_xy[1] = left_xy[0] = pair_xy - 1;
219 if (curr_mb_frame_flag) {
220 if (bottom) {
221 left_block[0]= 2;
222 left_block[1]= 2;
223 left_block[2]= 3;
224 left_block[3]= 3;
225 left_block[4]= 8;
226 left_block[5]= 11;
227 left_block[6]= 8;
228 left_block[7]= 11;
229 } else {
230 left_block[0]= 0;
231 left_block[1]= 0;
232 left_block[2]= 1;
233 left_block[3]= 1;
234 left_block[4]= 7;
235 left_block[5]= 10;
236 left_block[6]= 7;
237 left_block[7]= 10;
238 }
239 } else {
240 left_xy[1] += s->mb_stride;
241 //left_block[0]= 0;
242 left_block[1]= 2;
243 left_block[2]= 0;
244 left_block[3]= 2;
245 //left_block[4]= 7;
246 left_block[5]= 10;
247 left_block[6]= 7;
248 left_block[7]= 10;
249 }
250 }
0da71265
MN
251 }
252
826de46e
LLL
253 h->top_mb_xy = top_xy;
254 h->left_mb_xy[0] = left_xy[0];
255 h->left_mb_xy[1] = left_xy[1];
6ba71fc4 256 if(for_deblock){
717b1733
LM
257 topleft_type = 0;
258 topright_type = 0;
46f2f05f 259 top_type = h->slice_table[top_xy ] < 255 ? s->current_picture.mb_type[top_xy] : 0;
46f2f05f
MN
260 left_type[0] = h->slice_table[left_xy[0] ] < 255 ? s->current_picture.mb_type[left_xy[0]] : 0;
261 left_type[1] = h->slice_table[left_xy[1] ] < 255 ? s->current_picture.mb_type[left_xy[1]] : 0;
5d18eaad
LM
262
263 if(FRAME_MBAFF && !IS_INTRA(mb_type)){
264 int list;
265 int v = *(uint16_t*)&h->non_zero_count[mb_xy][14];
266 for(i=0; i<16; i++)
267 h->non_zero_count_cache[scan8[i]] = (v>>i)&1;
3425501d 268 for(list=0; list<h->list_count; list++){
5d18eaad
LM
269 if(USES_LIST(mb_type,list)){
270 uint32_t *src = (uint32_t*)s->current_picture.motion_val[list][h->mb2b_xy[mb_xy]];
271 uint32_t *dst = (uint32_t*)h->mv_cache[list][scan8[0]];
191e8ca7 272 int8_t *ref = &s->current_picture.ref_index[list][h->mb2b8_xy[mb_xy]];
5d18eaad
LM
273 for(i=0; i<4; i++, dst+=8, src+=h->b_stride){
274 dst[0] = src[0];
275 dst[1] = src[1];
276 dst[2] = src[2];
277 dst[3] = src[3];
278 }
279 *(uint32_t*)&h->ref_cache[list][scan8[ 0]] =
280 *(uint32_t*)&h->ref_cache[list][scan8[ 2]] = pack16to32(ref[0],ref[1])*0x0101;
281 ref += h->b8_stride;
282 *(uint32_t*)&h->ref_cache[list][scan8[ 8]] =
283 *(uint32_t*)&h->ref_cache[list][scan8[10]] = pack16to32(ref[0],ref[1])*0x0101;
284 }else{
285 fill_rectangle(&h-> mv_cache[list][scan8[ 0]], 4, 4, 8, 0, 4);
286 fill_rectangle(&h->ref_cache[list][scan8[ 0]], 4, 4, 8, (uint8_t)LIST_NOT_USED, 1);
287 }
288 }
289 }
46f2f05f
MN
290 }else{
291 topleft_type = h->slice_table[topleft_xy ] == h->slice_num ? s->current_picture.mb_type[topleft_xy] : 0;
292 top_type = h->slice_table[top_xy ] == h->slice_num ? s->current_picture.mb_type[top_xy] : 0;
293 topright_type= h->slice_table[topright_xy] == h->slice_num ? s->current_picture.mb_type[topright_xy]: 0;
294 left_type[0] = h->slice_table[left_xy[0] ] == h->slice_num ? s->current_picture.mb_type[left_xy[0]] : 0;
295 left_type[1] = h->slice_table[left_xy[1] ] == h->slice_num ? s->current_picture.mb_type[left_xy[1]] : 0;
296 }
0da71265
MN
297
298 if(IS_INTRA(mb_type)){
115329f1
DB
299 h->topleft_samples_available=
300 h->top_samples_available=
0da71265
MN
301 h->left_samples_available= 0xFFFF;
302 h->topright_samples_available= 0xEEEA;
303
304 if(!IS_INTRA(top_type) && (top_type==0 || h->pps.constrained_intra_pred)){
305 h->topleft_samples_available= 0xB3FF;
306 h->top_samples_available= 0x33FF;
307 h->topright_samples_available= 0x26EA;
308 }
309 for(i=0; i<2; i++){
310 if(!IS_INTRA(left_type[i]) && (left_type[i]==0 || h->pps.constrained_intra_pred)){
311 h->topleft_samples_available&= 0xDF5F;
312 h->left_samples_available&= 0x5F5F;
313 }
314 }
115329f1 315
0da71265
MN
316 if(!IS_INTRA(topleft_type) && (topleft_type==0 || h->pps.constrained_intra_pred))
317 h->topleft_samples_available&= 0x7FFF;
115329f1 318
0da71265
MN
319 if(!IS_INTRA(topright_type) && (topright_type==0 || h->pps.constrained_intra_pred))
320 h->topright_samples_available&= 0xFBFF;
115329f1 321
0da71265
MN
322 if(IS_INTRA4x4(mb_type)){
323 if(IS_INTRA4x4(top_type)){
324 h->intra4x4_pred_mode_cache[4+8*0]= h->intra4x4_pred_mode[top_xy][4];
325 h->intra4x4_pred_mode_cache[5+8*0]= h->intra4x4_pred_mode[top_xy][5];
326 h->intra4x4_pred_mode_cache[6+8*0]= h->intra4x4_pred_mode[top_xy][6];
327 h->intra4x4_pred_mode_cache[7+8*0]= h->intra4x4_pred_mode[top_xy][3];
328 }else{
329 int pred;
6fbcaaa0 330 if(!top_type || (IS_INTER(top_type) && h->pps.constrained_intra_pred))
0da71265 331 pred= -1;
6fbcaaa0
LLL
332 else{
333 pred= 2;
0da71265
MN
334 }
335 h->intra4x4_pred_mode_cache[4+8*0]=
336 h->intra4x4_pred_mode_cache[5+8*0]=
337 h->intra4x4_pred_mode_cache[6+8*0]=
338 h->intra4x4_pred_mode_cache[7+8*0]= pred;
339 }
340 for(i=0; i<2; i++){
341 if(IS_INTRA4x4(left_type[i])){
342 h->intra4x4_pred_mode_cache[3+8*1 + 2*8*i]= h->intra4x4_pred_mode[left_xy[i]][left_block[0+2*i]];
343 h->intra4x4_pred_mode_cache[3+8*2 + 2*8*i]= h->intra4x4_pred_mode[left_xy[i]][left_block[1+2*i]];
344 }else{
345 int pred;
6fbcaaa0 346 if(!left_type[i] || (IS_INTER(left_type[i]) && h->pps.constrained_intra_pred))
0da71265 347 pred= -1;
6fbcaaa0
LLL
348 else{
349 pred= 2;
0da71265
MN
350 }
351 h->intra4x4_pred_mode_cache[3+8*1 + 2*8*i]=
352 h->intra4x4_pred_mode_cache[3+8*2 + 2*8*i]= pred;
353 }
354 }
355 }
356 }
115329f1
DB
357
358
0da71265 359/*
115329f1
DB
3600 . T T. T T T T
3611 L . .L . . . .
3622 L . .L . . . .
3633 . T TL . . . .
3644 L . .L . . . .
3655 L . .. . . . .
0da71265
MN
366*/
367//FIXME constraint_intra_pred & partitioning & nnz (lets hope this is just a typo in the spec)
368 if(top_type){
6867a90b
LLL
369 h->non_zero_count_cache[4+8*0]= h->non_zero_count[top_xy][4];
370 h->non_zero_count_cache[5+8*0]= h->non_zero_count[top_xy][5];
371 h->non_zero_count_cache[6+8*0]= h->non_zero_count[top_xy][6];
53c05b1e 372 h->non_zero_count_cache[7+8*0]= h->non_zero_count[top_xy][3];
115329f1 373
6867a90b 374 h->non_zero_count_cache[1+8*0]= h->non_zero_count[top_xy][9];
53c05b1e 375 h->non_zero_count_cache[2+8*0]= h->non_zero_count[top_xy][8];
115329f1 376
6867a90b 377 h->non_zero_count_cache[1+8*3]= h->non_zero_count[top_xy][12];
53c05b1e 378 h->non_zero_count_cache[2+8*3]= h->non_zero_count[top_xy][11];
115329f1 379
0da71265 380 }else{
115329f1 381 h->non_zero_count_cache[4+8*0]=
0da71265
MN
382 h->non_zero_count_cache[5+8*0]=
383 h->non_zero_count_cache[6+8*0]=
384 h->non_zero_count_cache[7+8*0]=
115329f1 385
0da71265
MN
386 h->non_zero_count_cache[1+8*0]=
387 h->non_zero_count_cache[2+8*0]=
115329f1 388
0da71265 389 h->non_zero_count_cache[1+8*3]=
3981c385 390 h->non_zero_count_cache[2+8*3]= h->pps.cabac && !IS_INTRA(mb_type) ? 0 : 64;
115329f1 391
0da71265 392 }
826de46e 393
6867a90b
LLL
394 for (i=0; i<2; i++) {
395 if(left_type[i]){
396 h->non_zero_count_cache[3+8*1 + 2*8*i]= h->non_zero_count[left_xy[i]][left_block[0+2*i]];
397 h->non_zero_count_cache[3+8*2 + 2*8*i]= h->non_zero_count[left_xy[i]][left_block[1+2*i]];
398 h->non_zero_count_cache[0+8*1 + 8*i]= h->non_zero_count[left_xy[i]][left_block[4+2*i]];
399 h->non_zero_count_cache[0+8*4 + 8*i]= h->non_zero_count[left_xy[i]][left_block[5+2*i]];
6867a90b 400 }else{
115329f1
DB
401 h->non_zero_count_cache[3+8*1 + 2*8*i]=
402 h->non_zero_count_cache[3+8*2 + 2*8*i]=
403 h->non_zero_count_cache[0+8*1 + 8*i]=
6867a90b 404 h->non_zero_count_cache[0+8*4 + 8*i]= h->pps.cabac && !IS_INTRA(mb_type) ? 0 : 64;
826de46e
LLL
405 }
406 }
407
408 if( h->pps.cabac ) {
409 // top_cbp
410 if(top_type) {
411 h->top_cbp = h->cbp_table[top_xy];
412 } else if(IS_INTRA(mb_type)) {
413 h->top_cbp = 0x1C0;
414 } else {
415 h->top_cbp = 0;
416 }
417 // left_cbp
418 if (left_type[0]) {
419 h->left_cbp = h->cbp_table[left_xy[0]] & 0x1f0;
420 } else if(IS_INTRA(mb_type)) {
421 h->left_cbp = 0x1C0;
422 } else {
423 h->left_cbp = 0;
424 }
425 if (left_type[0]) {
426 h->left_cbp |= ((h->cbp_table[left_xy[0]]>>((left_block[0]&(~1))+1))&0x1) << 1;
427 }
428 if (left_type[1]) {
429 h->left_cbp |= ((h->cbp_table[left_xy[1]]>>((left_block[2]&(~1))+1))&0x1) << 3;
6867a90b 430 }
0da71265 431 }
6867a90b 432
0da71265 433#if 1
e2e5894a 434 if(IS_INTER(mb_type) || IS_DIRECT(mb_type)){
0da71265 435 int list;
3425501d 436 for(list=0; list<h->list_count; list++){
e2e5894a 437 if(!USES_LIST(mb_type, list) && !IS_DIRECT(mb_type) && !h->deblocking_filter){
0da71265
MN
438 /*if(!h->mv_cache_clean[list]){
439 memset(h->mv_cache [list], 0, 8*5*2*sizeof(int16_t)); //FIXME clean only input? clean at all?
440 memset(h->ref_cache[list], PART_NOT_AVAILABLE, 8*5*sizeof(int8_t));
441 h->mv_cache_clean[list]= 1;
442 }*/
5ad984c9 443 continue;
0da71265
MN
444 }
445 h->mv_cache_clean[list]= 0;
115329f1 446
53b19144 447 if(USES_LIST(top_type, list)){
0da71265
MN
448 const int b_xy= h->mb2b_xy[top_xy] + 3*h->b_stride;
449 const int b8_xy= h->mb2b8_xy[top_xy] + h->b8_stride;
450 *(uint32_t*)h->mv_cache[list][scan8[0] + 0 - 1*8]= *(uint32_t*)s->current_picture.motion_val[list][b_xy + 0];
451 *(uint32_t*)h->mv_cache[list][scan8[0] + 1 - 1*8]= *(uint32_t*)s->current_picture.motion_val[list][b_xy + 1];
452 *(uint32_t*)h->mv_cache[list][scan8[0] + 2 - 1*8]= *(uint32_t*)s->current_picture.motion_val[list][b_xy + 2];
453 *(uint32_t*)h->mv_cache[list][scan8[0] + 3 - 1*8]= *(uint32_t*)s->current_picture.motion_val[list][b_xy + 3];
454 h->ref_cache[list][scan8[0] + 0 - 1*8]=
455 h->ref_cache[list][scan8[0] + 1 - 1*8]= s->current_picture.ref_index[list][b8_xy + 0];
456 h->ref_cache[list][scan8[0] + 2 - 1*8]=
457 h->ref_cache[list][scan8[0] + 3 - 1*8]= s->current_picture.ref_index[list][b8_xy + 1];
458 }else{
115329f1
DB
459 *(uint32_t*)h->mv_cache [list][scan8[0] + 0 - 1*8]=
460 *(uint32_t*)h->mv_cache [list][scan8[0] + 1 - 1*8]=
461 *(uint32_t*)h->mv_cache [list][scan8[0] + 2 - 1*8]=
0da71265
MN
462 *(uint32_t*)h->mv_cache [list][scan8[0] + 3 - 1*8]= 0;
463 *(uint32_t*)&h->ref_cache[list][scan8[0] + 0 - 1*8]= ((top_type ? LIST_NOT_USED : PART_NOT_AVAILABLE)&0xFF)*0x01010101;
464 }
465
4672503d
LM
466 for(i=0; i<2; i++){
467 int cache_idx = scan8[0] - 1 + i*2*8;
468 if(USES_LIST(left_type[i], list)){
469 const int b_xy= h->mb2b_xy[left_xy[i]] + 3;
470 const int b8_xy= h->mb2b8_xy[left_xy[i]] + 1;
471 *(uint32_t*)h->mv_cache[list][cache_idx ]= *(uint32_t*)s->current_picture.motion_val[list][b_xy + h->b_stride*left_block[0+i*2]];
472 *(uint32_t*)h->mv_cache[list][cache_idx+8]= *(uint32_t*)s->current_picture.motion_val[list][b_xy + h->b_stride*left_block[1+i*2]];
473 h->ref_cache[list][cache_idx ]= s->current_picture.ref_index[list][b8_xy + h->b8_stride*(left_block[0+i*2]>>1)];
474 h->ref_cache[list][cache_idx+8]= s->current_picture.ref_index[list][b8_xy + h->b8_stride*(left_block[1+i*2]>>1)];
475 }else{
476 *(uint32_t*)h->mv_cache [list][cache_idx ]=
477 *(uint32_t*)h->mv_cache [list][cache_idx+8]= 0;
478 h->ref_cache[list][cache_idx ]=
479 h->ref_cache[list][cache_idx+8]= left_type[i] ? LIST_NOT_USED : PART_NOT_AVAILABLE;
480 }
0da71265
MN
481 }
482
ae08a563 483 if((for_deblock || (IS_DIRECT(mb_type) && !h->direct_spatial_mv_pred)) && !FRAME_MBAFF)
46f2f05f
MN
484 continue;
485
53b19144 486 if(USES_LIST(topleft_type, list)){
e2e5894a
LM
487 const int b_xy = h->mb2b_xy[topleft_xy] + 3 + 3*h->b_stride;
488 const int b8_xy= h->mb2b8_xy[topleft_xy] + 1 + h->b8_stride;
489 *(uint32_t*)h->mv_cache[list][scan8[0] - 1 - 1*8]= *(uint32_t*)s->current_picture.motion_val[list][b_xy];
490 h->ref_cache[list][scan8[0] - 1 - 1*8]= s->current_picture.ref_index[list][b8_xy];
491 }else{
492 *(uint32_t*)h->mv_cache[list][scan8[0] - 1 - 1*8]= 0;
493 h->ref_cache[list][scan8[0] - 1 - 1*8]= topleft_type ? LIST_NOT_USED : PART_NOT_AVAILABLE;
494 }
115329f1 495
53b19144 496 if(USES_LIST(topright_type, list)){
e2e5894a
LM
497 const int b_xy= h->mb2b_xy[topright_xy] + 3*h->b_stride;
498 const int b8_xy= h->mb2b8_xy[topright_xy] + h->b8_stride;
499 *(uint32_t*)h->mv_cache[list][scan8[0] + 4 - 1*8]= *(uint32_t*)s->current_picture.motion_val[list][b_xy];
500 h->ref_cache[list][scan8[0] + 4 - 1*8]= s->current_picture.ref_index[list][b8_xy];
501 }else{
502 *(uint32_t*)h->mv_cache [list][scan8[0] + 4 - 1*8]= 0;
503 h->ref_cache[list][scan8[0] + 4 - 1*8]= topright_type ? LIST_NOT_USED : PART_NOT_AVAILABLE;
504 }
e2e5894a 505
ae08a563 506 if((IS_SKIP(mb_type) || IS_DIRECT(mb_type)) && !FRAME_MBAFF)
717b1733 507 continue;
115329f1
DB
508
509 h->ref_cache[list][scan8[5 ]+1] =
510 h->ref_cache[list][scan8[7 ]+1] =
3b66c4c5 511 h->ref_cache[list][scan8[13]+1] = //FIXME remove past 3 (init somewhere else)
115329f1 512 h->ref_cache[list][scan8[4 ]] =
0da71265
MN
513 h->ref_cache[list][scan8[12]] = PART_NOT_AVAILABLE;
514 *(uint32_t*)h->mv_cache [list][scan8[5 ]+1]=
515 *(uint32_t*)h->mv_cache [list][scan8[7 ]+1]=
3b66c4c5 516 *(uint32_t*)h->mv_cache [list][scan8[13]+1]= //FIXME remove past 3 (init somewhere else)
0da71265
MN
517 *(uint32_t*)h->mv_cache [list][scan8[4 ]]=
518 *(uint32_t*)h->mv_cache [list][scan8[12]]= 0;
9e528114
LA
519
520 if( h->pps.cabac ) {
521 /* XXX beurk, Load mvd */
53b19144 522 if(USES_LIST(top_type, list)){
9e528114
LA
523 const int b_xy= h->mb2b_xy[top_xy] + 3*h->b_stride;
524 *(uint32_t*)h->mvd_cache[list][scan8[0] + 0 - 1*8]= *(uint32_t*)h->mvd_table[list][b_xy + 0];
525 *(uint32_t*)h->mvd_cache[list][scan8[0] + 1 - 1*8]= *(uint32_t*)h->mvd_table[list][b_xy + 1];
526 *(uint32_t*)h->mvd_cache[list][scan8[0] + 2 - 1*8]= *(uint32_t*)h->mvd_table[list][b_xy + 2];
527 *(uint32_t*)h->mvd_cache[list][scan8[0] + 3 - 1*8]= *(uint32_t*)h->mvd_table[list][b_xy + 3];
528 }else{
115329f1
DB
529 *(uint32_t*)h->mvd_cache [list][scan8[0] + 0 - 1*8]=
530 *(uint32_t*)h->mvd_cache [list][scan8[0] + 1 - 1*8]=
531 *(uint32_t*)h->mvd_cache [list][scan8[0] + 2 - 1*8]=
9e528114
LA
532 *(uint32_t*)h->mvd_cache [list][scan8[0] + 3 - 1*8]= 0;
533 }
53b19144 534 if(USES_LIST(left_type[0], list)){
9e528114
LA
535 const int b_xy= h->mb2b_xy[left_xy[0]] + 3;
536 *(uint32_t*)h->mvd_cache[list][scan8[0] - 1 + 0*8]= *(uint32_t*)h->mvd_table[list][b_xy + h->b_stride*left_block[0]];
537 *(uint32_t*)h->mvd_cache[list][scan8[0] - 1 + 1*8]= *(uint32_t*)h->mvd_table[list][b_xy + h->b_stride*left_block[1]];
538 }else{
539 *(uint32_t*)h->mvd_cache [list][scan8[0] - 1 + 0*8]=
540 *(uint32_t*)h->mvd_cache [list][scan8[0] - 1 + 1*8]= 0;
541 }
53b19144 542 if(USES_LIST(left_type[1], list)){
9e528114
LA
543 const int b_xy= h->mb2b_xy[left_xy[1]] + 3;
544 *(uint32_t*)h->mvd_cache[list][scan8[0] - 1 + 2*8]= *(uint32_t*)h->mvd_table[list][b_xy + h->b_stride*left_block[2]];
545 *(uint32_t*)h->mvd_cache[list][scan8[0] - 1 + 3*8]= *(uint32_t*)h->mvd_table[list][b_xy + h->b_stride*left_block[3]];
546 }else{
547 *(uint32_t*)h->mvd_cache [list][scan8[0] - 1 + 2*8]=
548 *(uint32_t*)h->mvd_cache [list][scan8[0] - 1 + 3*8]= 0;
549 }
550 *(uint32_t*)h->mvd_cache [list][scan8[5 ]+1]=
551 *(uint32_t*)h->mvd_cache [list][scan8[7 ]+1]=
3b66c4c5 552 *(uint32_t*)h->mvd_cache [list][scan8[13]+1]= //FIXME remove past 3 (init somewhere else)
9e528114
LA
553 *(uint32_t*)h->mvd_cache [list][scan8[4 ]]=
554 *(uint32_t*)h->mvd_cache [list][scan8[12]]= 0;
5ad984c9
LM
555
556 if(h->slice_type == B_TYPE){
557 fill_rectangle(&h->direct_cache[scan8[0]], 4, 4, 8, 0, 1);
558
559 if(IS_DIRECT(top_type)){
560 *(uint32_t*)&h->direct_cache[scan8[0] - 1*8]= 0x01010101;
561 }else if(IS_8X8(top_type)){
562 int b8_xy = h->mb2b8_xy[top_xy] + h->b8_stride;
563 h->direct_cache[scan8[0] + 0 - 1*8]= h->direct_table[b8_xy];
564 h->direct_cache[scan8[0] + 2 - 1*8]= h->direct_table[b8_xy + 1];
565 }else{
566 *(uint32_t*)&h->direct_cache[scan8[0] - 1*8]= 0;
567 }
115329f1 568
5d18eaad
LM
569 if(IS_DIRECT(left_type[0]))
570 h->direct_cache[scan8[0] - 1 + 0*8]= 1;
571 else if(IS_8X8(left_type[0]))
572 h->direct_cache[scan8[0] - 1 + 0*8]= h->direct_table[h->mb2b8_xy[left_xy[0]] + 1 + h->b8_stride*(left_block[0]>>1)];
573 else
574 h->direct_cache[scan8[0] - 1 + 0*8]= 0;
575
576 if(IS_DIRECT(left_type[1]))
5ad984c9 577 h->direct_cache[scan8[0] - 1 + 2*8]= 1;
5d18eaad
LM
578 else if(IS_8X8(left_type[1]))
579 h->direct_cache[scan8[0] - 1 + 2*8]= h->direct_table[h->mb2b8_xy[left_xy[1]] + 1 + h->b8_stride*(left_block[2]>>1)];
580 else
5ad984c9 581 h->direct_cache[scan8[0] - 1 + 2*8]= 0;
5d18eaad
LM
582 }
583 }
584
585 if(FRAME_MBAFF){
586#define MAP_MVS\
587 MAP_F2F(scan8[0] - 1 - 1*8, topleft_type)\
588 MAP_F2F(scan8[0] + 0 - 1*8, top_type)\
589 MAP_F2F(scan8[0] + 1 - 1*8, top_type)\
590 MAP_F2F(scan8[0] + 2 - 1*8, top_type)\
591 MAP_F2F(scan8[0] + 3 - 1*8, top_type)\
592 MAP_F2F(scan8[0] + 4 - 1*8, topright_type)\
593 MAP_F2F(scan8[0] - 1 + 0*8, left_type[0])\
594 MAP_F2F(scan8[0] - 1 + 1*8, left_type[0])\
595 MAP_F2F(scan8[0] - 1 + 2*8, left_type[1])\
596 MAP_F2F(scan8[0] - 1 + 3*8, left_type[1])
597 if(MB_FIELD){
598#define MAP_F2F(idx, mb_type)\
599 if(!IS_INTERLACED(mb_type) && h->ref_cache[list][idx] >= 0){\
600 h->ref_cache[list][idx] <<= 1;\
601 h->mv_cache[list][idx][1] /= 2;\
602 h->mvd_cache[list][idx][1] /= 2;\
603 }
604 MAP_MVS
605#undef MAP_F2F
606 }else{
607#define MAP_F2F(idx, mb_type)\
608 if(IS_INTERLACED(mb_type) && h->ref_cache[list][idx] >= 0){\
609 h->ref_cache[list][idx] >>= 1;\
610 h->mv_cache[list][idx][1] <<= 1;\
611 h->mvd_cache[list][idx][1] <<= 1;\
5ad984c9 612 }
5d18eaad
LM
613 MAP_MVS
614#undef MAP_F2F
5ad984c9 615 }
9e528114 616 }
0da71265 617 }
0da71265
MN
618 }
619#endif
43efd19a
LM
620
621 h->neighbor_transform_size= !!IS_8x8DCT(top_type) + !!IS_8x8DCT(left_type[0]);
0da71265
MN
622}
623
624static inline void write_back_intra_pred_mode(H264Context *h){
625 MpegEncContext * const s = &h->s;
7bc9090a 626 const int mb_xy= s->mb_x + s->mb_y*s->mb_stride;
0da71265
MN
627
628 h->intra4x4_pred_mode[mb_xy][0]= h->intra4x4_pred_mode_cache[7+8*1];
629 h->intra4x4_pred_mode[mb_xy][1]= h->intra4x4_pred_mode_cache[7+8*2];
630 h->intra4x4_pred_mode[mb_xy][2]= h->intra4x4_pred_mode_cache[7+8*3];
631 h->intra4x4_pred_mode[mb_xy][3]= h->intra4x4_pred_mode_cache[7+8*4];
632 h->intra4x4_pred_mode[mb_xy][4]= h->intra4x4_pred_mode_cache[4+8*4];
633 h->intra4x4_pred_mode[mb_xy][5]= h->intra4x4_pred_mode_cache[5+8*4];
634 h->intra4x4_pred_mode[mb_xy][6]= h->intra4x4_pred_mode_cache[6+8*4];
635}
636
637/**
638 * checks if the top & left blocks are available if needed & changes the dc mode so it only uses the available blocks.
639 */
640static inline int check_intra4x4_pred_mode(H264Context *h){
641 MpegEncContext * const s = &h->s;
642 static const int8_t top [12]= {-1, 0,LEFT_DC_PRED,-1,-1,-1,-1,-1, 0};
643 static const int8_t left[12]= { 0,-1, TOP_DC_PRED, 0,-1,-1,-1, 0,-1,DC_128_PRED};
644 int i;
115329f1 645
0da71265
MN
646 if(!(h->top_samples_available&0x8000)){
647 for(i=0; i<4; i++){
648 int status= top[ h->intra4x4_pred_mode_cache[scan8[0] + i] ];
649 if(status<0){
9b879566 650 av_log(h->s.avctx, AV_LOG_ERROR, "top block unavailable for requested intra4x4 mode %d at %d %d\n", status, s->mb_x, s->mb_y);
0da71265
MN
651 return -1;
652 } else if(status){
653 h->intra4x4_pred_mode_cache[scan8[0] + i]= status;
654 }
655 }
656 }
115329f1 657
0da71265
MN
658 if(!(h->left_samples_available&0x8000)){
659 for(i=0; i<4; i++){
660 int status= left[ h->intra4x4_pred_mode_cache[scan8[0] + 8*i] ];
661 if(status<0){
9b879566 662 av_log(h->s.avctx, AV_LOG_ERROR, "left block unavailable for requested intra4x4 mode %d at %d %d\n", status, s->mb_x, s->mb_y);
0da71265
MN
663 return -1;
664 } else if(status){
665 h->intra4x4_pred_mode_cache[scan8[0] + 8*i]= status;
666 }
667 }
668 }
669
670 return 0;
671} //FIXME cleanup like next
672
673/**
674 * checks if the top & left blocks are available if needed & changes the dc mode so it only uses the available blocks.
675 */
676static inline int check_intra_pred_mode(H264Context *h, int mode){
677 MpegEncContext * const s = &h->s;
678 static const int8_t top [7]= {LEFT_DC_PRED8x8, 1,-1,-1};
679 static const int8_t left[7]= { TOP_DC_PRED8x8,-1, 2,-1,DC_128_PRED8x8};
115329f1 680
43ff0714 681 if(mode > 6U) {
5175b937 682 av_log(h->s.avctx, AV_LOG_ERROR, "out of range intra chroma pred mode at %d %d\n", s->mb_x, s->mb_y);
7440fe83 683 return -1;
5175b937 684 }
115329f1 685
0da71265
MN
686 if(!(h->top_samples_available&0x8000)){
687 mode= top[ mode ];
688 if(mode<0){
9b879566 689 av_log(h->s.avctx, AV_LOG_ERROR, "top block unavailable for requested intra mode at %d %d\n", s->mb_x, s->mb_y);
0da71265
MN
690 return -1;
691 }
692 }
115329f1 693
0da71265
MN
694 if(!(h->left_samples_available&0x8000)){
695 mode= left[ mode ];
696 if(mode<0){
9b879566 697 av_log(h->s.avctx, AV_LOG_ERROR, "left block unavailable for requested intra mode at %d %d\n", s->mb_x, s->mb_y);
0da71265 698 return -1;
115329f1 699 }
0da71265
MN
700 }
701
702 return mode;
703}
704
705/**
706 * gets the predicted intra4x4 prediction mode.
707 */
708static inline int pred_intra_mode(H264Context *h, int n){
709 const int index8= scan8[n];
710 const int left= h->intra4x4_pred_mode_cache[index8 - 1];
711 const int top = h->intra4x4_pred_mode_cache[index8 - 8];
712 const int min= FFMIN(left, top);
713
a9c9a240 714 tprintf(h->s.avctx, "mode:%d %d min:%d\n", left ,top, min);
0da71265
MN
715
716 if(min<0) return DC_PRED;
717 else return min;
718}
719
720static inline void write_back_non_zero_count(H264Context *h){
721 MpegEncContext * const s = &h->s;
7bc9090a 722 const int mb_xy= s->mb_x + s->mb_y*s->mb_stride;
0da71265 723
6867a90b
LLL
724 h->non_zero_count[mb_xy][0]= h->non_zero_count_cache[7+8*1];
725 h->non_zero_count[mb_xy][1]= h->non_zero_count_cache[7+8*2];
726 h->non_zero_count[mb_xy][2]= h->non_zero_count_cache[7+8*3];
53c05b1e 727 h->non_zero_count[mb_xy][3]= h->non_zero_count_cache[7+8*4];
6867a90b
LLL
728 h->non_zero_count[mb_xy][4]= h->non_zero_count_cache[4+8*4];
729 h->non_zero_count[mb_xy][5]= h->non_zero_count_cache[5+8*4];
730 h->non_zero_count[mb_xy][6]= h->non_zero_count_cache[6+8*4];
115329f1 731
6867a90b 732 h->non_zero_count[mb_xy][9]= h->non_zero_count_cache[1+8*2];
53c05b1e 733 h->non_zero_count[mb_xy][8]= h->non_zero_count_cache[2+8*2];
6867a90b 734 h->non_zero_count[mb_xy][7]= h->non_zero_count_cache[2+8*1];
53c05b1e 735
6867a90b 736 h->non_zero_count[mb_xy][12]=h->non_zero_count_cache[1+8*5];
53c05b1e 737 h->non_zero_count[mb_xy][11]=h->non_zero_count_cache[2+8*5];
6867a90b 738 h->non_zero_count[mb_xy][10]=h->non_zero_count_cache[2+8*4];
5d18eaad
LM
739
740 if(FRAME_MBAFF){
741 // store all luma nnzs, for deblocking
742 int v = 0, i;
743 for(i=0; i<16; i++)
744 v += (!!h->non_zero_count_cache[scan8[i]]) << i;
745 *(uint16_t*)&h->non_zero_count[mb_xy][14] = v;
746 }
0da71265
MN
747}
748
749/**
750 * gets the predicted number of non zero coefficients.
751 * @param n block index
752 */
753static inline int pred_non_zero_count(H264Context *h, int n){
754 const int index8= scan8[n];
755 const int left= h->non_zero_count_cache[index8 - 1];
756 const int top = h->non_zero_count_cache[index8 - 8];
757 int i= left + top;
115329f1 758
0da71265
MN
759 if(i<64) i= (i+1)>>1;
760
a9c9a240 761 tprintf(h->s.avctx, "pred_nnz L%X T%X n%d s%d P%X\n", left, top, n, scan8[n], i&31);
0da71265
MN
762
763 return i&31;
764}
765
1924f3ce
MN
766static inline int fetch_diagonal_mv(H264Context *h, const int16_t **C, int i, int list, int part_width){
767 const int topright_ref= h->ref_cache[list][ i - 8 + part_width ];
a9c9a240 768 MpegEncContext *s = &h->s;
1924f3ce 769
5d18eaad
LM
770 /* there is no consistent mapping of mvs to neighboring locations that will
771 * make mbaff happy, so we can't move all this logic to fill_caches */
772 if(FRAME_MBAFF){
191e8ca7 773 const uint32_t *mb_types = s->current_picture_ptr->mb_type;
5d18eaad
LM
774 const int16_t *mv;
775 *(uint32_t*)h->mv_cache[list][scan8[0]-2] = 0;
776 *C = h->mv_cache[list][scan8[0]-2];
777
778 if(!MB_FIELD
779 && (s->mb_y&1) && i < scan8[0]+8 && topright_ref != PART_NOT_AVAILABLE){
780 int topright_xy = s->mb_x + (s->mb_y-1)*s->mb_stride + (i == scan8[0]+3);
781 if(IS_INTERLACED(mb_types[topright_xy])){
782#define SET_DIAG_MV(MV_OP, REF_OP, X4, Y4)\
783 const int x4 = X4, y4 = Y4;\
784 const int mb_type = mb_types[(x4>>2)+(y4>>2)*s->mb_stride];\
785 if(!USES_LIST(mb_type,list) && !IS_8X8(mb_type))\
786 return LIST_NOT_USED;\
787 mv = s->current_picture_ptr->motion_val[list][x4 + y4*h->b_stride];\
788 h->mv_cache[list][scan8[0]-2][0] = mv[0];\
789 h->mv_cache[list][scan8[0]-2][1] = mv[1] MV_OP;\
790 return s->current_picture_ptr->ref_index[list][(x4>>1) + (y4>>1)*h->b8_stride] REF_OP;
791
792 SET_DIAG_MV(*2, >>1, s->mb_x*4+(i&7)-4+part_width, s->mb_y*4-1);
793 }
794 }
795 if(topright_ref == PART_NOT_AVAILABLE
796 && ((s->mb_y&1) || i >= scan8[0]+8) && (i&7)==4
797 && h->ref_cache[list][scan8[0]-1] != PART_NOT_AVAILABLE){
798 if(!MB_FIELD
799 && IS_INTERLACED(mb_types[h->left_mb_xy[0]])){
800 SET_DIAG_MV(*2, >>1, s->mb_x*4-1, (s->mb_y|1)*4+(s->mb_y&1)*2+(i>>4)-1);
801 }
802 if(MB_FIELD
803 && !IS_INTERLACED(mb_types[h->left_mb_xy[0]])
804 && i >= scan8[0]+8){
805 // leftshift will turn LIST_NOT_USED into PART_NOT_AVAILABLE, but that's ok.
806 SET_DIAG_MV(>>1, <<1, s->mb_x*4-1, (s->mb_y&~1)*4 - 1 + ((i-scan8[0])>>3)*2);
807 }
808 }
809#undef SET_DIAG_MV
810 }
811
1924f3ce
MN
812 if(topright_ref != PART_NOT_AVAILABLE){
813 *C= h->mv_cache[list][ i - 8 + part_width ];
814 return topright_ref;
815 }else{
a9c9a240 816 tprintf(s->avctx, "topright MV not available\n");
95c26348 817
1924f3ce
MN
818 *C= h->mv_cache[list][ i - 8 - 1 ];
819 return h->ref_cache[list][ i - 8 - 1 ];
820 }
821}
822
0da71265
MN
823/**
824 * gets the predicted MV.
825 * @param n the block index
826 * @param part_width the width of the partition (4, 8,16) -> (1, 2, 4)
827 * @param mx the x component of the predicted motion vector
828 * @param my the y component of the predicted motion vector
829 */
830static inline void pred_motion(H264Context * const h, int n, int part_width, int list, int ref, int * const mx, int * const my){
0da71265
MN
831 const int index8= scan8[n];
832 const int top_ref= h->ref_cache[list][ index8 - 8 ];
0da71265
MN
833 const int left_ref= h->ref_cache[list][ index8 - 1 ];
834 const int16_t * const A= h->mv_cache[list][ index8 - 1 ];
835 const int16_t * const B= h->mv_cache[list][ index8 - 8 ];
1924f3ce
MN
836 const int16_t * C;
837 int diagonal_ref, match_count;
838
0da71265 839 assert(part_width==1 || part_width==2 || part_width==4);
1924f3ce 840
0da71265 841/* mv_cache
115329f1 842 B . . A T T T T
0da71265
MN
843 U . . L . . , .
844 U . . L . . . .
845 U . . L . . , .
846 . . . L . . . .
847*/
1924f3ce
MN
848
849 diagonal_ref= fetch_diagonal_mv(h, &C, index8, list, part_width);
850 match_count= (diagonal_ref==ref) + (top_ref==ref) + (left_ref==ref);
a9c9a240 851 tprintf(h->s.avctx, "pred_motion match_count=%d\n", match_count);
1924f3ce
MN
852 if(match_count > 1){ //most common
853 *mx= mid_pred(A[0], B[0], C[0]);
854 *my= mid_pred(A[1], B[1], C[1]);
855 }else if(match_count==1){
856 if(left_ref==ref){
857 *mx= A[0];
115329f1 858 *my= A[1];
1924f3ce
MN
859 }else if(top_ref==ref){
860 *mx= B[0];
115329f1 861 *my= B[1];
0da71265 862 }else{
1924f3ce 863 *mx= C[0];
115329f1 864 *my= C[1];
0da71265
MN
865 }
866 }else{
1924f3ce 867 if(top_ref == PART_NOT_AVAILABLE && diagonal_ref == PART_NOT_AVAILABLE && left_ref != PART_NOT_AVAILABLE){
0da71265 868 *mx= A[0];
115329f1 869 *my= A[1];
0da71265 870 }else{
1924f3ce
MN
871 *mx= mid_pred(A[0], B[0], C[0]);
872 *my= mid_pred(A[1], B[1], C[1]);
0da71265 873 }
0da71265 874 }
115329f1 875
a9c9a240 876 tprintf(h->s.avctx, "pred_motion (%2d %2d %2d) (%2d %2d %2d) (%2d %2d %2d) -> (%2d %2d %2d) at %2d %2d %d list %d\n", top_ref, B[0], B[1], diagonal_ref, C[0], C[1], left_ref, A[0], A[1], ref, *mx, *my, h->s.mb_x, h->s.mb_y, n, list);
0da71265
MN
877}
878
879/**
880 * gets the directionally predicted 16x8 MV.
881 * @param n the block index
882 * @param mx the x component of the predicted motion vector
883 * @param my the y component of the predicted motion vector
884 */
885static inline void pred_16x8_motion(H264Context * const h, int n, int list, int ref, int * const mx, int * const my){
0da71265
MN
886 if(n==0){
887 const int top_ref= h->ref_cache[list][ scan8[0] - 8 ];
888 const int16_t * const B= h->mv_cache[list][ scan8[0] - 8 ];
889
a9c9a240 890 tprintf(h->s.avctx, "pred_16x8: (%2d %2d %2d) at %2d %2d %d list %d\n", top_ref, B[0], B[1], h->s.mb_x, h->s.mb_y, n, list);
115329f1 891
0da71265
MN
892 if(top_ref == ref){
893 *mx= B[0];
894 *my= B[1];
895 return;
896 }
897 }else{
898 const int left_ref= h->ref_cache[list][ scan8[8] - 1 ];
899 const int16_t * const A= h->mv_cache[list][ scan8[8] - 1 ];
115329f1 900
a9c9a240 901 tprintf(h->s.avctx, "pred_16x8: (%2d %2d %2d) at %2d %2d %d list %d\n", left_ref, A[0], A[1], h->s.mb_x, h->s.mb_y, n, list);
0da71265
MN
902
903 if(left_ref == ref){
904 *mx= A[0];
905 *my= A[1];
906 return;
907 }
908 }
909
910 //RARE
911 pred_motion(h, n, 4, list, ref, mx, my);
912}
913
914/**
915 * gets the directionally predicted 8x16 MV.
916 * @param n the block index
917 * @param mx the x component of the predicted motion vector
918 * @param my the y component of the predicted motion vector
919 */
920static inline void pred_8x16_motion(H264Context * const h, int n, int list, int ref, int * const mx, int * const my){
0da71265
MN
921 if(n==0){
922 const int left_ref= h->ref_cache[list][ scan8[0] - 1 ];
923 const int16_t * const A= h->mv_cache[list][ scan8[0] - 1 ];
115329f1 924
a9c9a240 925 tprintf(h->s.avctx, "pred_8x16: (%2d %2d %2d) at %2d %2d %d list %d\n", left_ref, A[0], A[1], h->s.mb_x, h->s.mb_y, n, list);
0da71265
MN
926
927 if(left_ref == ref){
928 *mx= A[0];
929 *my= A[1];
930 return;
931 }
932 }else{
1924f3ce
MN
933 const int16_t * C;
934 int diagonal_ref;
935
936 diagonal_ref= fetch_diagonal_mv(h, &C, scan8[4], list, 2);
115329f1 937
a9c9a240 938 tprintf(h->s.avctx, "pred_8x16: (%2d %2d %2d) at %2d %2d %d list %d\n", diagonal_ref, C[0], C[1], h->s.mb_x, h->s.mb_y, n, list);
0da71265 939
115329f1 940 if(diagonal_ref == ref){
0da71265
MN
941 *mx= C[0];
942 *my= C[1];
943 return;
944 }
0da71265
MN
945 }
946
947 //RARE
948 pred_motion(h, n, 2, list, ref, mx, my);
949}
950
951static inline void pred_pskip_motion(H264Context * const h, int * const mx, int * const my){
0da71265
MN
952 const int top_ref = h->ref_cache[0][ scan8[0] - 8 ];
953 const int left_ref= h->ref_cache[0][ scan8[0] - 1 ];
954
a9c9a240 955 tprintf(h->s.avctx, "pred_pskip: (%d) (%d) at %2d %2d\n", top_ref, left_ref, h->s.mb_x, h->s.mb_y);
0da71265
MN
956
957 if(top_ref == PART_NOT_AVAILABLE || left_ref == PART_NOT_AVAILABLE
958 || (top_ref == 0 && *(uint32_t*)h->mv_cache[0][ scan8[0] - 8 ] == 0)
959 || (left_ref == 0 && *(uint32_t*)h->mv_cache[0][ scan8[0] - 1 ] == 0)){
115329f1 960
0da71265
MN
961 *mx = *my = 0;
962 return;
963 }
115329f1 964
0da71265
MN
965 pred_motion(h, 0, 4, 0, 0, mx, my);
966
967 return;
968}
969
5ad984c9
LM
970static inline void direct_dist_scale_factor(H264Context * const h){
971 const int poc = h->s.current_picture_ptr->poc;
972 const int poc1 = h->ref_list[1][0].poc;
973 int i;
974 for(i=0; i<h->ref_count[0]; i++){
975 int poc0 = h->ref_list[0][i].poc;
f66e4f5f 976 int td = av_clip(poc1 - poc0, -128, 127);
5ad984c9
LM
977 if(td == 0 /* FIXME || pic0 is a long-term ref */){
978 h->dist_scale_factor[i] = 256;
979 }else{
f66e4f5f 980 int tb = av_clip(poc - poc0, -128, 127);
c26abfa5 981 int tx = (16384 + (FFABS(td) >> 1)) / td;
f66e4f5f 982 h->dist_scale_factor[i] = av_clip((tb*tx + 32) >> 6, -1024, 1023);
5ad984c9
LM
983 }
984 }
5d18eaad
LM
985 if(FRAME_MBAFF){
986 for(i=0; i<h->ref_count[0]; i++){
987 h->dist_scale_factor_field[2*i] =
988 h->dist_scale_factor_field[2*i+1] = h->dist_scale_factor[i];
989 }
990 }
5ad984c9 991}
2f944356
LM
992static inline void direct_ref_list_init(H264Context * const h){
993 MpegEncContext * const s = &h->s;
994 Picture * const ref1 = &h->ref_list[1][0];
995 Picture * const cur = s->current_picture_ptr;
996 int list, i, j;
997 if(cur->pict_type == I_TYPE)
998 cur->ref_count[0] = 0;
999 if(cur->pict_type != B_TYPE)
1000 cur->ref_count[1] = 0;
1001 for(list=0; list<2; list++){
1002 cur->ref_count[list] = h->ref_count[list];
1003 for(j=0; j<h->ref_count[list]; j++)
1004 cur->ref_poc[list][j] = h->ref_list[list][j].poc;
1005 }
1006 if(cur->pict_type != B_TYPE || h->direct_spatial_mv_pred)
1007 return;
1008 for(list=0; list<2; list++){
1009 for(i=0; i<ref1->ref_count[list]; i++){
1010 const int poc = ref1->ref_poc[list][i];
171c4076 1011 h->map_col_to_list0[list][i] = 0; /* bogus; fills in for missing frames */
2f944356
LM
1012 for(j=0; j<h->ref_count[list]; j++)
1013 if(h->ref_list[list][j].poc == poc){
1014 h->map_col_to_list0[list][i] = j;
1015 break;
1016 }
1017 }
1018 }
5d18eaad
LM
1019 if(FRAME_MBAFF){
1020 for(list=0; list<2; list++){
1021 for(i=0; i<ref1->ref_count[list]; i++){
1022 j = h->map_col_to_list0[list][i];
1023 h->map_col_to_list0_field[list][2*i] = 2*j;
1024 h->map_col_to_list0_field[list][2*i+1] = 2*j+1;
1025 }
1026 }
1027 }
2f944356 1028}
5ad984c9
LM
1029
1030static inline void pred_direct_motion(H264Context * const h, int *mb_type){
1031 MpegEncContext * const s = &h->s;
1032 const int mb_xy = s->mb_x + s->mb_y*s->mb_stride;
1033 const int b8_xy = 2*s->mb_x + 2*s->mb_y*h->b8_stride;
1034 const int b4_xy = 4*s->mb_x + 4*s->mb_y*h->b_stride;
1035 const int mb_type_col = h->ref_list[1][0].mb_type[mb_xy];
1036 const int16_t (*l1mv0)[2] = (const int16_t (*)[2]) &h->ref_list[1][0].motion_val[0][b4_xy];
4866bd2b 1037 const int16_t (*l1mv1)[2] = (const int16_t (*)[2]) &h->ref_list[1][0].motion_val[1][b4_xy];
5ad984c9 1038 const int8_t *l1ref0 = &h->ref_list[1][0].ref_index[0][b8_xy];
2f944356 1039 const int8_t *l1ref1 = &h->ref_list[1][0].ref_index[1][b8_xy];
5ad984c9 1040 const int is_b8x8 = IS_8X8(*mb_type);
88e7a4d1 1041 unsigned int sub_mb_type;
5ad984c9
LM
1042 int i8, i4;
1043
5d18eaad 1044#define MB_TYPE_16x16_OR_INTRA (MB_TYPE_16x16|MB_TYPE_INTRA4x4|MB_TYPE_INTRA16x16|MB_TYPE_INTRA_PCM)
5ad984c9
LM
1045 if(IS_8X8(mb_type_col) && !h->sps.direct_8x8_inference_flag){
1046 /* FIXME save sub mb types from previous frames (or derive from MVs)
1047 * so we know exactly what block size to use */
1048 sub_mb_type = MB_TYPE_8x8|MB_TYPE_P0L0|MB_TYPE_P0L1|MB_TYPE_DIRECT2; /* B_SUB_4x4 */
3622988f 1049 *mb_type = MB_TYPE_8x8|MB_TYPE_L0L1;
5d18eaad 1050 }else if(!is_b8x8 && (mb_type_col & MB_TYPE_16x16_OR_INTRA)){
5ad984c9
LM
1051 sub_mb_type = MB_TYPE_16x16|MB_TYPE_P0L0|MB_TYPE_P0L1|MB_TYPE_DIRECT2; /* B_SUB_8x8 */
1052 *mb_type = MB_TYPE_16x16|MB_TYPE_P0L0|MB_TYPE_P0L1|MB_TYPE_DIRECT2; /* B_16x16 */
1053 }else{
1054 sub_mb_type = MB_TYPE_16x16|MB_TYPE_P0L0|MB_TYPE_P0L1|MB_TYPE_DIRECT2; /* B_SUB_8x8 */
3622988f 1055 *mb_type = MB_TYPE_8x8|MB_TYPE_L0L1;
5ad984c9
LM
1056 }
1057 if(!is_b8x8)
1058 *mb_type |= MB_TYPE_DIRECT2;
5d18eaad
LM
1059 if(MB_FIELD)
1060 *mb_type |= MB_TYPE_INTERLACED;
5ad984c9 1061
a9c9a240 1062 tprintf(s->avctx, "mb_type = %08x, sub_mb_type = %08x, is_b8x8 = %d, mb_type_col = %08x\n", *mb_type, sub_mb_type, is_b8x8, mb_type_col);
115329f1 1063
5ad984c9
LM
1064 if(h->direct_spatial_mv_pred){
1065 int ref[2];
1066 int mv[2][2];
1067 int list;
1068
5d18eaad
LM
1069 /* FIXME interlacing + spatial direct uses wrong colocated block positions */
1070
5ad984c9
LM
1071 /* ref = min(neighbors) */
1072 for(list=0; list<2; list++){
1073 int refa = h->ref_cache[list][scan8[0] - 1];
1074 int refb = h->ref_cache[list][scan8[0] - 8];
1075 int refc = h->ref_cache[list][scan8[0] - 8 + 4];
1076 if(refc == -2)
1077 refc = h->ref_cache[list][scan8[0] - 8 - 1];
1078 ref[list] = refa;
1079 if(ref[list] < 0 || (refb < ref[list] && refb >= 0))
1080 ref[list] = refb;
1081 if(ref[list] < 0 || (refc < ref[list] && refc >= 0))
1082 ref[list] = refc;
1083 if(ref[list] < 0)
1084 ref[list] = -1;
1085 }
1086
1087 if(ref[0] < 0 && ref[1] < 0){
1088 ref[0] = ref[1] = 0;
1089 mv[0][0] = mv[0][1] =
1090 mv[1][0] = mv[1][1] = 0;
1091 }else{
1092 for(list=0; list<2; list++){
1093 if(ref[list] >= 0)
1094 pred_motion(h, 0, 4, list, ref[list], &mv[list][0], &mv[list][1]);
1095 else
1096 mv[list][0] = mv[list][1] = 0;
1097 }
1098 }
1099
1100 if(ref[1] < 0){
1101 *mb_type &= ~MB_TYPE_P0L1;
1102 sub_mb_type &= ~MB_TYPE_P0L1;
1103 }else if(ref[0] < 0){
1104 *mb_type &= ~MB_TYPE_P0L0;
1105 sub_mb_type &= ~MB_TYPE_P0L0;
1106 }
1107
1108 if(IS_16X16(*mb_type)){
d19f5acb
MN
1109 int a=0, b=0;
1110
cec93959
LM
1111 fill_rectangle(&h->ref_cache[0][scan8[0]], 4, 4, 8, (uint8_t)ref[0], 1);
1112 fill_rectangle(&h->ref_cache[1][scan8[0]], 4, 4, 8, (uint8_t)ref[1], 1);
115329f1 1113 if(!IS_INTRA(mb_type_col)
c26abfa5
DB
1114 && ( (l1ref0[0] == 0 && FFABS(l1mv0[0][0]) <= 1 && FFABS(l1mv0[0][1]) <= 1)
1115 || (l1ref0[0] < 0 && l1ref1[0] == 0 && FFABS(l1mv1[0][0]) <= 1 && FFABS(l1mv1[0][1]) <= 1
bf4e3bd2 1116 && (h->x264_build>33 || !h->x264_build)))){
5ad984c9 1117 if(ref[0] > 0)
d19f5acb 1118 a= pack16to32(mv[0][0],mv[0][1]);
5ad984c9 1119 if(ref[1] > 0)
d19f5acb 1120 b= pack16to32(mv[1][0],mv[1][1]);
5ad984c9 1121 }else{
d19f5acb
MN
1122 a= pack16to32(mv[0][0],mv[0][1]);
1123 b= pack16to32(mv[1][0],mv[1][1]);
5ad984c9 1124 }
d19f5acb
MN
1125 fill_rectangle(&h->mv_cache[0][scan8[0]], 4, 4, 8, a, 4);
1126 fill_rectangle(&h->mv_cache[1][scan8[0]], 4, 4, 8, b, 4);
5ad984c9
LM
1127 }else{
1128 for(i8=0; i8<4; i8++){
1129 const int x8 = i8&1;
1130 const int y8 = i8>>1;
115329f1 1131
5ad984c9
LM
1132 if(is_b8x8 && !IS_DIRECT(h->sub_mb_type[i8]))
1133 continue;
1134 h->sub_mb_type[i8] = sub_mb_type;
115329f1 1135
5ad984c9
LM
1136 fill_rectangle(&h->mv_cache[0][scan8[i8*4]], 2, 2, 8, pack16to32(mv[0][0],mv[0][1]), 4);
1137 fill_rectangle(&h->mv_cache[1][scan8[i8*4]], 2, 2, 8, pack16to32(mv[1][0],mv[1][1]), 4);
cec93959
LM
1138 fill_rectangle(&h->ref_cache[0][scan8[i8*4]], 2, 2, 8, (uint8_t)ref[0], 1);
1139 fill_rectangle(&h->ref_cache[1][scan8[i8*4]], 2, 2, 8, (uint8_t)ref[1], 1);
115329f1 1140
5ad984c9 1141 /* col_zero_flag */
115329f1
DB
1142 if(!IS_INTRA(mb_type_col) && ( l1ref0[x8 + y8*h->b8_stride] == 0
1143 || (l1ref0[x8 + y8*h->b8_stride] < 0 && l1ref1[x8 + y8*h->b8_stride] == 0
bf4e3bd2 1144 && (h->x264_build>33 || !h->x264_build)))){
4866bd2b 1145 const int16_t (*l1mv)[2]= l1ref0[x8 + y8*h->b8_stride] == 0 ? l1mv0 : l1mv1;
f1f17e54
LM
1146 if(IS_SUB_8X8(sub_mb_type)){
1147 const int16_t *mv_col = l1mv[x8*3 + y8*3*h->b_stride];
c26abfa5 1148 if(FFABS(mv_col[0]) <= 1 && FFABS(mv_col[1]) <= 1){
f1f17e54
LM
1149 if(ref[0] == 0)
1150 fill_rectangle(&h->mv_cache[0][scan8[i8*4]], 2, 2, 8, 0, 4);
1151 if(ref[1] == 0)
1152 fill_rectangle(&h->mv_cache[1][scan8[i8*4]], 2, 2, 8, 0, 4);
1153 }
1154 }else
5ad984c9 1155 for(i4=0; i4<4; i4++){
4866bd2b 1156 const int16_t *mv_col = l1mv[x8*2 + (i4&1) + (y8*2 + (i4>>1))*h->b_stride];
c26abfa5 1157 if(FFABS(mv_col[0]) <= 1 && FFABS(mv_col[1]) <= 1){
5ad984c9
LM
1158 if(ref[0] == 0)
1159 *(uint32_t*)h->mv_cache[0][scan8[i8*4+i4]] = 0;
1160 if(ref[1] == 0)
1161 *(uint32_t*)h->mv_cache[1][scan8[i8*4+i4]] = 0;
1162 }
1163 }
1164 }
1165 }
1166 }
1167 }else{ /* direct temporal mv pred */
5d18eaad
LM
1168 const int *map_col_to_list0[2] = {h->map_col_to_list0[0], h->map_col_to_list0[1]};
1169 const int *dist_scale_factor = h->dist_scale_factor;
1170
1171 if(FRAME_MBAFF){
1172 if(IS_INTERLACED(*mb_type)){
1173 map_col_to_list0[0] = h->map_col_to_list0_field[0];
1174 map_col_to_list0[1] = h->map_col_to_list0_field[1];
1175 dist_scale_factor = h->dist_scale_factor_field;
1176 }
1177 if(IS_INTERLACED(*mb_type) != IS_INTERLACED(mb_type_col)){
1178 /* FIXME assumes direct_8x8_inference == 1 */
1179 const int pair_xy = s->mb_x + (s->mb_y&~1)*s->mb_stride;
1180 int mb_types_col[2];
1181 int y_shift;
1182
1183 *mb_type = MB_TYPE_8x8|MB_TYPE_L0L1
1184 | (is_b8x8 ? 0 : MB_TYPE_DIRECT2)
1185 | (*mb_type & MB_TYPE_INTERLACED);
1186 sub_mb_type = MB_TYPE_P0L0|MB_TYPE_P0L1|MB_TYPE_DIRECT2|MB_TYPE_16x16;
1187
1188 if(IS_INTERLACED(*mb_type)){
1189 /* frame to field scaling */
1190 mb_types_col[0] = h->ref_list[1][0].mb_type[pair_xy];
1191 mb_types_col[1] = h->ref_list[1][0].mb_type[pair_xy+s->mb_stride];
1192 if(s->mb_y&1){
1193 l1ref0 -= 2*h->b8_stride;
1194 l1ref1 -= 2*h->b8_stride;
1195 l1mv0 -= 4*h->b_stride;
1196 l1mv1 -= 4*h->b_stride;
1197 }
1198 y_shift = 0;
1199
1200 if( (mb_types_col[0] & MB_TYPE_16x16_OR_INTRA)
1201 && (mb_types_col[1] & MB_TYPE_16x16_OR_INTRA)
1202 && !is_b8x8)
1203 *mb_type |= MB_TYPE_16x8;
1204 else
1205 *mb_type |= MB_TYPE_8x8;
1206 }else{
1207 /* field to frame scaling */
1208 /* col_mb_y = (mb_y&~1) + (topAbsDiffPOC < bottomAbsDiffPOC ? 0 : 1)
1209 * but in MBAFF, top and bottom POC are equal */
1210 int dy = (s->mb_y&1) ? 1 : 2;
1211 mb_types_col[0] =
1212 mb_types_col[1] = h->ref_list[1][0].mb_type[pair_xy+s->mb_stride];
1213 l1ref0 += dy*h->b8_stride;
1214 l1ref1 += dy*h->b8_stride;
1215 l1mv0 += 2*dy*h->b_stride;
1216 l1mv1 += 2*dy*h->b_stride;
1217 y_shift = 2;
1218
1219 if((mb_types_col[0] & (MB_TYPE_16x16_OR_INTRA|MB_TYPE_16x8))
1220 && !is_b8x8)
1221 *mb_type |= MB_TYPE_16x16;
1222 else
1223 *mb_type |= MB_TYPE_8x8;
1224 }
1225
1226 for(i8=0; i8<4; i8++){
1227 const int x8 = i8&1;
1228 const int y8 = i8>>1;
1229 int ref0, scale;
1230 const int16_t (*l1mv)[2]= l1mv0;
1231
1232 if(is_b8x8 && !IS_DIRECT(h->sub_mb_type[i8]))
1233 continue;
1234 h->sub_mb_type[i8] = sub_mb_type;
1235
1236 fill_rectangle(&h->ref_cache[1][scan8[i8*4]], 2, 2, 8, 0, 1);
1237 if(IS_INTRA(mb_types_col[y8])){
1238 fill_rectangle(&h->ref_cache[0][scan8[i8*4]], 2, 2, 8, 0, 1);
1239 fill_rectangle(&h-> mv_cache[0][scan8[i8*4]], 2, 2, 8, 0, 4);
1240 fill_rectangle(&h-> mv_cache[1][scan8[i8*4]], 2, 2, 8, 0, 4);
1241 continue;
1242 }
1243
1244 ref0 = l1ref0[x8 + (y8*2>>y_shift)*h->b8_stride];
1245 if(ref0 >= 0)
1246 ref0 = map_col_to_list0[0][ref0*2>>y_shift];
1247 else{
1248 ref0 = map_col_to_list0[1][l1ref1[x8 + (y8*2>>y_shift)*h->b8_stride]*2>>y_shift];
1249 l1mv= l1mv1;
1250 }
1251 scale = dist_scale_factor[ref0];
1252 fill_rectangle(&h->ref_cache[0][scan8[i8*4]], 2, 2, 8, ref0, 1);
1253
1254 {
1255 const int16_t *mv_col = l1mv[x8*3 + (y8*6>>y_shift)*h->b_stride];
1256 int my_col = (mv_col[1]<<y_shift)/2;
1257 int mx = (scale * mv_col[0] + 128) >> 8;
1258 int my = (scale * my_col + 128) >> 8;
1259 fill_rectangle(&h->mv_cache[0][scan8[i8*4]], 2, 2, 8, pack16to32(mx,my), 4);
1260 fill_rectangle(&h->mv_cache[1][scan8[i8*4]], 2, 2, 8, pack16to32(mx-mv_col[0],my-my_col), 4);
1261 }
1262 }
1263 return;
1264 }
1265 }
1266
1267 /* one-to-one mv scaling */
1268
5ad984c9 1269 if(IS_16X16(*mb_type)){
fda51641
MN
1270 int ref, mv0, mv1;
1271
5ad984c9
LM
1272 fill_rectangle(&h->ref_cache[1][scan8[0]], 4, 4, 8, 0, 1);
1273 if(IS_INTRA(mb_type_col)){
fda51641 1274 ref=mv0=mv1=0;
5ad984c9 1275 }else{
5d18eaad
LM
1276 const int ref0 = l1ref0[0] >= 0 ? map_col_to_list0[0][l1ref0[0]]
1277 : map_col_to_list0[1][l1ref1[0]];
1278 const int scale = dist_scale_factor[ref0];
8583bef8 1279 const int16_t *mv_col = l1ref0[0] >= 0 ? l1mv0[0] : l1mv1[0];
5ad984c9 1280 int mv_l0[2];
5d18eaad
LM
1281 mv_l0[0] = (scale * mv_col[0] + 128) >> 8;
1282 mv_l0[1] = (scale * mv_col[1] + 128) >> 8;
fda51641
MN
1283 ref= ref0;
1284 mv0= pack16to32(mv_l0[0],mv_l0[1]);
1285 mv1= pack16to32(mv_l0[0]-mv_col[0],mv_l0[1]-mv_col[1]);
5ad984c9 1286 }
fda51641
MN
1287 fill_rectangle(&h->ref_cache[0][scan8[0]], 4, 4, 8, ref, 1);
1288 fill_rectangle(&h-> mv_cache[0][scan8[0]], 4, 4, 8, mv0, 4);
1289 fill_rectangle(&h-> mv_cache[1][scan8[0]], 4, 4, 8, mv1, 4);
5ad984c9
LM
1290 }else{
1291 for(i8=0; i8<4; i8++){
1292 const int x8 = i8&1;
1293 const int y8 = i8>>1;
5d18eaad 1294 int ref0, scale;
bf4e3bd2 1295 const int16_t (*l1mv)[2]= l1mv0;
8583bef8 1296
5ad984c9
LM
1297 if(is_b8x8 && !IS_DIRECT(h->sub_mb_type[i8]))
1298 continue;
1299 h->sub_mb_type[i8] = sub_mb_type;
5d18eaad 1300 fill_rectangle(&h->ref_cache[1][scan8[i8*4]], 2, 2, 8, 0, 1);
5ad984c9
LM
1301 if(IS_INTRA(mb_type_col)){
1302 fill_rectangle(&h->ref_cache[0][scan8[i8*4]], 2, 2, 8, 0, 1);
5ad984c9
LM
1303 fill_rectangle(&h-> mv_cache[0][scan8[i8*4]], 2, 2, 8, 0, 4);
1304 fill_rectangle(&h-> mv_cache[1][scan8[i8*4]], 2, 2, 8, 0, 4);
1305 continue;
1306 }
115329f1 1307
5ad984c9 1308 ref0 = l1ref0[x8 + y8*h->b8_stride];
2f944356 1309 if(ref0 >= 0)
5d18eaad 1310 ref0 = map_col_to_list0[0][ref0];
8583bef8 1311 else{
5d18eaad 1312 ref0 = map_col_to_list0[1][l1ref1[x8 + y8*h->b8_stride]];
8583bef8
MN
1313 l1mv= l1mv1;
1314 }
5d18eaad 1315 scale = dist_scale_factor[ref0];
115329f1 1316
5ad984c9 1317 fill_rectangle(&h->ref_cache[0][scan8[i8*4]], 2, 2, 8, ref0, 1);
f1f17e54
LM
1318 if(IS_SUB_8X8(sub_mb_type)){
1319 const int16_t *mv_col = l1mv[x8*3 + y8*3*h->b_stride];
5d18eaad
LM
1320 int mx = (scale * mv_col[0] + 128) >> 8;
1321 int my = (scale * mv_col[1] + 128) >> 8;
f1f17e54
LM
1322 fill_rectangle(&h->mv_cache[0][scan8[i8*4]], 2, 2, 8, pack16to32(mx,my), 4);
1323 fill_rectangle(&h->mv_cache[1][scan8[i8*4]], 2, 2, 8, pack16to32(mx-mv_col[0],my-mv_col[1]), 4);
1324 }else
5ad984c9 1325 for(i4=0; i4<4; i4++){
8583bef8 1326 const int16_t *mv_col = l1mv[x8*2 + (i4&1) + (y8*2 + (i4>>1))*h->b_stride];
5ad984c9 1327 int16_t *mv_l0 = h->mv_cache[0][scan8[i8*4+i4]];
5d18eaad
LM
1328 mv_l0[0] = (scale * mv_col[0] + 128) >> 8;
1329 mv_l0[1] = (scale * mv_col[1] + 128) >> 8;
5ad984c9
LM
1330 *(uint32_t*)h->mv_cache[1][scan8[i8*4+i4]] =
1331 pack16to32(mv_l0[0]-mv_col[0],mv_l0[1]-mv_col[1]);
1332 }
1333 }
1334 }
1335 }
1336}
1337
0da71265
MN
1338static inline void write_back_motion(H264Context *h, int mb_type){
1339 MpegEncContext * const s = &h->s;
0da71265
MN
1340 const int b_xy = 4*s->mb_x + 4*s->mb_y*h->b_stride;
1341 const int b8_xy= 2*s->mb_x + 2*s->mb_y*h->b8_stride;
1342 int list;
1343
2ea39252
LM
1344 if(!USES_LIST(mb_type, 0))
1345 fill_rectangle(&s->current_picture.ref_index[0][b8_xy], 2, 2, h->b8_stride, (uint8_t)LIST_NOT_USED, 1);
1346
3425501d 1347 for(list=0; list<h->list_count; list++){
0da71265 1348 int y;
53b19144 1349 if(!USES_LIST(mb_type, list))
5ad984c9 1350 continue;
115329f1 1351
0da71265
MN
1352 for(y=0; y<4; y++){
1353 *(uint64_t*)s->current_picture.motion_val[list][b_xy + 0 + y*h->b_stride]= *(uint64_t*)h->mv_cache[list][scan8[0]+0 + 8*y];
1354 *(uint64_t*)s->current_picture.motion_val[list][b_xy + 2 + y*h->b_stride]= *(uint64_t*)h->mv_cache[list][scan8[0]+2 + 8*y];
1355 }
9e528114 1356 if( h->pps.cabac ) {
e6e77eb6
LM
1357 if(IS_SKIP(mb_type))
1358 fill_rectangle(h->mvd_table[list][b_xy], 4, 4, h->b_stride, 0, 4);
1359 else
9e528114
LA
1360 for(y=0; y<4; y++){
1361 *(uint64_t*)h->mvd_table[list][b_xy + 0 + y*h->b_stride]= *(uint64_t*)h->mvd_cache[list][scan8[0]+0 + 8*y];
1362 *(uint64_t*)h->mvd_table[list][b_xy + 2 + y*h->b_stride]= *(uint64_t*)h->mvd_cache[list][scan8[0]+2 + 8*y];
1363 }
1364 }
53b19144
LM
1365
1366 {
191e8ca7 1367 int8_t *ref_index = &s->current_picture.ref_index[list][b8_xy];
53b19144
LM
1368 ref_index[0+0*h->b8_stride]= h->ref_cache[list][scan8[0]];
1369 ref_index[1+0*h->b8_stride]= h->ref_cache[list][scan8[4]];
1370 ref_index[0+1*h->b8_stride]= h->ref_cache[list][scan8[8]];
1371 ref_index[1+1*h->b8_stride]= h->ref_cache[list][scan8[12]];
0da71265
MN
1372 }
1373 }
115329f1 1374
5ad984c9
LM
1375 if(h->slice_type == B_TYPE && h->pps.cabac){
1376 if(IS_8X8(mb_type)){
53b19144
LM
1377 uint8_t *direct_table = &h->direct_table[b8_xy];
1378 direct_table[1+0*h->b8_stride] = IS_DIRECT(h->sub_mb_type[1]) ? 1 : 0;
1379 direct_table[0+1*h->b8_stride] = IS_DIRECT(h->sub_mb_type[2]) ? 1 : 0;
1380 direct_table[1+1*h->b8_stride] = IS_DIRECT(h->sub_mb_type[3]) ? 1 : 0;
5ad984c9
LM
1381 }
1382 }
0da71265
MN
1383}
1384
1385/**
1386 * Decodes a network abstraction layer unit.
1387 * @param consumed is the number of bytes used as input
1388 * @param length is the length of the array
3b66c4c5 1389 * @param dst_length is the number of decoded bytes FIXME here or a decode rbsp tailing?
115329f1 1390 * @returns decoded bytes, might be src+1 if no escapes
0da71265
MN
1391 */
1392static uint8_t *decode_nal(H264Context *h, uint8_t *src, int *dst_length, int *consumed, int length){
1393 int i, si, di;
1394 uint8_t *dst;
1395
bb270c08 1396// src[0]&0x80; //forbidden bit
0da71265
MN
1397 h->nal_ref_idc= src[0]>>5;
1398 h->nal_unit_type= src[0]&0x1F;
1399
1400 src++; length--;
115329f1 1401#if 0
0da71265
MN
1402 for(i=0; i<length; i++)
1403 printf("%2X ", src[i]);
1404#endif
1405 for(i=0; i+1<length; i+=2){
1406 if(src[i]) continue;
1407 if(i>0 && src[i-1]==0) i--;
1408 if(i+2<length && src[i+1]==0 && src[i+2]<=3){
1409 if(src[i+2]!=3){
1410 /* startcode, so we must be past the end */
1411 length=i;
1412 }
1413 break;
1414 }
1415 }
1416
1417 if(i>=length-1){ //no escaped 0
1418 *dst_length= length;
1419 *consumed= length+1; //+1 for the header
115329f1 1420 return src;
0da71265
MN
1421 }
1422
1423 h->rbsp_buffer= av_fast_realloc(h->rbsp_buffer, &h->rbsp_buffer_size, length);
1424 dst= h->rbsp_buffer;
1425
ac658be5
FOL
1426 if (dst == NULL){
1427 return NULL;
1428 }
1429
3b66c4c5 1430//printf("decoding esc\n");
0da71265 1431 si=di=0;
115329f1 1432 while(si<length){
0da71265
MN
1433 //remove escapes (very rare 1:2^22)
1434 if(si+2<length && src[si]==0 && src[si+1]==0 && src[si+2]<=3){
1435 if(src[si+2]==3){ //escape
1436 dst[di++]= 0;
1437 dst[di++]= 0;
1438 si+=3;
c8470cc1 1439 continue;
0da71265
MN
1440 }else //next start code
1441 break;
1442 }
1443
1444 dst[di++]= src[si++];
1445 }
1446
1447 *dst_length= di;
1448 *consumed= si + 1;//+1 for the header
90b5b51e 1449//FIXME store exact number of bits in the getbitcontext (it is needed for decoding)
0da71265
MN
1450 return dst;
1451}
1452
0da71265
MN
1453/**
1454 * identifies the exact end of the bitstream
1455 * @return the length of the trailing, or 0 if damaged
1456 */
a9c9a240 1457static int decode_rbsp_trailing(H264Context *h, uint8_t *src){
0da71265
MN
1458 int v= *src;
1459 int r;
1460
a9c9a240 1461 tprintf(h->s.avctx, "rbsp trailing %X\n", v);
0da71265
MN
1462
1463 for(r=1; r<9; r++){
1464 if(v&1) return r;
1465 v>>=1;
1466 }
1467 return 0;
1468}
1469
1470/**
1471 * idct tranforms the 16 dc values and dequantize them.
1472 * @param qp quantization parameter
1473 */
239ea04c 1474static void h264_luma_dc_dequant_idct_c(DCTELEM *block, int qp, int qmul){
0da71265
MN
1475#define stride 16
1476 int i;
1477 int temp[16]; //FIXME check if this is a good idea
1478 static const int x_offset[4]={0, 1*stride, 4* stride, 5*stride};
1479 static const int y_offset[4]={0, 2*stride, 8* stride, 10*stride};
1480
1481//memset(block, 64, 2*256);
1482//return;
1483 for(i=0; i<4; i++){
1484 const int offset= y_offset[i];
1485 const int z0= block[offset+stride*0] + block[offset+stride*4];
1486 const int z1= block[offset+stride*0] - block[offset+stride*4];
1487 const int z2= block[offset+stride*1] - block[offset+stride*5];
1488 const int z3= block[offset+stride*1] + block[offset+stride*5];
1489
1490 temp[4*i+0]= z0+z3;
1491 temp[4*i+1]= z1+z2;
1492 temp[4*i+2]= z1-z2;
1493 temp[4*i+3]= z0-z3;
1494 }
1495
1496 for(i=0; i<4; i++){
1497 const int offset= x_offset[i];
1498 const int z0= temp[4*0+i] + temp[4*2+i];
1499 const int z1= temp[4*0+i] - temp[4*2+i];
1500 const int z2= temp[4*1+i] - temp[4*3+i];
1501 const int z3= temp[4*1+i] + temp[4*3+i];
1502
239ea04c
LM
1503 block[stride*0 +offset]= ((((z0 + z3)*qmul + 128 ) >> 8)); //FIXME think about merging this into decode_resdual
1504 block[stride*2 +offset]= ((((z1 + z2)*qmul + 128 ) >> 8));
1505 block[stride*8 +offset]= ((((z1 - z2)*qmul + 128 ) >> 8));
1506 block[stride*10+offset]= ((((z0 - z3)*qmul + 128 ) >> 8));
0da71265
MN
1507 }
1508}
1509
e5017ab8 1510#if 0
0da71265
MN
1511/**
1512 * dct tranforms the 16 dc values.
1513 * @param qp quantization parameter ??? FIXME
1514 */
1515static void h264_luma_dc_dct_c(DCTELEM *block/*, int qp*/){
1516// const int qmul= dequant_coeff[qp][0];
1517 int i;
1518 int temp[16]; //FIXME check if this is a good idea
1519 static const int x_offset[4]={0, 1*stride, 4* stride, 5*stride};
1520 static const int y_offset[4]={0, 2*stride, 8* stride, 10*stride};
1521
1522 for(i=0; i<4; i++){
1523 const int offset= y_offset[i];
1524 const int z0= block[offset+stride*0] + block[offset+stride*4];
1525 const int z1= block[offset+stride*0] - block[offset+stride*4];
1526 const int z2= block[offset+stride*1] - block[offset+stride*5];
1527 const int z3= block[offset+stride*1] + block[offset+stride*5];
1528
1529 temp[4*i+0]= z0+z3;
1530 temp[4*i+1]= z1+z2;
1531 temp[4*i+2]= z1-z2;
1532 temp[4*i+3]= z0-z3;
1533 }
1534
1535 for(i=0; i<4; i++){
1536 const int offset= x_offset[i];
1537 const int z0= temp[4*0+i] + temp[4*2+i];
1538 const int z1= temp[4*0+i] - temp[4*2+i];
1539 const int z2= temp[4*1+i] - temp[4*3+i];
1540 const int z3= temp[4*1+i] + temp[4*3+i];
1541
1542 block[stride*0 +offset]= (z0 + z3)>>1;
1543 block[stride*2 +offset]= (z1 + z2)>>1;
1544 block[stride*8 +offset]= (z1 - z2)>>1;
1545 block[stride*10+offset]= (z0 - z3)>>1;
1546 }
1547}
e5017ab8
LA
1548#endif
1549
0da71265
MN
1550#undef xStride
1551#undef stride
1552
239ea04c 1553static void chroma_dc_dequant_idct_c(DCTELEM *block, int qp, int qmul){
0da71265
MN
1554 const int stride= 16*2;
1555 const int xStride= 16;
1556 int a,b,c,d,e;
1557
1558 a= block[stride*0 + xStride*0];
1559 b= block[stride*0 + xStride*1];
1560 c= block[stride*1 + xStride*0];
1561 d= block[stride*1 + xStride*1];
1562
1563 e= a-b;
1564 a= a+b;
1565 b= c-d;
1566 c= c+d;
1567
239ea04c
LM
1568 block[stride*0 + xStride*0]= ((a+c)*qmul) >> 7;
1569 block[stride*0 + xStride*1]= ((e+b)*qmul) >> 7;
1570 block[stride*1 + xStride*0]= ((a-c)*qmul) >> 7;
1571 block[stride*1 + xStride*1]= ((e-b)*qmul) >> 7;
0da71265
MN
1572}
1573
e5017ab8 1574#if 0
0da71265
MN
1575static void chroma_dc_dct_c(DCTELEM *block){
1576 const int stride= 16*2;
1577 const int xStride= 16;
1578 int a,b,c,d,e;
1579
1580 a= block[stride*0 + xStride*0];
1581 b= block[stride*0 + xStride*1];
1582 c= block[stride*1 + xStride*0];
1583 d= block[stride*1 + xStride*1];
1584
1585 e= a-b;
1586 a= a+b;
1587 b= c-d;
1588 c= c+d;
1589
1590 block[stride*0 + xStride*0]= (a+c);
1591 block[stride*0 + xStride*1]= (e+b);
1592 block[stride*1 + xStride*0]= (a-c);
1593 block[stride*1 + xStride*1]= (e-b);
1594}
e5017ab8 1595#endif
0da71265
MN
1596
1597/**
1598 * gets the chroma qp.
1599 */
6ba71fc4 1600static inline int get_chroma_qp(int chroma_qp_index_offset, int qscale){
115329f1 1601
f66e4f5f 1602 return chroma_qp[av_clip(qscale + chroma_qp_index_offset, 0, 51)];
0da71265
MN
1603}
1604
755bfeab 1605//FIXME need to check that this does not overflow signed 32 bit for low qp, i am not sure, it's very close
0afd2a92
DB
1606//FIXME check that gcc inlines this (and optimizes intra & separate_dc stuff away)
1607static inline int quantize_c(DCTELEM *block, uint8_t *scantable, int qscale, int intra, int separate_dc){
0da71265
MN
1608 int i;
1609 const int * const quant_table= quant_coeff[qscale];
1610 const int bias= intra ? (1<<QUANT_SHIFT)/3 : (1<<QUANT_SHIFT)/6;
1611 const unsigned int threshold1= (1<<QUANT_SHIFT) - bias - 1;
1612 const unsigned int threshold2= (threshold1<<1);
1613 int last_non_zero;
1614
0afd2a92 1615 if(separate_dc){
0da71265
MN
1616 if(qscale<=18){
1617 //avoid overflows
1618 const int dc_bias= intra ? (1<<(QUANT_SHIFT-2))/3 : (1<<(QUANT_SHIFT-2))/6;
1619 const unsigned int dc_threshold1= (1<<(QUANT_SHIFT-2)) - dc_bias - 1;
1620 const unsigned int dc_threshold2= (dc_threshold1<<1);
1621
1622 int level= block[0]*quant_coeff[qscale+18][0];
1623 if(((unsigned)(level+dc_threshold1))>dc_threshold2){
1624 if(level>0){
1625 level= (dc_bias + level)>>(QUANT_SHIFT-2);
1626 block[0]= level;
1627 }else{
1628 level= (dc_bias - level)>>(QUANT_SHIFT-2);
1629 block[0]= -level;
1630 }
1631// last_non_zero = i;
1632 }else{
1633 block[0]=0;
1634 }
1635 }else{
1636 const int dc_bias= intra ? (1<<(QUANT_SHIFT+1))/3 : (1<<(QUANT_SHIFT+1))/6;
1637 const unsigned int dc_threshold1= (1<<(QUANT_SHIFT+1)) - dc_bias - 1;
1638 const unsigned int dc_threshold2= (dc_threshold1<<1);
1639
1640 int level= block[0]*quant_table[0];
1641 if(((unsigned)(level+dc_threshold1))>dc_threshold2){
1642 if(level>0){
1643 level= (dc_bias + level)>>(QUANT_SHIFT+1);
1644 block[0]= level;
1645 }else{
1646 level= (dc_bias - level)>>(QUANT_SHIFT+1);
1647 block[0]= -level;
1648 }
1649// last_non_zero = i;
1650 }else{
1651 block[0]=0;
1652 }
1653 }
1654 last_non_zero= 0;
1655 i=1;
1656 }else{
1657 last_non_zero= -1;
1658 i=0;
1659 }
1660
1661 for(; i<16; i++){
1662 const int j= scantable[i];
1663 int level= block[j]*quant_table[j];
1664
1665// if( bias+level >= (1<<(QMAT_SHIFT - 3))
1666// || bias-level >= (1<<(QMAT_SHIFT - 3))){
1667 if(((unsigned)(level+threshold1))>threshold2){
1668 if(level>0){
1669 level= (bias + level)>>QUANT_SHIFT;
1670 block[j]= level;
1671 }else{
1672 level= (bias - level)>>QUANT_SHIFT;
1673 block[j]= -level;
1674 }
1675 last_non_zero = i;
1676 }else{
1677 block[j]=0;
1678 }
1679 }
1680
1681 return last_non_zero;
1682}
1683
1684static void pred4x4_vertical_c(uint8_t *src, uint8_t *topright, int stride){
1685 const uint32_t a= ((uint32_t*)(src-stride))[0];
1686 ((uint32_t*)(src+0*stride))[0]= a;
1687 ((uint32_t*)(src+1*stride))[0]= a;
1688 ((uint32_t*)(src+2*stride))[0]= a;
1689 ((uint32_t*)(src+3*stride))[0]= a;
1690}
1691
1692static void pred4x4_horizontal_c(uint8_t *src, uint8_t *topright, int stride){
1693 ((uint32_t*)(src+0*stride))[0]= src[-1+0*stride]*0x01010101;
1694 ((uint32_t*)(src+1*stride))[0]= src[-1+1*stride]*0x01010101;
1695 ((uint32_t*)(src+2*stride))[0]= src[-1+2*stride]*0x01010101;
1696 ((uint32_t*)(src+3*stride))[0]= src[-1+3*stride]*0x01010101;
1697}
1698
1699static void pred4x4_dc_c(uint8_t *src, uint8_t *topright, int stride){
1700 const int dc= ( src[-stride] + src[1-stride] + src[2-stride] + src[3-stride]
1701 + src[-1+0*stride] + src[-1+1*stride] + src[-1+2*stride] + src[-1+3*stride] + 4) >>3;
115329f1
DB
1702
1703 ((uint32_t*)(src+0*stride))[0]=
1704 ((uint32_t*)(src+1*stride))[0]=
1705 ((uint32_t*)(src+2*stride))[0]=
1706 ((uint32_t*)(src+3*stride))[0]= dc* 0x01010101;
0da71265
MN
1707}
1708
1709static void pred4x4_left_dc_c(uint8_t *src, uint8_t *topright, int stride){
1710 const int dc= ( src[-1+0*stride] + src[-1+1*stride] + src[-1+2*stride] + src[-1+3*stride] + 2) >>2;
115329f1
DB
1711
1712 ((uint32_t*)(src+0*stride))[0]=
1713 ((uint32_t*)(src+1*stride))[0]=
1714 ((uint32_t*)(src+2*stride))[0]=
1715 ((uint32_t*)(src+3*stride))[0]= dc* 0x01010101;
0da71265
MN
1716}
1717
1718static void pred4x4_top_dc_c(uint8_t *src, uint8_t *topright, int stride){
1719 const int dc= ( src[-stride] + src[1-stride] + src[2-stride] + src[3-stride] + 2) >>2;
115329f1
DB
1720
1721 ((uint32_t*)(src+0*stride))[0]=
1722 ((uint32_t*)(src+1*stride))[0]=
1723 ((uint32_t*)(src+2*stride))[0]=
1724 ((uint32_t*)(src+3*stride))[0]= dc* 0x01010101;
0da71265
MN
1725}
1726
1727static void pred4x4_128_dc_c(uint8_t *src, uint8_t *topright, int stride){
115329f1
DB
1728 ((uint32_t*)(src+0*stride))[0]=
1729 ((uint32_t*)(src+1*stride))[0]=
1730 ((uint32_t*)(src+2*stride))[0]=
0da71265
MN
1731 ((uint32_t*)(src+3*stride))[0]= 128U*0x01010101U;
1732}
1733
1734
1735#define LOAD_TOP_RIGHT_EDGE\
154e30f6
CEH
1736 const int av_unused t4= topright[0];\
1737 const int av_unused t5= topright[1];\
1738 const int av_unused t6= topright[2];\
1739 const int av_unused t7= topright[3];\
0da71265
MN
1740
1741#define LOAD_LEFT_EDGE\
154e30f6
CEH
1742 const int av_unused l0= src[-1+0*stride];\
1743 const int av_unused l1= src[-1+1*stride];\
1744 const int av_unused l2= src[-1+2*stride];\
1745 const int av_unused l3= src[-1+3*stride];\
0da71265
MN
1746
1747#define LOAD_TOP_EDGE\
154e30f6
CEH
1748 const int av_unused t0= src[ 0-1*stride];\
1749 const int av_unused t1= src[ 1-1*stride];\
1750 const int av_unused t2= src[ 2-1*stride];\
1751 const int av_unused t3= src[ 3-1*stride];\
0da71265
MN
1752
1753static void pred4x4_down_right_c(uint8_t *src, uint8_t *topright, int stride){
1754 const int lt= src[-1-1*stride];
1755 LOAD_TOP_EDGE
1756 LOAD_LEFT_EDGE
1757
115329f1 1758 src[0+3*stride]=(l3 + 2*l2 + l1 + 2)>>2;
0da71265 1759 src[0+2*stride]=
115329f1 1760 src[1+3*stride]=(l2 + 2*l1 + l0 + 2)>>2;
0da71265
MN
1761 src[0+1*stride]=
1762 src[1+2*stride]=
115329f1 1763 src[2+3*stride]=(l1 + 2*l0 + lt + 2)>>2;
0da71265
MN
1764 src[0+0*stride]=
1765 src[1+1*stride]=
1766 src[2+2*stride]=
115329f1 1767 src[3+3*stride]=(l0 + 2*lt + t0 + 2)>>2;
0da71265
MN
1768 src[1+0*stride]=
1769 src[2+1*stride]=
1770 src[3+2*stride]=(lt + 2*t0 + t1 + 2)>>2;
1771 src[2+0*stride]=
1772 src[3+1*stride]=(t0 + 2*t1 + t2 + 2)>>2;
1773 src[3+0*stride]=(t1 + 2*t2 + t3 + 2)>>2;
4cfbf61b 1774}
0da71265
MN
1775
1776static void pred4x4_down_left_c(uint8_t *src, uint8_t *topright, int stride){
115329f1
DB
1777 LOAD_TOP_EDGE
1778 LOAD_TOP_RIGHT_EDGE
1779// LOAD_LEFT_EDGE
0da71265
MN
1780
1781 src[0+0*stride]=(t0 + t2 + 2*t1 + 2)>>2;
1782 src[1+0*stride]=
1783 src[0+1*stride]=(t1 + t3 + 2*t2 + 2)>>2;
1784 src[2+0*stride]=
1785 src[1+1*stride]=
1786 src[0+2*stride]=(t2 + t4 + 2*t3 + 2)>>2;
1787 src[3+0*stride]=
1788 src[2+1*stride]=
1789 src[1+2*stride]=
1790 src[0+3*stride]=(t3 + t5 + 2*t4 + 2)>>2;
1791 src[3+1*stride]=
1792 src[2+2*stride]=
1793 src[1+3*stride]=(t4 + t6 + 2*t5 + 2)>>2;
1794 src[3+2*stride]=
1795 src[2+3*stride]=(t5 + t7 + 2*t6 + 2)>>2;
1796 src[3+3*stride]=(t6 + 3*t7 + 2)>>2;
4cfbf61b 1797}
0da71265
MN
1798
1799static void pred4x4_vertical_right_c(uint8_t *src, uint8_t *topright, int stride){
1800 const int lt= src[-1-1*stride];
115329f1
DB
1801 LOAD_TOP_EDGE
1802 LOAD_LEFT_EDGE
0da71265
MN
1803
1804 src[0+0*stride]=
1805 src[1+2*stride]=(lt + t0 + 1)>>1;
1806 src[1+0*stride]=
1807 src[2+2*stride]=(t0 + t1 + 1)>>1;
1808 src[2+0*stride]=
1809 src[3+2*stride]=(t1 + t2 + 1)>>1;
1810 src[3+0*stride]=(t2 + t3 + 1)>>1;
1811 src[0+1*stride]=
1812 src[1+3*stride]=(l0 + 2*lt + t0 + 2)>>2;
1813 src[1+1*stride]=
1814 src[2+3*stride]=(lt + 2*t0 + t1 + 2)>>2;
1815 src[2+1*stride]=
1816 src[3+3*stride]=(t0 + 2*t1 + t2 + 2)>>2;
1817 src[3+1*stride]=(t1 + 2*t2 + t3 + 2)>>2;
1818 src[0+2*stride]=(lt + 2*l0 + l1 + 2)>>2;
1819 src[0+3*stride]=(l0 + 2*l1 + l2 + 2)>>2;
4cfbf61b 1820}
0da71265
MN
1821
1822static void pred4x4_vertical_left_c(uint8_t *src, uint8_t *topright, int stride){
115329f1
DB
1823 LOAD_TOP_EDGE
1824 LOAD_TOP_RIGHT_EDGE
0da71265
MN
1825
1826 src[0+0*stride]=(t0 + t1 + 1)>>1;
1827 src[1+0*stride]=
1828 src[0+2*stride]=(t1 + t2 + 1)>>1;
1829 src[2+0*stride]=
1830 src[1+2*stride]=(t2 + t3 + 1)>>1;
1831 src[3+0*stride]=
1832 src[2+2*stride]=(t3 + t4+ 1)>>1;
1833 src[3+2*stride]=(t4 + t5+ 1)>>1;
1834 src[0+1*stride]=(t0 + 2*t1 + t2 + 2)>>2;
1835 src[1+1*stride]=
1836 src[0+3*stride]=(t1 + 2*t2 + t3 + 2)>>2;
1837 src[2+1*stride]=
1838 src[1+3*stride]=(t2 + 2*t3 + t4 + 2)>>2;
1839 src[3+1*stride]=
1840 src[2+3*stride]=(t3 + 2*t4 + t5 + 2)>>2;
1841 src[3+3*stride]=(t4 + 2*t5 + t6 + 2)>>2;
4cfbf61b 1842}
0da71265
MN
1843
1844static void pred4x4_horizontal_up_c(uint8_t *src, uint8_t *topright, int stride){
115329f1 1845 LOAD_LEFT_EDGE
0da71265
MN
1846
1847 src[0+0*stride]=(l0 + l1 + 1)>>1;
1848 src[1+0*stride]=(l0 + 2*l1 + l2 + 2)>>2;
1849 src[2+0*stride]=
1850 src[0+1*stride]=(l1 + l2 + 1)>>1;
1851 src[3+0*stride]=
1852 src[1+1*stride]=(l1 + 2*l2 + l3 + 2)>>2;
1853 src[2+1*stride]=
1854 src[0+2*stride]=(l2 + l3 + 1)>>1;
1855 src[3+1*stride]=
1856 src[1+2*stride]=(l2 + 2*l3 + l3 + 2)>>2;
1857 src[3+2*stride]=
1858 src[1+3*stride]=
1859 src[0+3*stride]=
1860 src[2+2*stride]=
1861 src[2+3*stride]=
1862 src[3+3*stride]=l3;
4cfbf61b 1863}
115329f1 1864
0da71265
MN
1865static void pred4x4_horizontal_down_c(uint8_t *src, uint8_t *topright, int stride){
1866 const int lt= src[-1-1*stride];
115329f1
DB
1867 LOAD_TOP_EDGE
1868 LOAD_LEFT_EDGE
0da71265
MN
1869
1870 src[0+0*stride]=
1871 src[2+1*stride]=(lt + l0 + 1)>>1;
1872 src[1+0*stride]=
1873 src[3+1*stride]=(l0 + 2*lt + t0 + 2)>>2;
1874 src[2+0*stride]=(lt + 2*t0 + t1 + 2)>>2;
1875 src[3+0*stride]=(t0 + 2*t1 + t2 + 2)>>2;
1876 src[0+1*stride]=
1877 src[2+2*stride]=(l0 + l1 + 1)>>1;
1878 src[1+1*stride]=
1879 src[3+2*stride]=(lt + 2*l0 + l1 + 2)>>2;
1880 src[0+2*stride]=
1881 src[2+3*stride]=(l1 + l2+ 1)>>1;
1882 src[1+2*stride]=
1883 src[3+3*stride]=(l0 + 2*l1 + l2 + 2)>>2;
1884 src[0+3*stride]=(l2 + l3 + 1)>>1;
1885 src[1+3*stride]=(l1 + 2*l2 + l3 + 2)>>2;
4cfbf61b 1886}
0da71265 1887
a074f143 1888void ff_pred16x16_vertical_c(uint8_t *src, int stride){
0da71265
MN
1889 int i;
1890 const uint32_t a= ((uint32_t*)(src-stride))[0];
1891 const uint32_t b= ((uint32_t*)(src-stride))[1];
1892 const uint32_t c= ((uint32_t*)(src-stride))[2];
1893 const uint32_t d= ((uint32_t*)(src-stride))[3];
115329f1 1894
0da71265
MN
1895 for(i=0; i<16; i++){
1896 ((uint32_t*)(src+i*stride))[0]= a;
1897 ((uint32_t*)(src+i*stride))[1]= b;
1898 ((uint32_t*)(src+i*stride))[2]= c;
1899 ((uint32_t*)(src+i*stride))[3]= d;
1900 }
1901}
1902
a074f143 1903void ff_pred16x16_horizontal_c(uint8_t *src, int stride){
0da71265
MN
1904 int i;
1905
1906 for(i=0; i<16; i++){
1907 ((uint32_t*)(src+i*stride))[0]=
1908 ((uint32_t*)(src+i*stride))[1]=
1909 ((uint32_t*)(src+i*stride))[2]=
1910 ((uint32_t*)(src+i*stride))[3]= src[-1+i*stride]*0x01010101;
1911 }
1912}
1913
a074f143 1914void ff_pred16x16_dc_c(uint8_t *src, int stride){
0da71265
MN
1915 int i, dc=0;
1916
1917 for(i=0;i<16; i++){
1918 dc+= src[-1+i*stride];
1919 }
115329f1 1920
0da71265
MN
1921 for(i=0;i<16; i++){
1922 dc+= src[i-stride];
1923 }
1924
1925 dc= 0x01010101*((dc + 16)>>5);
1926
1927 for(i=0; i<16; i++){
1928 ((uint32_t*)(src+i*stride))[0]=
1929 ((uint32_t*)(src+i*stride))[1]=
1930 ((uint32_t*)(src+i*stride))[2]=
1931 ((uint32_t*)(src+i*stride))[3]= dc;
1932 }
1933}
1934
cb1a74cf 1935void ff_pred16x16_left_dc_c(uint8_t *src, int stride){
0da71265
MN
1936 int i, dc=0;
1937
1938 for(i=0;i<16; i++){
1939 dc+= src[-1+i*stride];
1940 }
115329f1 1941
0da71265
MN
1942 dc= 0x01010101*((dc + 8)>>4);
1943
1944 for(i=0; i<16; i++){
1945 ((uint32_t*)(src+i*stride))[0]=
1946 ((uint32_t*)(src+i*stride))[1]=
1947 ((uint32_t*)(src+i*stride))[2]=
1948 ((uint32_t*)(src+i*stride))[3]= dc;
1949 }
1950}
1951
cb1a74cf 1952void ff_pred16x16_top_dc_c(uint8_t *src, int stride){
0da71265
MN
1953 int i, dc=0;
1954
1955 for(i=0;i<16; i++){
1956 dc+= src[i-stride];
1957 }
1958 dc= 0x01010101*((dc + 8)>>4);
1959
1960 for(i=0; i<16; i++){
1961 ((uint32_t*)(src+i*stride))[0]=
1962 ((uint32_t*)(src+i*stride))[1]=
1963 ((uint32_t*)(src+i*stride))[2]=
1964 ((uint32_t*)(src+i*stride))[3]= dc;
1965 }
1966}
1967
a074f143 1968void ff_pred16x16_128_dc_c(uint8_t *src, int stride){
0da71265
MN
1969 int i;
1970
1971 for(i=0; i<16; i++){
1972 ((uint32_t*)(src+i*stride))[0]=
1973 ((uint32_t*)(src+i*stride))[1]=
1974 ((uint32_t*)(src+i*stride))[2]=
1975 ((uint32_t*)(src+i*stride))[3]= 0x01010101U*128U;
1976 }
1977}
1978
8b82a956 1979static inline void pred16x16_plane_compat_c(uint8_t *src, int stride, const int svq3){
30f73fc7
MN
1980 int i, j, k;
1981 int a;
55fde95e 1982 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;
30f73fc7
MN
1983 const uint8_t * const src0 = src+7-stride;
1984 const uint8_t *src1 = src+8*stride-1;
1985 const uint8_t *src2 = src1-2*stride; // == src+6*stride-1;
1986 int H = src0[1] - src0[-1];
1987 int V = src1[0] - src2[ 0];
1988 for(k=2; k<=8; ++k) {
1989 src1 += stride; src2 -= stride;
1990 H += k*(src0[k] - src0[-k]);
1991 V += k*(src1[0] - src2[ 0]);
1992 }
8b82a956
MN
1993 if(svq3){
1994 H = ( 5*(H/4) ) / 16;
1995 V = ( 5*(V/4) ) / 16;
2e26c8d2
MM
1996
1997 /* required for 100% accuracy */
1998 i = H; H = V; V = i;
8b82a956
MN
1999 }else{
2000 H = ( 5*H+32 ) >> 6;
2001 V = ( 5*V+32 ) >> 6;
2002 }
30f73fc7
MN
2003
2004 a = 16*(src1[0] + src2[16] + 1) - 7*(V+H);
2005 for(j=16; j>0; --j) {
2006 int b = a;
2007 a += V;
2008 for(i=-16; i<0; i+=4) {
2009 src[16+i] = cm[ (b ) >> 5 ];
2010 src[17+i] = cm[ (b+ H) >> 5 ];
2011 src[18+i] = cm[ (b+2*H) >> 5 ];
2012 src[19+i] = cm[ (b+3*H) >> 5 ];
2013 b += 4*H;
0da71265 2014 }
30f73fc7
MN
2015 src += stride;
2016 }
0da71265
MN
2017}
2018
a074f143 2019void ff_pred16x16_plane_c(uint8_t *src, int stride){
8b82a956
MN
2020 pred16x16_plane_compat_c(src, stride, 0);
2021}
2022
a074f143 2023void ff_pred8x8_vertical_c(uint8_t *src, int stride){
0da71265
MN
2024 int i;
2025 const uint32_t a= ((uint32_t*)(src-stride))[0];
2026 const uint32_t b= ((uint32_t*)(src-stride))[1];
115329f1 2027
0da71265
MN
2028 for(i=0; i<8; i++){
2029 ((uint32_t*)(src+i*stride))[0]= a;
2030 ((uint32_t*)(src+i*stride))[1]= b;
2031 }
2032}
2033
a074f143 2034void ff_pred8x8_horizontal_c(uint8_t *src, int stride){
0da71265
MN
2035 int i;
2036
2037 for(i=0; i<8; i++){
2038 ((uint32_t*)(src+i*stride))[0]=
2039 ((uint32_t*)(src+i*stride))[1]= src[-1+i*stride]*0x01010101;
2040 }
2041}
2042
a074f143 2043void ff_pred8x8_128_dc_c(uint8_t *src, int stride){
0da71265
MN
2044 int i;
2045
43efd19a 2046 for(i=0; i<8; i++){
115329f1 2047 ((uint32_t*)(src+i*stride))[0]=
0da71265
MN
2048 ((uint32_t*)(src+i*stride))[1]= 0x01010101U*128U;
2049 }
2050}
2051
cb1a74cf 2052void ff_pred8x8_left_dc_c(uint8_t *src, int stride){
0da71265
MN
2053 int i;
2054 int dc0, dc2;
2055
2056 dc0=dc2=0;
2057 for(i=0;i<4; i++){
2058 dc0+= src[-1+i*stride];
2059 dc2+= src[-1+(i+4)*stride];
2060 }
2061 dc0= 0x01010101*((dc0 + 2)>>2);
2062 dc2= 0x01010101*((dc2 + 2)>>2);
2063
2064 for(i=0; i<4; i++){
2065 ((uint32_t*)(src+i*stride))[0]=
2066 ((uint32_t*)(src+i*stride))[1]= dc0;
2067 }
2068 for(i=4; i<8; i++){
2069 ((uint32_t*)(src+i*stride))[0]=
2070 ((uint32_t*)(src+i*stride))[1]= dc2;
2071 }
2072}
2073
cb1a74cf 2074void ff_pred8x8_top_dc_c(uint8_t *src, int stride){
0da71265
MN
2075 int i;
2076 int dc0, dc1;
2077
2078 dc0=dc1=0;
2079 for(i=0;i<4; i++){
2080 dc0+= src[i-stride];
2081 dc1+= src[4+i-stride];
2082 }
2083 dc0= 0x01010101*((dc0 + 2)>>2);
2084 dc1= 0x01010101*((dc1 + 2)>>2);
2085
2086 for(i=0; i<4; i++){
2087 ((uint32_t*)(src+i*stride))[0]= dc0;
2088 ((uint32_t*)(src+i*stride))[1]= dc1;
2089 }
2090 for(i=4; i<8; i++){
2091 ((uint32_t*)(src+i*stride))[0]= dc0;
2092 ((uint32_t*)(src+i*stride))[1]= dc1;
2093 }
2094}
2095
2096
a074f143 2097void ff_pred8x8_dc_c(uint8_t *src, int stride){
0da71265
MN
2098 int i;
2099 int dc0, dc1, dc2, dc3;
2100
2101 dc0=dc1=dc2=0;
2102 for(i=0;i<4; i++){
2103 dc0+= src[-1+i*stride] + src[i-stride];
2104 dc1+= src[4+i-stride];
2105 dc2+= src[-1+(i+4)*stride];
2106 }
2107 dc3= 0x01010101*((dc1 + dc2 + 4)>>3);
2108 dc0= 0x01010101*((dc0 + 4)>>3);
2109 dc1= 0x01010101*((dc1 + 2)>>2);
2110 dc2= 0x01010101*((dc2 + 2)>>2);
2111
2112 for(i=0; i<4; i++){
2113 ((uint32_t*)(src+i*stride))[0]= dc0;
2114 ((uint32_t*)(src+i*stride))[1]= dc1;
2115 }
2116 for(i=4; i<8; i++){
2117 ((uint32_t*)(src+i*stride))[0]= dc2;
2118 ((uint32_t*)(src+i*stride))[1]= dc3;
2119 }
2120}
2121
a074f143 2122void ff_pred8x8_plane_c(uint8_t *src, int stride){
30f73fc7
MN
2123 int j, k;
2124 int a;
55fde95e 2125 uint8_t *cm = ff_cropTbl + MAX_NEG_CROP;
30f73fc7
MN
2126 const uint8_t * const src0 = src+3-stride;
2127 const uint8_t *src1 = src+4*stride-1;
2128 const uint8_t *src2 = src1-2*stride; // == src+2*stride-1;
2129 int H = src0[1] - src0[-1];
2130 int V = src1[0] - src2[ 0];
2131 for(k=2; k<=4; ++k) {
2132 src1 += stride; src2 -= stride;
2133 H += k*(src0[k] - src0[-k]);
2134 V += k*(src1[0] - src2[ 0]);
2135 }
2136 H = ( 17*H+16 ) >> 5;
2137 V = ( 17*V+16 ) >> 5;
2138
2139 a = 16*(src1[0] + src2[8]+1) - 3*(V+H);
2140 for(j=8; j>0; --j) {
2141 int b = a;
2142 a += V;
2143 src[0] = cm[ (b ) >> 5 ];
2144 src[1] = cm[ (b+ H) >> 5 ];
2145 src[2] = cm[ (b+2*H) >> 5 ];
2146 src[3] = cm[ (b+3*H) >> 5 ];
2147 src[4] = cm[ (b+4*H) >> 5 ];
2148 src[5] = cm[ (b+5*H) >> 5 ];
2149 src[6] = cm[ (b+6*H) >> 5 ];
2150 src[7] = cm[ (b+7*H) >> 5 ];
2151 src += stride;
2152 }
0da71265
MN
2153}
2154
43efd19a
LM
2155#define SRC(x,y) src[(x)+(y)*stride]
2156#define PL(y) \
2157 const int l##y = (SRC(-1,y-1) + 2*SRC(-1,y) + SRC(-1,y+1) + 2) >> 2;
2158#define PREDICT_8x8_LOAD_LEFT \
2159 const int l0 = ((has_topleft ? SRC(-1,-1) : SRC(-1,0)) \
2160 + 2*SRC(-1,0) + SRC(-1,1) + 2) >> 2; \
2161 PL(1) PL(2) PL(3) PL(4) PL(5) PL(6) \
154e30f6 2162 const int l7 av_unused = (SRC(-1,6) + 3*SRC(-1,7) + 2) >> 2
43efd19a
LM
2163
2164#define PT(x) \
2165 const int t##x = (SRC(x-1,-1) + 2*SRC(x,-1) + SRC(x+1,-1) + 2) >> 2;
2166#define PREDICT_8x8_LOAD_TOP \
2167 const int t0 = ((has_topleft ? SRC(-1,-1) : SRC(0,-1)) \
2168 + 2*SRC(0,-1) + SRC(1,-1) + 2) >> 2; \
2169 PT(1) PT(2) PT(3) PT(4) PT(5) PT(6) \
154e30f6 2170 const int t7 av_unused = ((has_topright ? SRC(8,-1) : SRC(7,-1)) \
fb3ca6ba 2171 + 2*SRC(7,-1) + SRC(6,-1) + 2) >> 2
43efd19a
LM
2172
2173#define PTR(x) \
2174 t##x = (SRC(x-1,-1) + 2*SRC(x,-1) + SRC(x+1,-1) + 2) >> 2;
2175#define PREDICT_8x8_LOAD_TOPRIGHT \
2176 int t8, t9, t10, t11, t12, t13, t14, t15; \
2177 if(has_topright) { \
2178 PTR(8) PTR(9) PTR(10) PTR(11) PTR(12) PTR(13) PTR(14) \
2179 t15 = (SRC(14,-1) + 3*SRC(15,-1) + 2) >> 2; \
2180 } else t8=t9=t10=t11=t12=t13=t14=t15= SRC(7,-1);
2181
2182#define PREDICT_8x8_LOAD_TOPLEFT \
fb3ca6ba 2183 const int lt = (SRC(-1,0) + 2*SRC(-1,-1) + SRC(0,-1) + 2) >> 2
43efd19a
LM
2184
2185#define PREDICT_8x8_DC(v) \
2186 int y; \
2187 for( y = 0; y < 8; y++ ) { \
2188 ((uint32_t*)src)[0] = \
2189 ((uint32_t*)src)[1] = v; \
2190 src += stride; \
2191 }
2192
2193static void pred8x8l_128_dc_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2194{
2195 PREDICT_8x8_DC(0x80808080);
2196}
2197static void pred8x8l_left_dc_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2198{
2199 PREDICT_8x8_LOAD_LEFT;
2200 const uint32_t dc = ((l0+l1+l2+l3+l4+l5+l6+l7+4) >> 3) * 0x01010101;
2201 PREDICT_8x8_DC(dc);
2202}
2203static void pred8x8l_top_dc_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2204{
2205 PREDICT_8x8_LOAD_TOP;
2206 const uint32_t dc = ((t0+t1+t2+t3+t4+t5+t6+t7+4) >> 3) * 0x01010101;
2207 PREDICT_8x8_DC(dc);
2208}
2209static void pred8x8l_dc_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2210{
2211 PREDICT_8x8_LOAD_LEFT;
2212 PREDICT_8x8_LOAD_TOP;
2213 const uint32_t dc = ((l0+l1+l2+l3+l4+l5+l6+l7
2214 +t0+t1+t2+t3+t4+t5+t6+t7+8) >> 4) * 0x01010101;
2215 PREDICT_8x8_DC(dc);
2216}
2217static void pred8x8l_horizontal_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2218{
2219 PREDICT_8x8_LOAD_LEFT;
2220#define ROW(y) ((uint32_t*)(src+y*stride))[0] =\
2221 ((uint32_t*)(src+y*stride))[1] = 0x01010101 * l##y
2222 ROW(0); ROW(1); ROW(2); ROW(3); ROW(4); ROW(5); ROW(6); ROW(7);
2223#undef ROW
2224}
2225static void pred8x8l_vertical_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2226{
2227 int y;
2228 PREDICT_8x8_LOAD_TOP;
2229 src[0] = t0;
2230 src[1] = t1;
2231 src[2] = t2;
2232 src[3] = t3;
2233 src[4] = t4;
2234 src[5] = t5;
2235 src[6] = t6;
2236 src[7] = t7;
2237 for( y = 1; y < 8; y++ )
2238 *(uint64_t*)(src+y*stride) = *(uint64_t*)src;
2239}
2240static void pred8x8l_down_left_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2241{
2242 PREDICT_8x8_LOAD_TOP;
2243 PREDICT_8x8_LOAD_TOPRIGHT;
2244 SRC(0,0)= (t0 + 2*t1 + t2 + 2) >> 2;
2245 SRC(0,1)=SRC(1,0)= (t1 + 2*t2 + t3 + 2) >> 2;
2246 SRC(0,2)=SRC(1,1)=SRC(2,0)= (t2 + 2*t3 + t4 + 2) >> 2;
2247 SRC(0,3)=SRC(1,2)=SRC(2,1)=SRC(3,0)= (t3 + 2*t4 + t5 + 2) >> 2;
2248 SRC(0,4)=SRC(1,3)=SRC(2,2)=SRC(3,1)=SRC(4,0)= (t4 + 2*t5 + t6 + 2) >> 2;
2249 SRC(0,5)=SRC(1,4)=SRC(2,3)=SRC(3,2)=SRC(4,1)=SRC(5,0)= (t5 + 2*t6 + t7 + 2) >> 2;
2250 SRC(0,6)=SRC(1,5)=SRC(2,4)=SRC(3,3)=SRC(4,2)=SRC(5,1)=SRC(6,0)= (t6 + 2*t7 + t8 + 2) >> 2;
2251 SRC(0,7)=SRC(1,6)=SRC(2,5)=SRC(3,4)=SRC(4,3)=SRC(5,2)=SRC(6,1)=SRC(7,0)= (t7 + 2*t8 + t9 + 2) >> 2;
2252 SRC(1,7)=SRC(2,6)=SRC(3,5)=SRC(4,4)=SRC(5,3)=SRC(6,2)=SRC(7,1)= (t8 + 2*t9 + t10 + 2) >> 2;
2253 SRC(2,7)=SRC(3,6)=SRC(4,5)=SRC(5,4)=SRC(6,3)=SRC(7,2)= (t9 + 2*t10 + t11 + 2) >> 2;
2254 SRC(3,7)=SRC(4,6)=SRC(5,5)=SRC(6,4)=SRC(7,3)= (t10 + 2*t11 + t12 + 2) >> 2;
2255 SRC(4,7)=SRC(5,6)=SRC(6,5)=SRC(7,4)= (t11 + 2*t12 + t13 + 2) >> 2;
2256 SRC(5,7)=SRC(6,6)=SRC(7,5)= (t12 + 2*t13 + t14 + 2) >> 2;
2257 SRC(6,7)=SRC(7,6)= (t13 + 2*t14 + t15 + 2) >> 2;
2258 SRC(7,7)= (t14 + 3*t15 + 2) >> 2;
2259}
2260static void pred8x8l_down_right_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2261{
2262 PREDICT_8x8_LOAD_TOP;
2263 PREDICT_8x8_LOAD_LEFT;
2264 PREDICT_8x8_LOAD_TOPLEFT;
2265 SRC(0,7)= (l7 + 2*l6 + l5 + 2) >> 2;
2266 SRC(0,6)=SRC(1,7)= (l6 + 2*l5 + l4 + 2) >> 2;
2267 SRC(0,5)=SRC(1,6)=SRC(2,7)= (l5 + 2*l4 + l3 + 2) >> 2;
2268 SRC(0,4)=SRC(1,5)=SRC(2,6)=SRC(3,7)= (l4 + 2*l3 + l2 + 2) >> 2;
2269 SRC(0,3)=SRC(1,4)=SRC(2,5)=SRC(3,6)=SRC(4,7)= (l3 + 2*l2 + l1 + 2) >> 2;
2270 SRC(0,2)=SRC(1,3)=SRC(2,4)=SRC(3,5)=SRC(4,6)=SRC(5,7)= (l2 + 2*l1 + l0 + 2) >> 2;
2271 SRC(0,1)=SRC(1,2)=SRC(2,3)=SRC(3,4)=SRC(4,5)=SRC(5,6)=SRC(6,7)= (l1 + 2*l0 + lt + 2) >> 2;
2272 SRC(0,0)=SRC(1,1)=SRC(2,2)=SRC(3,3)=SRC(4,4)=SRC(5,5)=SRC(6,6)=SRC(7,7)= (l0 + 2*lt + t0 + 2) >> 2;
2273 SRC(1,0)=SRC(2,1)=SRC(3,2)=SRC(4,3)=SRC(5,4)=SRC(6,5)=SRC(7,6)= (lt + 2*t0 + t1 + 2) >> 2;
2274 SRC(2,0)=SRC(3,1)=SRC(4,2)=SRC(5,3)=SRC(6,4)=SRC(7,5)= (t0 + 2*t1 + t2 + 2) >> 2;
2275 SRC(3,0)=SRC(4,1)=SRC(5,2)=SRC(6,3)=SRC(7,4)= (t1 + 2*t2 + t3 + 2) >> 2;
2276 SRC(4,0)=SRC(5,1)=SRC(6,2)=SRC(7,3)= (t2 + 2*t3 + t4 + 2) >> 2;
2277 SRC(5,0)=SRC(6,1)=SRC(7,2)= (t3 + 2*t4 + t5 + 2) >> 2;
2278 SRC(6,0)=SRC(7,1)= (t4 + 2*t5 + t6 + 2) >> 2;
2279 SRC(7,0)= (t5 + 2*t6 + t7 + 2) >> 2;
115329f1 2280
43efd19a
LM
2281}
2282static void pred8x8l_vertical_right_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2283{
2284 PREDICT_8x8_LOAD_TOP;
2285 PREDICT_8x8_LOAD_LEFT;
2286 PREDICT_8x8_LOAD_TOPLEFT;
2287 SRC(0,6)= (l5 + 2*l4 + l3 + 2) >> 2;
2288 SRC(0,7)= (l6 + 2*l5 + l4 + 2) >> 2;
2289 SRC(0,4)=SRC(1,6)= (l3 + 2*l2 + l1 + 2) >> 2;
2290 SRC(0,5)=SRC(1,7)= (l4 + 2*l3 + l2 + 2) >> 2;
2291 SRC(0,2)=SRC(1,4)=SRC(2,6)= (l1 + 2*l0 + lt + 2) >> 2;
2292 SRC(0,3)=SRC(1,5)=SRC(2,7)= (l2 + 2*l1 + l0 + 2) >> 2;
2293 SRC(0,1)=SRC(1,3)=SRC(2,5)=SRC(3,7)= (l0 + 2*lt + t0 + 2) >> 2;
2294 SRC(0,0)=SRC(1,2)=SRC(2,4)=SRC(3,6)= (lt + t0 + 1) >> 1;
2295 SRC(1,1)=SRC(2,3)=SRC(3,5)=SRC(4,7)= (lt + 2*t0 + t1 + 2) >> 2;
2296 SRC(1,0)=SRC(2,2)=SRC(3,4)=SRC(4,6)= (t0 + t1 + 1) >> 1;
2297 SRC(2,1)=SRC(3,3)=SRC(4,5)=SRC(5,7)= (t0 + 2*t1 + t2 + 2) >> 2;
2298 SRC(2,0)=SRC(3,2)=SRC(4,4)=SRC(5,6)= (t1 + t2 + 1) >> 1;
2299 SRC(3,1)=SRC(4,3)=SRC(5,5)=SRC(6,7)= (t1 + 2*t2 + t3 + 2) >> 2;
2300 SRC(3,0)=SRC(4,2)=SRC(5,4)=SRC(6,6)= (t2 + t3 + 1) >> 1;
2301 SRC(4,1)=SRC(5,3)=SRC(6,5)=SRC(7,7)= (t2 + 2*t3 + t4 + 2) >> 2;
2302 SRC(4,0)=SRC(5,2)=SRC(6,4)=SRC(7,6)= (t3 + t4 + 1) >> 1;
2303 SRC(5,1)=SRC(6,3)=SRC(7,5)= (t3 + 2*t4 + t5 + 2) >> 2;
2304 SRC(5,0)=SRC(6,2)=SRC(7,4)= (t4 + t5 + 1) >> 1;
2305 SRC(6,1)=SRC(7,3)= (t4 + 2*t5 + t6 + 2) >> 2;
2306 SRC(6,0)=SRC(7,2)= (t5 + t6 + 1) >> 1;
2307 SRC(7,1)= (t5 + 2*t6 + t7 + 2) >> 2;
2308 SRC(7,0)= (t6 + t7 + 1) >> 1;
2309}
2310static void pred8x8l_horizontal_down_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2311{
2312 PREDICT_8x8_LOAD_TOP;
2313 PREDICT_8x8_LOAD_LEFT;
2314 PREDICT_8x8_LOAD_TOPLEFT;
2315 SRC(0,7)= (l6 + l7 + 1) >> 1;
2316 SRC(1,7)= (l5 + 2*l6 + l7 + 2) >> 2;
2317 SRC(0,6)=SRC(2,7)= (l5 + l6 + 1) >> 1;
2318 SRC(1,6)=SRC(3,7)= (l4 + 2*l5 + l6 + 2) >> 2;
2319 SRC(0,5)=SRC(2,6)=SRC(4,7)= (l4 + l5 + 1) >> 1;
2320 SRC(1,5)=SRC(3,6)=SRC(5,7)= (l3 + 2*l4 + l5 + 2) >> 2;
2321 SRC(0,4)=SRC(2,5)=SRC(4,6)=SRC(6,7)= (l3 + l4 + 1) >> 1;
2322 SRC(1,4)=SRC(3,5)=SRC(5,6)=SRC(7,7)= (l2 + 2*l3 + l4 + 2) >> 2;
2323 SRC(0,3)=SRC(2,4)=SRC(4,5)=SRC(6,6)= (l2 + l3 + 1) >> 1;
2324 SRC(1,3)=SRC(3,4)=SRC(5,5)=SRC(7,6)= (l1 + 2*l2 + l3 + 2) >> 2;
2325 SRC(0,2)=SRC(2,3)=SRC(4,4)=SRC(6,5)= (l1 + l2 + 1) >> 1;
2326 SRC(1,2)=SRC(3,3)=SRC(5,4)=SRC(7,5)= (l0 + 2*l1 + l2 + 2) >> 2;
2327 SRC(0,1)=SRC(2,2)=SRC(4,3)=SRC(6,4)= (l0 + l1 + 1) >> 1;
2328 SRC(1,1)=SRC(3,2)=SRC(5,3)=SRC(7,4)= (lt + 2*l0 + l1 + 2) >> 2;
2329 SRC(0,0)=SRC(2,1)=SRC(4,2)=SRC(6,3)= (lt + l0 + 1) >> 1;
2330 SRC(1,0)=SRC(3,1)=SRC(5,2)=SRC(7,3)= (l0 + 2*lt + t0 + 2) >> 2;
2331 SRC(2,0)=SRC(4,1)=SRC(6,2)= (t1 + 2*t0 + lt + 2) >> 2;
2332 SRC(3,0)=SRC(5,1)=SRC(7,2)= (t2 + 2*t1 + t0 + 2) >> 2;
2333 SRC(4,0)=SRC(6,1)= (t3 + 2*t2 + t1 + 2) >> 2;
2334 SRC(5,0)=SRC(7,1)= (t4 + 2*t3 + t2 + 2) >> 2;
2335 SRC(6,0)= (t5 + 2*t4 + t3 + 2) >> 2;
2336 SRC(7,0)= (t6 + 2*t5 + t4 + 2) >> 2;
2337}
2338static void pred8x8l_vertical_left_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2339{
2340 PREDICT_8x8_LOAD_TOP;
2341 PREDICT_8x8_LOAD_TOPRIGHT;
2342 SRC(0,0)= (t0 + t1 + 1) >> 1;
2343 SRC(0,1)= (t0 + 2*t1 + t2 + 2) >> 2;
2344 SRC(0,2)=SRC(1,0)= (t1 + t2 + 1) >> 1;
2345 SRC(0,3)=SRC(1,1)= (t1 + 2*t2 + t3 + 2) >> 2;
2346 SRC(0,4)=SRC(1,2)=SRC(2,0)= (t2 + t3 + 1) >> 1;
2347 SRC(0,5)=SRC(1,3)=SRC(2,1)= (t2 + 2*t3 + t4 + 2) >> 2;
2348 SRC(0,6)=SRC(1,4)=SRC(2,2)=SRC(3,0)= (t3 + t4 + 1) >> 1;
2349 SRC(0,7)=SRC(1,5)=SRC(2,3)=SRC(3,1)= (t3 + 2*t4 + t5 + 2) >> 2;
2350 SRC(1,6)=SRC(2,4)=SRC(3,2)=SRC(4,0)= (t4 + t5 + 1) >> 1;
2351 SRC(1,7)=SRC(2,5)=SRC(3,3)=SRC(4,1)= (t4 + 2*t5 + t6 + 2) >> 2;
2352 SRC(2,6)=SRC(3,4)=SRC(4,2)=SRC(5,0)= (t5 + t6 + 1) >> 1;
2353 SRC(2,7)=SRC(3,5)=SRC(4,3)=SRC(5,1)= (t5 + 2*t6 + t7 + 2) >> 2;
2354 SRC(3,6)=SRC(4,4)=SRC(5,2)=SRC(6,0)= (t6 + t7 + 1) >> 1;
2355 SRC(3,7)=SRC(4,5)=SRC(5,3)=SRC(6,1)= (t6 + 2*t7 + t8 + 2) >> 2;
2356 SRC(4,6)=SRC(5,4)=SRC(6,2)=SRC(7,0)= (t7 + t8 + 1) >> 1;
2357 SRC(4,7)=SRC(5,5)=SRC(6,3)=SRC(7,1)= (t7 + 2*t8 + t9 + 2) >> 2;
2358 SRC(5,6)=SRC(6,4)=SRC(7,2)= (t8 + t9 + 1) >> 1;
2359 SRC(5,7)=SRC(6,5)=SRC(7,3)= (t8 + 2*t9 + t10 + 2) >> 2;
2360 SRC(6,6)=SRC(7,4)= (t9 + t10 + 1) >> 1;
2361 SRC(6,7)=SRC(7,5)= (t9 + 2*t10 + t11 + 2) >> 2;
2362 SRC(7,6)= (t10 + t11 + 1) >> 1;
2363 SRC(7,7)= (t10 + 2*t11 + t12 + 2) >> 2;
2364}
2365static void pred8x8l_horizontal_up_c(uint8_t *src, int has_topleft, int has_topright, int stride)
2366{
2367 PREDICT_8x8_LOAD_LEFT;
2368 SRC(0,0)= (l0 + l1 + 1) >> 1;
2369 SRC(1,0)= (l0 + 2*l1 + l2 + 2) >> 2;
2370 SRC(0,1)=SRC(2,0)= (l1 + l2 + 1) >> 1;
2371 SRC(1,1)=SRC(3,0)= (l1 + 2*l2 + l3 + 2) >> 2;
2372 SRC(0,2)=SRC(2,1)=SRC(4,0)= (l2 + l3 + 1) >> 1;
2373 SRC(1,2)=SRC(3,1)=SRC(5,0)= (l2 + 2*l3 + l4 + 2) >> 2;
2374 SRC(0,3)=SRC(2,2)=SRC(4,1)=SRC(6,0)= (l3 + l4 + 1) >> 1;
2375 SRC(1,3)=SRC(3,2)=SRC(5,1)=SRC(7,0)= (l3 + 2*l4 + l5 + 2) >> 2;
2376 SRC(0,4)=SRC(2,3)=SRC(4,2)=SRC(6,1)= (l4 + l5 + 1) >> 1;
2377 SRC(1,4)=SRC(3,3)=SRC(5,2)=SRC(7,1)= (l4 + 2*l5 + l6 + 2) >> 2;
2378 SRC(0,5)=SRC(2,4)=SRC(4,3)=SRC(6,2)= (l5 + l6 + 1) >> 1;
2379 SRC(1,5)=SRC(3,4)=SRC(5,3)=SRC(7,2)= (l5 + 2*l6 + l7 + 2) >> 2;
2380 SRC(0,6)=SRC(2,5)=SRC(4,4)=SRC(6,3)= (l6 + l7 + 1) >> 1;
2381 SRC(1,6)=SRC(3,5)=SRC(5,4)=SRC(7,3)= (l6 + 3*l7 + 2) >> 2;
2382 SRC(0,7)=SRC(1,7)=SRC(2,6)=SRC(2,7)=SRC(3,6)=
2383 SRC(3,7)=SRC(4,5)=SRC(4,6)=SRC(4,7)=SRC(5,5)=
2384 SRC(5,6)=SRC(5,7)=SRC(6,4)=SRC(6,5)=SRC(6,6)=
2385 SRC(6,7)=SRC(7,4)=SRC(7,5)=SRC(7,6)=SRC(7,7)= l7;
2386}
2387#undef PREDICT_8x8_LOAD_LEFT
2388#undef PREDICT_8x8_LOAD_TOP
2389#undef PREDICT_8x8_LOAD_TOPLEFT
2390#undef PREDICT_8x8_LOAD_TOPRIGHT
2391#undef PREDICT_8x8_DC
2392#undef PTR
2393#undef PT
2394#undef PL
2395#undef SRC
2396
0da71265
MN
2397static inline void mc_dir_part(H264Context *h, Picture *pic, int n, int square, int chroma_height, int delta, int list,
2398 uint8_t *dest_y, uint8_t *dest_cb, uint8_t *dest_cr,
2399 int src_x_offset, int src_y_offset,
2400 qpel_mc_func *qpix_op, h264_chroma_mc_func chroma_op){
2401 MpegEncContext * const s = &h->s;
2402 const int mx= h->mv_cache[list][ scan8[n] ][0] + src_x_offset*8;
5d18eaad 2403 int my= h->mv_cache[list][ scan8[n] ][1] + src_y_offset*8;
0da71265 2404 const int luma_xy= (mx&3) + ((my&3)<<2);
5d18eaad
LM
2405 uint8_t * src_y = pic->data[0] + (mx>>2) + (my>>2)*h->mb_linesize;
2406 uint8_t * src_cb, * src_cr;
2407 int extra_width= h->emu_edge_width;
2408 int extra_height= h->emu_edge_height;
0da71265
MN
2409 int emu=0;
2410 const int full_mx= mx>>2;
2411 const int full_my= my>>2;
fbd312fd 2412 const int pic_width = 16*s->mb_width;
5d18eaad 2413 const int pic_height = 16*s->mb_height >> MB_MBAFF;
115329f1 2414
2f29af39 2415 if(!pic->data[0]) //FIXME this is unacceptable, some senseable error concealment must be done for missing reference frames
171c4076 2416 return;
115329f1 2417
0da71265
MN
2418 if(mx&7) extra_width -= 3;
2419 if(my&7) extra_height -= 3;
115329f1
DB
2420
2421 if( full_mx < 0-extra_width
2422 || full_my < 0-extra_height
2423 || full_mx + 16/*FIXME*/ > pic_width + extra_width
fbd312fd 2424 || full_my + 16/*FIXME*/ > pic_height + extra_height){
5d18eaad
LM
2425 ff_emulated_edge_mc(s->edge_emu_buffer, src_y - 2 - 2*h->mb_linesize, h->mb_linesize, 16+5, 16+5/*FIXME*/, full_mx-2, full_my-2, pic_width, pic_height);
2426 src_y= s->edge_emu_buffer + 2 + 2*h->mb_linesize;
0da71265
MN
2427 emu=1;
2428 }
115329f1 2429
5d18eaad 2430 qpix_op[luma_xy](dest_y, src_y, h->mb_linesize); //FIXME try variable height perhaps?
0da71265 2431 if(!square){
5d18eaad 2432 qpix_op[luma_xy](dest_y + delta, src_y + delta, h->mb_linesize);
0da71265 2433 }
115329f1 2434
0da71265 2435 if(s->flags&CODEC_FLAG_GRAY) return;
115329f1 2436
5d18eaad
LM
2437 if(MB_MBAFF){
2438 // chroma offset when predicting from a field of opposite parity
2439 my += 2 * ((s->mb_y & 1) - (h->ref_cache[list][scan8[n]] & 1));
2440 emu |= (my>>3) < 0 || (my>>3) + 8 >= (pic_height>>1);
2441 }
2442 src_cb= pic->data[1] + (mx>>3) + (my>>3)*h->mb_uvlinesize;
2443 src_cr= pic->data[2] + (mx>>3) + (my>>3)*h->mb_uvlinesize;
2444
0da71265 2445 if(emu){
5d18eaad 2446 ff_emulated_edge_mc(s->edge_emu_buffer, src_cb, h->mb_uvlinesize, 9, 9/*FIXME*/, (mx>>3), (my>>3), pic_width>>1, pic_height>>1);
0da71265
MN
2447 src_cb= s->edge_emu_buffer;
2448 }
5d18eaad 2449 chroma_op(dest_cb, src_cb, h->mb_uvlinesize, chroma_height, mx&7, my&7);
0da71265
MN
2450
2451 if(emu){
5d18eaad 2452 ff_emulated_edge_mc(s->edge_emu_buffer, src_cr, h->mb_uvlinesize, 9, 9/*FIXME*/, (mx>>3), (my>>3), pic_width>>1, pic_height>>1);
0da71265
MN
2453 src_cr= s->edge_emu_buffer;
2454 }
5d18eaad 2455 chroma_op(dest_cr, src_cr, h->mb_uvlinesize, chroma_height, mx&7, my&7);
0da71265
MN
2456}
2457
9f2d1b4f 2458static inline void mc_part_std(H264Context *h, int n, int square, int chroma_height, int delta,
0da71265
MN
2459 uint8_t *dest_y, uint8_t *dest_cb, uint8_t *dest_cr,
2460 int x_offset, int y_offset,
2461 qpel_mc_func *qpix_put, h264_chroma_mc_func chroma_put,
2462 qpel_mc_func *qpix_avg, h264_chroma_mc_func chroma_avg,
2463 int list0, int list1){
2464 MpegEncContext * const s = &h->s;
2465 qpel_mc_func *qpix_op= qpix_put;
2466 h264_chroma_mc_func chroma_op= chroma_put;
115329f1 2467
5d18eaad
LM
2468 dest_y += 2*x_offset + 2*y_offset*h-> mb_linesize;
2469 dest_cb += x_offset + y_offset*h->mb_uvlinesize;
2470 dest_cr += x_offset + y_offset*h->mb_uvlinesize;
0da71265 2471 x_offset += 8*s->mb_x;
5d18eaad 2472 y_offset += 8*(s->mb_y >> MB_MBAFF);
115329f1 2473
0da71265 2474 if(list0){
1924f3ce 2475 Picture *ref= &h->ref_list[0][ h->ref_cache[0][ scan8[n] ] ];
0da71265
MN
2476 mc_dir_part(h, ref, n, square, chroma_height, delta, 0,
2477 dest_y, dest_cb, dest_cr, x_offset, y_offset,
2478 qpix_op, chroma_op);
2479
2480 qpix_op= qpix_avg;
2481 chroma_op= chroma_avg;
2482 }
2483
2484 if(list1){
1924f3ce 2485 Picture *ref= &h->ref_list[1][ h->ref_cache[1][ scan8[n] ] ];
0da71265
MN
2486 mc_dir_part(h, ref, n, square, chroma_height, delta, 1,
2487 dest_y, dest_cb, dest_cr, x_offset, y_offset,
2488 qpix_op, chroma_op);
2489 }
2490}
2491
9f2d1b4f
LM
2492static inline void mc_part_weighted(H264Context *h, int n, int square, int chroma_height, int delta,
2493 uint8_t *dest_y, uint8_t *dest_cb, uint8_t *dest_cr,
2494 int x_offset, int y_offset,
2495 qpel_mc_func *qpix_put, h264_chroma_mc_func chroma_put,
2496 h264_weight_func luma_weight_op, h264_weight_func chroma_weight_op,
2497 h264_biweight_func luma_weight_avg, h264_biweight_func chroma_weight_avg,
2498 int list0, int list1){
2499 MpegEncContext * const s = &h->s;
2500
5d18eaad
LM
2501 dest_y += 2*x_offset + 2*y_offset*h-> mb_linesize;
2502 dest_cb += x_offset + y_offset*h->mb_uvlinesize;
2503 dest_cr += x_offset + y_offset*h->mb_uvlinesize;
9f2d1b4f 2504 x_offset += 8*s->mb_x;
5d18eaad 2505 y_offset += 8*(s->mb_y >> MB_MBAFF);
115329f1 2506
9f2d1b4f
LM
2507 if(list0 && list1){
2508 /* don't optimize for luma-only case, since B-frames usually
2509 * use implicit weights => chroma too. */
2510 uint8_t *tmp_cb = s->obmc_scratchpad;
5d18eaad
LM
2511 uint8_t *tmp_cr = s->obmc_scratchpad + 8;
2512 uint8_t *tmp_y = s->obmc_scratchpad + 8*h->mb_uvlinesize;
9f2d1b4f
LM
2513 int refn0 = h->ref_cache[0][ scan8[n] ];
2514 int refn1 = h->ref_cache[1][ scan8[n] ];
2515
2516 mc_dir_part(h, &h->ref_list[0][refn0], n, square, chroma_height, delta, 0,
2517 dest_y, dest_cb, dest_cr,
2518 x_offset, y_offset, qpix_put, chroma_put);
2519 mc_dir_part(h, &h->ref_list[1][refn1], n, square, chroma_height, delta, 1,
2520 tmp_y, tmp_cb, tmp_cr,
2521 x_offset, y_offset, qpix_put, chroma_put);
2522
2523 if(h->use_weight == 2){
2524 int weight0 = h->implicit_weight[refn0][refn1];
2525 int weight1 = 64 - weight0;
5d18eaad
LM
2526 luma_weight_avg( dest_y, tmp_y, h-> mb_linesize, 5, weight0, weight1, 0);
2527 chroma_weight_avg(dest_cb, tmp_cb, h->mb_uvlinesize, 5, weight0, weight1, 0);
2528 chroma_weight_avg(dest_cr, tmp_cr, h->mb_uvlinesize, 5, weight0, weight1, 0);
9f2d1b4f 2529 }else{
5d18eaad 2530 luma_weight_avg(dest_y, tmp_y, h->mb_linesize, h->luma_log2_weight_denom,
115329f1 2531 h->luma_weight[0][refn0], h->luma_weight[1][refn1],
e8b56208 2532 h->luma_offset[0][refn0] + h->luma_offset[1][refn1]);
5d18eaad 2533 chroma_weight_avg(dest_cb, tmp_cb, h->mb_uvlinesize, h->chroma_log2_weight_denom,
115329f1 2534 h->chroma_weight[0][refn0][0], h->chroma_weight[1][refn1][0],
e8b56208 2535 h->chroma_offset[0][refn0][0] + h->chroma_offset[1][refn1][0]);
5d18eaad 2536 chroma_weight_avg(dest_cr, tmp_cr, h->mb_uvlinesize, h->chroma_log2_weight_denom,
115329f1 2537 h->chroma_weight[0][refn0][1], h->chroma_weight[1][refn1][1],
e8b56208 2538 h->chroma_offset[0][refn0][1] + h->chroma_offset[1][refn1][1]);
9f2d1b4f
LM
2539 }
2540 }else{
2541 int list = list1 ? 1 : 0;
2542 int refn = h->ref_cache[list][ scan8[n] ];
2543 Picture *ref= &h->ref_list[list][refn];
2544 mc_dir_part(h, ref, n, square, chroma_height, delta, list,
2545 dest_y, dest_cb, dest_cr, x_offset, y_offset,
2546 qpix_put, chroma_put);
2547
5d18eaad 2548 luma_weight_op(dest_y, h->mb_linesize, h->luma_log2_weight_denom,
9f2d1b4f
LM
2549 h->luma_weight[list][refn], h->luma_offset[list][refn]);
2550 if(h->use_weight_chroma){
5d18eaad 2551 chroma_weight_op(dest_cb, h->mb_uvlinesize, h->chroma_log2_weight_denom,
9f2d1b4f 2552 h->chroma_weight[list][refn][0], h->chroma_offset[list][refn][0]);
5d18eaad 2553 chroma_weight_op(dest_cr, h->mb_uvlinesize, h->chroma_log2_weight_denom,
9f2d1b4f
LM
2554 h->chroma_weight[list][refn][1], h->chroma_offset[list][refn][1]);
2555 }
2556 }
2557}
2558
2559static inline void mc_part(H264Context *h, int n, int square, int chroma_height, int delta,
2560 uint8_t *dest_y, uint8_t *dest_cb, uint8_t *dest_cr,
2561 int x_offset, int y_offset,
2562 qpel_mc_func *qpix_put, h264_chroma_mc_func chroma_put,
2563 qpel_mc_func *qpix_avg, h264_chroma_mc_func chroma_avg,
115329f1 2564 h264_weight_func *weight_op, h264_biweight_func *weight_avg,
9f2d1b4f
LM
2565 int list0, int list1){
2566 if((h->use_weight==2 && list0 && list1
2567 && (h->implicit_weight[ h->ref_cache[0][scan8[n]] ][ h->ref_cache[1][scan8[n]] ] != 32))
2568 || h->use_weight==1)
2569 mc_part_weighted(h, n, square, chroma_height, delta, dest_y, dest_cb, dest_cr,
2570 x_offset, y_offset, qpix_put, chroma_put,
2571 weight_op[0], weight_op[3], weight_avg[0], weight_avg[3], list0, list1);
2572 else
2573 mc_part_std(h, n, square, chroma_height, delta, dest_y, dest_cb, dest_cr,
2574 x_offset, y_offset, qpix_put, chroma_put, qpix_avg, chroma_avg, list0, list1);
2575}
2576
513fbd8e
LM
2577static inline void prefetch_motion(H264Context *h, int list){
2578 /* fetch pixels for estimated mv 4 macroblocks ahead
2579 * optimized for 64byte cache lines */
2580 MpegEncContext * const s = &h->s;
2581 const int refn = h->ref_cache[list][scan8[0]];
2582 if(refn >= 0){
2583 const int mx= (h->mv_cache[list][scan8[0]][0]>>2) + 16*s->mb_x + 8;
2584 const int my= (h->mv_cache[list][scan8[0]][1]>>2) + 16*s->mb_y;
2585 uint8_t **src= h->ref_list[list][refn].data;
5d18eaad 2586 int off= mx + (my + (s->mb_x&3)*4)*h->mb_linesize + 64;
513fbd8e
LM
2587 s->dsp.prefetch(src[0]+off, s->linesize, 4);
2588 off= (mx>>1) + ((my>>1) + (s->mb_x&7))*s->uvlinesize + 64;
2589 s->dsp.prefetch(src[1]+off, src[2]-src[1], 2);
2590 }
2591}
2592
0da71265
MN
2593static void hl_motion(H264Context *h, uint8_t *dest_y, uint8_t *dest_cb, uint8_t *dest_cr,
2594 qpel_mc_func (*qpix_put)[16], h264_chroma_mc_func (*chroma_put),
9f2d1b4f
LM
2595 qpel_mc_func (*qpix_avg)[16], h264_chroma_mc_func (*chroma_avg),
2596 h264_weight_func *weight_op, h264_biweight_func *weight_avg){
0da71265 2597 MpegEncContext * const s = &h->s;
7bc9090a 2598 const int mb_xy= s->mb_x + s->mb_y*s->mb_stride;
0da71265 2599 const int mb_type= s->current_picture.mb_type[mb_xy];
115329f1 2600
0da71265 2601 assert(IS_INTER(mb_type));
115329f1 2602
513fbd8e
LM
2603 prefetch_motion(h, 0);
2604
0da71265
MN
2605 if(IS_16X16(mb_type)){
2606 mc_part(h, 0, 1, 8, 0, dest_y, dest_cb, dest_cr, 0, 0,
2607 qpix_put[0], chroma_put[0], qpix_avg[0], chroma_avg[0],
9f2d1b4f 2608 &weight_op[0], &weight_avg[0],
0da71265
MN
2609 IS_DIR(mb_type, 0, 0), IS_DIR(mb_type, 0, 1));
2610 }else if(IS_16X8(mb_type)){
2611 mc_part(h, 0, 0, 4, 8, dest_y, dest_cb, dest_cr, 0, 0,
2612 qpix_put[1], chroma_put[0], qpix_avg[1], chroma_avg[0],
9f2d1b4f 2613 &weight_op[1], &weight_avg[1],
0da71265
MN
2614 IS_DIR(mb_type, 0, 0), IS_DIR(mb_type, 0, 1));
2615 mc_part(h, 8, 0, 4, 8, dest_y, dest_cb, dest_cr, 0, 4,
2616 qpix_put[1], chroma_put[0], qpix_avg[1], chroma_avg[0],
9f2d1b4f 2617 &weight_op[1], &weight_avg[1],
0da71265
MN
2618 IS_DIR(mb_type, 1, 0), IS_DIR(mb_type, 1, 1));
2619 }else if(IS_8X16(mb_type)){
5d18eaad 2620 mc_part(h, 0, 0, 8, 8*h->mb_linesize, dest_y, dest_cb, dest_cr, 0, 0,
0da71265 2621 qpix_put[1], chroma_put[1], qpix_avg[1], chroma_avg[1],
9f2d1b4f 2622 &weight_op[2], &weight_avg[2],
0da71265 2623 IS_DIR(mb_type, 0, 0), IS_DIR(mb_type, 0, 1));
5d18eaad 2624 mc_part(h, 4, 0, 8, 8*h->mb_linesize, dest_y, dest_cb, dest_cr, 4, 0,
0da71265 2625 qpix_put[1], chroma_put[1], qpix_avg[1], chroma_avg[1],
9f2d1b4f 2626 &weight_op[2], &weight_avg[2],
0da71265
MN
2627 IS_DIR(mb_type, 1, 0), IS_DIR(mb_type, 1, 1));
2628 }else{
2629 int i;
115329f1 2630
0da71265
MN
2631 assert(IS_8X8(mb_type));
2632
2633 for(i=0; i<4; i++){
2634 const int sub_mb_type= h->sub_mb_type[i];
2635 const int n= 4*i;
2636 int x_offset= (i&1)<<2;
2637 int y_offset= (i&2)<<1;
2638
2639 if(IS_SUB_8X8(sub_mb_type)){
2640 mc_part(h, n, 1, 4, 0, dest_y, dest_cb, dest_cr, x_offset, y_offset,
2641 qpix_put[1], chroma_put[1], qpix_avg[1], chroma_avg[1],
9f2d1b4f 2642 &weight_op[3], &weight_avg[3],
0da71265
MN
2643 IS_DIR(sub_mb_type, 0, 0), IS_DIR(sub_mb_type, 0, 1));
2644 }else if(IS_SUB_8X4(sub_mb_type)){
2645 mc_part(h, n , 0, 2, 4, dest_y, dest_cb, dest_cr, x_offset, y_offset,
2646 qpix_put[2], chroma_put[1], qpix_avg[2], chroma_avg[1],
9f2d1b4f 2647 &weight_op[4], &weight_avg[4],
0da71265
MN
2648 IS_DIR(sub_mb_type, 0, 0), IS_DIR(sub_mb_type, 0, 1));
2649 mc_part(h, n+2, 0, 2, 4, dest_y, dest_cb, dest_cr, x_offset, y_offset+2,
2650 qpix_put[2], chroma_put[1], qpix_avg[2], chroma_avg[1],
9f2d1b4f 2651 &weight_op[4], &weight_avg[4],
0da71265
MN
2652 IS_DIR(sub_mb_type, 0, 0), IS_DIR(sub_mb_type, 0, 1));
2653 }else if(IS_SUB_4X8(sub_mb_type)){
5d18eaad 2654 mc_part(h, n , 0, 4, 4*h->mb_linesize, dest_y, dest_cb, dest_cr, x_offset, y_offset,
0da71265 2655 qpix_put[2], chroma_put[2], qpix_avg[2], chroma_avg[2],
9f2d1b4f 2656 &weight_op[5], &weight_avg[5],
0da71265 2657 IS_DIR(sub_mb_type, 0, 0), IS_DIR(sub_mb_type, 0, 1));
5d18eaad 2658 mc_part(h, n+1, 0, 4, 4*h->mb_linesize, dest_y, dest_cb, dest_cr, x_offset+2, y_offset,
0da71265 2659 qpix_put[2], chroma_put[2], qpix_avg[2], chroma_avg[2],
9f2d1b4f 2660 &weight_op[5], &weight_avg[5],
0da71265
MN
2661 IS_DIR(sub_mb_type, 0, 0), IS_DIR(sub_mb_type, 0, 1));
2662 }else{
2663 int j;
2664 assert(IS_SUB_4X4(sub_mb_type));
2665 for(j=0; j<4; j++){
2666 int sub_x_offset= x_offset + 2*(j&1);
2667 int sub_y_offset= y_offset + (j&2);
2668 mc_part(h, n+j, 1, 2, 0, dest_y, dest_cb, dest_cr, sub_x_offset, sub_y_offset,
2669 qpix_put[2], chroma_put[2], qpix_avg[2], chroma_avg[2],
9f2d1b4f 2670 &weight_op[6], &weight_avg[6],
0da71265
MN
2671 IS_DIR(sub_mb_type, 0, 0), IS_DIR(sub_mb_type, 0, 1));
2672 }
2673 }
2674 }
2675 }
513fbd8e
LM
2676
2677 prefetch_motion(h, 1);
0da71265
MN
2678}
2679
2b100ab2 2680static void decode_init_vlc(void){
0da71265
MN
2681 static int done = 0;
2682
2683 if (!done) {
2684 int i;
2685 done = 1;
2686
115329f1 2687 init_vlc(&chroma_dc_coeff_token_vlc, CHROMA_DC_COEFF_TOKEN_VLC_BITS, 4*5,
0da71265 2688 &chroma_dc_coeff_token_len [0], 1, 1,
073c2593 2689 &chroma_dc_coeff_token_bits[0], 1, 1, 1);
0da71265
MN
2690
2691 for(i=0; i<4; i++){
115329f1 2692 init_vlc(&coeff_token_vlc[i], COEFF_TOKEN_VLC_BITS, 4*17,
0da71265 2693 &coeff_token_len [i][0], 1, 1,
073c2593 2694 &coeff_token_bits[i][0], 1, 1, 1);
0da71265
MN
2695 }
2696
2697 for(i=0; i<3; i++){
2698 init_vlc(&chroma_dc_total_zeros_vlc[i], CHROMA_DC_TOTAL_ZEROS_VLC_BITS, 4,
2699 &chroma_dc_total_zeros_len [i][0], 1, 1,
073c2593 2700 &chroma_dc_total_zeros_bits[i][0], 1, 1, 1);
0da71265
MN
2701 }
2702 for(i=0; i<15; i++){
115329f1 2703 init_vlc(&total_zeros_vlc[i], TOTAL_ZEROS_VLC_BITS, 16,
0da71265 2704 &total_zeros_len [i][0], 1, 1,
073c2593 2705 &total_zeros_bits[i][0], 1, 1, 1);
0da71265
MN
2706 }
2707
2708 for(i=0; i<6; i++){
115329f1 2709 init_vlc(&run_vlc[i], RUN_VLC_BITS, 7,
0da71265 2710 &run_len [i][0], 1, 1,
073c2593 2711 &run_bits[i][0], 1, 1, 1);
0da71265 2712 }
115329f1 2713 init_vlc(&run7_vlc, RUN7_VLC_BITS, 16,
0da71265 2714 &run_len [6][0], 1, 1,
073c2593 2715 &run_bits[6][0], 1, 1, 1);
0da71265
MN
2716 }
2717}
2718
2719/**
2720 * Sets the intra prediction function pointers.
2721 */
2722static void init_pred_ptrs(H264Context *h){
2723// MpegEncContext * const s = &h->s;
2724
2725 h->pred4x4[VERT_PRED ]= pred4x4_vertical_c;
2726 h->pred4x4[HOR_PRED ]= pred4x4_horizontal_c;
2727 h->pred4x4[DC_PRED ]= pred4x4_dc_c;
2728 h->pred4x4[DIAG_DOWN_LEFT_PRED ]= pred4x4_down_left_c;
2729 h->pred4x4[DIAG_DOWN_RIGHT_PRED]= pred4x4_down_right_c;
2730 h->pred4x4[VERT_RIGHT_PRED ]= pred4x4_vertical_right_c;
2731 h->pred4x4[HOR_DOWN_PRED ]= pred4x4_horizontal_down_c;
2732 h->pred4x4[VERT_LEFT_PRED ]= pred4x4_vertical_left_c;
2733 h->pred4x4[HOR_UP_PRED ]= pred4x4_horizontal_up_c;
2734 h->pred4x4[LEFT_DC_PRED ]= pred4x4_left_dc_c;
2735 h->pred4x4[TOP_DC_PRED ]= pred4x4_top_dc_c;
2736 h->pred4x4[DC_128_PRED ]= pred4x4_128_dc_c;
2737
43efd19a
LM
2738 h->pred8x8l[VERT_PRED ]= pred8x8l_vertical_c;
2739 h->pred8x8l[HOR_PRED ]= pred8x8l_horizontal_c;
2740 h->pred8x8l[DC_PRED ]= pred8x8l_dc_c;
2741 h->pred8x8l[DIAG_DOWN_LEFT_PRED ]= pred8x8l_down_left_c;
2742 h->pred8x8l[DIAG_DOWN_RIGHT_PRED]= pred8x8l_down_right_c;
2743 h->pred8x8l[VERT_RIGHT_PRED ]= pred8x8l_vertical_right_c;
2744 h->pred8x8l[HOR_DOWN_PRED ]= pred8x8l_horizontal_down_c;
2745 h->pred8x8l[VERT_LEFT_PRED ]= pred8x8l_vertical_left_c;
2746 h->pred8x8l[HOR_UP_PRED ]= pred8x8l_horizontal_up_c;
2747 h->pred8x8l[LEFT_DC_PRED ]= pred8x8l_left_dc_c;
2748 h->pred8x8l[TOP_DC_PRED ]= pred8x8l_top_dc_c;
2749 h->pred8x8l[DC_128_PRED ]= pred8x8l_128_dc_c;
2750
a074f143
PI
2751 h->pred8x8[DC_PRED8x8 ]= ff_pred8x8_dc_c;
2752 h->pred8x8[VERT_PRED8x8 ]= ff_pred8x8_vertical_c;
2753 h->pred8x8[HOR_PRED8x8 ]= ff_pred8x8_horizontal_c;
2754 h->pred8x8[PLANE_PRED8x8 ]= ff_pred8x8_plane_c;
cb1a74cf
PI
2755 h->pred8x8[LEFT_DC_PRED8x8]= ff_pred8x8_left_dc_c;
2756 h->pred8x8[TOP_DC_PRED8x8 ]= ff_pred8x8_top_dc_c;
a074f143 2757 h->pred8x8[DC_128_PRED8x8 ]= ff_pred8x8_128_dc_c;
0da71265 2758
a074f143
PI
2759 h->pred16x16[DC_PRED8x8 ]= ff_pred16x16_dc_c;
2760 h->pred16x16[VERT_PRED8x8 ]= ff_pred16x16_vertical_c;
2761 h->pred16x16[HOR_PRED8x8 ]= ff_pred16x16_horizontal_c;
2762 h->pred16x16[PLANE_PRED8x8 ]= ff_pred16x16_plane_c;
cb1a74cf
PI
2763 h->pred16x16[LEFT_DC_PRED8x8]= ff_pred16x16_left_dc_c;
2764 h->pred16x16[TOP_DC_PRED8x8 ]= ff_pred16x16_top_dc_c;
a074f143 2765 h->pred16x16[DC_128_PRED8x8 ]= ff_pred16x16_128_dc_c;
0da71265
MN
2766}
2767
0da71265 2768static void free_tables(H264Context *h){
7978debd 2769 int i;
0da71265 2770 av_freep(&h->intra4x4_pred_mode);
e5017ab8
LA
2771 av_freep(&h->chroma_pred_mode_table);
2772 av_freep(&h->cbp_table);
9e528114
LA
2773 av_freep(&h->mvd_table[0]);
2774 av_freep(&h->mvd_table[1]);
5ad984c9 2775 av_freep(&h->direct_table);
0da71265
MN
2776 av_freep(&h->non_zero_count);
2777 av_freep(&h->slice_table_base);
6ba71fc4
LLL
2778 av_freep(&h->top_borders[1]);
2779 av_freep(&h->top_borders[0]);
0da71265 2780 h->slice_table= NULL;
e5017ab8 2781
0da71265
MN
2782 av_freep(&h->mb2b_xy);
2783 av_freep(&h->mb2b8_xy);
9f2d1b4f
LM
2784
2785 av_freep(&h->s.obmc_scratchpad);
7978debd
2786
2787 for(i = 0; i < MAX_SPS_COUNT; i++)
2788 av_freep(h->sps_buffers + i);
2789
2790 for(i = 0; i < MAX_PPS_COUNT; i++)
2791 av_freep(h->pps_buffers + i);
0da71265
MN
2792}
2793
239ea04c
LM
2794static void init_dequant8_coeff_table(H264Context *h){
2795 int i,q,x;
548a1c8a 2796 const int transpose = (h->s.dsp.h264_idct8_add != ff_h264_idct8_add_c); //FIXME ugly
239ea04c
LM
2797 h->dequant8_coeff[0] = h->dequant8_buffer[0];
2798 h->dequant8_coeff[1] = h->dequant8_buffer[1];
2799
2800 for(i=0; i<2; i++ ){
2801 if(i && !memcmp(h->pps.scaling_matrix8[0], h->pps.scaling_matrix8[1], 64*sizeof(uint8_t))){
2802 h->dequant8_coeff[1] = h->dequant8_buffer[0];
2803 break;
2804 }
2805
2806 for(q=0; q<52; q++){
acd8d10f
PI
2807 int shift = ff_div6[q];
2808 int idx = ff_rem6[q];
239ea04c 2809 for(x=0; x<64; x++)
548a1c8a
LM
2810 h->dequant8_coeff[i][q][transpose ? (x>>3)|((x&7)<<3) : x] =
2811 ((uint32_t)dequant8_coeff_init[idx][ dequant8_coeff_init_scan[((x>>1)&12) | (x&3)] ] *
2812 h->pps.scaling_matrix8[i][x]) << shift;
239ea04c
LM
2813 }
2814 }
2815}
2816
2817static void init_dequant4_coeff_table(H264Context *h){
2818 int i,j,q,x;
ab2e3e2c 2819 const int transpose = (h->s.dsp.h264_idct_add != ff_h264_idct_add_c); //FIXME ugly
239ea04c
LM
2820 for(i=0; i<6; i++ ){
2821 h->dequant4_coeff[i] = h->dequant4_buffer[i];
2822 for(j=0; j<i; j++){
2823 if(!memcmp(h->pps.scaling_matrix4[j], h->pps.scaling_matrix4[i], 16*sizeof(uint8_t))){
2824 h->dequant4_coeff[i] = h->dequant4_buffer[j];
2825 break;
2826 }
2827 }
2828 if(j<i)
2829 continue;
2830
2831 for(q=0; q<52; q++){
acd8d10f
PI
2832 int shift = ff_div6[q] + 2;
2833 int idx = ff_rem6[q];
239ea04c 2834 for(x=0; x<16; x++)
ab2e3e2c
LM
2835 h->dequant4_coeff[i][q][transpose ? (x>>2)|((x<<2)&0xF) : x] =
2836 ((uint32_t)dequant4_coeff_init[idx][(x&1) + ((x>>2)&1)] *
239ea04c
LM
2837 h->pps.scaling_matrix4[i][x]) << shift;
2838 }
2839 }
2840}
2841
2842static void init_dequant_tables(H264Context *h){
2843 int i,x;
2844 init_dequant4_coeff_table(h);
2845 if(h->pps.transform_8x8_mode)
2846 init_dequant8_coeff_table(h);
2847 if(h->sps.transform_bypass){
2848 for(i=0; i<6; i++)
2849 for(x=0; x<16; x++)
2850 h->dequant4_coeff[i][0][x] = 1<<6;
2851 if(h->pps.transform_8x8_mode)
2852 for(i=0; i<2; i++)
2853 for(x=0; x<64; x++)
2854 h->dequant8_coeff[i][0][x] = 1<<6;
2855 }
2856}
2857
2858
0da71265
MN
2859/**
2860 * allocates tables.
3b66c4c5 2861 * needs width/height
0da71265
MN
2862 */
2863static int alloc_tables(H264Context *h){
2864 MpegEncContext * const s = &h->s;
7bc9090a 2865 const int big_mb_num= s->mb_stride * (s->mb_height+1);
239ea04c 2866 int x,y;
0da71265
MN
2867
2868 CHECKED_ALLOCZ(h->intra4x4_pred_mode, big_mb_num * 8 * sizeof(uint8_t))
e5017ab8 2869
53c05b1e 2870 CHECKED_ALLOCZ(h->non_zero_count , big_mb_num * 16 * sizeof(uint8_t))
5d18eaad 2871 CHECKED_ALLOCZ(h->slice_table_base , (big_mb_num+s->mb_stride) * sizeof(uint8_t))
6ba71fc4
LLL
2872 CHECKED_ALLOCZ(h->top_borders[0] , s->mb_width * (16+8+8) * sizeof(uint8_t))
2873 CHECKED_ALLOCZ(h->top_borders[1] , s->mb_width * (16+8+8) * sizeof(uint8_t))
5d0e4cb8 2874 CHECKED_ALLOCZ(h->cbp_table, big_mb_num * sizeof(uint16_t))
0da71265 2875
e5017ab8
LA
2876 if( h->pps.cabac ) {
2877 CHECKED_ALLOCZ(h->chroma_pred_mode_table, big_mb_num * sizeof(uint8_t))
9e528114
LA
2878 CHECKED_ALLOCZ(h->mvd_table[0], 32*big_mb_num * sizeof(uint16_t));
2879 CHECKED_ALLOCZ(h->mvd_table[1], 32*big_mb_num * sizeof(uint16_t));
5ad984c9 2880 CHECKED_ALLOCZ(h->direct_table, 32*big_mb_num * sizeof(uint8_t));
e5017ab8
LA
2881 }
2882
5d18eaad
LM
2883 memset(h->slice_table_base, -1, (big_mb_num+s->mb_stride) * sizeof(uint8_t));
2884 h->slice_table= h->slice_table_base + s->mb_stride*2 + 1;
0da71265 2885
a55f20bd
LM
2886 CHECKED_ALLOCZ(h->mb2b_xy , big_mb_num * sizeof(uint32_t));
2887 CHECKED_ALLOCZ(h->mb2b8_xy , big_mb_num * sizeof(uint32_t));
0da71265
MN
2888 for(y=0; y<s->mb_height; y++){
2889 for(x=0; x<s->mb_width; x++){
7bc9090a 2890 const int mb_xy= x + y*s->mb_stride;
0da71265
MN
2891 const int b_xy = 4*x + 4*y*h->b_stride;
2892 const int b8_xy= 2*x + 2*y*h->b8_stride;
115329f1 2893
0da71265
MN
2894 h->mb2b_xy [mb_xy]= b_xy;
2895 h->mb2b8_xy[mb_xy]= b8_xy;
2896 }
2897 }
9f2d1b4f 2898
9c6221ae
GV
2899 s->obmc_scratchpad = NULL;
2900
56edbd81
LM
2901 if(!h->dequant4_coeff[0])
2902 init_dequant_tables(h);
2903
0da71265
MN
2904 return 0;
2905fail:
2906 free_tables(h);
2907 return -1;
2908}
2909
2910static void common_init(H264Context *h){
2911 MpegEncContext * const s = &h->s;
0da71265
MN
2912
2913 s->width = s->avctx->width;
2914 s->height = s->avctx->height;
2915 s->codec_id= s->avctx->codec->id;
115329f1 2916
0da71265
MN
2917 init_pred_ptrs(h);
2918
239ea04c 2919 h->dequant_coeff_pps= -1;
9a41c2c7 2920 s->unrestricted_mv=1;
0da71265 2921 s->decode=1; //FIXME
56edbd81
LM
2922
2923 memset(h->pps.scaling_matrix4, 16, 6*16*sizeof(uint8_t));
2924 memset(h->pps.scaling_matrix8, 16, 2*64*sizeof(uint8_t));
0da71265
MN
2925}
2926
2927static int decode_init(AVCodecContext *avctx){
2928 H264Context *h= avctx->priv_data;
2929 MpegEncContext * const s = &h->s;
2930
3edcacde 2931 MPV_decode_defaults(s);
115329f1 2932
0da71265
MN
2933 s->avctx = avctx;
2934 common_init(h);
2935
2936 s->out_format = FMT_H264;
2937 s->workaround_bugs= avctx->workaround_bugs;
2938
2939 // set defaults
0da71265
MN
2940// s->decode_mb= ff_h263_decode_mb;
2941 s->low_delay= 1;
2942 avctx->pix_fmt= PIX_FMT_YUV420P;
2943
c2212338 2944 decode_init_vlc();
115329f1 2945
26165f99
MR
2946 if(avctx->extradata_size > 0 && avctx->extradata &&
2947 *(char *)avctx->extradata == 1){
4770b1b4
RT
2948 h->is_avc = 1;
2949 h->got_avcC = 0;
26165f99
MR
2950 } else {
2951 h->is_avc = 0;
4770b1b4
RT
2952 }
2953
0da71265
MN
2954 return 0;
2955}
2956
af8aa846 2957static int frame_start(H264Context *h){
0da71265
MN
2958 MpegEncContext * const s = &h->s;
2959 int i;
2960
af8aa846
MN
2961 if(MPV_frame_start(s, s->avctx) < 0)
2962 return -1;
0da71265 2963 ff_er_frame_start(s);
0da71265
MN
2964
2965 assert(s->linesize && s->uvlinesize);
2966
2967 for(i=0; i<16; i++){
2968 h->block_offset[i]= 4*((scan8[i] - scan8[0])&7) + 4*s->linesize*((scan8[i] - scan8[0])>>3);
6867a90b 2969 h->block_offset[24+i]= 4*((scan8[i] - scan8[0])&7) + 8*s->linesize*((scan8[i] - scan8[0])>>3);
0da71265
MN
2970 }
2971 for(i=0; i<4; i++){
2972 h->block_offset[16+i]=
2973 h->block_offset[20+i]= 4*((scan8[i] - scan8[0])&7) + 4*s->uvlinesize*((scan8[i] - scan8[0])>>3);
6867a90b
LLL
2974 h->block_offset[24+16+i]=
2975 h->block_offset[24+20+i]= 4*((scan8[i] - scan8[0])&7) + 8*s->uvlinesize*((scan8[i] - scan8[0])>>3);
0da71265
MN
2976 }
2977
934b0821
LM
2978 /* can't be in alloc_tables because linesize isn't known there.
2979 * FIXME: redo bipred weight to not require extra buffer? */
2980 if(!s->obmc_scratchpad)
5d18eaad
LM
2981 s->obmc_scratchpad = av_malloc(16*2*s->linesize + 8*2*s->uvlinesize);
2982
2983 /* some macroblocks will be accessed before they're available */
2984 if(FRAME_MBAFF)
2985 memset(h->slice_table, -1, (s->mb_height*s->mb_stride-1) * sizeof(uint8_t));
934b0821 2986
0da71265 2987// s->decode= (s->flags&CODEC_FLAG_PSNR) || !s->encoding || s->current_picture.reference /*|| h->contains_intra*/ || 1;
af8aa846 2988 return 0;
0da71265
MN
2989}
2990
93cc10fa 2991static inline void backup_mb_border(H264Context *h, uint8_t *src_y, uint8_t *src_cb, uint8_t *src_cr, int linesize, int uvlinesize, int simple){
53c05b1e
MN
2992 MpegEncContext * const s = &h->s;
2993 int i;
115329f1 2994
53c05b1e
MN
2995 src_y -= linesize;
2996 src_cb -= uvlinesize;
2997 src_cr -= uvlinesize;
2998
3b66c4c5 2999 // There are two lines saved, the line above the the top macroblock of a pair,
6867a90b 3000 // and the line above the bottom macroblock
6ba71fc4 3001 h->left_border[0]= h->top_borders[0][s->mb_x][15];
53c05b1e
MN
3002 for(i=1; i<17; i++){
3003 h->left_border[i]= src_y[15+i* linesize];
3004 }
115329f1 3005
6ba71fc4
LLL
3006 *(uint64_t*)(h->top_borders[0][s->mb_x]+0)= *(uint64_t*)(src_y + 16*linesize);
3007 *(uint64_t*)(h->top_borders[0][s->mb_x]+8)= *(uint64_t*)(src_y +8+16*linesize);
53c05b1e 3008
93cc10fa 3009 if(simple || !(s->flags&CODEC_FLAG_GRAY)){
6ba71fc4
LLL
3010 h->left_border[17 ]= h->top_borders[0][s->mb_x][16+7];
3011 h->left_border[17+9]= h->top_borders[0][s->mb_x][24+7];
53c05b1e
MN
3012 for(i=1; i<9; i++){
3013 h->left_border[i+17 ]= src_cb[7+i*uvlinesize];
3014 h->left_border[i+17+9]= src_cr[7+i*uvlinesize];
3015 }
6ba71fc4
LLL
3016 *(uint64_t*)(h->top_borders[0][s->mb_x]+16)= *(uint64_t*)(src_cb+8*uvlinesize);
3017 *(uint64_t*)(h->top_borders[0][s->mb_x]+24)= *(uint64_t*)(src_cr+8*uvlinesize);
53c05b1e
MN
3018 }
3019}
3020
93cc10fa 3021static inline void xchg_mb_border(H264Context *h, uint8_t *src_y, uint8_t *src_cb, uint8_t *src_cr, int linesize, int uvlinesize, int xchg, int simple){
53c05b1e
MN
3022 MpegEncContext * const s = &h->s;
3023 int temp8, i;
3024 uint64_t temp64;
d89dc06a
LM
3025 int deblock_left = (s->mb_x > 0);
3026 int deblock_top = (s->mb_y > 0);
53c05b1e
MN
3027
3028 src_y -= linesize + 1;
3029 src_cb -= uvlinesize + 1;
3030 src_cr -= uvlinesize + 1;
3031
3032#define XCHG(a,b,t,xchg)\
3033t= a;\
3034if(xchg)\
3035 a= b;\
3036b= t;
d89dc06a
LM
3037
3038 if(deblock_left){
3039 for(i = !deblock_top; i<17; i++){
3040 XCHG(h->left_border[i ], src_y [i* linesize], temp8, xchg);
3041 }
3042 }
3043
3044 if(deblock_top){
6ba71fc4
LLL
3045 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+0), *(uint64_t*)(src_y +1), temp64, xchg);
3046 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+8), *(uint64_t*)(src_y +9), temp64, 1);
cad4368a 3047 if(s->mb_x+1 < s->mb_width){
43efd19a
LM
3048 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x+1]), *(uint64_t*)(src_y +17), temp64, 1);
3049 }
53c05b1e 3050 }
53c05b1e 3051
93cc10fa 3052 if(simple || !(s->flags&CODEC_FLAG_GRAY)){
d89dc06a
LM
3053 if(deblock_left){
3054 for(i = !deblock_top; i<9; i++){
3055 XCHG(h->left_border[i+17 ], src_cb[i*uvlinesize], temp8, xchg);
3056 XCHG(h->left_border[i+17+9], src_cr[i*uvlinesize], temp8, xchg);
3057 }
3058 }
3059 if(deblock_top){
6ba71fc4
LLL
3060 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+16), *(uint64_t*)(src_cb+1), temp64, 1);
3061 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+24), *(uint64_t*)(src_cr+1), temp64, 1);
3062 }
3063 }
3064}
3065
3066static inline void backup_pair_border(H264Context *h, uint8_t *src_y, uint8_t *src_cb, uint8_t *src_cr, int linesize, int uvlinesize){
3067 MpegEncContext * const s = &h->s;
3068 int i;
115329f1 3069
6ba71fc4
LLL
3070 src_y -= 2 * linesize;
3071 src_cb -= 2 * uvlinesize;
3072 src_cr -= 2 * uvlinesize;
3073
3b66c4c5 3074 // There are two lines saved, the line above the the top macroblock of a pair,
6ba71fc4
LLL
3075 // and the line above the bottom macroblock
3076 h->left_border[0]= h->top_borders[0][s->mb_x][15];
3077 h->left_border[1]= h->top_borders[1][s->mb_x][15];
3078 for(i=2; i<34; i++){
3079 h->left_border[i]= src_y[15+i* linesize];
3080 }
115329f1 3081
6ba71fc4
LLL
3082 *(uint64_t*)(h->top_borders[0][s->mb_x]+0)= *(uint64_t*)(src_y + 32*linesize);
3083 *(uint64_t*)(h->top_borders[0][s->mb_x]+8)= *(uint64_t*)(src_y +8+32*linesize);
3084 *(uint64_t*)(h->top_borders[1][s->mb_x]+0)= *(uint64_t*)(src_y + 33*linesize);
3085 *(uint64_t*)(h->top_borders[1][s->mb_x]+8)= *(uint64_t*)(src_y +8+33*linesize);
3086
3087 if(!(s->flags&CODEC_FLAG_GRAY)){
3088 h->left_border[34 ]= h->top_borders[0][s->mb_x][16+7];
3089 h->left_border[34+ 1]= h->top_borders[1][s->mb_x][16+7];
3090 h->left_border[34+18 ]= h->top_borders[0][s->mb_x][24+7];
3091 h->left_border[34+18+1]= h->top_borders[1][s->mb_x][24+7];
3092 for(i=2; i<18; i++){
3093 h->left_border[i+34 ]= src_cb[7+i*uvlinesize];
3094 h->left_border[i+34+18]= src_cr[7+i*uvlinesize];
3095 }
3096 *(uint64_t*)(h->top_borders[0][s->mb_x]+16)= *(uint64_t*)(src_cb+16*uvlinesize);
3097 *(uint64_t*)(h->top_borders[0][s->mb_x]+24)= *(uint64_t*)(src_cr+16*uvlinesize);
3098 *(uint64_t*)(h->top_borders[1][s->mb_x]+16)= *(uint64_t*)(src_cb+17*uvlinesize);
3099 *(uint64_t*)(h->top_borders[1][s->mb_x]+24)= *(uint64_t*)(src_cr+17*uvlinesize);
3100 }
3101}
3102
3103static inline void xchg_pair_border(H264Context *h, uint8_t *src_y, uint8_t *src_cb, uint8_t *src_cr, int linesize, int uvlinesize, int xchg){
3104 MpegEncContext * const s = &h->s;
3105 int temp8, i;
3106 uint64_t temp64;
3107 int deblock_left = (s->mb_x > 0);
5d18eaad 3108 int deblock_top = (s->mb_y > 1);
6ba71fc4 3109
a9c9a240 3110 tprintf(s->avctx, "xchg_pair_border: src_y:%p src_cb:%p src_cr:%p ls:%d uvls:%d\n", src_y, src_cb, src_cr, linesize, uvlinesize);
6ba71fc4
LLL
3111
3112 src_y -= 2 * linesize + 1;
3113 src_cb -= 2 * uvlinesize + 1;
3114 src_cr -= 2 * uvlinesize + 1;
3115
3116#define XCHG(a,b,t,xchg)\
3117t= a;\
3118if(xchg)\
3119 a= b;\
3120b= t;
3121
3122 if(deblock_left){
3123 for(i = (!deblock_top)<<1; i<34; i++){
3124 XCHG(h->left_border[i ], src_y [i* linesize], temp8, xchg);
3125 }
3126 }
3127
3128 if(deblock_top){
3129 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+0), *(uint64_t*)(src_y +1), temp64, xchg);
3130 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+8), *(uint64_t*)(src_y +9), temp64, 1);
3131 XCHG(*(uint64_t*)(h->top_borders[1][s->mb_x]+0), *(uint64_t*)(src_y +1 +linesize), temp64, xchg);
3132 XCHG(*(uint64_t*)(h->top_borders[1][s->mb_x]+8), *(uint64_t*)(src_y +9 +linesize), temp64, 1);
5d18eaad
LM
3133 if(s->mb_x+1 < s->mb_width){
3134 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x+1]), *(uint64_t*)(src_y +17), temp64, 1);
3135 XCHG(*(uint64_t*)(h->top_borders[1][s->mb_x+1]), *(uint64_t*)(src_y +17 +linesize), temp64, 1);
3136 }
6ba71fc4
LLL
3137 }
3138
3139 if(!(s->flags&CODEC_FLAG_GRAY)){
3140 if(deblock_left){
3141 for(i = (!deblock_top) << 1; i<18; i++){
3142 XCHG(h->left_border[i+34 ], src_cb[i*uvlinesize], temp8, xchg);
3143 XCHG(h->left_border[i+34+18], src_cr[i*uvlinesize], temp8, xchg);
3144 }
3145 }
3146 if(deblock_top){
3147 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+16), *(uint64_t*)(src_cb+1), temp64, 1);
3148 XCHG(*(uint64_t*)(h->top_borders[0][s->mb_x]+24), *(uint64_t*)(src_cr+1), temp64, 1);
3149 XCHG(*(uint64_t*)(h->top_borders[1][s->mb_x]+16), *(uint64_t*)(src_cb+1 +uvlinesize), temp64, 1);
3150 XCHG(*(uint64_t*)(h->top_borders[1][s->mb_x]+24), *(uint64_t*)(src_cr+1 +uvlinesize), temp64, 1);
53c05b1e 3151 }
53c05b1e
MN
3152 }
3153}
3154
5a6a6cc7 3155static av_always_inline void hl_decode_mb_internal(H264Context *h, int simple){
0da71265
MN
3156 MpegEncContext * const s = &h->s;
3157 const int mb_x= s->mb_x;
3158 const int mb_y= s->mb_y;
7bc9090a 3159 const int mb_xy= mb_x + mb_y*s->mb_stride;
0da71265
MN
3160 const int mb_type= s->current_picture.mb_type[mb_xy];
3161 uint8_t *dest_y, *dest_cb, *dest_cr;
3162 int linesize, uvlinesize /*dct_offset*/;
3163 int i;
6867a90b 3164 int *block_offset = &h->block_offset[0];
6ba71fc4 3165 const unsigned int bottom = mb_y & 1;
bd91fee3 3166 const int transform_bypass = (s->qscale == 0 && h->sps.transform_bypass), is_h264 = (simple || s->codec_id == CODEC_ID_H264);
36940eca 3167 void (*idct_add)(uint8_t *dst, DCTELEM *block, int stride);
ef9d1d15 3168 void (*idct_dc_add)(uint8_t *dst, DCTELEM *block, int stride);
0da71265 3169
0da71265
MN
3170 dest_y = s->current_picture.data[0] + (mb_y * 16* s->linesize ) + mb_x * 16;
3171 dest_cb = s->current_picture.data[1] + (mb_y * 8 * s->uvlinesize) + mb_x * 8;
3172 dest_cr = s->current_picture.data[2] + (mb_y * 8 * s->uvlinesize) + mb_x * 8;
3173
a957c27b
LM
3174 s->dsp.prefetch(dest_y + (s->mb_x&3)*4*s->linesize + 64, s->linesize, 4);
3175 s->dsp.prefetch(dest_cb + (s->mb_x&7)*s->uvlinesize + 64, dest_cr - dest_cb, 2);
3176
bd91fee3 3177 if (!simple && MB_FIELD) {
5d18eaad
LM
3178 linesize = h->mb_linesize = s->linesize * 2;
3179 uvlinesize = h->mb_uvlinesize = s->uvlinesize * 2;
6867a90b 3180 block_offset = &h->block_offset[24];
0da71265
MN
3181 if(mb_y&1){ //FIXME move out of this func?
3182 dest_y -= s->linesize*15;
6867a90b
LLL
3183 dest_cb-= s->uvlinesize*7;
3184 dest_cr-= s->uvlinesize*7;
0da71265 3185 }
5d18eaad
LM
3186 if(FRAME_MBAFF) {
3187 int list;
3425501d 3188 for(list=0; list<h->list_count; list++){
5d18eaad
LM
3189 if(!USES_LIST(mb_type, list))
3190 continue;
3191 if(IS_16X16(mb_type)){
3192 int8_t *ref = &h->ref_cache[list][scan8[0]];
3193 fill_rectangle(ref, 4, 4, 8, 16+*ref^(s->mb_y&1), 1);
3194 }else{
3195 for(i=0; i<16; i+=4){
3196 //FIXME can refs be smaller than 8x8 when !direct_8x8_inference ?
3197 int ref = h->ref_cache[list][scan8[i]];
3198 if(ref >= 0)
3199 fill_rectangle(&h->ref_cache[list][scan8[i]], 2, 2, 8, 16+ref^(s->mb_y&1), 1);
3200 }
3201 }
3202 }
3203 }
0da71265 3204 } else {
5d18eaad
LM
3205 linesize = h->mb_linesize = s->linesize;
3206 uvlinesize = h->mb_uvlinesize = s->uvlinesize;
0da71265
MN
3207// dct_offset = s->linesize * 16;
3208 }
115329f1 3209
ef9d1d15
LM
3210 if(transform_bypass){
3211 idct_dc_add =
3212 idct_add = IS_8x8DCT(mb_type) ? s->dsp.add_pixels8 : s->dsp.add_pixels4;
3213 }else if(IS_8x8DCT(mb_type)){
3214 idct_dc_add = s->dsp.h264_idct8_dc_add;
3215 idct_add = s->dsp.h264_idct8_add;
3216 }else{
3217 idct_dc_add = s->dsp.h264_idct_dc_add;
3218 idct_add = s->dsp.h264_idct_add;
3219 }
0da71265 3220
bd91fee3 3221 if(!simple && FRAME_MBAFF && h->deblocking_filter && IS_INTRA(mb_type)
5d18eaad
LM
3222 && (!bottom || !IS_INTRA(s->current_picture.mb_type[mb_xy-s->mb_stride]))){
3223 int mbt_y = mb_y&~1;
3224 uint8_t *top_y = s->current_picture.data[0] + (mbt_y * 16* s->linesize ) + mb_x * 16;
3225 uint8_t *top_cb = s->current_picture.data[1] + (mbt_y * 8 * s->uvlinesize) + mb_x * 8;
3226 uint8_t *top_cr = s->current_picture.data[2] + (mbt_y * 8 * s->uvlinesize) + mb_x * 8;
3227 xchg_pair_border(h, top_y, top_cb, top_cr, s->linesize, s->uvlinesize, 1);
3228 }
3229
bd91fee3 3230 if (!simple && IS_INTRA_PCM(mb_type)) {
6fbcaaa0
LLL
3231 unsigned int x, y;
3232
3233 // The pixels are stored in h->mb array in the same order as levels,
3234 // copy them in output in the correct order.
3235 for(i=0; i<16; i++) {
3236 for (y=0; y<4; y++) {
3237 for (x=0; x<4; x++) {
6867a90b 3238 *(dest_y + block_offset[i] + y*linesize + x) = h->mb[i*16+y*4+x];
6fbcaaa0
LLL
3239 }
3240 }
3241 }
3242 for(i=16; i<16+4; i++) {
3243 for (y=0; y<4; y++) {
3244 for (x=0; x<4; x++) {
6867a90b 3245 *(dest_cb + block_offset[i] + y*uvlinesize + x) = h->mb[i*16+y*4+x];
6fbcaaa0
LLL
3246 }
3247 }
3248 }
3249 for(i=20; i<20+4; i++) {
3250 for (y=0; y<4; y++) {
3251 for (x=0; x<4; x++) {
6867a90b 3252 *(dest_cr + block_offset[i] + y*uvlinesize + x) = h->mb[i*16+y*4+x];
6fbcaaa0
LLL
3253 }
3254 }
3255 }
e7e09b49
LLL
3256 } else {
3257 if(IS_INTRA(mb_type)){
bd91fee3 3258 if(h->deblocking_filter && (simple || !FRAME_MBAFF))
93cc10fa 3259 xchg_mb_border(h, dest_y, dest_cb, dest_cr, linesize, uvlinesize, 1, simple);
53c05b1e 3260
bd91fee3 3261 if(simple || !(s->flags&CODEC_FLAG_GRAY)){
e7e09b49
LLL
3262 h->pred8x8[ h->chroma_pred_mode ](dest_cb, uvlinesize);
3263 h->pred8x8[ h->chroma_pred_mode ](dest_cr, uvlinesize);
3264 }
0da71265 3265
e7e09b49 3266 if(IS_INTRA4x4(mb_type)){
bd91fee3 3267 if(simple || !s->encoding){
43efd19a
LM
3268 if(IS_8x8DCT(mb_type)){
3269 for(i=0; i<16; i+=4){
3270 uint8_t * const ptr= dest_y + block_offset[i];
3271 const int dir= h->intra4x4_pred_mode_cache[ scan8[i] ];
ef9d1d15 3272 const int nnz = h->non_zero_count_cache[ scan8[i] ];
43efd19a 3273 h->pred8x8l[ dir ](ptr, (h->topleft_samples_available<<i)&0x8000,
4672503d 3274 (h->topright_samples_available<<i)&0x4000, linesize);
ef9d1d15
LM
3275 if(nnz){
3276 if(nnz == 1 && h->mb[i*16])
3277 idct_dc_add(ptr, h->mb + i*16, linesize);
3278 else
3279 idct_add(ptr, h->mb + i*16, linesize);
3280 }
43efd19a
LM
3281 }
3282 }else
e7e09b49 3283 for(i=0; i<16; i++){
6867a90b 3284 uint8_t * const ptr= dest_y + block_offset[i];
e7e09b49
LLL
3285 uint8_t *topright;
3286 const int dir= h->intra4x4_pred_mode_cache[ scan8[i] ];
ef9d1d15 3287 int nnz, tr;
e7e09b49
LLL
3288
3289 if(dir == DIAG_DOWN_LEFT_PRED || dir == VERT_LEFT_PRED){
3290 const int topright_avail= (h->topright_samples_available<<i)&0x8000;
6867a90b 3291 assert(mb_y || linesize <= block_offset[i]);
e7e09b49
LLL
3292 if(!topright_avail){
3293 tr= ptr[3 - linesize]*0x01010101;
3294 topright= (uint8_t*) &tr;
115329f1 3295 }else
e7e09b49 3296 topright= ptr + 4 - linesize;
a9799653 3297 }else
e7e09b49
LLL
3298 topright= NULL;
3299
3300 h->pred4x4[ dir ](ptr, topright, linesize);
ef9d1d15
LM
3301 nnz = h->non_zero_count_cache[ scan8[i] ];
3302 if(nnz){
bd91fee3 3303 if(is_h264){
ef9d1d15
LM
3304 if(nnz == 1 && h->mb[i*16])
3305 idct_dc_add(ptr, h->mb + i*16, linesize);
3306 else
3307 idct_add(ptr, h->mb + i*16, linesize);
3308 }else
e7e09b49
LLL
3309 svq3_add_idct_c(ptr, h->mb + i*16, linesize, s->qscale, 0);
3310 }
8b82a956 3311 }
0da71265 3312 }
e7e09b49
LLL
3313 }else{
3314 h->pred16x16[ h->intra16x16_pred_mode ](dest_y , linesize);
bd91fee3 3315 if(is_h264){
36940eca 3316 if(!transform_bypass)
239ea04c 3317 h264_luma_dc_dequant_idct_c(h->mb, s->qscale, h->dequant4_coeff[IS_INTRA(mb_type) ? 0:3][s->qscale][0]);
36940eca 3318 }else
e7e09b49 3319 svq3_luma_dc_dequant_idct_c(h->mb, s->qscale);
0da71265 3320 }
bd91fee3 3321 if(h->deblocking_filter && (simple || !FRAME_MBAFF))
93cc10fa 3322 xchg_mb_border(h, dest_y, dest_cb, dest_cr, linesize, uvlinesize, 0, simple);
bd91fee3 3323 }else if(is_h264){
e7e09b49 3324 hl_motion(h, dest_y, dest_cb, dest_cr,
2833fc46
LM
3325 s->me.qpel_put, s->dsp.put_h264_chroma_pixels_tab,
3326 s->me.qpel_avg, s->dsp.avg_h264_chroma_pixels_tab,
e7e09b49 3327 s->dsp.weight_h264_pixels_tab, s->dsp.biweight_h264_pixels_tab);
0da71265 3328 }
e7e09b49
LLL
3329
3330
3331 if(!IS_INTRA4x4(mb_type)){
bd91fee3 3332 if(is_h264){
ef9d1d15
LM
3333 if(IS_INTRA16x16(mb_type)){
3334 for(i=0; i<16; i++){
3335 if(h->non_zero_count_cache[ scan8[i] ])
3336 idct_add(dest_y + block_offset[i], h->mb + i*16, linesize);
3337 else if(h->mb[i*16])
3338 idct_dc_add(dest_y + block_offset[i], h->mb + i*16, linesize);
3339 }
3340 }else{
3341 const int di = IS_8x8DCT(mb_type) ? 4 : 1;
3342 for(i=0; i<16; i+=di){
3343 int nnz = h->non_zero_count_cache[ scan8[i] ];
3344 if(nnz){
3345 if(nnz==1 && h->mb[i*16])
3346 idct_dc_add(dest_y + block_offset[i], h->mb + i*16, linesize);
3347 else
3348 idct_add(dest_y + block_offset[i], h->mb + i*16, linesize);
3349 }
e7e09b49 3350 }
4704097a 3351 }
e7e09b49
LLL
3352 }else{
3353 for(i=0; i<16; i++){
3354 if(h->non_zero_count_cache[ scan8[i] ] || h->mb[i*16]){ //FIXME benchmark weird rule, & below
6867a90b 3355 uint8_t * const ptr= dest_y + block_offset[i];
e7e09b49
LLL
3356 svq3_add_idct_c(ptr, h->mb + i*16, linesize, s->qscale, IS_INTRA(mb_type) ? 1 : 0);
3357 }
4704097a 3358 }
0da71265
MN
3359 }
3360 }
0da71265 3361
bd91fee3 3362 if(simple || !(s->flags&CODEC_FLAG_GRAY)){
ef9d1d15
LM
3363 uint8_t *dest[2] = {dest_cb, dest_cr};
3364 if(transform_bypass){
3365 idct_add = idct_dc_add = s->dsp.add_pixels4;
3366 }else{
3367 idct_add = s->dsp.h264_idct_add;
3368 idct_dc_add = s->dsp.h264_idct_dc_add;
239ea04c
LM
3369 chroma_dc_dequant_idct_c(h->mb + 16*16, h->chroma_qp, h->dequant4_coeff[IS_INTRA(mb_type) ? 1:4][h->chroma_qp][0]);
3370 chroma_dc_dequant_idct_c(h->mb + 16*16+4*16, h->chroma_qp, h->dequant4_coeff[IS_INTRA(mb_type) ? 2:5][h->chroma_qp][0]);
36940eca 3371 }
bd91fee3 3372 if(is_h264){
ef9d1d15
LM
3373 for(i=16; i<16+8; i++){
3374 if(h->non_zero_count_cache[ scan8[i] ])
3375 idct_add(dest[(i&4)>>2] + block_offset[i], h->mb + i*16, uvlinesize);
3376 else if(h->mb[i*16])
3377 idct_dc_add(dest[(i&4)>>2] + block_offset[i], h->mb + i*16, uvlinesize);
4704097a 3378 }
e7e09b49 3379 }else{
ef9d1d15 3380 for(i=16; i<16+8; i++){
e7e09b49 3381 if(h->non_zero_count_cache[ scan8[i] ] || h->mb[i*16]){
ef9d1d15 3382 uint8_t * const ptr= dest[(i&4)>>2] + block_offset[i];
e7e09b49
LLL
3383 svq3_add_idct_c(ptr, h->mb + i*16, uvlinesize, chroma_qp[s->qscale + 12] - 12, 2);
3384 }
4704097a 3385 }
0da71265
MN
3386 }
3387 }
3388 }
53c05b1e 3389 if(h->deblocking_filter) {
bd91fee3 3390 if (!simple && FRAME_MBAFF) {
5d18eaad
LM
3391 //FIXME try deblocking one mb at a time?
3392 // the reduction in load/storing mvs and such might outweigh the extra backup/xchg_border
6ba71fc4
LLL
3393 const int mb_y = s->mb_y - 1;
3394 uint8_t *pair_dest_y, *pair_dest_cb, *pair_dest_cr;
3395 const int mb_xy= mb_x + mb_y*s->mb_stride;
3396 const int mb_type_top = s->current_picture.mb_type[mb_xy];
3397 const int mb_type_bottom= s->current_picture.mb_type[mb_xy+s->mb_stride];
6ba71fc4
LLL
3398 if (!bottom) return;
3399 pair_dest_y = s->current_picture.data[0] + (mb_y * 16* s->linesize ) + mb_x * 16;
3400 pair_dest_cb = s->current_picture.data[1] + (mb_y * 8 * s->uvlinesize) + mb_x * 8;
3401 pair_dest_cr = s->current_picture.data[2] + (mb_y * 8 * s->uvlinesize) + mb_x * 8;
3402
5d18eaad
LM
3403 if(IS_INTRA(mb_type_top | mb_type_bottom))
3404 xchg_pair_border(h, pair_dest_y, pair_dest_cb, pair_dest_cr, s->linesize, s->uvlinesize, 0);
3405
6ba71fc4 3406 backup_pair_border(h, pair_dest_y, pair_dest_cb, pair_dest_cr, s->linesize, s->uvlinesize);
5d18eaad