Remove graphparser.h header, move AVFilterInOut and
[libav.git] / ffplay.c
1 /*
2 * FFplay : Simple Media Player based on the FFmpeg libraries
3 * Copyright (c) 2003 Fabrice Bellard
4 *
5 * This file is part of FFmpeg.
6 *
7 * FFmpeg is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.1 of the License, or (at your option) any later version.
11 *
12 * FFmpeg is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with FFmpeg; if not, write to the Free Software
19 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20 */
21
22 #define _XOPEN_SOURCE 600
23
24 #include "config.h"
25 #include <inttypes.h>
26 #include <math.h>
27 #include <limits.h>
28 #include "libavutil/avstring.h"
29 #include "libavutil/colorspace.h"
30 #include "libavutil/pixdesc.h"
31 #include "libavcore/imgutils.h"
32 #include "libavcore/parseutils.h"
33 #include "libavcore/samplefmt.h"
34 #include "libavformat/avformat.h"
35 #include "libavdevice/avdevice.h"
36 #include "libswscale/swscale.h"
37 #include "libavcodec/audioconvert.h"
38 #include "libavcodec/opt.h"
39 #include "libavcodec/avfft.h"
40
41 #if CONFIG_AVFILTER
42 # include "libavfilter/avfilter.h"
43 # include "libavfilter/avfiltergraph.h"
44 #endif
45
46 #include "cmdutils.h"
47
48 #include <SDL.h>
49 #include <SDL_thread.h>
50
51 #ifdef __MINGW32__
52 #undef main /* We don't want SDL to override our main() */
53 #endif
54
55 #include <unistd.h>
56 #include <assert.h>
57
58 const char program_name[] = "FFplay";
59 const int program_birth_year = 2003;
60
61 //#define DEBUG_SYNC
62
63 #define MAX_QUEUE_SIZE (15 * 1024 * 1024)
64 #define MIN_AUDIOQ_SIZE (20 * 16 * 1024)
65 #define MIN_FRAMES 5
66
67 /* SDL audio buffer size, in samples. Should be small to have precise
68 A/V sync as SDL does not have hardware buffer fullness info. */
69 #define SDL_AUDIO_BUFFER_SIZE 1024
70
71 /* no AV sync correction is done if below the AV sync threshold */
72 #define AV_SYNC_THRESHOLD 0.01
73 /* no AV correction is done if too big error */
74 #define AV_NOSYNC_THRESHOLD 10.0
75
76 #define FRAME_SKIP_FACTOR 0.05
77
78 /* maximum audio speed change to get correct sync */
79 #define SAMPLE_CORRECTION_PERCENT_MAX 10
80
81 /* we use about AUDIO_DIFF_AVG_NB A-V differences to make the average */
82 #define AUDIO_DIFF_AVG_NB 20
83
84 /* NOTE: the size must be big enough to compensate the hardware audio buffersize size */
85 #define SAMPLE_ARRAY_SIZE (2*65536)
86
87 static int sws_flags = SWS_BICUBIC;
88
89 typedef struct PacketQueue {
90 AVPacketList *first_pkt, *last_pkt;
91 int nb_packets;
92 int size;
93 int abort_request;
94 SDL_mutex *mutex;
95 SDL_cond *cond;
96 } PacketQueue;
97
98 #define VIDEO_PICTURE_QUEUE_SIZE 2
99 #define SUBPICTURE_QUEUE_SIZE 4
100
101 typedef struct VideoPicture {
102 double pts; ///<presentation time stamp for this picture
103 double target_clock; ///<av_gettime() time at which this should be displayed ideally
104 int64_t pos; ///<byte position in file
105 SDL_Overlay *bmp;
106 int width, height; /* source height & width */
107 int allocated;
108 enum PixelFormat pix_fmt;
109
110 #if CONFIG_AVFILTER
111 AVFilterBufferRef *picref;
112 #endif
113 } VideoPicture;
114
115 typedef struct SubPicture {
116 double pts; /* presentation time stamp for this picture */
117 AVSubtitle sub;
118 } SubPicture;
119
120 enum {
121 AV_SYNC_AUDIO_MASTER, /* default choice */
122 AV_SYNC_VIDEO_MASTER,
123 AV_SYNC_EXTERNAL_CLOCK, /* synchronize to an external clock */
124 };
125
126 typedef struct VideoState {
127 SDL_Thread *parse_tid;
128 SDL_Thread *video_tid;
129 SDL_Thread *refresh_tid;
130 AVInputFormat *iformat;
131 int no_background;
132 int abort_request;
133 int paused;
134 int last_paused;
135 int seek_req;
136 int seek_flags;
137 int64_t seek_pos;
138 int64_t seek_rel;
139 int read_pause_return;
140 AVFormatContext *ic;
141 int dtg_active_format;
142
143 int audio_stream;
144
145 int av_sync_type;
146 double external_clock; /* external clock base */
147 int64_t external_clock_time;
148
149 double audio_clock;
150 double audio_diff_cum; /* used for AV difference average computation */
151 double audio_diff_avg_coef;
152 double audio_diff_threshold;
153 int audio_diff_avg_count;
154 AVStream *audio_st;
155 PacketQueue audioq;
156 int audio_hw_buf_size;
157 /* samples output by the codec. we reserve more space for avsync
158 compensation */
159 DECLARE_ALIGNED(16,uint8_t,audio_buf1)[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
160 DECLARE_ALIGNED(16,uint8_t,audio_buf2)[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
161 uint8_t *audio_buf;
162 unsigned int audio_buf_size; /* in bytes */
163 int audio_buf_index; /* in bytes */
164 AVPacket audio_pkt_temp;
165 AVPacket audio_pkt;
166 enum SampleFormat audio_src_fmt;
167 AVAudioConvert *reformat_ctx;
168
169 int show_audio; /* if true, display audio samples */
170 int16_t sample_array[SAMPLE_ARRAY_SIZE];
171 int sample_array_index;
172 int last_i_start;
173 RDFTContext *rdft;
174 int rdft_bits;
175 FFTSample *rdft_data;
176 int xpos;
177
178 SDL_Thread *subtitle_tid;
179 int subtitle_stream;
180 int subtitle_stream_changed;
181 AVStream *subtitle_st;
182 PacketQueue subtitleq;
183 SubPicture subpq[SUBPICTURE_QUEUE_SIZE];
184 int subpq_size, subpq_rindex, subpq_windex;
185 SDL_mutex *subpq_mutex;
186 SDL_cond *subpq_cond;
187
188 double frame_timer;
189 double frame_last_pts;
190 double frame_last_delay;
191 double video_clock; ///<pts of last decoded frame / predicted pts of next decoded frame
192 int video_stream;
193 AVStream *video_st;
194 PacketQueue videoq;
195 double video_current_pts; ///<current displayed pts (different from video_clock if frame fifos are used)
196 double video_current_pts_drift; ///<video_current_pts - time (av_gettime) at which we updated video_current_pts - used to have running video pts
197 int64_t video_current_pos; ///<current displayed file pos
198 VideoPicture pictq[VIDEO_PICTURE_QUEUE_SIZE];
199 int pictq_size, pictq_rindex, pictq_windex;
200 SDL_mutex *pictq_mutex;
201 SDL_cond *pictq_cond;
202 #if !CONFIG_AVFILTER
203 struct SwsContext *img_convert_ctx;
204 #endif
205
206 // QETimer *video_timer;
207 char filename[1024];
208 int width, height, xleft, ytop;
209
210 PtsCorrectionContext pts_ctx;
211
212 #if CONFIG_AVFILTER
213 AVFilterContext *out_video_filter; ///<the last filter in the video chain
214 #endif
215
216 float skip_frames;
217 float skip_frames_index;
218 int refresh;
219 } VideoState;
220
221 static void show_help(void);
222 static int audio_write_get_buf_size(VideoState *is);
223
224 /* options specified by the user */
225 static AVInputFormat *file_iformat;
226 static const char *input_filename;
227 static const char *window_title;
228 static int fs_screen_width;
229 static int fs_screen_height;
230 static int screen_width = 0;
231 static int screen_height = 0;
232 static int frame_width = 0;
233 static int frame_height = 0;
234 static enum PixelFormat frame_pix_fmt = PIX_FMT_NONE;
235 static int audio_disable;
236 static int video_disable;
237 static int wanted_stream[AVMEDIA_TYPE_NB]={
238 [AVMEDIA_TYPE_AUDIO]=-1,
239 [AVMEDIA_TYPE_VIDEO]=-1,
240 [AVMEDIA_TYPE_SUBTITLE]=-1,
241 };
242 static int seek_by_bytes=-1;
243 static int display_disable;
244 static int show_status = 1;
245 static int av_sync_type = AV_SYNC_AUDIO_MASTER;
246 static int64_t start_time = AV_NOPTS_VALUE;
247 static int64_t duration = AV_NOPTS_VALUE;
248 static int debug = 0;
249 static int debug_mv = 0;
250 static int step = 0;
251 static int thread_count = 1;
252 static int workaround_bugs = 1;
253 static int fast = 0;
254 static int genpts = 0;
255 static int lowres = 0;
256 static int idct = FF_IDCT_AUTO;
257 static enum AVDiscard skip_frame= AVDISCARD_DEFAULT;
258 static enum AVDiscard skip_idct= AVDISCARD_DEFAULT;
259 static enum AVDiscard skip_loop_filter= AVDISCARD_DEFAULT;
260 static int error_recognition = FF_ER_CAREFUL;
261 static int error_concealment = 3;
262 static int decoder_reorder_pts= -1;
263 static int autoexit;
264 static int exit_on_keydown;
265 static int exit_on_mousedown;
266 static int loop=1;
267 static int framedrop=1;
268
269 static int rdftspeed=20;
270 #if CONFIG_AVFILTER
271 static char *vfilters = NULL;
272 #endif
273
274 /* current context */
275 static int is_full_screen;
276 static VideoState *cur_stream;
277 static int64_t audio_callback_time;
278
279 static AVPacket flush_pkt;
280
281 #define FF_ALLOC_EVENT (SDL_USEREVENT)
282 #define FF_REFRESH_EVENT (SDL_USEREVENT + 1)
283 #define FF_QUIT_EVENT (SDL_USEREVENT + 2)
284
285 static SDL_Surface *screen;
286
287 static int packet_queue_put(PacketQueue *q, AVPacket *pkt);
288
289 /* packet queue handling */
290 static void packet_queue_init(PacketQueue *q)
291 {
292 memset(q, 0, sizeof(PacketQueue));
293 q->mutex = SDL_CreateMutex();
294 q->cond = SDL_CreateCond();
295 packet_queue_put(q, &flush_pkt);
296 }
297
298 static void packet_queue_flush(PacketQueue *q)
299 {
300 AVPacketList *pkt, *pkt1;
301
302 SDL_LockMutex(q->mutex);
303 for(pkt = q->first_pkt; pkt != NULL; pkt = pkt1) {
304 pkt1 = pkt->next;
305 av_free_packet(&pkt->pkt);
306 av_freep(&pkt);
307 }
308 q->last_pkt = NULL;
309 q->first_pkt = NULL;
310 q->nb_packets = 0;
311 q->size = 0;
312 SDL_UnlockMutex(q->mutex);
313 }
314
315 static void packet_queue_end(PacketQueue *q)
316 {
317 packet_queue_flush(q);
318 SDL_DestroyMutex(q->mutex);
319 SDL_DestroyCond(q->cond);
320 }
321
322 static int packet_queue_put(PacketQueue *q, AVPacket *pkt)
323 {
324 AVPacketList *pkt1;
325
326 /* duplicate the packet */
327 if (pkt!=&flush_pkt && av_dup_packet(pkt) < 0)
328 return -1;
329
330 pkt1 = av_malloc(sizeof(AVPacketList));
331 if (!pkt1)
332 return -1;
333 pkt1->pkt = *pkt;
334 pkt1->next = NULL;
335
336
337 SDL_LockMutex(q->mutex);
338
339 if (!q->last_pkt)
340
341 q->first_pkt = pkt1;
342 else
343 q->last_pkt->next = pkt1;
344 q->last_pkt = pkt1;
345 q->nb_packets++;
346 q->size += pkt1->pkt.size + sizeof(*pkt1);
347 /* XXX: should duplicate packet data in DV case */
348 SDL_CondSignal(q->cond);
349
350 SDL_UnlockMutex(q->mutex);
351 return 0;
352 }
353
354 static void packet_queue_abort(PacketQueue *q)
355 {
356 SDL_LockMutex(q->mutex);
357
358 q->abort_request = 1;
359
360 SDL_CondSignal(q->cond);
361
362 SDL_UnlockMutex(q->mutex);
363 }
364
365 /* return < 0 if aborted, 0 if no packet and > 0 if packet. */
366 static int packet_queue_get(PacketQueue *q, AVPacket *pkt, int block)
367 {
368 AVPacketList *pkt1;
369 int ret;
370
371 SDL_LockMutex(q->mutex);
372
373 for(;;) {
374 if (q->abort_request) {
375 ret = -1;
376 break;
377 }
378
379 pkt1 = q->first_pkt;
380 if (pkt1) {
381 q->first_pkt = pkt1->next;
382 if (!q->first_pkt)
383 q->last_pkt = NULL;
384 q->nb_packets--;
385 q->size -= pkt1->pkt.size + sizeof(*pkt1);
386 *pkt = pkt1->pkt;
387 av_free(pkt1);
388 ret = 1;
389 break;
390 } else if (!block) {
391 ret = 0;
392 break;
393 } else {
394 SDL_CondWait(q->cond, q->mutex);
395 }
396 }
397 SDL_UnlockMutex(q->mutex);
398 return ret;
399 }
400
401 static inline void fill_rectangle(SDL_Surface *screen,
402 int x, int y, int w, int h, int color)
403 {
404 SDL_Rect rect;
405 rect.x = x;
406 rect.y = y;
407 rect.w = w;
408 rect.h = h;
409 SDL_FillRect(screen, &rect, color);
410 }
411
412 #if 0
413 /* draw only the border of a rectangle */
414 void fill_border(VideoState *s, int x, int y, int w, int h, int color)
415 {
416 int w1, w2, h1, h2;
417
418 /* fill the background */
419 w1 = x;
420 if (w1 < 0)
421 w1 = 0;
422 w2 = s->width - (x + w);
423 if (w2 < 0)
424 w2 = 0;
425 h1 = y;
426 if (h1 < 0)
427 h1 = 0;
428 h2 = s->height - (y + h);
429 if (h2 < 0)
430 h2 = 0;
431 fill_rectangle(screen,
432 s->xleft, s->ytop,
433 w1, s->height,
434 color);
435 fill_rectangle(screen,
436 s->xleft + s->width - w2, s->ytop,
437 w2, s->height,
438 color);
439 fill_rectangle(screen,
440 s->xleft + w1, s->ytop,
441 s->width - w1 - w2, h1,
442 color);
443 fill_rectangle(screen,
444 s->xleft + w1, s->ytop + s->height - h2,
445 s->width - w1 - w2, h2,
446 color);
447 }
448 #endif
449
450 #define ALPHA_BLEND(a, oldp, newp, s)\
451 ((((oldp << s) * (255 - (a))) + (newp * (a))) / (255 << s))
452
453 #define RGBA_IN(r, g, b, a, s)\
454 {\
455 unsigned int v = ((const uint32_t *)(s))[0];\
456 a = (v >> 24) & 0xff;\
457 r = (v >> 16) & 0xff;\
458 g = (v >> 8) & 0xff;\
459 b = v & 0xff;\
460 }
461
462 #define YUVA_IN(y, u, v, a, s, pal)\
463 {\
464 unsigned int val = ((const uint32_t *)(pal))[*(const uint8_t*)(s)];\
465 a = (val >> 24) & 0xff;\
466 y = (val >> 16) & 0xff;\
467 u = (val >> 8) & 0xff;\
468 v = val & 0xff;\
469 }
470
471 #define YUVA_OUT(d, y, u, v, a)\
472 {\
473 ((uint32_t *)(d))[0] = (a << 24) | (y << 16) | (u << 8) | v;\
474 }
475
476
477 #define BPP 1
478
479 static void blend_subrect(AVPicture *dst, const AVSubtitleRect *rect, int imgw, int imgh)
480 {
481 int wrap, wrap3, width2, skip2;
482 int y, u, v, a, u1, v1, a1, w, h;
483 uint8_t *lum, *cb, *cr;
484 const uint8_t *p;
485 const uint32_t *pal;
486 int dstx, dsty, dstw, dsth;
487
488 dstw = av_clip(rect->w, 0, imgw);
489 dsth = av_clip(rect->h, 0, imgh);
490 dstx = av_clip(rect->x, 0, imgw - dstw);
491 dsty = av_clip(rect->y, 0, imgh - dsth);
492 lum = dst->data[0] + dsty * dst->linesize[0];
493 cb = dst->data[1] + (dsty >> 1) * dst->linesize[1];
494 cr = dst->data[2] + (dsty >> 1) * dst->linesize[2];
495
496 width2 = ((dstw + 1) >> 1) + (dstx & ~dstw & 1);
497 skip2 = dstx >> 1;
498 wrap = dst->linesize[0];
499 wrap3 = rect->pict.linesize[0];
500 p = rect->pict.data[0];
501 pal = (const uint32_t *)rect->pict.data[1]; /* Now in YCrCb! */
502
503 if (dsty & 1) {
504 lum += dstx;
505 cb += skip2;
506 cr += skip2;
507
508 if (dstx & 1) {
509 YUVA_IN(y, u, v, a, p, pal);
510 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
511 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
512 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
513 cb++;
514 cr++;
515 lum++;
516 p += BPP;
517 }
518 for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
519 YUVA_IN(y, u, v, a, p, pal);
520 u1 = u;
521 v1 = v;
522 a1 = a;
523 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
524
525 YUVA_IN(y, u, v, a, p + BPP, pal);
526 u1 += u;
527 v1 += v;
528 a1 += a;
529 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
530 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
531 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
532 cb++;
533 cr++;
534 p += 2 * BPP;
535 lum += 2;
536 }
537 if (w) {
538 YUVA_IN(y, u, v, a, p, pal);
539 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
540 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
541 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
542 p++;
543 lum++;
544 }
545 p += wrap3 - dstw * BPP;
546 lum += wrap - dstw - dstx;
547 cb += dst->linesize[1] - width2 - skip2;
548 cr += dst->linesize[2] - width2 - skip2;
549 }
550 for(h = dsth - (dsty & 1); h >= 2; h -= 2) {
551 lum += dstx;
552 cb += skip2;
553 cr += skip2;
554
555 if (dstx & 1) {
556 YUVA_IN(y, u, v, a, p, pal);
557 u1 = u;
558 v1 = v;
559 a1 = a;
560 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
561 p += wrap3;
562 lum += wrap;
563 YUVA_IN(y, u, v, a, p, pal);
564 u1 += u;
565 v1 += v;
566 a1 += a;
567 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
568 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
569 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
570 cb++;
571 cr++;
572 p += -wrap3 + BPP;
573 lum += -wrap + 1;
574 }
575 for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
576 YUVA_IN(y, u, v, a, p, pal);
577 u1 = u;
578 v1 = v;
579 a1 = a;
580 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
581
582 YUVA_IN(y, u, v, a, p + BPP, pal);
583 u1 += u;
584 v1 += v;
585 a1 += a;
586 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
587 p += wrap3;
588 lum += wrap;
589
590 YUVA_IN(y, u, v, a, p, pal);
591 u1 += u;
592 v1 += v;
593 a1 += a;
594 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
595
596 YUVA_IN(y, u, v, a, p + BPP, pal);
597 u1 += u;
598 v1 += v;
599 a1 += a;
600 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
601
602 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 2);
603 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 2);
604
605 cb++;
606 cr++;
607 p += -wrap3 + 2 * BPP;
608 lum += -wrap + 2;
609 }
610 if (w) {
611 YUVA_IN(y, u, v, a, p, pal);
612 u1 = u;
613 v1 = v;
614 a1 = a;
615 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
616 p += wrap3;
617 lum += wrap;
618 YUVA_IN(y, u, v, a, p, pal);
619 u1 += u;
620 v1 += v;
621 a1 += a;
622 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
623 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u1, 1);
624 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v1, 1);
625 cb++;
626 cr++;
627 p += -wrap3 + BPP;
628 lum += -wrap + 1;
629 }
630 p += wrap3 + (wrap3 - dstw * BPP);
631 lum += wrap + (wrap - dstw - dstx);
632 cb += dst->linesize[1] - width2 - skip2;
633 cr += dst->linesize[2] - width2 - skip2;
634 }
635 /* handle odd height */
636 if (h) {
637 lum += dstx;
638 cb += skip2;
639 cr += skip2;
640
641 if (dstx & 1) {
642 YUVA_IN(y, u, v, a, p, pal);
643 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
644 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
645 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
646 cb++;
647 cr++;
648 lum++;
649 p += BPP;
650 }
651 for(w = dstw - (dstx & 1); w >= 2; w -= 2) {
652 YUVA_IN(y, u, v, a, p, pal);
653 u1 = u;
654 v1 = v;
655 a1 = a;
656 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
657
658 YUVA_IN(y, u, v, a, p + BPP, pal);
659 u1 += u;
660 v1 += v;
661 a1 += a;
662 lum[1] = ALPHA_BLEND(a, lum[1], y, 0);
663 cb[0] = ALPHA_BLEND(a1 >> 2, cb[0], u, 1);
664 cr[0] = ALPHA_BLEND(a1 >> 2, cr[0], v, 1);
665 cb++;
666 cr++;
667 p += 2 * BPP;
668 lum += 2;
669 }
670 if (w) {
671 YUVA_IN(y, u, v, a, p, pal);
672 lum[0] = ALPHA_BLEND(a, lum[0], y, 0);
673 cb[0] = ALPHA_BLEND(a >> 2, cb[0], u, 0);
674 cr[0] = ALPHA_BLEND(a >> 2, cr[0], v, 0);
675 }
676 }
677 }
678
679 static void free_subpicture(SubPicture *sp)
680 {
681 avsubtitle_free(&sp->sub);
682 }
683
684 static void video_image_display(VideoState *is)
685 {
686 VideoPicture *vp;
687 SubPicture *sp;
688 AVPicture pict;
689 float aspect_ratio;
690 int width, height, x, y;
691 SDL_Rect rect;
692 int i;
693
694 vp = &is->pictq[is->pictq_rindex];
695 if (vp->bmp) {
696 #if CONFIG_AVFILTER
697 if (vp->picref->video->pixel_aspect.num == 0)
698 aspect_ratio = 0;
699 else
700 aspect_ratio = av_q2d(vp->picref->video->pixel_aspect);
701 #else
702
703 /* XXX: use variable in the frame */
704 if (is->video_st->sample_aspect_ratio.num)
705 aspect_ratio = av_q2d(is->video_st->sample_aspect_ratio);
706 else if (is->video_st->codec->sample_aspect_ratio.num)
707 aspect_ratio = av_q2d(is->video_st->codec->sample_aspect_ratio);
708 else
709 aspect_ratio = 0;
710 #endif
711 if (aspect_ratio <= 0.0)
712 aspect_ratio = 1.0;
713 aspect_ratio *= (float)vp->width / (float)vp->height;
714 /* if an active format is indicated, then it overrides the
715 mpeg format */
716 #if 0
717 if (is->video_st->codec->dtg_active_format != is->dtg_active_format) {
718 is->dtg_active_format = is->video_st->codec->dtg_active_format;
719 printf("dtg_active_format=%d\n", is->dtg_active_format);
720 }
721 #endif
722 #if 0
723 switch(is->video_st->codec->dtg_active_format) {
724 case FF_DTG_AFD_SAME:
725 default:
726 /* nothing to do */
727 break;
728 case FF_DTG_AFD_4_3:
729 aspect_ratio = 4.0 / 3.0;
730 break;
731 case FF_DTG_AFD_16_9:
732 aspect_ratio = 16.0 / 9.0;
733 break;
734 case FF_DTG_AFD_14_9:
735 aspect_ratio = 14.0 / 9.0;
736 break;
737 case FF_DTG_AFD_4_3_SP_14_9:
738 aspect_ratio = 14.0 / 9.0;
739 break;
740 case FF_DTG_AFD_16_9_SP_14_9:
741 aspect_ratio = 14.0 / 9.0;
742 break;
743 case FF_DTG_AFD_SP_4_3:
744 aspect_ratio = 4.0 / 3.0;
745 break;
746 }
747 #endif
748
749 if (is->subtitle_st)
750 {
751 if (is->subpq_size > 0)
752 {
753 sp = &is->subpq[is->subpq_rindex];
754
755 if (vp->pts >= sp->pts + ((float) sp->sub.start_display_time / 1000))
756 {
757 SDL_LockYUVOverlay (vp->bmp);
758
759 pict.data[0] = vp->bmp->pixels[0];
760 pict.data[1] = vp->bmp->pixels[2];
761 pict.data[2] = vp->bmp->pixels[1];
762
763 pict.linesize[0] = vp->bmp->pitches[0];
764 pict.linesize[1] = vp->bmp->pitches[2];
765 pict.linesize[2] = vp->bmp->pitches[1];
766
767 for (i = 0; i < sp->sub.num_rects; i++)
768 blend_subrect(&pict, sp->sub.rects[i],
769 vp->bmp->w, vp->bmp->h);
770
771 SDL_UnlockYUVOverlay (vp->bmp);
772 }
773 }
774 }
775
776
777 /* XXX: we suppose the screen has a 1.0 pixel ratio */
778 height = is->height;
779 width = ((int)rint(height * aspect_ratio)) & ~1;
780 if (width > is->width) {
781 width = is->width;
782 height = ((int)rint(width / aspect_ratio)) & ~1;
783 }
784 x = (is->width - width) / 2;
785 y = (is->height - height) / 2;
786 if (!is->no_background) {
787 /* fill the background */
788 // fill_border(is, x, y, width, height, QERGB(0x00, 0x00, 0x00));
789 } else {
790 is->no_background = 0;
791 }
792 rect.x = is->xleft + x;
793 rect.y = is->ytop + y;
794 rect.w = width;
795 rect.h = height;
796 SDL_DisplayYUVOverlay(vp->bmp, &rect);
797 } else {
798 #if 0
799 fill_rectangle(screen,
800 is->xleft, is->ytop, is->width, is->height,
801 QERGB(0x00, 0x00, 0x00));
802 #endif
803 }
804 }
805
806 static inline int compute_mod(int a, int b)
807 {
808 a = a % b;
809 if (a >= 0)
810 return a;
811 else
812 return a + b;
813 }
814
815 static void video_audio_display(VideoState *s)
816 {
817 int i, i_start, x, y1, y, ys, delay, n, nb_display_channels;
818 int ch, channels, h, h2, bgcolor, fgcolor;
819 int16_t time_diff;
820 int rdft_bits, nb_freq;
821
822 for(rdft_bits=1; (1<<rdft_bits)<2*s->height; rdft_bits++)
823 ;
824 nb_freq= 1<<(rdft_bits-1);
825
826 /* compute display index : center on currently output samples */
827 channels = s->audio_st->codec->channels;
828 nb_display_channels = channels;
829 if (!s->paused) {
830 int data_used= s->show_audio==1 ? s->width : (2*nb_freq);
831 n = 2 * channels;
832 delay = audio_write_get_buf_size(s);
833 delay /= n;
834
835 /* to be more precise, we take into account the time spent since
836 the last buffer computation */
837 if (audio_callback_time) {
838 time_diff = av_gettime() - audio_callback_time;
839 delay -= (time_diff * s->audio_st->codec->sample_rate) / 1000000;
840 }
841
842 delay += 2*data_used;
843 if (delay < data_used)
844 delay = data_used;
845
846 i_start= x = compute_mod(s->sample_array_index - delay * channels, SAMPLE_ARRAY_SIZE);
847 if(s->show_audio==1){
848 h= INT_MIN;
849 for(i=0; i<1000; i+=channels){
850 int idx= (SAMPLE_ARRAY_SIZE + x - i) % SAMPLE_ARRAY_SIZE;
851 int a= s->sample_array[idx];
852 int b= s->sample_array[(idx + 4*channels)%SAMPLE_ARRAY_SIZE];
853 int c= s->sample_array[(idx + 5*channels)%SAMPLE_ARRAY_SIZE];
854 int d= s->sample_array[(idx + 9*channels)%SAMPLE_ARRAY_SIZE];
855 int score= a-d;
856 if(h<score && (b^c)<0){
857 h= score;
858 i_start= idx;
859 }
860 }
861 }
862
863 s->last_i_start = i_start;
864 } else {
865 i_start = s->last_i_start;
866 }
867
868 bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
869 if(s->show_audio==1){
870 fill_rectangle(screen,
871 s->xleft, s->ytop, s->width, s->height,
872 bgcolor);
873
874 fgcolor = SDL_MapRGB(screen->format, 0xff, 0xff, 0xff);
875
876 /* total height for one channel */
877 h = s->height / nb_display_channels;
878 /* graph height / 2 */
879 h2 = (h * 9) / 20;
880 for(ch = 0;ch < nb_display_channels; ch++) {
881 i = i_start + ch;
882 y1 = s->ytop + ch * h + (h / 2); /* position of center line */
883 for(x = 0; x < s->width; x++) {
884 y = (s->sample_array[i] * h2) >> 15;
885 if (y < 0) {
886 y = -y;
887 ys = y1 - y;
888 } else {
889 ys = y1;
890 }
891 fill_rectangle(screen,
892 s->xleft + x, ys, 1, y,
893 fgcolor);
894 i += channels;
895 if (i >= SAMPLE_ARRAY_SIZE)
896 i -= SAMPLE_ARRAY_SIZE;
897 }
898 }
899
900 fgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0xff);
901
902 for(ch = 1;ch < nb_display_channels; ch++) {
903 y = s->ytop + ch * h;
904 fill_rectangle(screen,
905 s->xleft, y, s->width, 1,
906 fgcolor);
907 }
908 SDL_UpdateRect(screen, s->xleft, s->ytop, s->width, s->height);
909 }else{
910 nb_display_channels= FFMIN(nb_display_channels, 2);
911 if(rdft_bits != s->rdft_bits){
912 av_rdft_end(s->rdft);
913 av_free(s->rdft_data);
914 s->rdft = av_rdft_init(rdft_bits, DFT_R2C);
915 s->rdft_bits= rdft_bits;
916 s->rdft_data= av_malloc(4*nb_freq*sizeof(*s->rdft_data));
917 }
918 {
919 FFTSample *data[2];
920 for(ch = 0;ch < nb_display_channels; ch++) {
921 data[ch] = s->rdft_data + 2*nb_freq*ch;
922 i = i_start + ch;
923 for(x = 0; x < 2*nb_freq; x++) {
924 double w= (x-nb_freq)*(1.0/nb_freq);
925 data[ch][x]= s->sample_array[i]*(1.0-w*w);
926 i += channels;
927 if (i >= SAMPLE_ARRAY_SIZE)
928 i -= SAMPLE_ARRAY_SIZE;
929 }
930 av_rdft_calc(s->rdft, data[ch]);
931 }
932 //least efficient way to do this, we should of course directly access it but its more than fast enough
933 for(y=0; y<s->height; y++){
934 double w= 1/sqrt(nb_freq);
935 int a= sqrt(w*sqrt(data[0][2*y+0]*data[0][2*y+0] + data[0][2*y+1]*data[0][2*y+1]));
936 int b= (nb_display_channels == 2 ) ? sqrt(w*sqrt(data[1][2*y+0]*data[1][2*y+0]
937 + data[1][2*y+1]*data[1][2*y+1])) : a;
938 a= FFMIN(a,255);
939 b= FFMIN(b,255);
940 fgcolor = SDL_MapRGB(screen->format, a, b, (a+b)/2);
941
942 fill_rectangle(screen,
943 s->xpos, s->height-y, 1, 1,
944 fgcolor);
945 }
946 }
947 SDL_UpdateRect(screen, s->xpos, s->ytop, 1, s->height);
948 s->xpos++;
949 if(s->xpos >= s->width)
950 s->xpos= s->xleft;
951 }
952 }
953
954 static int video_open(VideoState *is){
955 int flags = SDL_HWSURFACE|SDL_ASYNCBLIT|SDL_HWACCEL;
956 int w,h;
957
958 if(is_full_screen) flags |= SDL_FULLSCREEN;
959 else flags |= SDL_RESIZABLE;
960
961 if (is_full_screen && fs_screen_width) {
962 w = fs_screen_width;
963 h = fs_screen_height;
964 } else if(!is_full_screen && screen_width){
965 w = screen_width;
966 h = screen_height;
967 #if CONFIG_AVFILTER
968 }else if (is->out_video_filter && is->out_video_filter->inputs[0]){
969 w = is->out_video_filter->inputs[0]->w;
970 h = is->out_video_filter->inputs[0]->h;
971 #else
972 }else if (is->video_st && is->video_st->codec->width){
973 w = is->video_st->codec->width;
974 h = is->video_st->codec->height;
975 #endif
976 } else {
977 w = 640;
978 h = 480;
979 }
980 if(screen && is->width == screen->w && screen->w == w
981 && is->height== screen->h && screen->h == h)
982 return 0;
983
984 #ifndef __APPLE__
985 screen = SDL_SetVideoMode(w, h, 0, flags);
986 #else
987 /* setting bits_per_pixel = 0 or 32 causes blank video on OS X */
988 screen = SDL_SetVideoMode(w, h, 24, flags);
989 #endif
990 if (!screen) {
991 fprintf(stderr, "SDL: could not set video mode - exiting\n");
992 return -1;
993 }
994 if (!window_title)
995 window_title = input_filename;
996 SDL_WM_SetCaption(window_title, window_title);
997
998 is->width = screen->w;
999 is->height = screen->h;
1000
1001 return 0;
1002 }
1003
1004 /* display the current picture, if any */
1005 static void video_display(VideoState *is)
1006 {
1007 if(!screen)
1008 video_open(cur_stream);
1009 if (is->audio_st && is->show_audio)
1010 video_audio_display(is);
1011 else if (is->video_st)
1012 video_image_display(is);
1013 }
1014
1015 static int refresh_thread(void *opaque)
1016 {
1017 VideoState *is= opaque;
1018 while(!is->abort_request){
1019 SDL_Event event;
1020 event.type = FF_REFRESH_EVENT;
1021 event.user.data1 = opaque;
1022 if(!is->refresh){
1023 is->refresh=1;
1024 SDL_PushEvent(&event);
1025 }
1026 usleep(is->audio_st && is->show_audio ? rdftspeed*1000 : 5000); //FIXME ideally we should wait the correct time but SDLs event passing is so slow it would be silly
1027 }
1028 return 0;
1029 }
1030
1031 /* get the current audio clock value */
1032 static double get_audio_clock(VideoState *is)
1033 {
1034 double pts;
1035 int hw_buf_size, bytes_per_sec;
1036 pts = is->audio_clock;
1037 hw_buf_size = audio_write_get_buf_size(is);
1038 bytes_per_sec = 0;
1039 if (is->audio_st) {
1040 bytes_per_sec = is->audio_st->codec->sample_rate *
1041 2 * is->audio_st->codec->channels;
1042 }
1043 if (bytes_per_sec)
1044 pts -= (double)hw_buf_size / bytes_per_sec;
1045 return pts;
1046 }
1047
1048 /* get the current video clock value */
1049 static double get_video_clock(VideoState *is)
1050 {
1051 if (is->paused) {
1052 return is->video_current_pts;
1053 } else {
1054 return is->video_current_pts_drift + av_gettime() / 1000000.0;
1055 }
1056 }
1057
1058 /* get the current external clock value */
1059 static double get_external_clock(VideoState *is)
1060 {
1061 int64_t ti;
1062 ti = av_gettime();
1063 return is->external_clock + ((ti - is->external_clock_time) * 1e-6);
1064 }
1065
1066 /* get the current master clock value */
1067 static double get_master_clock(VideoState *is)
1068 {
1069 double val;
1070
1071 if (is->av_sync_type == AV_SYNC_VIDEO_MASTER) {
1072 if (is->video_st)
1073 val = get_video_clock(is);
1074 else
1075 val = get_audio_clock(is);
1076 } else if (is->av_sync_type == AV_SYNC_AUDIO_MASTER) {
1077 if (is->audio_st)
1078 val = get_audio_clock(is);
1079 else
1080 val = get_video_clock(is);
1081 } else {
1082 val = get_external_clock(is);
1083 }
1084 return val;
1085 }
1086
1087 /* seek in the stream */
1088 static void stream_seek(VideoState *is, int64_t pos, int64_t rel, int seek_by_bytes)
1089 {
1090 if (!is->seek_req) {
1091 is->seek_pos = pos;
1092 is->seek_rel = rel;
1093 is->seek_flags &= ~AVSEEK_FLAG_BYTE;
1094 if (seek_by_bytes)
1095 is->seek_flags |= AVSEEK_FLAG_BYTE;
1096 is->seek_req = 1;
1097 }
1098 }
1099
1100 /* pause or resume the video */
1101 static void stream_pause(VideoState *is)
1102 {
1103 if (is->paused) {
1104 is->frame_timer += av_gettime() / 1000000.0 + is->video_current_pts_drift - is->video_current_pts;
1105 if(is->read_pause_return != AVERROR(ENOSYS)){
1106 is->video_current_pts = is->video_current_pts_drift + av_gettime() / 1000000.0;
1107 }
1108 is->video_current_pts_drift = is->video_current_pts - av_gettime() / 1000000.0;
1109 }
1110 is->paused = !is->paused;
1111 }
1112
1113 static double compute_target_time(double frame_current_pts, VideoState *is)
1114 {
1115 double delay, sync_threshold, diff;
1116
1117 /* compute nominal delay */
1118 delay = frame_current_pts - is->frame_last_pts;
1119 if (delay <= 0 || delay >= 10.0) {
1120 /* if incorrect delay, use previous one */
1121 delay = is->frame_last_delay;
1122 } else {
1123 is->frame_last_delay = delay;
1124 }
1125 is->frame_last_pts = frame_current_pts;
1126
1127 /* update delay to follow master synchronisation source */
1128 if (((is->av_sync_type == AV_SYNC_AUDIO_MASTER && is->audio_st) ||
1129 is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
1130 /* if video is slave, we try to correct big delays by
1131 duplicating or deleting a frame */
1132 diff = get_video_clock(is) - get_master_clock(is);
1133
1134 /* skip or repeat frame. We take into account the
1135 delay to compute the threshold. I still don't know
1136 if it is the best guess */
1137 sync_threshold = FFMAX(AV_SYNC_THRESHOLD, delay);
1138 if (fabs(diff) < AV_NOSYNC_THRESHOLD) {
1139 if (diff <= -sync_threshold)
1140 delay = 0;
1141 else if (diff >= sync_threshold)
1142 delay = 2 * delay;
1143 }
1144 }
1145 is->frame_timer += delay;
1146 #if defined(DEBUG_SYNC)
1147 printf("video: delay=%0.3f actual_delay=%0.3f pts=%0.3f A-V=%f\n",
1148 delay, actual_delay, frame_current_pts, -diff);
1149 #endif
1150
1151 return is->frame_timer;
1152 }
1153
1154 /* called to display each frame */
1155 static void video_refresh_timer(void *opaque)
1156 {
1157 VideoState *is = opaque;
1158 VideoPicture *vp;
1159
1160 SubPicture *sp, *sp2;
1161
1162 if (is->video_st) {
1163 retry:
1164 if (is->pictq_size == 0) {
1165 //nothing to do, no picture to display in the que
1166 } else {
1167 double time= av_gettime()/1000000.0;
1168 double next_target;
1169 /* dequeue the picture */
1170 vp = &is->pictq[is->pictq_rindex];
1171
1172 if(time < vp->target_clock)
1173 return;
1174 /* update current video pts */
1175 is->video_current_pts = vp->pts;
1176 is->video_current_pts_drift = is->video_current_pts - time;
1177 is->video_current_pos = vp->pos;
1178 if(is->pictq_size > 1){
1179 VideoPicture *nextvp= &is->pictq[(is->pictq_rindex+1)%VIDEO_PICTURE_QUEUE_SIZE];
1180 assert(nextvp->target_clock >= vp->target_clock);
1181 next_target= nextvp->target_clock;
1182 }else{
1183 next_target= vp->target_clock + is->video_clock - vp->pts; //FIXME pass durations cleanly
1184 }
1185 if(framedrop && time > next_target){
1186 is->skip_frames *= 1.0 + FRAME_SKIP_FACTOR;
1187 if(is->pictq_size > 1 || time > next_target + 0.5){
1188 /* update queue size and signal for next picture */
1189 if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1190 is->pictq_rindex = 0;
1191
1192 SDL_LockMutex(is->pictq_mutex);
1193 is->pictq_size--;
1194 SDL_CondSignal(is->pictq_cond);
1195 SDL_UnlockMutex(is->pictq_mutex);
1196 goto retry;
1197 }
1198 }
1199
1200 if(is->subtitle_st) {
1201 if (is->subtitle_stream_changed) {
1202 SDL_LockMutex(is->subpq_mutex);
1203
1204 while (is->subpq_size) {
1205 free_subpicture(&is->subpq[is->subpq_rindex]);
1206
1207 /* update queue size and signal for next picture */
1208 if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1209 is->subpq_rindex = 0;
1210
1211 is->subpq_size--;
1212 }
1213 is->subtitle_stream_changed = 0;
1214
1215 SDL_CondSignal(is->subpq_cond);
1216 SDL_UnlockMutex(is->subpq_mutex);
1217 } else {
1218 if (is->subpq_size > 0) {
1219 sp = &is->subpq[is->subpq_rindex];
1220
1221 if (is->subpq_size > 1)
1222 sp2 = &is->subpq[(is->subpq_rindex + 1) % SUBPICTURE_QUEUE_SIZE];
1223 else
1224 sp2 = NULL;
1225
1226 if ((is->video_current_pts > (sp->pts + ((float) sp->sub.end_display_time / 1000)))
1227 || (sp2 && is->video_current_pts > (sp2->pts + ((float) sp2->sub.start_display_time / 1000))))
1228 {
1229 free_subpicture(sp);
1230
1231 /* update queue size and signal for next picture */
1232 if (++is->subpq_rindex == SUBPICTURE_QUEUE_SIZE)
1233 is->subpq_rindex = 0;
1234
1235 SDL_LockMutex(is->subpq_mutex);
1236 is->subpq_size--;
1237 SDL_CondSignal(is->subpq_cond);
1238 SDL_UnlockMutex(is->subpq_mutex);
1239 }
1240 }
1241 }
1242 }
1243
1244 /* display picture */
1245 video_display(is);
1246
1247 /* update queue size and signal for next picture */
1248 if (++is->pictq_rindex == VIDEO_PICTURE_QUEUE_SIZE)
1249 is->pictq_rindex = 0;
1250
1251 SDL_LockMutex(is->pictq_mutex);
1252 is->pictq_size--;
1253 SDL_CondSignal(is->pictq_cond);
1254 SDL_UnlockMutex(is->pictq_mutex);
1255 }
1256 } else if (is->audio_st) {
1257 /* draw the next audio frame */
1258
1259 /* if only audio stream, then display the audio bars (better
1260 than nothing, just to test the implementation */
1261
1262 /* display picture */
1263 video_display(is);
1264 }
1265 if (show_status) {
1266 static int64_t last_time;
1267 int64_t cur_time;
1268 int aqsize, vqsize, sqsize;
1269 double av_diff;
1270
1271 cur_time = av_gettime();
1272 if (!last_time || (cur_time - last_time) >= 30000) {
1273 aqsize = 0;
1274 vqsize = 0;
1275 sqsize = 0;
1276 if (is->audio_st)
1277 aqsize = is->audioq.size;
1278 if (is->video_st)
1279 vqsize = is->videoq.size;
1280 if (is->subtitle_st)
1281 sqsize = is->subtitleq.size;
1282 av_diff = 0;
1283 if (is->audio_st && is->video_st)
1284 av_diff = get_audio_clock(is) - get_video_clock(is);
1285 printf("%7.2f A-V:%7.3f s:%3.1f aq=%5dKB vq=%5dKB sq=%5dB f=%"PRId64"/%"PRId64" \r",
1286 get_master_clock(is), av_diff, FFMAX(is->skip_frames-1, 0), aqsize / 1024, vqsize / 1024, sqsize, is->pts_ctx.num_faulty_dts, is->pts_ctx.num_faulty_pts);
1287 fflush(stdout);
1288 last_time = cur_time;
1289 }
1290 }
1291 }
1292
1293 static void stream_close(VideoState *is)
1294 {
1295 VideoPicture *vp;
1296 int i;
1297 /* XXX: use a special url_shutdown call to abort parse cleanly */
1298 is->abort_request = 1;
1299 SDL_WaitThread(is->parse_tid, NULL);
1300 SDL_WaitThread(is->refresh_tid, NULL);
1301
1302 /* free all pictures */
1303 for(i=0;i<VIDEO_PICTURE_QUEUE_SIZE; i++) {
1304 vp = &is->pictq[i];
1305 #if CONFIG_AVFILTER
1306 if (vp->picref) {
1307 avfilter_unref_buffer(vp->picref);
1308 vp->picref = NULL;
1309 }
1310 #endif
1311 if (vp->bmp) {
1312 SDL_FreeYUVOverlay(vp->bmp);
1313 vp->bmp = NULL;
1314 }
1315 }
1316 SDL_DestroyMutex(is->pictq_mutex);
1317 SDL_DestroyCond(is->pictq_cond);
1318 SDL_DestroyMutex(is->subpq_mutex);
1319 SDL_DestroyCond(is->subpq_cond);
1320 #if !CONFIG_AVFILTER
1321 if (is->img_convert_ctx)
1322 sws_freeContext(is->img_convert_ctx);
1323 #endif
1324 av_free(is);
1325 }
1326
1327 static void do_exit(void)
1328 {
1329 if (cur_stream) {
1330 stream_close(cur_stream);
1331 cur_stream = NULL;
1332 }
1333 uninit_opts();
1334 #if CONFIG_AVFILTER
1335 avfilter_uninit();
1336 #endif
1337 if (show_status)
1338 printf("\n");
1339 SDL_Quit();
1340 av_log(NULL, AV_LOG_QUIET, "");
1341 exit(0);
1342 }
1343
1344 /* allocate a picture (needs to do that in main thread to avoid
1345 potential locking problems */
1346 static void alloc_picture(void *opaque)
1347 {
1348 VideoState *is = opaque;
1349 VideoPicture *vp;
1350
1351 vp = &is->pictq[is->pictq_windex];
1352
1353 if (vp->bmp)
1354 SDL_FreeYUVOverlay(vp->bmp);
1355
1356 #if CONFIG_AVFILTER
1357 if (vp->picref)
1358 avfilter_unref_buffer(vp->picref);
1359 vp->picref = NULL;
1360
1361 vp->width = is->out_video_filter->inputs[0]->w;
1362 vp->height = is->out_video_filter->inputs[0]->h;
1363 vp->pix_fmt = is->out_video_filter->inputs[0]->format;
1364 #else
1365 vp->width = is->video_st->codec->width;
1366 vp->height = is->video_st->codec->height;
1367 vp->pix_fmt = is->video_st->codec->pix_fmt;
1368 #endif
1369
1370 vp->bmp = SDL_CreateYUVOverlay(vp->width, vp->height,
1371 SDL_YV12_OVERLAY,
1372 screen);
1373 if (!vp->bmp || vp->bmp->pitches[0] < vp->width) {
1374 /* SDL allocates a buffer smaller than requested if the video
1375 * overlay hardware is unable to support the requested size. */
1376 fprintf(stderr, "Error: the video system does not support an image\n"
1377 "size of %dx%d pixels. Try using -lowres or -vf \"scale=w:h\"\n"
1378 "to reduce the image size.\n", vp->width, vp->height );
1379 do_exit();
1380 }
1381
1382 SDL_LockMutex(is->pictq_mutex);
1383 vp->allocated = 1;
1384 SDL_CondSignal(is->pictq_cond);
1385 SDL_UnlockMutex(is->pictq_mutex);
1386 }
1387
1388 /**
1389 *
1390 * @param pts the dts of the pkt / pts of the frame and guessed if not known
1391 */
1392 static int queue_picture(VideoState *is, AVFrame *src_frame, double pts, int64_t pos)
1393 {
1394 VideoPicture *vp;
1395 int dst_pix_fmt;
1396 #if CONFIG_AVFILTER
1397 AVPicture pict_src;
1398 #endif
1399 /* wait until we have space to put a new picture */
1400 SDL_LockMutex(is->pictq_mutex);
1401
1402 if(is->pictq_size>=VIDEO_PICTURE_QUEUE_SIZE && !is->refresh)
1403 is->skip_frames= FFMAX(1.0 - FRAME_SKIP_FACTOR, is->skip_frames * (1.0-FRAME_SKIP_FACTOR));
1404
1405 while (is->pictq_size >= VIDEO_PICTURE_QUEUE_SIZE &&
1406 !is->videoq.abort_request) {
1407 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1408 }
1409 SDL_UnlockMutex(is->pictq_mutex);
1410
1411 if (is->videoq.abort_request)
1412 return -1;
1413
1414 vp = &is->pictq[is->pictq_windex];
1415
1416 /* alloc or resize hardware picture buffer */
1417 if (!vp->bmp ||
1418 #if CONFIG_AVFILTER
1419 vp->width != is->out_video_filter->inputs[0]->w ||
1420 vp->height != is->out_video_filter->inputs[0]->h) {
1421 #else
1422 vp->width != is->video_st->codec->width ||
1423 vp->height != is->video_st->codec->height) {
1424 #endif
1425 SDL_Event event;
1426
1427 vp->allocated = 0;
1428
1429 /* the allocation must be done in the main thread to avoid
1430 locking problems */
1431 event.type = FF_ALLOC_EVENT;
1432 event.user.data1 = is;
1433 SDL_PushEvent(&event);
1434
1435 /* wait until the picture is allocated */
1436 SDL_LockMutex(is->pictq_mutex);
1437 while (!vp->allocated && !is->videoq.abort_request) {
1438 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1439 }
1440 SDL_UnlockMutex(is->pictq_mutex);
1441
1442 if (is->videoq.abort_request)
1443 return -1;
1444 }
1445
1446 /* if the frame is not skipped, then display it */
1447 if (vp->bmp) {
1448 AVPicture pict;
1449 #if CONFIG_AVFILTER
1450 if(vp->picref)
1451 avfilter_unref_buffer(vp->picref);
1452 vp->picref = src_frame->opaque;
1453 #endif
1454
1455 /* get a pointer on the bitmap */
1456 SDL_LockYUVOverlay (vp->bmp);
1457
1458 dst_pix_fmt = PIX_FMT_YUV420P;
1459 memset(&pict,0,sizeof(AVPicture));
1460 pict.data[0] = vp->bmp->pixels[0];
1461 pict.data[1] = vp->bmp->pixels[2];
1462 pict.data[2] = vp->bmp->pixels[1];
1463
1464 pict.linesize[0] = vp->bmp->pitches[0];
1465 pict.linesize[1] = vp->bmp->pitches[2];
1466 pict.linesize[2] = vp->bmp->pitches[1];
1467
1468 #if CONFIG_AVFILTER
1469 pict_src.data[0] = src_frame->data[0];
1470 pict_src.data[1] = src_frame->data[1];
1471 pict_src.data[2] = src_frame->data[2];
1472
1473 pict_src.linesize[0] = src_frame->linesize[0];
1474 pict_src.linesize[1] = src_frame->linesize[1];
1475 pict_src.linesize[2] = src_frame->linesize[2];
1476
1477 //FIXME use direct rendering
1478 av_picture_copy(&pict, &pict_src,
1479 vp->pix_fmt, vp->width, vp->height);
1480 #else
1481 sws_flags = av_get_int(sws_opts, "sws_flags", NULL);
1482 is->img_convert_ctx = sws_getCachedContext(is->img_convert_ctx,
1483 vp->width, vp->height, vp->pix_fmt, vp->width, vp->height,
1484 dst_pix_fmt, sws_flags, NULL, NULL, NULL);
1485 if (is->img_convert_ctx == NULL) {
1486 fprintf(stderr, "Cannot initialize the conversion context\n");
1487 exit(1);
1488 }
1489 sws_scale(is->img_convert_ctx, src_frame->data, src_frame->linesize,
1490 0, vp->height, pict.data, pict.linesize);
1491 #endif
1492 /* update the bitmap content */
1493 SDL_UnlockYUVOverlay(vp->bmp);
1494
1495 vp->pts = pts;
1496 vp->pos = pos;
1497
1498 /* now we can update the picture count */
1499 if (++is->pictq_windex == VIDEO_PICTURE_QUEUE_SIZE)
1500 is->pictq_windex = 0;
1501 SDL_LockMutex(is->pictq_mutex);
1502 vp->target_clock= compute_target_time(vp->pts, is);
1503
1504 is->pictq_size++;
1505 SDL_UnlockMutex(is->pictq_mutex);
1506 }
1507 return 0;
1508 }
1509
1510 /**
1511 * compute the exact PTS for the picture if it is omitted in the stream
1512 * @param pts1 the dts of the pkt / pts of the frame
1513 */
1514 static int output_picture2(VideoState *is, AVFrame *src_frame, double pts1, int64_t pos)
1515 {
1516 double frame_delay, pts;
1517
1518 pts = pts1;
1519
1520 if (pts != 0) {
1521 /* update video clock with pts, if present */
1522 is->video_clock = pts;
1523 } else {
1524 pts = is->video_clock;
1525 }
1526 /* update video clock for next frame */
1527 frame_delay = av_q2d(is->video_st->codec->time_base);
1528 /* for MPEG2, the frame can be repeated, so we update the
1529 clock accordingly */
1530 frame_delay += src_frame->repeat_pict * (frame_delay * 0.5);
1531 is->video_clock += frame_delay;
1532
1533 #if defined(DEBUG_SYNC) && 0
1534 printf("frame_type=%c clock=%0.3f pts=%0.3f\n",
1535 av_get_pict_type_char(src_frame->pict_type), pts, pts1);
1536 #endif
1537 return queue_picture(is, src_frame, pts, pos);
1538 }
1539
1540 static int get_video_frame(VideoState *is, AVFrame *frame, int64_t *pts, AVPacket *pkt)
1541 {
1542 int len1, got_picture, i;
1543
1544 if (packet_queue_get(&is->videoq, pkt, 1) < 0)
1545 return -1;
1546
1547 if(pkt->data == flush_pkt.data){
1548 avcodec_flush_buffers(is->video_st->codec);
1549
1550 SDL_LockMutex(is->pictq_mutex);
1551 //Make sure there are no long delay timers (ideally we should just flush the que but thats harder)
1552 for(i=0; i<VIDEO_PICTURE_QUEUE_SIZE; i++){
1553 is->pictq[i].target_clock= 0;
1554 }
1555 while (is->pictq_size && !is->videoq.abort_request) {
1556 SDL_CondWait(is->pictq_cond, is->pictq_mutex);
1557 }
1558 is->video_current_pos= -1;
1559 SDL_UnlockMutex(is->pictq_mutex);
1560
1561 init_pts_correction(&is->pts_ctx);
1562 is->frame_last_pts= AV_NOPTS_VALUE;
1563 is->frame_last_delay = 0;
1564 is->frame_timer = (double)av_gettime() / 1000000.0;
1565 is->skip_frames= 1;
1566 is->skip_frames_index= 0;
1567 return 0;
1568 }
1569
1570 /* NOTE: ipts is the PTS of the _first_ picture beginning in
1571 this packet, if any */
1572 is->video_st->codec->reordered_opaque= pkt->pts;
1573 len1 = avcodec_decode_video2(is->video_st->codec,
1574 frame, &got_picture,
1575 pkt);
1576
1577 if (got_picture) {
1578 if (decoder_reorder_pts == -1) {
1579 *pts = guess_correct_pts(&is->pts_ctx, frame->reordered_opaque, pkt->dts);
1580 } else if (decoder_reorder_pts) {
1581 *pts = frame->reordered_opaque;
1582 } else {
1583 *pts = pkt->dts;
1584 }
1585
1586 if (*pts == AV_NOPTS_VALUE) {
1587 *pts = 0;
1588 }
1589 }
1590
1591 // if (len1 < 0)
1592 // break;
1593 if (got_picture){
1594 is->skip_frames_index += 1;
1595 if(is->skip_frames_index >= is->skip_frames){
1596 is->skip_frames_index -= FFMAX(is->skip_frames, 1.0);
1597 return 1;
1598 }
1599
1600 }
1601 return 0;
1602 }
1603
1604 #if CONFIG_AVFILTER
1605 typedef struct {
1606 VideoState *is;
1607 AVFrame *frame;
1608 int use_dr1;
1609 } FilterPriv;
1610
1611 static int input_get_buffer(AVCodecContext *codec, AVFrame *pic)
1612 {
1613 AVFilterContext *ctx = codec->opaque;
1614 AVFilterBufferRef *ref;
1615 int perms = AV_PERM_WRITE;
1616 int i, w, h, stride[4];
1617 unsigned edge;
1618
1619 if(pic->buffer_hints & FF_BUFFER_HINTS_VALID) {
1620 if(pic->buffer_hints & FF_BUFFER_HINTS_READABLE) perms |= AV_PERM_READ;
1621 if(pic->buffer_hints & FF_BUFFER_HINTS_PRESERVE) perms |= AV_PERM_PRESERVE;
1622 if(pic->buffer_hints & FF_BUFFER_HINTS_REUSABLE) perms |= AV_PERM_REUSE2;
1623 }
1624 if(pic->reference) perms |= AV_PERM_READ | AV_PERM_PRESERVE;
1625
1626 w = codec->width;
1627 h = codec->height;
1628 avcodec_align_dimensions2(codec, &w, &h, stride);
1629 edge = codec->flags & CODEC_FLAG_EMU_EDGE ? 0 : avcodec_get_edge_width();
1630 w += edge << 1;
1631 h += edge << 1;
1632
1633 if(!(ref = avfilter_get_video_buffer(ctx->outputs[0], perms, w, h)))
1634 return -1;
1635
1636 ref->video->w = codec->width;
1637 ref->video->h = codec->height;
1638 for(i = 0; i < 4; i ++) {
1639 unsigned hshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_w : 0;
1640 unsigned vshift = (i == 1 || i == 2) ? av_pix_fmt_descriptors[ref->format].log2_chroma_h : 0;
1641
1642 if (ref->data[i]) {
1643 ref->data[i] += (edge >> hshift) + ((edge * ref->linesize[i]) >> vshift);
1644 }
1645 pic->data[i] = ref->data[i];
1646 pic->linesize[i] = ref->linesize[i];
1647 }
1648 pic->opaque = ref;
1649 pic->age = INT_MAX;
1650 pic->type = FF_BUFFER_TYPE_USER;
1651 pic->reordered_opaque = codec->reordered_opaque;
1652 return 0;
1653 }
1654
1655 static void input_release_buffer(AVCodecContext *codec, AVFrame *pic)
1656 {
1657 memset(pic->data, 0, sizeof(pic->data));
1658 avfilter_unref_buffer(pic->opaque);
1659 }
1660
1661 static int input_reget_buffer(AVCodecContext *codec, AVFrame *pic)
1662 {
1663 AVFilterBufferRef *ref = pic->opaque;
1664
1665 if (pic->data[0] == NULL) {
1666 pic->buffer_hints |= FF_BUFFER_HINTS_READABLE;
1667 return codec->get_buffer(codec, pic);
1668 }
1669
1670 if ((codec->width != ref->video->w) || (codec->height != ref->video->h) ||
1671 (codec->pix_fmt != ref->format)) {
1672 av_log(codec, AV_LOG_ERROR, "Picture properties changed.\n");
1673 return -1;
1674 }
1675
1676 pic->reordered_opaque = codec->reordered_opaque;
1677 return 0;
1678 }
1679
1680 static int input_init(AVFilterContext *ctx, const char *args, void *opaque)
1681 {
1682 FilterPriv *priv = ctx->priv;
1683 AVCodecContext *codec;
1684 if(!opaque) return -1;
1685
1686 priv->is = opaque;
1687 codec = priv->is->video_st->codec;
1688 codec->opaque = ctx;
1689 if(codec->codec->capabilities & CODEC_CAP_DR1) {
1690 priv->use_dr1 = 1;
1691 codec->get_buffer = input_get_buffer;
1692 codec->release_buffer = input_release_buffer;
1693 codec->reget_buffer = input_reget_buffer;
1694 }
1695
1696 priv->frame = avcodec_alloc_frame();
1697
1698 return 0;
1699 }
1700
1701 static void input_uninit(AVFilterContext *ctx)
1702 {
1703 FilterPriv *priv = ctx->priv;
1704 av_free(priv->frame);
1705 }
1706
1707 static int input_request_frame(AVFilterLink *link)
1708 {
1709 FilterPriv *priv = link->src->priv;
1710 AVFilterBufferRef *picref;
1711 int64_t pts = 0;
1712 AVPacket pkt;
1713 int ret;
1714
1715 while (!(ret = get_video_frame(priv->is, priv->frame, &pts, &pkt)))
1716 av_free_packet(&pkt);
1717 if (ret < 0)
1718 return -1;
1719
1720 if(priv->use_dr1) {
1721 picref = avfilter_ref_buffer(priv->frame->opaque, ~0);
1722 } else {
1723 picref = avfilter_get_video_buffer(link, AV_PERM_WRITE, link->w, link->h);
1724 av_image_copy(picref->data, picref->linesize,
1725 priv->frame->data, priv->frame->linesize,
1726 picref->format, link->w, link->h);
1727 }
1728 av_free_packet(&pkt);
1729
1730 picref->pts = pts;
1731 picref->pos = pkt.pos;
1732 picref->video->pixel_aspect = priv->is->video_st->codec->sample_aspect_ratio;
1733 avfilter_start_frame(link, picref);
1734 avfilter_draw_slice(link, 0, link->h, 1);
1735 avfilter_end_frame(link);
1736
1737 return 0;
1738 }
1739
1740 static int input_query_formats(AVFilterContext *ctx)
1741 {
1742 FilterPriv *priv = ctx->priv;
1743 enum PixelFormat pix_fmts[] = {
1744 priv->is->video_st->codec->pix_fmt, PIX_FMT_NONE
1745 };
1746
1747 avfilter_set_common_formats(ctx, avfilter_make_format_list(pix_fmts));
1748 return 0;
1749 }
1750
1751 static int input_config_props(AVFilterLink *link)
1752 {
1753 FilterPriv *priv = link->src->priv;
1754 AVCodecContext *c = priv->is->video_st->codec;
1755
1756 link->w = c->width;
1757 link->h = c->height;
1758 link->time_base = priv->is->video_st->time_base;
1759
1760 return 0;
1761 }
1762
1763 static AVFilter input_filter =
1764 {
1765 .name = "ffplay_input",
1766
1767 .priv_size = sizeof(FilterPriv),
1768
1769 .init = input_init,
1770 .uninit = input_uninit,
1771
1772 .query_formats = input_query_formats,
1773
1774 .inputs = (AVFilterPad[]) {{ .name = NULL }},
1775 .outputs = (AVFilterPad[]) {{ .name = "default",
1776 .type = AVMEDIA_TYPE_VIDEO,
1777 .request_frame = input_request_frame,
1778 .config_props = input_config_props, },
1779 { .name = NULL }},
1780 };
1781
1782 #endif /* CONFIG_AVFILTER */
1783
1784 static int video_thread(void *arg)
1785 {
1786 VideoState *is = arg;
1787 AVFrame *frame= avcodec_alloc_frame();
1788 int64_t pts_int;
1789 double pts;
1790 int ret;
1791
1792 #if CONFIG_AVFILTER
1793 int64_t pos;
1794 char sws_flags_str[128];
1795 FFSinkContext ffsink_ctx = { .pix_fmt = PIX_FMT_YUV420P };
1796 AVFilterContext *filt_src = NULL, *filt_out = NULL;
1797 AVFilterGraph *graph = avfilter_graph_alloc();
1798 snprintf(sws_flags_str, sizeof(sws_flags_str), "flags=%d", sws_flags);
1799 graph->scale_sws_opts = av_strdup(sws_flags_str);
1800
1801 if (avfilter_open(&filt_src, &input_filter, "src") < 0) goto the_end;
1802 if (avfilter_open(&filt_out, &ffsink , "out") < 0) goto the_end;
1803
1804 if(avfilter_init_filter(filt_src, NULL, is)) goto the_end;
1805 if(avfilter_init_filter(filt_out, NULL, &ffsink_ctx)) goto the_end;
1806
1807
1808 if(vfilters) {
1809 AVFilterInOut *outputs = av_malloc(sizeof(AVFilterInOut));
1810 AVFilterInOut *inputs = av_malloc(sizeof(AVFilterInOut));
1811
1812 outputs->name = av_strdup("in");
1813 outputs->filter_ctx = filt_src;
1814 outputs->pad_idx = 0;
1815 outputs->next = NULL;
1816
1817 inputs->name = av_strdup("out");
1818 inputs->filter_ctx = filt_out;
1819 inputs->pad_idx = 0;
1820 inputs->next = NULL;
1821
1822 if (avfilter_graph_parse(graph, vfilters, inputs, outputs, NULL) < 0)
1823 goto the_end;
1824 av_freep(&vfilters);
1825 } else {
1826 if(avfilter_link(filt_src, 0, filt_out, 0) < 0) goto the_end;
1827 }
1828 avfilter_graph_add_filter(graph, filt_src);
1829 avfilter_graph_add_filter(graph, filt_out);
1830
1831 if (avfilter_graph_config(graph, NULL) < 0)
1832 goto the_end;
1833
1834 is->out_video_filter = filt_out;
1835 #endif
1836
1837 for(;;) {
1838 #if !CONFIG_AVFILTER
1839 AVPacket pkt;
1840 #else
1841 AVFilterBufferRef *picref;
1842 AVRational tb;
1843 #endif
1844 while (is->paused && !is->videoq.abort_request)
1845 SDL_Delay(10);
1846 #if CONFIG_AVFILTER
1847 ret = get_filtered_video_frame(filt_out, frame, &picref, &tb);
1848 if (picref) {
1849 pts_int = picref->pts;
1850 pos = picref->pos;
1851 frame->opaque = picref;
1852 }
1853
1854 if (av_cmp_q(tb, is->video_st->time_base)) {
1855 int64_t pts1 = pts_int;
1856 pts_int = av_rescale_q(pts_int, tb, is->video_st->time_base);
1857 av_log(NULL, AV_LOG_DEBUG, "video_thread(): "
1858 "tb:%d/%d pts:%"PRId64" -> tb:%d/%d pts:%"PRId64"\n",
1859 tb.num, tb.den, pts1,
1860 is->video_st->time_base.num, is->video_st->time_base.den, pts_int);
1861 }
1862 #else
1863 ret = get_video_frame(is, frame, &pts_int, &pkt);
1864 #endif
1865
1866 if (ret < 0) goto the_end;
1867
1868 if (!ret)
1869 continue;
1870
1871 pts = pts_int*av_q2d(is->video_st->time_base);
1872
1873 #if CONFIG_AVFILTER
1874 ret = output_picture2(is, frame, pts, pos);
1875 #else
1876 ret = output_picture2(is, frame, pts, pkt.pos);
1877 av_free_packet(&pkt);
1878 #endif
1879 if (ret < 0)
1880 goto the_end;
1881
1882 if (step)
1883 if (cur_stream)
1884 stream_pause(cur_stream);
1885 }
1886 the_end:
1887 #if CONFIG_AVFILTER
1888 avfilter_graph_free(graph);
1889 av_freep(&graph);
1890 #endif
1891 av_free(frame);
1892 return 0;
1893 }
1894
1895 static int subtitle_thread(void *arg)
1896 {
1897 VideoState *is = arg;
1898 SubPicture *sp;
1899 AVPacket pkt1, *pkt = &pkt1;
1900 int len1, got_subtitle;
1901 double pts;
1902 int i, j;
1903 int r, g, b, y, u, v, a;
1904
1905 for(;;) {
1906 while (is->paused && !is->subtitleq.abort_request) {
1907 SDL_Delay(10);
1908 }
1909 if (packet_queue_get(&is->subtitleq, pkt, 1) < 0)
1910 break;
1911
1912 if(pkt->data == flush_pkt.data){
1913 avcodec_flush_buffers(is->subtitle_st->codec);
1914 continue;
1915 }
1916 SDL_LockMutex(is->subpq_mutex);
1917 while (is->subpq_size >= SUBPICTURE_QUEUE_SIZE &&
1918 !is->subtitleq.abort_request) {
1919 SDL_CondWait(is->subpq_cond, is->subpq_mutex);
1920 }
1921 SDL_UnlockMutex(is->subpq_mutex);
1922
1923 if (is->subtitleq.abort_request)
1924 goto the_end;
1925
1926 sp = &is->subpq[is->subpq_windex];
1927
1928 /* NOTE: ipts is the PTS of the _first_ picture beginning in
1929 this packet, if any */
1930 pts = 0;
1931 if (pkt->pts != AV_NOPTS_VALUE)
1932 pts = av_q2d(is->subtitle_st->time_base)*pkt->pts;
1933
1934 len1 = avcodec_decode_subtitle2(is->subtitle_st->codec,
1935 &sp->sub, &got_subtitle,
1936 pkt);
1937 // if (len1 < 0)
1938 // break;
1939 if (got_subtitle && sp->sub.format == 0) {
1940 sp->pts = pts;
1941
1942 for (i = 0; i < sp->sub.num_rects; i++)
1943 {
1944 for (j = 0; j < sp->sub.rects[i]->nb_colors; j++)
1945 {
1946 RGBA_IN(r, g, b, a, (uint32_t*)sp->sub.rects[i]->pict.data[1] + j);
1947 y = RGB_TO_Y_CCIR(r, g, b);
1948 u = RGB_TO_U_CCIR(r, g, b, 0);
1949 v = RGB_TO_V_CCIR(r, g, b, 0);
1950 YUVA_OUT((uint32_t*)sp->sub.rects[i]->pict.data[1] + j, y, u, v, a);
1951 }
1952 }
1953
1954 /* now we can update the picture count */
1955 if (++is->subpq_windex == SUBPICTURE_QUEUE_SIZE)
1956 is->subpq_windex = 0;
1957 SDL_LockMutex(is->subpq_mutex);
1958 is->subpq_size++;
1959 SDL_UnlockMutex(is->subpq_mutex);
1960 }
1961 av_free_packet(pkt);
1962 // if (step)
1963 // if (cur_stream)
1964 // stream_pause(cur_stream);
1965 }
1966 the_end:
1967 return 0;
1968 }
1969
1970 /* copy samples for viewing in editor window */
1971 static void update_sample_display(VideoState *is, short *samples, int samples_size)
1972 {
1973 int size, len, channels;
1974
1975 channels = is->audio_st->codec->channels;
1976
1977 size = samples_size / sizeof(short);
1978 while (size > 0) {
1979 len = SAMPLE_ARRAY_SIZE - is->sample_array_index;
1980 if (len > size)
1981 len = size;
1982 memcpy(is->sample_array + is->sample_array_index, samples, len * sizeof(short));
1983 samples += len;
1984 is->sample_array_index += len;
1985 if (is->sample_array_index >= SAMPLE_ARRAY_SIZE)
1986 is->sample_array_index = 0;
1987 size -= len;
1988 }
1989 }
1990
1991 /* return the new audio buffer size (samples can be added or deleted
1992 to get better sync if video or external master clock) */
1993 static int synchronize_audio(VideoState *is, short *samples,
1994 int samples_size1, double pts)
1995 {
1996 int n, samples_size;
1997 double ref_clock;
1998
1999 n = 2 * is->audio_st->codec->channels;
2000 samples_size = samples_size1;
2001
2002 /* if not master, then we try to remove or add samples to correct the clock */
2003 if (((is->av_sync_type == AV_SYNC_VIDEO_MASTER && is->video_st) ||
2004 is->av_sync_type == AV_SYNC_EXTERNAL_CLOCK)) {
2005 double diff, avg_diff;
2006 int wanted_size, min_size, max_size, nb_samples;
2007
2008 ref_clock = get_master_clock(is);
2009 diff = get_audio_clock(is) - ref_clock;
2010
2011 if (diff < AV_NOSYNC_THRESHOLD) {
2012 is->audio_diff_cum = diff + is->audio_diff_avg_coef * is->audio_diff_cum;
2013 if (is->audio_diff_avg_count < AUDIO_DIFF_AVG_NB) {
2014 /* not enough measures to have a correct estimate */
2015 is->audio_diff_avg_count++;
2016 } else {
2017 /* estimate the A-V difference */
2018 avg_diff = is->audio_diff_cum * (1.0 - is->audio_diff_avg_coef);
2019
2020 if (fabs(avg_diff) >= is->audio_diff_threshold) {
2021 wanted_size = samples_size + ((int)(diff * is->audio_st->codec->sample_rate) * n);
2022 nb_samples = samples_size / n;
2023
2024 min_size = ((nb_samples * (100 - SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
2025 max_size = ((nb_samples * (100 + SAMPLE_CORRECTION_PERCENT_MAX)) / 100) * n;
2026 if (wanted_size < min_size)
2027 wanted_size = min_size;
2028 else if (wanted_size > max_size)
2029 wanted_size = max_size;
2030
2031 /* add or remove samples to correction the synchro */
2032 if (wanted_size < samples_size) {
2033 /* remove samples */
2034 samples_size = wanted_size;
2035 } else if (wanted_size > samples_size) {
2036 uint8_t *samples_end, *q;
2037 int nb;
2038
2039 /* add samples */
2040 nb = (samples_size - wanted_size);
2041 samples_end = (uint8_t *)samples + samples_size - n;
2042 q = samples_end + n;
2043 while (nb > 0) {
2044 memcpy(q, samples_end, n);
2045 q += n;
2046 nb -= n;
2047 }
2048 samples_size = wanted_size;
2049 }
2050 }
2051 #if 0
2052 printf("diff=%f adiff=%f sample_diff=%d apts=%0.3f vpts=%0.3f %f\n",
2053 diff, avg_diff, samples_size - samples_size1,
2054 is->audio_clock, is->video_clock, is->audio_diff_threshold);
2055 #endif
2056 }
2057 } else {
2058 /* too big difference : may be initial PTS errors, so
2059 reset A-V filter */
2060 is->audio_diff_avg_count = 0;
2061 is->audio_diff_cum = 0;
2062 }
2063 }
2064
2065 return samples_size;
2066 }
2067
2068 /* decode one audio frame and returns its uncompressed size */
2069 static int audio_decode_frame(VideoState *is, double *pts_ptr)
2070 {
2071 AVPacket *pkt_temp = &is->audio_pkt_temp;
2072 AVPacket *pkt = &is->audio_pkt;
2073 AVCodecContext *dec= is->audio_st->codec;
2074 int n, len1, data_size;
2075 double pts;
2076
2077 for(;;) {
2078 /* NOTE: the audio packet can contain several frames */
2079 while (pkt_temp->size > 0) {
2080 data_size = sizeof(is->audio_buf1);
2081 len1 = avcodec_decode_audio3(dec,
2082 (int16_t *)is->audio_buf1, &data_size,
2083 pkt_temp);
2084 if (len1 < 0) {
2085 /* if error, we skip the frame */
2086 pkt_temp->size = 0;
2087 break;
2088 }
2089
2090 pkt_temp->data += len1;
2091 pkt_temp->size -= len1;
2092 if (data_size <= 0)
2093 continue;
2094
2095 if (dec->sample_fmt != is->audio_src_fmt) {
2096 if (is->reformat_ctx)
2097 av_audio_convert_free(is->reformat_ctx);
2098 is->reformat_ctx= av_audio_convert_alloc(SAMPLE_FMT_S16, 1,
2099 dec->sample_fmt, 1, NULL, 0);
2100 if (!is->reformat_ctx) {
2101 fprintf(stderr, "Cannot convert %s sample format to %s sample format\n",
2102 av_get_sample_fmt_name(dec->sample_fmt),
2103 av_get_sample_fmt_name(SAMPLE_FMT_S16));
2104 break;
2105 }
2106 is->audio_src_fmt= dec->sample_fmt;
2107 }
2108
2109 if (is->reformat_ctx) {
2110 const void *ibuf[6]= {is->audio_buf1};
2111 void *obuf[6]= {is->audio_buf2};
2112 int istride[6]= {av_get_bits_per_sample_fmt(dec->sample_fmt)/8};
2113 int ostride[6]= {2};
2114 int len= data_size/istride[0];
2115 if (av_audio_convert(is->reformat_ctx, obuf, ostride, ibuf, istride, len)<0) {
2116 printf("av_audio_convert() failed\n");
2117 break;
2118 }
2119 is->audio_buf= is->audio_buf2;
2120 /* FIXME: existing code assume that data_size equals framesize*channels*2
2121 remove this legacy cruft */
2122 data_size= len*2;
2123 }else{
2124 is->audio_buf= is->audio_buf1;
2125 }
2126
2127 /* if no pts, then compute it */
2128 pts = is->audio_clock;
2129 *pts_ptr = pts;
2130 n = 2 * dec->channels;
2131 is->audio_clock += (double)data_size /
2132 (double)(n * dec->sample_rate);
2133 #if defined(DEBUG_SYNC)
2134 {
2135 static double last_clock;
2136 printf("audio: delay=%0.3f clock=%0.3f pts=%0.3f\n",
2137 is->audio_clock - last_clock,
2138 is->audio_clock, pts);
2139 last_clock = is->audio_clock;
2140 }
2141 #endif
2142 return data_size;
2143 }
2144
2145 /* free the current packet */
2146 if (pkt->data)
2147 av_free_packet(pkt);
2148
2149 if (is->paused || is->audioq.abort_request) {
2150 return -1;
2151 }
2152
2153 /* read next packet */
2154 if (packet_queue_get(&is->audioq, pkt, 1) < 0)
2155 return -1;
2156 if(pkt->data == flush_pkt.data){
2157 avcodec_flush_buffers(dec);
2158 continue;
2159 }
2160
2161 pkt_temp->data = pkt->data;
2162 pkt_temp->size = pkt->size;
2163
2164 /* if update the audio clock with the pts */
2165 if (pkt->pts != AV_NOPTS_VALUE) {
2166 is->audio_clock = av_q2d(is->audio_st->time_base)*pkt->pts;
2167 }
2168 }
2169 }
2170
2171 /* get the current audio output buffer size, in samples. With SDL, we
2172 cannot have a precise information */
2173 static int audio_write_get_buf_size(VideoState *is)
2174 {
2175 return is->audio_buf_size - is->audio_buf_index;
2176 }
2177
2178
2179 /* prepare a new audio buffer */
2180 static void sdl_audio_callback(void *opaque, Uint8 *stream, int len)
2181 {
2182 VideoState *is = opaque;
2183 int audio_size, len1;
2184 double pts;
2185
2186 audio_callback_time = av_gettime();
2187
2188 while (len > 0) {
2189 if (is->audio_buf_index >= is->audio_buf_size) {
2190 audio_size = audio_decode_frame(is, &pts);
2191 if (audio_size < 0) {
2192 /* if error, just output silence */
2193 is->audio_buf = is->audio_buf1;
2194 is->audio_buf_size = 1024;
2195 memset(is->audio_buf, 0, is->audio_buf_size);
2196 } else {
2197 if (is->show_audio)
2198 update_sample_display(is, (int16_t *)is->audio_buf, audio_size);
2199 audio_size = synchronize_audio(is, (int16_t *)is->audio_buf, audio_size,
2200 pts);
2201 is->audio_buf_size = audio_size;
2202 }
2203 is->audio_buf_index = 0;
2204 }
2205 len1 = is->audio_buf_size - is->audio_buf_index;
2206 if (len1 > len)
2207 len1 = len;
2208 memcpy(stream, (uint8_t *)is->audio_buf + is->audio_buf_index, len1);
2209 len -= len1;
2210 stream += len1;
2211 is->audio_buf_index += len1;
2212 }
2213 }
2214
2215 /* open a given stream. Return 0 if OK */
2216 static int stream_component_open(VideoState *is, int stream_index)
2217 {
2218 AVFormatContext *ic = is->ic;
2219 AVCodecContext *avctx;
2220 AVCodec *codec;
2221 SDL_AudioSpec wanted_spec, spec;
2222
2223 if (stream_index < 0 || stream_index >= ic->nb_streams)
2224 return -1;
2225 avctx = ic->streams[stream_index]->codec;
2226
2227 /* prepare audio output */
2228 if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2229 if (avctx->channels > 0) {
2230 avctx->request_channels = FFMIN(2, avctx->channels);
2231 } else {
2232 avctx->request_channels = 2;
2233 }
2234 }
2235
2236 codec = avcodec_find_decoder(avctx->codec_id);
2237 avctx->debug_mv = debug_mv;
2238 avctx->debug = debug;
2239 avctx->workaround_bugs = workaround_bugs;
2240 avctx->lowres = lowres;
2241 if(lowres) avctx->flags |= CODEC_FLAG_EMU_EDGE;
2242 avctx->idct_algo= idct;
2243 if(fast) avctx->flags2 |= CODEC_FLAG2_FAST;
2244 avctx->skip_frame= skip_frame;
2245 avctx->skip_idct= skip_idct;
2246 avctx->skip_loop_filter= skip_loop_filter;
2247 avctx->error_recognition= error_recognition;
2248 avctx->error_concealment= error_concealment;
2249 avcodec_thread_init(avctx, thread_count);
2250
2251 set_context_opts(avctx, avcodec_opts[avctx->codec_type], 0, codec);
2252
2253 if (!codec ||
2254 avcodec_open(avctx, codec) < 0)
2255 return -1;
2256
2257 /* prepare audio output */
2258 if (avctx->codec_type == AVMEDIA_TYPE_AUDIO) {
2259 wanted_spec.freq = avctx->sample_rate;
2260 wanted_spec.format = AUDIO_S16SYS;
2261 wanted_spec.channels = avctx->channels;
2262 wanted_spec.silence = 0;
2263 wanted_spec.samples = SDL_AUDIO_BUFFER_SIZE;
2264 wanted_spec.callback = sdl_audio_callback;
2265 wanted_spec.userdata = is;
2266 if (SDL_OpenAudio(&wanted_spec, &spec) < 0) {
2267 fprintf(stderr, "SDL_OpenAudio: %s\n", SDL_GetError());
2268 return -1;
2269 }
2270 is->audio_hw_buf_size = spec.size;
2271 is->audio_src_fmt= SAMPLE_FMT_S16;
2272 }
2273
2274 ic->streams[stream_index]->discard = AVDISCARD_DEFAULT;
2275 switch(avctx->codec_type) {
2276 case AVMEDIA_TYPE_AUDIO:
2277 is->audio_stream = stream_index;
2278 is->audio_st = ic->streams[stream_index];
2279 is->audio_buf_size = 0;
2280 is->audio_buf_index = 0;
2281
2282 /* init averaging filter */
2283 is->audio_diff_avg_coef = exp(log(0.01) / AUDIO_DIFF_AVG_NB);
2284 is->audio_diff_avg_count = 0;
2285 /* since we do not have a precise anough audio fifo fullness,
2286 we correct audio sync only if larger than this threshold */
2287 is->audio_diff_threshold = 2.0 * SDL_AUDIO_BUFFER_SIZE / avctx->sample_rate;
2288
2289 memset(&is->audio_pkt, 0, sizeof(is->audio_pkt));
2290 packet_queue_init(&is->audioq);
2291 SDL_PauseAudio(0);
2292 break;
2293 case AVMEDIA_TYPE_VIDEO:
2294 is->video_stream = stream_index;
2295 is->video_st = ic->streams[stream_index];
2296
2297 // is->video_current_pts_time = av_gettime();
2298
2299 packet_queue_init(&is->videoq);
2300 is->video_tid = SDL_CreateThread(video_thread, is);
2301 break;
2302 case AVMEDIA_TYPE_SUBTITLE:
2303 is->subtitle_stream = stream_index;
2304 is->subtitle_st = ic->streams[stream_index];
2305 packet_queue_init(&is->subtitleq);
2306
2307 is->subtitle_tid = SDL_CreateThread(subtitle_thread, is);
2308 break;
2309 default:
2310 break;
2311 }
2312 return 0;
2313 }
2314
2315 static void stream_component_close(VideoState *is, int stream_index)
2316 {
2317 AVFormatContext *ic = is->ic;
2318 AVCodecContext *avctx;
2319
2320 if (stream_index < 0 || stream_index >= ic->nb_streams)
2321 return;
2322 avctx = ic->streams[stream_index]->codec;
2323
2324 switch(avctx->codec_type) {
2325 case AVMEDIA_TYPE_AUDIO:
2326 packet_queue_abort(&is->audioq);
2327
2328 SDL_CloseAudio();
2329
2330 packet_queue_end(&is->audioq);
2331 if (is->reformat_ctx)
2332 av_audio_convert_free(is->reformat_ctx);
2333 is->reformat_ctx = NULL;
2334 break;
2335 case AVMEDIA_TYPE_VIDEO:
2336 packet_queue_abort(&is->videoq);
2337
2338 /* note: we also signal this mutex to make sure we deblock the
2339 video thread in all cases */
2340 SDL_LockMutex(is->pictq_mutex);
2341 SDL_CondSignal(is->pictq_cond);
2342 SDL_UnlockMutex(is->pictq_mutex);
2343
2344 SDL_WaitThread(is->video_tid, NULL);
2345
2346 packet_queue_end(&is->videoq);
2347 break;
2348 case AVMEDIA_TYPE_SUBTITLE:
2349 packet_queue_abort(&is->subtitleq);
2350
2351 /* note: we also signal this mutex to make sure we deblock the
2352 video thread in all cases */
2353 SDL_LockMutex(is->subpq_mutex);
2354 is->subtitle_stream_changed = 1;
2355
2356 SDL_CondSignal(is->subpq_cond);
2357 SDL_UnlockMutex(is->subpq_mutex);
2358
2359 SDL_WaitThread(is->subtitle_tid, NULL);
2360
2361 packet_queue_end(&is->subtitleq);
2362 break;
2363 default:
2364 break;
2365 }
2366
2367 ic->streams[stream_index]->discard = AVDISCARD_ALL;
2368 avcodec_close(avctx);
2369 switch(avctx->codec_type) {
2370 case AVMEDIA_TYPE_AUDIO:
2371 is->audio_st = NULL;
2372 is->audio_stream = -1;
2373 break;
2374 case AVMEDIA_TYPE_VIDEO:
2375 is->video_st = NULL;
2376 is->video_stream = -1;
2377 break;
2378 case AVMEDIA_TYPE_SUBTITLE:
2379 is->subtitle_st = NULL;
2380 is->subtitle_stream = -1;
2381 break;
2382 default:
2383 break;
2384 }
2385 }
2386
2387 /* since we have only one decoding thread, we can use a global
2388 variable instead of a thread local variable */
2389 static VideoState *global_video_state;
2390
2391 static int decode_interrupt_cb(void)
2392 {
2393 return (global_video_state && global_video_state->abort_request);
2394 }
2395
2396 /* this thread gets the stream from the disk or the network */
2397 static int decode_thread(void *arg)
2398 {
2399 VideoState *is = arg;
2400 AVFormatContext *ic;
2401 int err, i, ret;
2402 int st_index[AVMEDIA_TYPE_NB];
2403 int st_count[AVMEDIA_TYPE_NB]={0};
2404 int st_best_packet_count[AVMEDIA_TYPE_NB];
2405 AVPacket pkt1, *pkt = &pkt1;
2406 AVFormatParameters params, *ap = &params;
2407 int eof=0;
2408 int pkt_in_play_range = 0;
2409
2410 ic = avformat_alloc_context();
2411
2412 memset(st_index, -1, sizeof(st_index));
2413 memset(st_best_packet_count, -1, sizeof(st_best_packet_count));
2414 is->video_stream = -1;
2415 is->audio_stream = -1;
2416 is->subtitle_stream = -1;
2417
2418 global_video_state = is;
2419 url_set_interrupt_cb(decode_interrupt_cb);
2420
2421 memset(ap, 0, sizeof(*ap));
2422
2423 ap->prealloced_context = 1;
2424 ap->width = frame_width;
2425 ap->height= frame_height;
2426 ap->time_base= (AVRational){1, 25};
2427 ap->pix_fmt = frame_pix_fmt;
2428
2429 set_context_opts(ic, avformat_opts, AV_OPT_FLAG_DECODING_PARAM, NULL);
2430
2431 err = av_open_input_file(&ic, is->filename, is->iformat, 0, ap);
2432 if (err < 0) {
2433 print_error(is->filename, err);
2434 ret = -1;
2435 goto fail;
2436 }
2437 is->ic = ic;
2438
2439 if(genpts)
2440 ic->flags |= AVFMT_FLAG_GENPTS;
2441
2442 err = av_find_stream_info(ic);
2443 if (err < 0) {
2444 fprintf(stderr, "%s: could not find codec parameters\n", is->filename);
2445 ret = -1;
2446 goto fail;
2447 }
2448 if(ic->pb)
2449 ic->pb->eof_reached= 0; //FIXME hack, ffplay maybe should not use url_feof() to test for the end
2450
2451 if(seek_by_bytes<0)
2452 seek_by_bytes= !!(ic->iformat->flags & AVFMT_TS_DISCONT);
2453
2454 /* if seeking requested, we execute it */
2455 if (start_time != AV_NOPTS_VALUE) {
2456 int64_t timestamp;
2457
2458 timestamp = start_time;
2459 /* add the stream start time */
2460 if (ic->start_time != AV_NOPTS_VALUE)
2461 timestamp += ic->start_time;
2462 ret = avformat_seek_file(ic, -1, INT64_MIN, timestamp, INT64_MAX, 0);
2463 if (ret < 0) {
2464 fprintf(stderr, "%s: could not seek to position %0.3f\n",
2465 is->filename, (double)timestamp / AV_TIME_BASE);
2466 }
2467 }
2468
2469 for(i = 0; i < ic->nb_streams; i++) {
2470 AVStream *st= ic->streams[i];
2471 AVCodecContext *avctx = st->codec;
2472 ic->streams[i]->discard = AVDISCARD_ALL;
2473 if(avctx->codec_type >= (unsigned)AVMEDIA_TYPE_NB)
2474 continue;
2475 if(st_count[avctx->codec_type]++ != wanted_stream[avctx->codec_type] && wanted_stream[avctx->codec_type] >= 0)
2476 continue;
2477
2478 if(st_best_packet_count[avctx->codec_type] >= st->codec_info_nb_frames)
2479 continue;
2480 st_best_packet_count[avctx->codec_type]= st->codec_info_nb_frames;
2481
2482 switch(avctx->codec_type) {
2483 case AVMEDIA_TYPE_AUDIO:
2484 if (!audio_disable)
2485 st_index[AVMEDIA_TYPE_AUDIO] = i;
2486 break;
2487 case AVMEDIA_TYPE_VIDEO:
2488 case AVMEDIA_TYPE_SUBTITLE:
2489 if (!video_disable)
2490 st_index[avctx->codec_type] = i;
2491 break;
2492 default:
2493 break;
2494 }
2495 }
2496 if (show_status) {
2497 dump_format(ic, 0, is->filename, 0);
2498 }
2499
2500 /* open the streams */
2501 if (st_index[AVMEDIA_TYPE_AUDIO] >= 0) {
2502 stream_component_open(is, st_index[AVMEDIA_TYPE_AUDIO]);
2503 }
2504
2505 ret=-1;
2506 if (st_index[AVMEDIA_TYPE_VIDEO] >= 0) {
2507 ret= stream_component_open(is, st_index[AVMEDIA_TYPE_VIDEO]);
2508 }
2509 is->refresh_tid = SDL_CreateThread(refresh_thread, is);
2510 if(ret<0) {
2511 if (!display_disable)
2512 is->show_audio = 2;
2513 }
2514
2515 if (st_index[AVMEDIA_TYPE_SUBTITLE] >= 0) {
2516 stream_component_open(is, st_index[AVMEDIA_TYPE_SUBTITLE]);
2517 }
2518
2519 if (is->video_stream < 0 && is->audio_stream < 0) {
2520 fprintf(stderr, "%s: could not open codecs\n", is->filename);
2521 ret = -1;
2522 goto fail;
2523 }
2524
2525 for(;;) {
2526 if (is->abort_request)
2527 break;
2528 if (is->paused != is->last_paused) {
2529 is->last_paused = is->paused;
2530 if (is->paused)
2531 is->read_pause_return= av_read_pause(ic);
2532 else
2533 av_read_play(ic);
2534 }
2535 #if CONFIG_RTSP_DEMUXER
2536 if (is->paused && !strcmp(ic->iformat->name, "rtsp")) {
2537 /* wait 10 ms to avoid trying to get another packet */
2538 /* XXX: horrible */
2539 SDL_Delay(10);
2540 continue;
2541 }
2542 #endif
2543 if (is->seek_req) {
2544 int64_t seek_target= is->seek_pos;
2545 int64_t seek_min= is->seek_rel > 0 ? seek_target - is->seek_rel + 2: INT64_MIN;
2546 int64_t seek_max= is->seek_rel < 0 ? seek_target - is->seek_rel - 2: INT64_MAX;
2547 //FIXME the +-2 is due to rounding being not done in the correct direction in generation
2548 // of the seek_pos/seek_rel variables
2549
2550 ret = avformat_seek_file(is->ic, -1, seek_min, seek_target, seek_max, is->seek_flags);
2551 if (ret < 0) {
2552 fprintf(stderr, "%s: error while seeking\n", is->ic->filename);
2553 }else{
2554 if (is->audio_stream >= 0) {
2555 packet_queue_flush(&is->audioq);
2556 packet_queue_put(&is->audioq, &flush_pkt);
2557 }
2558 if (is->subtitle_stream >= 0) {
2559 packet_queue_flush(&is->subtitleq);
2560 packet_queue_put(&is->subtitleq, &flush_pkt);
2561 }
2562 if (is->video_stream >= 0) {
2563 packet_queue_flush(&is->videoq);
2564 packet_queue_put(&is->videoq, &flush_pkt);
2565 }
2566 }
2567 is->seek_req = 0;
2568 eof= 0;
2569 }
2570
2571 /* if the queue are full, no need to read more */
2572 if ( is->audioq.size + is->videoq.size + is->subtitleq.size > MAX_QUEUE_SIZE
2573 || ( (is->audioq .size > MIN_AUDIOQ_SIZE || is->audio_stream<0)
2574 && (is->videoq .nb_packets > MIN_FRAMES || is->video_stream<0)
2575 && (is->subtitleq.nb_packets > MIN_FRAMES || is->subtitle_stream<0))) {
2576 /* wait 10 ms */
2577 SDL_Delay(10);
2578 continue;
2579 }
2580 if(eof) {
2581 if(is->video_stream >= 0){
2582 av_init_packet(pkt);
2583 pkt->data=NULL;
2584 pkt->size=0;
2585 pkt->stream_index= is->video_stream;
2586 packet_queue_put(&is->videoq, pkt);
2587 }
2588 SDL_Delay(10);
2589 if(is->audioq.size + is->videoq.size + is->subtitleq.size ==0){
2590 if(loop!=1 && (!loop || --loop)){
2591 stream_seek(cur_stream, start_time != AV_NOPTS_VALUE ? start_time : 0, 0, 0);
2592 }else if(autoexit){
2593 ret=AVERROR_EOF;
2594 goto fail;
2595 }
2596 }
2597 continue;
2598 }
2599 ret = av_read_frame(ic, pkt);
2600 if (ret < 0) {
2601 if (ret == AVERROR_EOF || url_feof(ic->pb))
2602 eof=1;
2603 if (url_ferror(ic->pb))
2604 break;
2605 SDL_Delay(100); /* wait for user event */
2606 continue;
2607 }
2608 /* check if packet is in play range specified by user, then queue, otherwise discard */
2609 pkt_in_play_range = duration == AV_NOPTS_VALUE ||
2610 (pkt->pts - ic->streams[pkt->stream_index]->start_time) *
2611 av_q2d(ic->streams[pkt->stream_index]->time_base) -
2612 (double)(start_time != AV_NOPTS_VALUE ? start_time : 0)/1000000
2613 <= ((double)duration/1000000);
2614 if (pkt->stream_index == is->audio_stream && pkt_in_play_range) {
2615 packet_queue_put(&is->audioq, pkt);
2616 } else if (pkt->stream_index == is->video_stream && pkt_in_play_range) {
2617 packet_queue_put(&is->videoq, pkt);
2618 } else if (pkt->stream_index == is->subtitle_stream && pkt_in_play_range) {
2619 packet_queue_put(&is->subtitleq, pkt);
2620 } else {
2621 av_free_packet(pkt);
2622 }
2623 }
2624 /* wait until the end */
2625 while (!is->abort_request) {
2626 SDL_Delay(100);
2627 }
2628
2629 ret = 0;
2630 fail:
2631 /* disable interrupting */
2632 global_video_state = NULL;
2633
2634 /* close each stream */
2635 if (is->audio_stream >= 0)
2636 stream_component_close(is, is->audio_stream);
2637 if (is->video_stream >= 0)
2638 stream_component_close(is, is->video_stream);
2639 if (is->subtitle_stream >= 0)
2640 stream_component_close(is, is->subtitle_stream);
2641 if (is->ic) {
2642 av_close_input_file(is->ic);
2643 is->ic = NULL; /* safety */
2644 }
2645 url_set_interrupt_cb(NULL);
2646
2647 if (ret != 0) {
2648 SDL_Event event;
2649
2650 event.type = FF_QUIT_EVENT;
2651 event.user.data1 = is;
2652 SDL_PushEvent(&event);
2653 }
2654 return 0;
2655 }
2656
2657 static VideoState *stream_open(const char *filename, AVInputFormat *iformat)
2658 {
2659 VideoState *is;
2660
2661 is = av_mallocz(sizeof(VideoState));
2662 if (!is)
2663 return NULL;
2664 av_strlcpy(is->filename, filename, sizeof(is->filename));
2665 is->iformat = iformat;
2666 is->ytop = 0;
2667 is->xleft = 0;
2668
2669 /* start video display */
2670 is->pictq_mutex = SDL_CreateMutex();
2671 is->pictq_cond = SDL_CreateCond();
2672
2673 is->subpq_mutex = SDL_CreateMutex();
2674 is->subpq_cond = SDL_CreateCond();
2675
2676 is->av_sync_type = av_sync_type;
2677 is->parse_tid = SDL_CreateThread(decode_thread, is);
2678 if (!is->parse_tid) {
2679 av_free(is);
2680 return NULL;
2681 }
2682 return is;
2683 }
2684
2685 static void stream_cycle_channel(VideoState *is, int codec_type)
2686 {
2687 AVFormatContext *ic = is->ic;
2688 int start_index, stream_index;
2689 AVStream *st;
2690
2691 if (codec_type == AVMEDIA_TYPE_VIDEO)
2692 start_index = is->video_stream;
2693 else if (codec_type == AVMEDIA_TYPE_AUDIO)
2694 start_index = is->audio_stream;
2695 else
2696 start_index = is->subtitle_stream;
2697 if (start_index < (codec_type == AVMEDIA_TYPE_SUBTITLE ? -1 : 0))
2698 return;
2699 stream_index = start_index;
2700 for(;;) {
2701 if (++stream_index >= is->ic->nb_streams)
2702 {
2703 if (codec_type == AVMEDIA_TYPE_SUBTITLE)
2704 {
2705 stream_index = -1;
2706 goto the_end;
2707 } else
2708 stream_index = 0;
2709 }
2710 if (stream_index == start_index)
2711 return;
2712 st = ic->streams[stream_index];
2713 if (st->codec->codec_type == codec_type) {
2714 /* check that parameters are OK */
2715 switch(codec_type) {
2716 case AVMEDIA_TYPE_AUDIO:
2717 if (st->codec->sample_rate != 0 &&
2718 st->codec->channels != 0)
2719 goto the_end;
2720 break;
2721 case AVMEDIA_TYPE_VIDEO:
2722 case AVMEDIA_TYPE_SUBTITLE:
2723 goto the_end;
2724 default:
2725 break;
2726 }
2727 }
2728 }
2729 the_end:
2730 stream_component_close(is, start_index);
2731 stream_component_open(is, stream_index);
2732 }
2733
2734
2735 static void toggle_full_screen(void)
2736 {
2737 is_full_screen = !is_full_screen;
2738 if (!fs_screen_width) {
2739 /* use default SDL method */
2740 // SDL_WM_ToggleFullScreen(screen);
2741 }
2742 video_open(cur_stream);
2743 }
2744
2745 static void toggle_pause(void)
2746 {
2747 if (cur_stream)
2748 stream_pause(cur_stream);
2749 step = 0;
2750 }
2751
2752 static void step_to_next_frame(void)
2753 {
2754 if (cur_stream) {
2755 /* if the stream is paused unpause it, then step */
2756 if (cur_stream->paused)
2757 stream_pause(cur_stream);
2758 }
2759 step = 1;
2760 }
2761
2762 static void toggle_audio_display(void)
2763 {
2764 if (cur_stream) {
2765 int bgcolor = SDL_MapRGB(screen->format, 0x00, 0x00, 0x00);
2766 cur_stream->show_audio = (cur_stream->show_audio + 1) % 3;
2767 fill_rectangle(screen,
2768 cur_stream->xleft, cur_stream->ytop, cur_stream->width, cur_stream->height,
2769 bgcolor);
2770 SDL_UpdateRect(screen, cur_stream->xleft, cur_stream->ytop, cur_stream->width, cur_stream->height);
2771 }
2772 }
2773
2774 /* handle an event sent by the GUI */
2775 static void event_loop(void)
2776 {
2777 SDL_Event event;
2778 double incr, pos, frac;
2779
2780 for(;;) {
2781 double x;
2782 SDL_WaitEvent(&event);
2783 switch(event.type) {
2784 case SDL_KEYDOWN:
2785 if (exit_on_keydown) {
2786 do_exit();
2787 break;
2788 }
2789 switch(event.key.keysym.sym) {
2790 case SDLK_ESCAPE:
2791 case SDLK_q:
2792 do_exit();
2793 break;
2794 case SDLK_f:
2795 toggle_full_screen();
2796 break;
2797 case SDLK_p:
2798 case SDLK_SPACE:
2799 toggle_pause();
2800 break;
2801 case SDLK_s: //S: Step to next frame
2802 step_to_next_frame();
2803 break;
2804 case SDLK_a:
2805 if (cur_stream)
2806 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_AUDIO);
2807 break;
2808 case SDLK_v:
2809 if (cur_stream)
2810 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_VIDEO);
2811 break;
2812 case SDLK_t:
2813 if (cur_stream)
2814 stream_cycle_channel(cur_stream, AVMEDIA_TYPE_SUBTITLE);
2815 break;
2816 case SDLK_w:
2817 toggle_audio_display();
2818 break;
2819 case SDLK_LEFT:
2820 incr = -10.0;
2821 goto do_seek;
2822 case SDLK_RIGHT:
2823 incr = 10.0;
2824 goto do_seek;
2825 case SDLK_UP:
2826 incr = 60.0;
2827 goto do_seek;
2828 case SDLK_DOWN:
2829 incr = -60.0;
2830 do_seek:
2831 if (cur_stream) {
2832 if (seek_by_bytes) {
2833 if (cur_stream->video_stream >= 0 && cur_stream->video_current_pos>=0){
2834 pos= cur_stream->video_current_pos;
2835 }else if(cur_stream->audio_stream >= 0 && cur_stream->audio_pkt.pos>=0){
2836 pos= cur_stream->audio_pkt.pos;
2837 }else
2838 pos = url_ftell(cur_stream->ic->pb);
2839 if (cur_stream->ic->bit_rate)
2840 incr *= cur_stream->ic->bit_rate / 8.0;
2841 else
2842 incr *= 180000.0;
2843 pos += incr;
2844 stream_seek(cur_stream, pos, incr, 1);
2845 } else {
2846 pos = get_master_clock(cur_stream);
2847 pos += incr;
2848 stream_seek(cur_stream, (int64_t)(pos * AV_TIME_BASE), (int64_t)(incr * AV_TIME_BASE), 0);
2849 }
2850 }
2851 break;
2852 default:
2853 break;
2854 }
2855 break;
2856 case SDL_MOUSEBUTTONDOWN:
2857 if (exit_on_mousedown) {
2858 do_exit();
2859 break;
2860 }
2861 case SDL_MOUSEMOTION:
2862 if(event.type ==SDL_MOUSEBUTTONDOWN){
2863 x= event.button.x;
2864 }else{
2865 if(event.motion.state != SDL_PRESSED)
2866 break;
2867 x= event.motion.x;
2868 }
2869 if (cur_stream) {
2870 if(seek_by_bytes || cur_stream->ic->duration<=0){
2871 uint64_t size= url_fsize(cur_stream->ic->pb);
2872 stream_seek(cur_stream, size*x/cur_stream->width, 0, 1);
2873 }else{
2874 int64_t ts;
2875 int ns, hh, mm, ss;
2876 int tns, thh, tmm, tss;
2877 tns = cur_stream->ic->duration/1000000LL;
2878 thh = tns/3600;
2879 tmm = (tns%3600)/60;
2880 tss = (tns%60);
2881 frac = x/cur_stream->width;
2882 ns = frac*tns;
2883 hh = ns/3600;
2884 mm = (ns%3600)/60;
2885 ss = (ns%60);
2886 fprintf(stderr, "Seek to %2.0f%% (%2d:%02d:%02d) of total duration (%2d:%02d:%02d) \n", frac*100,
2887 hh, mm, ss, thh, tmm, tss);
2888 ts = frac*cur_stream->ic->duration;
2889 if (cur_stream->ic->start_time != AV_NOPTS_VALUE)
2890 ts += cur_stream->ic->start_time;
2891 stream_seek(cur_stream, ts, 0, 0);
2892 }
2893 }
2894 break;
2895 case SDL_VIDEORESIZE:
2896 if (cur_stream) {
2897 screen = SDL_SetVideoMode(event.resize.w, event.resize.h, 0,
2898 SDL_HWSURFACE|SDL_RESIZABLE|SDL_ASYNCBLIT|SDL_HWACCEL);
2899 screen_width = cur_stream->width = event.resize.w;
2900 screen_height= cur_stream->height= event.resize.h;
2901 }
2902 break;
2903 case SDL_QUIT:
2904 case FF_QUIT_EVENT:
2905 do_exit();
2906 break;
2907 case FF_ALLOC_EVENT:
2908 video_open(event.user.data1);
2909 alloc_picture(event.user.data1);
2910 break;
2911 case FF_REFRESH_EVENT:
2912 video_refresh_timer(event.user.data1);
2913 cur_stream->refresh=0;
2914 break;
2915 default:
2916 break;
2917 }
2918 }
2919 }
2920
2921 static void opt_frame_size(const char *arg)
2922 {
2923 if (av_parse_video_size(&frame_width, &frame_height, arg) < 0) {
2924 fprintf(stderr, "Incorrect frame size\n");
2925 exit(1);
2926 }
2927 if ((frame_width % 2) != 0 || (frame_height % 2) != 0) {
2928 fprintf(stderr, "Frame size must be a multiple of 2\n");
2929 exit(1);
2930 }
2931 }
2932
2933 static int opt_width(const char *opt, const char *arg)
2934 {
2935 screen_width = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2936 return 0;
2937 }
2938
2939 static int opt_height(const char *opt, const char *arg)
2940 {
2941 screen_height = parse_number_or_die(opt, arg, OPT_INT64, 1, INT_MAX);
2942 return 0;
2943 }
2944
2945 static void opt_format(const char *arg)
2946 {
2947 file_iformat = av_find_input_format(arg);
2948 if (!file_iformat) {
2949 fprintf(stderr, "Unknown input format: %s\n", arg);
2950 exit(1);
2951 }
2952 }
2953
2954 static void opt_frame_pix_fmt(const char *arg)
2955 {
2956 frame_pix_fmt = av_get_pix_fmt(arg);
2957 }
2958
2959 static int opt_sync(const char *opt, const char *arg)
2960 {
2961 if (!strcmp(arg, "audio"))
2962 av_sync_type = AV_SYNC_AUDIO_MASTER;
2963 else if (!strcmp(arg, "video"))
2964 av_sync_type = AV_SYNC_VIDEO_MASTER;
2965 else if (!strcmp(arg, "ext"))
2966 av_sync_type = AV_SYNC_EXTERNAL_CLOCK;
2967 else {
2968 fprintf(stderr, "Unknown value for %s: %s\n", opt, arg);
2969 exit(1);
2970 }
2971 return 0;
2972 }
2973
2974 static int opt_seek(const char *opt, const char *arg)
2975 {
2976 start_time = parse_time_or_die(opt, arg, 1);
2977 return 0;
2978 }
2979
2980 static int opt_duration(const char *opt, const char *arg)
2981 {
2982 duration = parse_time_or_die(opt, arg, 1);
2983 return 0;
2984 }
2985
2986 static int opt_debug(const char *opt, const char *arg)
2987 {
2988 av_log_set_level(99);
2989 debug = parse_number_or_die(opt, arg, OPT_INT64, 0, INT_MAX);
2990 return 0;
2991 }
2992
2993 static int opt_vismv(const char *opt, const char *arg)
2994 {
2995 debug_mv = parse_number_or_die(opt, arg, OPT_INT64, INT_MIN, INT_MAX);
2996 return 0;
2997 }
2998
2999 static int opt_thread_count(const char *opt, const char *arg)
3000 {
3001 thread_count= parse_number_or_die(opt, arg, OPT_INT64, 0, INT_MAX);
3002 #if !HAVE_THREADS
3003 fprintf(stderr, "Warning: not compiled with thread support, using thread emulation\n");
3004 #endif
3005 return 0;
3006 }
3007
3008 static const OptionDef options[] = {
3009 #include "cmdutils_common_opts.h"
3010 { "x", HAS_ARG | OPT_FUNC2, {(void*)opt_width}, "force displayed width", "width" },
3011 { "y", HAS_ARG | OPT_FUNC2, {(void*)opt_height}, "force displayed height", "height" },
3012 { "s", HAS_ARG | OPT_VIDEO, {(void*)opt_frame_size}, "set frame size (WxH or abbreviation)", "size" },
3013 { "fs", OPT_BOOL, {(void*)&is_full_screen}, "force full screen" },
3014 { "an", OPT_BOOL, {(void*)&audio_disable}, "disable audio" },
3015 { "vn", OPT_BOOL, {(void*)&video_disable}, "disable video" },
3016 { "ast", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_AUDIO]}, "select desired audio stream", "stream_number" },
3017 { "vst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_VIDEO]}, "select desired video stream", "stream_number" },
3018 { "sst", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&wanted_stream[AVMEDIA_TYPE_SUBTITLE]}, "select desired subtitle stream", "stream_number" },
3019 { "ss", HAS_ARG | OPT_FUNC2, {(void*)&opt_seek}, "seek to a given position in seconds", "pos" },
3020 { "t", HAS_ARG | OPT_FUNC2, {(void*)&opt_duration}, "play \"duration\" seconds of audio/video", "duration" },
3021 { "bytes", OPT_INT | HAS_ARG, {(void*)&seek_by_bytes}, "seek by bytes 0=off 1=on -1=auto", "val" },
3022 { "nodisp", OPT_BOOL, {(void*)&display_disable}, "disable graphical display" },
3023 { "f", HAS_ARG, {(void*)opt_format}, "force format", "fmt" },
3024 { "pix_fmt", HAS_ARG | OPT_EXPERT | OPT_VIDEO, {(void*)opt_frame_pix_fmt}, "set pixel format", "format" },
3025 { "stats", OPT_BOOL | OPT_EXPERT, {(void*)&show_status}, "show status", "" },
3026 { "debug", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_debug}, "print specific debug info", "" },
3027 { "bug", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&workaround_bugs}, "workaround bugs", "" },
3028 { "vismv", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_vismv}, "visualize motion vectors", "" },
3029 { "fast", OPT_BOOL | OPT_EXPERT, {(void*)&fast}, "non spec compliant optimizations", "" },
3030 { "genpts", OPT_BOOL | OPT_EXPERT, {(void*)&genpts}, "generate pts", "" },
3031 { "drp", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&decoder_reorder_pts}, "let decoder reorder pts 0=off 1=on -1=auto", ""},
3032 { "lowres", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&lowres}, "", "" },
3033 { "skiploop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_loop_filter}, "", "" },
3034 { "skipframe", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_frame}, "", "" },
3035 { "skipidct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&skip_idct}, "", "" },
3036 { "idct", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&idct}, "set idct algo", "algo" },
3037 { "er", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_recognition}, "set error detection threshold (0-4)", "threshold" },
3038 { "ec", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&error_concealment}, "set error concealment options", "bit_mask" },
3039 { "sync", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_sync}, "set audio-video sync. type (type=audio/video/ext)", "type" },
3040 { "threads", HAS_ARG | OPT_FUNC2 | OPT_EXPERT, {(void*)opt_thread_count}, "thread count", "count" },
3041 { "autoexit", OPT_BOOL | OPT_EXPERT, {(void*)&autoexit}, "exit at the end", "" },
3042 { "exitonkeydown", OPT_BOOL | OPT_EXPERT, {(void*)&exit_on_keydown}, "exit on key down", "" },
3043 { "exitonmousedown", OPT_BOOL | OPT_EXPERT, {(void*)&exit_on_mousedown}, "exit on mouse down", "" },
3044 { "loop", OPT_INT | HAS_ARG | OPT_EXPERT, {(void*)&loop}, "set number of times the playback shall be looped", "loop count" },
3045 { "framedrop", OPT_BOOL | OPT_EXPERT, {(void*)&framedrop}, "drop frames when cpu is too slow", "" },
3046 { "window_title", OPT_STRING | HAS_ARG, {(void*)&window_title}, "set window title", "window title" },
3047 #if CONFIG_AVFILTER
3048 { "vf", OPT_STRING | HAS_ARG, {(void*)&vfilters}, "video filters", "filter list" },
3049 #endif
3050 { "rdftspeed", OPT_INT | HAS_ARG| OPT_AUDIO | OPT_EXPERT, {(void*)&rdftspeed}, "rdft speed", "msecs" },
3051 { "default", OPT_FUNC2 | HAS_ARG | OPT_AUDIO | OPT_VIDEO | OPT_EXPERT, {(void*)opt_default}, "generic catch all option", "" },
3052 { NULL, },
3053 };
3054
3055 static void show_usage(void)
3056 {
3057 printf("Simple media player\n");
3058 printf("usage: ffplay [options] input_file\n");
3059 printf("\n");
3060 }
3061
3062 static void show_help(void)
3063 {
3064 av_log_set_callback(log_callback_help);
3065 show_usage();
3066 show_help_options(options, "Main options:\n",
3067 OPT_EXPERT, 0);
3068 show_help_options(options, "\nAdvanced options:\n",
3069 OPT_EXPERT, OPT_EXPERT);
3070 printf("\n");
3071 av_opt_show2(avcodec_opts[0], NULL,
3072 AV_OPT_FLAG_DECODING_PARAM, 0);
3073 printf("\n");
3074 av_opt_show2(avformat_opts, NULL,
3075 AV_OPT_FLAG_DECODING_PARAM, 0);
3076 #if !CONFIG_AVFILTER
3077 printf("\n");
3078 av_opt_show2(sws_opts, NULL,
3079 AV_OPT_FLAG_ENCODING_PARAM, 0);
3080 #endif
3081 printf("\nWhile playing:\n"
3082 "q, ESC quit\n"
3083 "f toggle full screen\n"
3084 "p, SPC pause\n"
3085 "a cycle audio channel\n"
3086 "v cycle video channel\n"
3087 "t cycle subtitle channel\n"
3088 "w show audio waves\n"
3089 "s activate frame-step mode\n"
3090 "left/right seek backward/forward 10 seconds\n"
3091 "down/up seek backward/forward 1 minute\n"
3092 "mouse click seek to percentage in file corresponding to fraction of width\n"
3093 );
3094 }
3095
3096 static void opt_input_file(const char *filename)
3097 {
3098 if (input_filename) {
3099 fprintf(stderr, "Argument '%s' provided as input filename, but '%s' was already specified.\n",
3100 filename, input_filename);
3101 exit(1);
3102 }
3103 if (!strcmp(filename, "-"))
3104 filename = "pipe:";
3105 input_filename = filename;
3106 }
3107
3108 /* Called from the main */
3109 int main(int argc, char **argv)
3110 {
3111 int flags;
3112
3113 av_log_set_flags(AV_LOG_SKIP_REPEATED);
3114
3115 /* register all codecs, demux and protocols */
3116 avcodec_register_all();
3117 #if CONFIG_AVDEVICE
3118 avdevice_register_all();
3119 #endif
3120 #if CONFIG_AVFILTER
3121 avfilter_register_all();
3122 #endif
3123 av_register_all();
3124
3125 init_opts();
3126
3127 show_banner();
3128
3129 parse_options(argc, argv, options, opt_input_file);
3130
3131 if (!input_filename) {
3132 show_usage();
3133 fprintf(stderr, "An input file must be specified\n");
3134 fprintf(stderr, "Use -h to get full help or, even better, run 'man ffplay'\n");
3135 exit(1);
3136 }
3137
3138 if (display_disable) {
3139 video_disable = 1;
3140 }
3141 flags = SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER;
3142 #if !defined(__MINGW32__) && !defined(__APPLE__)
3143 flags |= SDL_INIT_EVENTTHREAD; /* Not supported on Windows or Mac OS X */
3144 #endif
3145 if (SDL_Init (flags)) {
3146 fprintf(stderr, "Could not initialize SDL - %s\n", SDL_GetError());
3147 exit(1);
3148 }
3149
3150 if (!display_disable) {
3151 #if HAVE_SDL_VIDEO_SIZE
3152 const SDL_VideoInfo *vi = SDL_GetVideoInfo();
3153 fs_screen_width = vi->current_w;
3154 fs_screen_height = vi->current_h;
3155 #endif
3156 }
3157
3158 SDL_EventState(SDL_ACTIVEEVENT, SDL_IGNORE);
3159 SDL_EventState(SDL_SYSWMEVENT, SDL_IGNORE);
3160 SDL_EventState(SDL_USEREVENT, SDL_IGNORE);
3161
3162 av_init_packet(&flush_pkt);
3163 flush_pkt.data= "FLUSH";
3164
3165 cur_stream = stream_open(input_filename, file_iformat);
3166
3167 event_loop();
3168
3169 /* never returns */
3170
3171 return 0;
3172 }