#undef NDEBUG
#include <assert.h>
+#define PERM_RWP AV_PERM_WRITE | AV_PERM_PRESERVE | AV_PERM_REUSE
+
#define CHECK(j)\
{ int score = FFABS(cur[mrefs-1+(j)] - cur[prefs-1-(j)])\
+ FFABS(cur[mrefs +(j)] - cur[prefs -(j)])\
int temporal_diff0 = FFABS(prev2[0] - next2[0]); \
int temporal_diff1 =(FFABS(prev[mrefs] - c) + FFABS(prev[prefs] - e) )>>1; \
int temporal_diff2 =(FFABS(next[mrefs] - c) + FFABS(next[prefs] - e) )>>1; \
- int diff = FFMAX3(temporal_diff0>>1, temporal_diff1, temporal_diff2); \
- int spatial_pred = (c+e)>>1; \
- int spatial_score = FFABS(cur[mrefs-1] - cur[prefs-1]) + FFABS(c-e) \
- + FFABS(cur[mrefs+1] - cur[prefs+1]) - 1; \
+ int diff = FFMAX3(temporal_diff0 >> 1, temporal_diff1, temporal_diff2); \
+ int spatial_pred = (c+e) >> 1; \
+ int spatial_score = FFABS(cur[mrefs - 1] - cur[prefs - 1]) + FFABS(c-e) \
+ + FFABS(cur[mrefs + 1] - cur[prefs + 1]) - 1; \
\
CHECK(-1) CHECK(-2) }} }} \
CHECK( 1) CHECK( 2) }} }} \
\
if (mode < 2) { \
- int b = (prev2[2*mrefs] + next2[2*mrefs])>>1; \
- int f = (prev2[2*prefs] + next2[2*prefs])>>1; \
- int max = FFMAX3(d-e, d-c, FFMIN(b-c, f-e)); \
- int min = FFMIN3(d-e, d-c, FFMAX(b-c, f-e)); \
+ int b = (prev2[2 * mrefs] + next2[2 * mrefs])>>1; \
+ int f = (prev2[2 * prefs] + next2[2 * prefs])>>1; \
+ int max = FFMAX3(d - e, d - c, FFMIN(b - c, f - e)); \
+ int min = FFMIN3(d - e, d - c, FFMAX(b - c, f - e)); \
\
diff = FFMAX3(diff, min, -max); \
} \
static void filter_line_c_16bit(uint16_t *dst,
uint16_t *prev, uint16_t *cur, uint16_t *next,
- int w, int prefs, int mrefs, int parity, int mode)
+ int w, int prefs, int mrefs, int parity,
+ int mode)
{
int x;
uint16_t *prev2 = parity ? prev : cur ;
for (y = 0; y < h; y++) {
if ((y ^ parity) & 1) {
- uint8_t *prev = &yadif->prev->data[i][y*refs];
- uint8_t *cur = &yadif->cur ->data[i][y*refs];
- uint8_t *next = &yadif->next->data[i][y*refs];
- uint8_t *dst = &dstpic->data[i][y*dstpic->linesize[i]];
- int mode = y==1 || y+2==h ? 2 : yadif->mode;
- yadif->filter_line(dst, prev, cur, next, w, y+1<h ? refs : -refs, y ? -refs : refs, parity ^ tff, mode);
+ uint8_t *prev = &yadif->prev->data[i][y * refs];
+ uint8_t *cur = &yadif->cur ->data[i][y * refs];
+ uint8_t *next = &yadif->next->data[i][y * refs];
+ uint8_t *dst = &dstpic->data[i][y * dstpic->linesize[i]];
+ int mode = y == 1 || y + 2 == h ? 2 : yadif->mode;
+ yadif->filter_line(dst, prev, cur, next, w,
+ y + 1 < h ? refs : -refs,
+ y ? -refs : refs,
+ parity ^ tff, mode);
} else {
- memcpy(&dstpic->data[i][y*dstpic->linesize[i]],
- &yadif->cur->data[i][y*refs], w*df);
+ memcpy(&dstpic->data[i][y * dstpic->linesize[i]],
+ &yadif->cur->data[i][y * refs], w * df);
}
}
}
emms_c();
}
-static AVFilterBufferRef *get_video_buffer(AVFilterLink *link, int perms, int w, int h)
+static AVFilterBufferRef *get_video_buffer(AVFilterLink *link, int perms,
+ int w, int h)
{
AVFilterBufferRef *picref;
- int width = FFALIGN(w, 32);
- int height= FFALIGN(h+2, 32);
+ int width = FFALIGN(w, 32);
+ int height = FFALIGN(h + 2, 32);
int i;
picref = ff_default_get_video_buffer(link, perms, width, height);
static int return_frame(AVFilterContext *ctx, int is_second)
{
YADIFContext *yadif = ctx->priv;
- AVFilterLink *link= ctx->outputs[0];
+ AVFilterLink *link = ctx->outputs[0];
int tff, ret;
if (yadif->parity == -1) {
tff = yadif->cur->video->interlaced ?
- yadif->cur->video->top_field_first : 1;
+ yadif->cur->video->top_field_first : 1;
} else {
- tff = yadif->parity^1;
+ tff = yadif->parity ^ 1;
}
if (is_second) {
- yadif->out = ff_get_video_buffer(link, AV_PERM_WRITE | AV_PERM_PRESERVE |
- AV_PERM_REUSE, link->w, link->h);
+ yadif->out = ff_get_video_buffer(link, PERM_RWP, link->w, link->h);
if (!yadif->out)
return AVERROR(ENOMEM);
!(yadif->prev = avfilter_ref_buffer(yadif->cur, AV_PERM_READ)))
return AVERROR(ENOMEM);
- yadif->out = ff_get_video_buffer(ctx->outputs[0], AV_PERM_WRITE | AV_PERM_PRESERVE |
- AV_PERM_REUSE, link->w, link->h);
+ yadif->out = ff_get_video_buffer(ctx->outputs[0], PERM_RWP,
+ link->w, link->h);
if (!yadif->out)
return AVERROR(ENOMEM);
avfilter_copy_buffer_ref_props(yadif->out, yadif->cur);
yadif->out->video->interlaced = 0;
+
if (yadif->out->pts != AV_NOPTS_VALUE)
yadif->out->pts *= 2;
+
return ff_start_frame(ctx->outputs[0], yadif->out);
}
ret = ff_request_frame(link->src->inputs[0]);
if (ret == AVERROR_EOF && yadif->next) {
- AVFilterBufferRef *next = avfilter_ref_buffer(yadif->next, AV_PERM_READ);
+ AVFilterBufferRef *next =
+ avfilter_ref_buffer(yadif->next, AV_PERM_READ);
+
if (!next)
return AVERROR(ENOMEM);
if (val <= 0)
return val;
- if (val==1 && !yadif->next) { //FIXME change API to not requre this red tape
+ //FIXME change API to not requre this red tape
+ if (val == 1 && !yadif->next) {
if ((ret = ff_request_frame(link->src->inputs[0])) < 0)
return ret;
val = ff_poll_frame(link->src->inputs[0]);
yadif->auto_enable = 0;
yadif->csp = NULL;
- if (args) sscanf(args, "%d:%d:%d", &yadif->mode, &yadif->parity, &yadif->auto_enable);
+ if (args)
+ sscanf(args, "%d:%d:%d",
+ &yadif->mode, &yadif->parity, &yadif->auto_enable);
yadif->filter_line = filter_line_c;
if (HAVE_MMX)
ff_yadif_init_x86(yadif);
- av_log(ctx, AV_LOG_VERBOSE, "mode:%d parity:%d auto_enable:%d\n", yadif->mode, yadif->parity, yadif->auto_enable);
+ av_log(ctx, AV_LOG_VERBOSE, "mode:%d parity:%d auto_enable:%d\n",
+ yadif->mode, yadif->parity, yadif->auto_enable);
return 0;
}