mirror of
https://github.com/FFmpeg/FFmpeg.git
synced 2026-02-04 14:30:55 +08:00
Merge remote-tracking branch 'qatar/master'
* qatar/master: yuv4mpeg: return proper error codes. Give all anonymously typedeffed structs in headers a name fate: Add parseutils test parseutils-test: Drop random colors from parsing test vf_pad/scale: use double precision for aspect ratios. build: error on variable-length arrays ppc: swscale: rework yuv2planeX_altivec() ppc: fmtconvert: kill VLA in float_to_int16_interleave_altivec() x86: dsputil: kill VLA in gmc_mmx() libspeexenc: Updated commentary to reflect recent changes libspeexenc: Add an option for enabling DTX doc/APIchanges: fill in missing dates and hashes. lavr: bump major to 1 and declare it stable. lavr: change the type of the data buffers to uint8_t**. lavc: deprecate the audio resampling API. Conflicts: cmdutils.h configure doc/APIchanges ffplay.c libavcodec/dwt.h libavcodec/libspeexenc.c libavfilter/vf_pad.c libavfilter/vf_scale.c libavformat/asf.h tests/fate/libavutil.mak tests/ref/fate/parseutils Merged-by: Michael Niedermayer <michaelni@gmx.at>
This commit is contained in:
@@ -127,7 +127,7 @@ static int request_frame(AVFilterLink *link)
|
||||
nb_samples);
|
||||
if (!buf)
|
||||
return AVERROR(ENOMEM);
|
||||
ret = avresample_convert(s->avr, (void**)buf->extended_data,
|
||||
ret = avresample_convert(s->avr, buf->extended_data,
|
||||
buf->linesize[0], nb_samples, NULL, 0, 0);
|
||||
if (ret <= 0) {
|
||||
avfilter_unref_bufferp(&buf);
|
||||
@@ -143,7 +143,7 @@ static int request_frame(AVFilterLink *link)
|
||||
|
||||
static int write_to_fifo(ASyncContext *s, AVFilterBufferRef *buf)
|
||||
{
|
||||
int ret = avresample_convert(s->avr, NULL, 0, 0, (void**)buf->extended_data,
|
||||
int ret = avresample_convert(s->avr, NULL, 0, 0, buf->extended_data,
|
||||
buf->linesize[0], buf->audio->nb_samples);
|
||||
avfilter_unref_buffer(buf);
|
||||
return ret;
|
||||
@@ -204,7 +204,7 @@ static int filter_samples(AVFilterLink *inlink, AVFilterBufferRef *buf)
|
||||
goto fail;
|
||||
}
|
||||
|
||||
avresample_read(s->avr, (void**)buf_out->extended_data, out_size);
|
||||
avresample_read(s->avr, buf_out->extended_data, out_size);
|
||||
buf_out->pts = s->pts;
|
||||
|
||||
if (delta > 0) {
|
||||
@@ -224,7 +224,7 @@ static int filter_samples(AVFilterLink *inlink, AVFilterBufferRef *buf)
|
||||
avresample_read(s->avr, NULL, avresample_available(s->avr));
|
||||
|
||||
s->pts = pts - avresample_get_delay(s->avr);
|
||||
ret = avresample_convert(s->avr, NULL, 0, 0, (void**)buf->extended_data,
|
||||
ret = avresample_convert(s->avr, NULL, 0, 0, buf->extended_data,
|
||||
buf->linesize[0], buf->audio->nb_samples);
|
||||
|
||||
fail:
|
||||
|
||||
@@ -149,7 +149,7 @@ static int request_frame(AVFilterLink *outlink)
|
||||
if (!buf)
|
||||
return AVERROR(ENOMEM);
|
||||
|
||||
ret = avresample_convert(s->avr, (void**)buf->extended_data,
|
||||
ret = avresample_convert(s->avr, buf->extended_data,
|
||||
buf->linesize[0], nb_samples,
|
||||
NULL, 0, 0);
|
||||
if (ret <= 0) {
|
||||
@@ -186,9 +186,9 @@ static int filter_samples(AVFilterLink *inlink, AVFilterBufferRef *buf)
|
||||
goto fail;
|
||||
}
|
||||
|
||||
ret = avresample_convert(s->avr, (void**)buf_out->extended_data,
|
||||
ret = avresample_convert(s->avr, buf_out->extended_data,
|
||||
buf_out->linesize[0], nb_samples,
|
||||
(void**)buf->extended_data, buf->linesize[0],
|
||||
buf->extended_data, buf->linesize[0],
|
||||
buf->audio->nb_samples);
|
||||
if (ret < 0) {
|
||||
avfilter_unref_buffer(buf_out);
|
||||
|
||||
@@ -25,7 +25,7 @@
|
||||
#include "avfilter.h"
|
||||
|
||||
/// Holds instance-specific information for gradfun.
|
||||
typedef struct {
|
||||
typedef struct GradFunContext {
|
||||
int thresh; ///< threshold for gradient algorithm
|
||||
int radius; ///< blur radius
|
||||
int chroma_w; ///< width of the chroma planes
|
||||
|
||||
@@ -126,9 +126,9 @@ static int config_input(AVFilterLink *inlink)
|
||||
var_values[VAR_IN_H] = var_values[VAR_IH] = inlink->h;
|
||||
var_values[VAR_OUT_W] = var_values[VAR_OW] = NAN;
|
||||
var_values[VAR_OUT_H] = var_values[VAR_OH] = NAN;
|
||||
var_values[VAR_A] = (float) inlink->w / inlink->h;
|
||||
var_values[VAR_A] = (double) inlink->w / inlink->h;
|
||||
var_values[VAR_SAR] = inlink->sample_aspect_ratio.num ?
|
||||
(float) inlink->sample_aspect_ratio.num / inlink->sample_aspect_ratio.den : 1;
|
||||
(double) inlink->sample_aspect_ratio.num / inlink->sample_aspect_ratio.den : 1;
|
||||
var_values[VAR_DAR] = var_values[VAR_A] * var_values[VAR_SAR];
|
||||
var_values[VAR_HSUB] = 1 << pad->draw.hsub_max;
|
||||
var_values[VAR_VSUB] = 1 << pad->draw.vsub_max;
|
||||
|
||||
@@ -172,9 +172,9 @@ static int config_props(AVFilterLink *outlink)
|
||||
var_values[VAR_IN_H] = var_values[VAR_IH] = inlink->h;
|
||||
var_values[VAR_OUT_W] = var_values[VAR_OW] = NAN;
|
||||
var_values[VAR_OUT_H] = var_values[VAR_OH] = NAN;
|
||||
var_values[VAR_A] = (float) inlink->w / inlink->h;
|
||||
var_values[VAR_A] = (double) inlink->w / inlink->h;
|
||||
var_values[VAR_SAR] = inlink->sample_aspect_ratio.num ?
|
||||
(float) inlink->sample_aspect_ratio.num / inlink->sample_aspect_ratio.den : 1;
|
||||
(double) inlink->sample_aspect_ratio.num / inlink->sample_aspect_ratio.den : 1;
|
||||
var_values[VAR_DAR] = var_values[VAR_A] * var_values[VAR_SAR];
|
||||
var_values[VAR_HSUB] = 1<<av_pix_fmt_descriptors[inlink->format].log2_chroma_w;
|
||||
var_values[VAR_VSUB] = 1<<av_pix_fmt_descriptors[inlink->format].log2_chroma_h;
|
||||
|
||||
@@ -22,7 +22,7 @@
|
||||
#include "libavutil/pixdesc.h"
|
||||
#include "avfilter.h"
|
||||
|
||||
typedef struct {
|
||||
typedef struct YADIFContext {
|
||||
/**
|
||||
* 0: send 1 frame for each frame
|
||||
* 1: send 1 frame for each field
|
||||
|
||||
Reference in New Issue
Block a user