Go to the documentation of this file.
94 #define OFFSET(x) offsetof(ShowWavesContext, x)
95 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
109 {
"colors",
"set channels colors",
OFFSET(colors),
AV_OPT_TYPE_STRING, {.str =
"red|green|blue|yellow|orange|lime|pink|magenta|brown" }, 0, 0,
FLAGS },
221 buf[
h * linesize + 0] +=
color[0];
222 buf[
h * linesize + 1] +=
color[1];
223 buf[
h * linesize + 2] +=
color[2];
224 buf[
h * linesize + 3] +=
color[3];
233 buf[
h * linesize + 0] =
color[0];
234 buf[
h * linesize + 1] =
color[1];
235 buf[
h * linesize + 2] =
color[2];
236 buf[
h * linesize + 3] =
color[3];
249 for (k = start; k <
end; k++) {
250 buf[k * linesize + 0] +=
color[0];
251 buf[k * linesize + 1] +=
color[1];
252 buf[k * linesize + 2] +=
color[2];
253 buf[k * linesize + 3] +=
color[3];
266 for (k = start; k <
end; k++) {
267 buf[k * linesize + 0] =
color[0];
268 buf[k * linesize + 1] =
color[1];
269 buf[k * linesize + 2] =
color[2];
270 buf[k * linesize + 3] =
color[3];
280 buf[
h * linesize + 0] +=
color[0];
281 buf[
h * linesize + 1] +=
color[1];
282 buf[
h * linesize + 2] +=
color[2];
283 buf[
h * linesize + 3] +=
color[3];
284 if (*prev_y &&
h != *prev_y) {
289 for (k = start + 1; k <
end; k++) {
290 buf[k * linesize + 0] +=
color[0];
291 buf[k * linesize + 1] +=
color[1];
292 buf[k * linesize + 2] +=
color[2];
293 buf[k * linesize + 3] +=
color[3];
306 buf[
h * linesize + 0] =
color[0];
307 buf[
h * linesize + 1] =
color[1];
308 buf[
h * linesize + 2] =
color[2];
309 buf[
h * linesize + 3] =
color[3];
310 if (*prev_y &&
h != *prev_y) {
315 for (k = start + 1; k <
end; k++) {
316 buf[k * linesize + 0] =
color[0];
317 buf[k * linesize + 1] =
color[1];
318 buf[k * linesize + 2] =
color[2];
319 buf[k * linesize + 3] =
color[3];
331 const int start = (
height -
h) / 2;
332 const int end = start +
h;
333 for (k = start; k <
end; k++) {
334 buf[k * linesize + 0] +=
color[0];
335 buf[k * linesize + 1] +=
color[1];
336 buf[k * linesize + 2] +=
color[2];
337 buf[k * linesize + 3] +=
color[3];
345 const int start = (
height -
h) / 2;
346 const int end = start +
h;
347 for (k = start; k <
end; k++) {
348 buf[k * linesize + 0] =
color[0];
349 buf[k * linesize + 1] =
color[1];
350 buf[k * linesize + 2] =
color[2];
351 buf[k * linesize + 3] =
color[3];
360 buf[
h * linesize] +=
color[0];
372 for (k = start; k <
end; k++)
373 buf[k * linesize] +=
color[0];
382 buf[
h * linesize] +=
color[0];
383 if (*prev_y &&
h != *prev_y) {
388 for (k = start + 1; k <
end; k++)
389 buf[k * linesize] +=
color[0];
400 const int start = (
height -
h) / 2;
401 const int end = start +
h;
402 for (k = start; k <
end; k++)
403 buf[k * linesize] +=
color[0];
412 char *colors, *saveptr =
NULL;
427 outlink->
w = showwaves->
w;
428 outlink->
h = showwaves->
h;
437 switch (outlink->
format) {
439 switch (showwaves->
mode) {
450 switch (showwaves->
mode) {
462 switch (showwaves->
scale) {
464 switch (showwaves->
mode) {
474 switch (showwaves->
mode) {
484 switch (showwaves->
mode) {
494 switch (showwaves->
mode) {
520 uint8_t fg[4] = { 0xff, 0xff, 0xff, 0xff };
528 showwaves->
fg[4*ch + 0] = fg[0] * x / 255.;
529 showwaves->
fg[4*ch + 1] = fg[1] * x / 255.;
530 showwaves->
fg[4*ch + 2] = fg[2] * x / 255.;
531 showwaves->
fg[4*ch + 3] = fg[3] * x / 255.;
535 showwaves->
fg[4 * ch + 0] = x;
563 int64_t n = 0, column_max_samples = showwaves->
total_samples / outlink->
w;
564 int64_t remaining_samples = showwaves->
total_samples - (column_max_samples * outlink->
w);
565 int64_t last_column_samples = column_max_samples + remaining_samples;
570 const int linesize =
out->linesize[0];
571 const int pixstep = showwaves->
pixstep;
573 int64_t *sum = showwaves->
sum;
575 if (column_max_samples == 0) {
580 av_log(
ctx,
AV_LOG_DEBUG,
"Create frame averaging %"PRId64
" samples per column\n", column_max_samples);
587 const int16_t *p = (
const int16_t *)
frame->data[0];
590 int64_t max_samples = col == outlink->
w - 1 ? last_column_samples: column_max_samples;
596 if (n == max_samples) {
598 int16_t
sample = sum[ch] / max_samples;
603 buf += ch*ch_height*linesize;
606 showwaves->
draw_sample(buf, ch_height, linesize, &showwaves->
buf_idy[ch], &showwaves->
fg[ch * 4],
h);
646 out->width = outlink->
w;
647 out->height = outlink->
h;
651 for (j = 0; j < outlink->
h; j++)
652 memset(
out->data[0] + j*
out->linesize[0], 0, outlink->
w * showwaves->
pixstep);
661 if (!strcmp(
ctx->filter->name,
"showwavespic")) {
669 #if CONFIG_SHOWWAVES_FILTER
678 int16_t *p = (int16_t *)insamples->
data[0];
681 const int pixstep = showwaves->
pixstep;
682 const int n = showwaves->
n;
686 for (
i = 0;
i < nb_samples;
i++) {
695 const int linesize = outpicref->
linesize[0];
699 buf += j*ch_height*linesize;
700 h = showwaves->
get_h(*p++, ch_height);
702 &showwaves->
buf_idy[j], &showwaves->
fg[j * 4],
h);
710 if (showwaves->
buf_idx == showwaves->
w ||
728 const int nb_samples = showwaves->
n * outlink->
w;
737 return showwaves_filter_frame(
inlink,
in);
769 .
inputs = showwaves_inputs,
772 .priv_class = &showwaves_class,
775 #endif // CONFIG_SHOWWAVES_FILTER
777 #if CONFIG_SHOWWAVESPIC_FILTER
779 #define OFFSET(x) offsetof(ShowWavesContext, x)
780 #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
782 static const AVOption showwavespic_options[] = {
786 {
"colors",
"set channels colors",
OFFSET(colors),
AV_OPT_TYPE_STRING, {.str =
"red|green|blue|yellow|orange|lime|pink|magenta|brown" }, 0, 0,
FLAGS },
819 int16_t *p = (int16_t *)insamples->
data[0];
835 f->frame = insamples;
858 .config_props = showwavespic_config_input,
859 .filter_frame = showwavespic_filter_frame,
864 static const AVFilterPad showwavespic_outputs[] = {
875 .
name =
"showwavespic",
881 .
inputs = showwavespic_inputs,
882 .
outputs = showwavespic_outputs,
883 .priv_class = &showwavespic_class,
886 #endif // CONFIG_SHOWWAVESPIC_FILTER
AVFrame * ff_get_video_buffer(AVFilterLink *link, int w, int h)
Request a picture buffer with a specific set of permissions.
A list of supported channel layouts.
AVPixelFormat
Pixel format.
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
static int get_cbrt_h(int16_t sample, int height)
#define FFSWAP(type, a, b)
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
static enum AVSampleFormat sample_fmts[]
enum MovChannelLayoutTag * layouts
#define AVERROR_EOF
End of file.
int av_parse_color(uint8_t *rgba_color, const char *color_string, int slen, void *log_ctx)
Put the RGBA values that correspond to color_string in rgba_color.
@ AV_OPT_TYPE_VIDEO_RATE
offset must point to AVRational
AVRational av_div_q(AVRational b, AVRational c)
Divide one rational by another.
int(* get_h)(int16_t sample, int height)
AVFilterFormats * in_formats
Lists of formats and channel layouts supported by the input and output filters respectively.
The exact code depends on how similar the blocks are and how related they are to the and needs to apply these operations to the correct inlink or outlink if there are several Macros are available to factor that when no extra processing is inlink
static void draw_sample_point_rgba_scale(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
static av_cold int end(AVCodecContext *avctx)
This structure describes decoded (raw) audio or video data.
int ff_request_frame(AVFilterLink *link)
Request an input frame from the filter at the other end of the link.
void * av_mallocz_array(size_t nmemb, size_t size)
#define AV_LOG_VERBOSE
Detailed information.
static int get_cbrt_h2(int16_t sample, int height)
static const AVOption showwaves_options[]
AVFILTER_DEFINE_CLASS(showwaves)
const char * name
Filter name.
static int get_sqrt_h2(int16_t sample, int height)
A link between two filters.
#define FF_FILTER_FORWARD_STATUS_BACK(outlink, inlink)
Forward the status on an output link to an input link.
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
static void draw_sample_line_rgba_scale(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
static void draw_sample_cline_rgba_full(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
static int get_log_h2(int16_t sample, int height)
void * priv
private data for use by the filter
static int get_sqrt_h(int16_t sample, int height)
static int get_lin_h2(int16_t sample, int height)
static int get_lin_h(int16_t sample, int height)
A filter pad used for either input or output.
static void draw_sample_point_gray(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
static void draw_sample_p2p_gray(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
struct frame_node * last_frame
static double av_q2d(AVRational a)
Convert an AVRational to a double.
struct frame_node * audio_frames
char * av_strtok(char *s, const char *delim, char **saveptr)
Split the string into several tokens which can be accessed by successive calls to av_strtok().
AVRational sample_aspect_ratio
agreed upon sample aspect ratio
#define av_assert0(cond)
assert() equivalent, that is always enabled.
static const AVFilterPad outputs[]
AVRational frame_rate
Frame rate of the stream on the link, or 1/0 if unknown or variable; if left to 0/0,...
static enum AVPixelFormat pix_fmts[]
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
static av_cold int init(AVFilterContext *ctx)
int64_t av_rescale_q(int64_t a, AVRational bq, AVRational cq)
Rescale a 64-bit integer by 2 rational numbers.
@ AV_PIX_FMT_RGBA
packed RGBA 8:8:8:8, 32bpp, RGBARGBA...
#define FFABS(a)
Absolute value, Note, INT_MIN / INT64_MIN result in undefined behavior as they are not representable ...
static int config_output(AVFilterLink *outlink)
static int push_frame(AVFilterLink *outlink)
Describe the class of an AVClass context structure.
int ff_inlink_consume_samples(AVFilterLink *link, unsigned min, unsigned max, AVFrame **rframe)
Take samples from the link's FIFO and update the link's stats.
Rational number (pair of numerator and denominator).
static int get_log_h(int16_t sample, int height)
filter_frame For filters that do not use the activate() callback
@ AV_OPT_TYPE_IMAGE_SIZE
offset must point to two consecutive integers
AVFilterLink ** inputs
array of pointers to input links
static av_cold void uninit(AVFilterContext *ctx)
static int request_frame(AVFilterLink *outlink)
static int query_formats(AVFilterContext *ctx)
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several inputs
@ AV_PIX_FMT_GRAY8
Y , 8bpp.
int draw_mode
ShowWavesDrawMode.
static void draw_sample_cline_gray(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
AVFilter ff_avf_showwavespic
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
static void draw_sample_p2p_rgba_full(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
int format
agreed upon media format
static const uint32_t color[16+AV_CLASS_CATEGORY_NB]
static AVRational av_make_q(int num, int den)
Create an AVRational.
AVFilterContext * src
source filter
FF_FILTER_FORWARD_WANTED(outlink, inlink)
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(const uint8_t *) pi - 0x80) *(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(const int16_t *) pi >> 8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t, *(const int16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(const int32_t *) pi >> 24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t, *(const int32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(const float *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(const float *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(const float *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(const double *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(const double *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(const double *) pi *(1U<< 31)))) #define SET_CONV_FUNC_GROUP(ofmt, ifmt) static void set_generic_function(AudioConvert *ac) { } void ff_audio_convert_free(AudioConvert **ac) { if(! *ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);} AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enum AVSampleFormat out_fmt, enum AVSampleFormat in_fmt, int channels, int sample_rate, int apply_map) { AudioConvert *ac;int in_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) return NULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method !=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt) > 2) { ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc) { av_free(ac);return NULL;} return ac;} in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar) { ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar ? ac->channels :1;} else if(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;else ac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);return ac;} int ff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in) { int use_generic=1;int len=in->nb_samples;int p;if(ac->dc) { av_log(ac->avr, AV_LOG_TRACE, "%d samples - audio_convert: %s to %s (dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));return ff_convert_dither(ac-> in
static int push_single_pic(AVFilterLink *outlink)
int nb_samples
number of audio samples (per channel) described by this frame
#define i(width, name, range_min, range_max)
int w
agreed upon image width
#define av_malloc_array(a, b)
AVSampleFormat
Audio sample formats.
static void draw_sample_line_gray(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
@ AV_SAMPLE_FMT_S16
signed 16 bits
static void draw_sample_p2p_rgba_scale(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
const char * name
Pad name.
int64_t av_rescale(int64_t a, int64_t b, int64_t c)
Rescale a 64-bit integer with rounding to nearest.
void(* draw_sample)(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
int h
agreed upon image height
static void draw_sample_cline_rgba_scale(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
AVRational time_base
Define the time base used by the PTS of the frames/samples which will pass through this link.
static int alloc_out_frame(ShowWavesContext *showwaves, const int16_t *p, const AVFilterLink *inlink, AVFilterLink *outlink, const AVFrame *in)
int ff_outlink_get_status(AVFilterLink *link)
Get the status on an output link.
char * av_strdup(const char *s)
Duplicate a string.
FF_FILTER_FORWARD_STATUS(inlink, outlink)
#define AVERROR_BUG
Internal bug, also see AVERROR_BUG2.
int linesize[AV_NUM_DATA_POINTERS]
For video, size in bytes of each picture line.
AVFilter ff_avf_showwaves
static void draw_sample_line_rgba_full(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)
static void draw_sample_point_rgba_full(uint8_t *buf, int height, int linesize, int16_t *prev_y, const uint8_t color[4], int h)