Go to the documentation of this file.
155 #define YAE_ATEMPO_MIN 0.5
156 #define YAE_ATEMPO_MAX 100.0
158 #define OFFSET(x) offsetof(ATempoContext, x)
161 {
"tempo",
"set tempo scale factor",
178 return &atempo->
frag[(atempo->
nfrag + 1) % 2];
247 #define RE_MALLOC_OR_FAIL(field, field_size) \
250 field = av_malloc(field_size); \
252 yae_release_buffers(atempo); \
253 return AVERROR(ENOMEM); \
267 uint32_t nlevels = 0;
283 if (pot < atempo->
window) {
322 double t = (double)
i / (
double)(atempo->
window - 1);
323 double h = 0.5 * (1.0 - cos(2.0 *
M_PI * t));
324 atempo->
hann[
i] = (float)
h;
336 double tempo =
av_strtod(arg_tempo, &tail);
352 atempo->
tempo = tempo;
360 #define yae_init_xdat(scalar_type, scalar_max) \
362 const uint8_t *src_end = src + \
363 frag->nsamples * atempo->channels * sizeof(scalar_type); \
365 FFTSample *xdat = frag->xdat; \
368 if (atempo->channels == 1) { \
369 for (; src < src_end; xdat++) { \
370 tmp = *(const scalar_type *)src; \
371 src += sizeof(scalar_type); \
373 *xdat = (FFTSample)tmp; \
376 FFTSample s, max, ti, si; \
379 for (; src < src_end; xdat++) { \
380 tmp = *(const scalar_type *)src; \
381 src += sizeof(scalar_type); \
383 max = (FFTSample)tmp; \
384 s = FFMIN((FFTSample)scalar_max, \
385 (FFTSample)fabsf(max)); \
387 for (i = 1; i < atempo->channels; i++) { \
388 tmp = *(const scalar_type *)src; \
389 src += sizeof(scalar_type); \
391 ti = (FFTSample)tmp; \
392 si = FFMIN((FFTSample)scalar_max, \
393 (FFTSample)fabsf(ti)); \
445 const int read_size = stop_here - atempo->
position[0];
447 if (stop_here <= atempo->position[0]) {
454 while (atempo->
position[0] < stop_here &&
src < src_end) {
455 int src_samples = (src_end -
src) / atempo->
stride;
458 int nsamples =
FFMIN(read_size, src_samples);
462 nsamples =
FFMIN(nsamples, atempo->
ring);
520 int64_t missing,
start, zeros;
523 int i0, i1, n0, n1, na, nb;
526 if (src_ref &&
yae_load_data(atempo, src_ref, src_end, stop_here) != 0) {
533 stop_here - atempo->
position[0] : 0;
536 missing < (int64_t)atempo->
window ?
537 (uint32_t)(atempo->
window - missing) : 0;
551 memset(dst, 0, zeros * atempo->
stride);
552 dst += zeros * atempo->
stride;
555 if (zeros == nsamples) {
573 i1 = i0 < na ? 0 : i0 - na;
575 n0 = i0 < na ?
FFMIN(na - i0, (
int)(nsamples - zeros)) : 0;
576 n1 = nsamples - zeros - n0;
580 dst += n0 * atempo->
stride;
595 const double fragment_step = atempo->
tempo * (double)(atempo->
window / 2);
634 for (
i = 1;
i <
window;
i++, xa++, xb++, xc++) {
635 xc->
re = (xa->re * xb->re + xa->im * xb->im);
636 xc->
im = (xa->im * xb->re - xa->re * xb->im);
657 int best_offset = -drift;
681 for (
i = i0;
i < i1;
i++, xcorr++) {
688 if (metric > best_metric) {
689 best_metric = metric;
708 const double prev_output_position =
712 const double ideal_output_position =
715 const int drift = (
int)(prev_output_position - ideal_output_position);
717 const int delta_max = atempo->
window / 2;
741 #define yae_blend(scalar_type) \
743 const scalar_type *aaa = (const scalar_type *)a; \
744 const scalar_type *bbb = (const scalar_type *)b; \
746 scalar_type *out = (scalar_type *)dst; \
747 scalar_type *out_end = (scalar_type *)dst_end; \
750 for (i = 0; i < overlap && out < out_end; \
751 i++, atempo->position[1]++, wa++, wb++) { \
756 for (j = 0; j < atempo->channels; \
757 j++, aaa++, bbb++, out++) { \
758 float t0 = (float)*aaa; \
759 float t1 = (float)*bbb; \
762 frag->position[0] + i < 0 ? \
764 (scalar_type)(t0 * w0 + t1 * w1); \
767 dst = (uint8_t *)out; \
792 const int64_t overlap = stop_here - start_here;
794 const int64_t ia = start_here - prev->
position[1];
795 const int64_t
ib = start_here - frag->
position[1];
797 const float *wa = atempo->
hann + ia;
798 const float *wb = atempo->
hann +
ib;
807 overlap <= frag->nsamples);
853 if (!atempo->
nfrag) {
926 if (!atempo->
nfrag) {
960 while (atempo->
position[1] < overlap_end) {
976 av_assert0(start_here <= stop_here && frag->position[1] <= start_here);
981 src_size = (
int)(stop_here - start_here) * atempo->
stride;
982 dst_size = dst_end - dst;
983 nbytes =
FFMIN(src_size, dst_size);
985 memcpy(dst,
src, nbytes);
1099 int n_out = (
int)(0.5 + ((
double)n_in) / atempo->
tempo);
1109 while (
src < src_end) {
1149 int n_max = atempo->
ring;
1153 while (err ==
AVERROR(EAGAIN)) {
1222 .priv_class = &atempo_class,
AVFrame * ff_get_audio_buffer(AVFilterLink *link, int nb_samples)
Request an audio samples buffer with a specific set of permissions.
A list of supported channel layouts.
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
static int push_samples(ATempoContext *atempo, AVFilterLink *outlink, int n_out)
static int config_props(AVFilterLink *inlink)
int ff_filter_frame(AVFilterLink *link, AVFrame *frame)
Send a frame of data to the next filter.
static enum AVSampleFormat sample_fmts[]
enum MovChannelLayoutTag * layouts
#define AVERROR_EOF
End of file.
The exact code depends on how similar the blocks are and how related they are to the and needs to apply these operations to the correct inlink or outlink if there are several Macros are available to factor that when no extra processing is inlink
static void yae_downmix(ATempoContext *atempo, AudioFragment *frag)
Initialize complex data buffer of a given audio fragment with down-mixed mono data of appropriate sca...
void av_frame_free(AVFrame **frame)
Free the frame and any dynamically allocated objects in it, e.g.
static av_cold int end(AVCodecContext *avctx)
static int yae_load_data(ATempoContext *atempo, const uint8_t **src_ref, const uint8_t *src_end, int64_t stop_here)
Populate the internal data buffer on as-needed basis.
This structure describes decoded (raw) audio or video data.
int64_t pts
Presentation timestamp in time_base units (time when frame should be shown to user).
int ff_request_frame(AVFilterLink *link)
Request an input frame from the filter at the other end of the link.
const char * name
Filter name.
A link between two filters.
#define AV_OPT_FLAG_FILTERING_PARAM
a generic parameter which can be set by the user for filtering
static av_cold int init(AVFilterContext *ctx)
uint8_t * data[AV_NUM_DATA_POINTERS]
pointer to the picture/channel planes.
static SDL_Window * window
static int query_formats(AVFilterContext *ctx)
static const AVFilterPad atempo_outputs[]
A filter pad used for either input or output.
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
static void yae_apply(ATempoContext *atempo, const uint8_t **src_ref, const uint8_t *src_end, uint8_t **dst_ref, uint8_t *dst_end)
Feed as much data to the filter as it is able to consume and receive as much processed data in the de...
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample format(the sample packing is implied by the sample format) and sample rate. The lists are not just lists
#define yae_init_xdat(scalar_type, scalar_max)
A helper macro for initializing complex data buffer with scalar data of a given type.
#define av_assert0(cond)
assert() equivalent, that is always enabled.
static const AVFilterPad outputs[]
RDFTContext * real_to_complex
int64_t av_rescale_q(int64_t a, AVRational bq, AVRational cq)
Rescale a 64-bit integer by 2 rational numbers.
void av_rdft_calc(RDFTContext *s, FFTSample *data)
#define AV_OPT_FLAG_AUDIO_PARAM
#define RE_MALLOC_OR_FAIL(field, field_size)
RDFTContext * complex_to_real
static int yae_align(AudioFragment *frag, const AudioFragment *prev, const int window, const int delta_max, const int drift, FFTSample *correlation, RDFTContext *complex_to_real)
Calculate alignment offset for given fragment relative to the previous fragment.
Describe the class of an AVClass context structure.
int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
Copy only "metadata" fields from src to dst.
Rational number (pair of numerator and denominator).
AVFILTER_DEFINE_CLASS(atempo)
static int filter_frame(AVFilterLink *inlink, AVFrame *src_buffer)
static AudioFragment * yae_curr_frag(ATempoContext *atempo)
static int yae_reset(ATempoContext *atempo, enum AVSampleFormat format, int sample_rate, int channels)
Prepare filter for processing audio data of given format, sample rate and number of channels.
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several inputs
static void correlation(int32_t *corr, int32_t *ener, int16_t *buffer, int16_t lag, int16_t blen, int16_t srange, int16_t scale)
static int process_command(AVFilterContext *ctx, const char *cmd, const char *arg, char *res, int res_len, int flags)
static int yae_overlap_add(ATempoContext *atempo, uint8_t **dst_ref, uint8_t *dst_end)
Blend the overlap region of previous and current audio fragment and output the results to the given d...
static int yae_load_frag(ATempoContext *atempo, const uint8_t **src_ref, const uint8_t *src_end)
Populate current audio fragment data buffer.
RDFTContext * av_rdft_init(int nbits, enum RDFTransformType trans)
Set up a real FFT.
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
static AudioFragment * yae_prev_frag(ATempoContext *atempo)
int sample_rate
Sample rate of the audio data.
#define AV_NOPTS_VALUE
Undefined timestamp value.
AVFilterContext * src
source filter
The reader does not expect b to be semantically here and if the code is changed by maybe adding a a division or other the signedness will almost certainly be mistaken To avoid this confusion a new type was SUINT is the C unsigned type but it holds a signed int to use the same example SUINT a
it s the only field you need to keep assuming you have a context There is some magic you don t need to care about around this just let it vf offset
static int yae_set_tempo(AVFilterContext *ctx, const char *arg_tempo)
#define yae_blend(scalar_type)
A helper macro for blending the overlap region of previous and current audio fragment.
int sample_rate
samples per second
int nb_samples
number of audio samples (per channel) described by this frame
#define i(width, name, range_min, range_max)
int av_get_bytes_per_sample(enum AVSampleFormat sample_fmt)
Return number of bytes per sample.
static void yae_advance_to_next_frag(ATempoContext *atempo)
Prepare for loading next audio fragment.
static void yae_clear(ATempoContext *atempo)
Reset filter to initial state, do not deallocate existing local buffers.
AVSampleFormat
Audio sample formats.
@ AV_SAMPLE_FMT_S16
signed 16 bits
const char * name
Pad name.
static const AVFilterPad atempo_inputs[]
static int yae_flush(ATempoContext *atempo, uint8_t **dst_ref, uint8_t *dst_end)
Flush any buffered data from the filter.
static const AVOption atempo_options[]
double av_strtod(const char *numstr, char **tail)
Parse the string in numstr and return its value as a double.
static int request_frame(AVFilterLink *outlink)
static void yae_xcorr_via_rdft(FFTSample *xcorr, RDFTContext *complex_to_real, const FFTComplex *xa, const FFTComplex *xb, const int window)
Calculate cross-correlation via rDFT.
enum AVSampleFormat format
static int yae_adjust_position(ATempoContext *atempo)
Adjust current fragment position for better alignment with previous fragment.
FilterState
Filter state machine states.
AVRational time_base
Define the time base used by the PTS of the frames/samples which will pass through this link.
static void yae_release_buffers(ATempoContext *atempo)
Reset filter to initial state and deallocate all buffers.
static av_cold void uninit(AVFilterContext *ctx)
#define flags(name, subs,...)
void av_rdft_end(RDFTContext *s)
@ AV_SAMPLE_FMT_DBL
double
@ AV_SAMPLE_FMT_S32
signed 32 bits
A fragment of audio waveform.