Go to the documentation of this file.
59 return RA_PIXEL_RANGE_FULL;
64 return RA_PIXEL_RANGE_FULL;
67 return RA_PIXEL_RANGE_LIMITED;
78 return RA_CHROMA_SAMPLING_CS420;
83 return RA_CHROMA_SAMPLING_CS422;
88 return RA_CHROMA_SAMPLING_CS444;
98 return RA_CHROMA_SAMPLE_POSITION_VERTICAL;
100 return RA_CHROMA_SAMPLE_POSITION_COLOCATED;
102 return RA_CHROMA_SAMPLE_POSITION_UNKNOWN;
109 RaData* buf = rav1e_twopass_out(
ctx->ctx);
115 ctx->pass_pos + buf->len);
117 rav1e_data_unref(buf);
122 memcpy(
ctx->pass_data +
ctx->pass_pos, buf->data, buf->len);
123 ctx->pass_pos += buf->len;
127 memcpy(
ctx->pass_data, buf->data, buf->len);
131 rav1e_data_unref(buf);
140 rav1e_data_unref(buf);
150 while (
ret > 0 &&
ctx->pass_size -
ctx->pass_pos > 0) {
151 ret = rav1e_twopass_in(
ctx->ctx,
ctx->pass_data +
ctx->pass_pos,
ctx->pass_size);
165 rav1e_context_unref(
ctx->ctx);
179 RaConfig *cfg =
NULL;
183 cfg = rav1e_config_default();
195 rav1e_config_set_time_base(cfg, (RaRational) {
199 rav1e_config_set_time_base(cfg, (RaRational) {
218 ctx->pass_size = (strlen(avctx->
stats_in) * 3) / 4;
220 if (!
ctx->pass_data) {
227 if (
ctx->pass_size < 0) {
240 "not found. This is a bug, please report it.\n");
267 int parse_ret = rav1e_config_parse(cfg, en->
key, en->
value);
273 rret = rav1e_config_parse_int(cfg,
"width", avctx->
width);
280 rret = rav1e_config_parse_int(cfg,
"height", avctx->
height);
287 rret = rav1e_config_parse_int(cfg,
"threads", avctx->
thread_count);
291 if (
ctx->speed >= 0) {
292 rret = rav1e_config_parse_int(cfg,
"speed",
ctx->speed);
301 if (
ctx->tiles > 0) {
302 rret = rav1e_config_parse_int(cfg,
"tiles",
ctx->tiles);
309 if (
ctx->tile_rows > 0) {
310 rret = rav1e_config_parse_int(cfg,
"tile_rows",
ctx->tile_rows);
317 if (
ctx->tile_cols > 0) {
318 rret = rav1e_config_parse_int(cfg,
"tile_cols",
ctx->tile_cols);
327 rret = rav1e_config_parse_int(cfg,
"key_frame_interval", avctx->
gop_size);
336 rret = rav1e_config_parse_int(cfg,
"min_key_frame_interval", avctx->
keyint_min);
345 int max_quantizer = avctx->
qmax >= 0 ? avctx->
qmax : 255;
347 rret = rav1e_config_parse_int(cfg,
"quantizer", max_quantizer);
354 if (avctx->
qmin >= 0) {
355 rret = rav1e_config_parse_int(cfg,
"min_quantizer", avctx->
qmin);
363 rret = rav1e_config_parse_int(cfg,
"bitrate", avctx->
bit_rate);
369 }
else if (
ctx->quantizer >= 0) {
373 rret = rav1e_config_parse_int(cfg,
"quantizer",
ctx->quantizer);
381 rret = rav1e_config_set_pixel_format(cfg,
desc->comp[0].depth,
392 rret = rav1e_config_set_color_description(cfg, (RaMatrixCoefficients) avctx->
colorspace,
394 (RaTransferCharacteristics) avctx->
color_trc);
403 ctx->ctx = rav1e_context_new(cfg);
414 rav1e_config_unref(cfg);
422 RaFrame *rframe =
NULL;
428 rframe = rav1e_frame_new(
ctx->ctx);
434 for (
int i = 0;
i <
desc->nb_components;
i++) {
436 int bytes =
desc->comp[0].depth == 8 ? 1 : 2;
437 rav1e_frame_fill_plane(rframe,
i,
frame->data[
i],
439 frame->linesize[
i], bytes);
443 ret = rav1e_send_frame(
ctx->ctx, rframe);
445 rav1e_frame_unref(rframe);
448 case RA_ENCODER_STATUS_SUCCESS:
450 case RA_ENCODER_STATUS_ENOUGH_DATA:
452 case RA_ENCODER_STATUS_FAILURE:
466 RaPacket *rpkt =
NULL;
481 ret = rav1e_receive_packet(
ctx->ctx, &rpkt);
483 case RA_ENCODER_STATUS_SUCCESS:
485 case RA_ENCODER_STATUS_LIMIT_REACHED:
492 case RA_ENCODER_STATUS_ENCODED:
496 case RA_ENCODER_STATUS_NEED_MORE_DATA:
502 case RA_ENCODER_STATUS_FAILURE:
506 av_log(avctx,
AV_LOG_ERROR,
"Unknown return code %d from rav1e_receive_packet: %s\n",
ret, rav1e_status_to_str(
ret));
513 rav1e_packet_unref(rpkt);
517 memcpy(
pkt->
data, rpkt->data, rpkt->len);
519 if (rpkt->frame_type == RA_FRAME_TYPE_KEY)
523 rav1e_packet_unref(rpkt);
544 #define OFFSET(x) offsetof(librav1eContext, x)
545 #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
553 {
"rav1e-params",
"set the rav1e configuration using a :-separated list of key=value parameters",
OFFSET(rav1e_opts),
AV_OPT_TYPE_DICT, { 0 }, 0, 0,
VE },
560 {
"keyint_min",
"0" },
599 .priv_class = &
class,
604 .wrapper_name =
"librav1e",
void av_packet_unref(AVPacket *pkt)
Wipe the packet.
#define AV_LOG_WARNING
Something somehow does not look correct.
AVPixelFormat
Pixel format.
int keyint_min
minimum GOP size
Filter the word “frame” indicates either a video frame or a group of audio as stored in an AVFrame structure Format for each input and each output the list of supported formats For video that means pixel format For audio that means channel sample they are references to shared objects When the negotiation mechanism computes the intersection of the formats supported at each end of a all references to both lists are replaced with a reference to the intersection And when a single format is eventually chosen for a link amongst the remaining all references to the list are updated That means that if a filter requires that its input and output have the same format amongst a supported all it has to do is use a reference to the same list of formats query_formats can leave some formats unset and return AVERROR(EAGAIN) to cause the negotiation mechanism toagain later. That can be used by filters with complex requirements to use the format negotiated on one link to set the formats supported on another. Frame references ownership and permissions
enum AVColorSpace colorspace
YUV colorspace type.
const AVPixFmtDescriptor * av_pix_fmt_desc_get(enum AVPixelFormat pix_fmt)
#define AVERROR_EOF
End of file.
int avcodec_parameters_from_context(AVCodecParameters *par, const AVCodecContext *codec)
Fill the parameters struct based on the values from the supplied codec context.
int err_recognition
Error recognition; may misdetect some more or less valid parts as errors.
int av_bsf_init(AVBSFContext *ctx)
Prepare the filter for use, after all the parameters and options have been set.
static av_cold int end(AVCodecContext *avctx)
This structure describes decoded (raw) audio or video data.
enum AVColorTransferCharacteristic color_trc
Color Transfer Characteristic.
@ AVCOL_RANGE_JPEG
the normal 2^n-1 "JPEG" YUV ranges
static int set_stats(AVCodecContext *avctx)
#define AV_PIX_FMT_YUV420P10
#define AV_DICT_IGNORE_SUFFIX
Return first entry in a dictionary whose first part corresponds to the search key,...
filter_frame For filters that do not use the this method is called when a frame is pushed to the filter s input It can be called at any time except in a reentrant way If the input frame is enough to produce then the filter should push the output frames on the output link immediately As an exception to the previous rule if the input frame is enough to produce several output frames then the filter needs output only at least one per link The additional frames can be left buffered in the filter
#define AVERROR_UNKNOWN
Unknown error, typically from an external library.
int qmax
maximum quantizer
#define AV_PKT_FLAG_KEY
The packet contains a keyframe.
The bitstream filter state.
#define AV_CODEC_FLAG_GLOBAL_HEADER
Place global headers in extradata instead of every keyframe.
const AVBitStreamFilter * av_bsf_get_by_name(const char *name)
int thread_count
thread count is used to decide how many independent tasks should be passed to execute()
static const AVCodecDefault defaults[]
int flags
AV_CODEC_FLAG_*.
#define AV_PIX_FMT_YUV444P10
enum AVColorPrimaries color_primaries
Chromaticity coordinates of the source primaries.
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
AVDictionaryEntry * av_dict_get(const AVDictionary *m, const char *key, const AVDictionaryEntry *prev, int flags)
Get a dictionary entry with matching key.
@ AV_PIX_FMT_YUVJ422P
planar YUV 4:2:2, 16bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV422P and setting col...
void * av_fast_realloc(void *ptr, unsigned int *size, size_t min_size)
Reallocate the given buffer if it is not large enough, otherwise do nothing.
char * stats_in
pass2 encoding statistics input buffer Concatenated stuff from stats_out of pass1 should be placed he...
int av_new_packet(AVPacket *pkt, int size)
Allocate the payload of a packet and initialize its fields with default values.
static av_cold int librav1e_encode_init(AVCodecContext *avctx)
static const AVCodecDefault librav1e_defaults[]
int ticks_per_frame
For some codecs, the time base is closer to the field rate than the frame rate.
#define av_assert0(cond)
assert() equivalent, that is always enabled.
static enum AVPixelFormat pix_fmts[]
static enum AVPixelFormat pix_fmt
@ AV_PIX_FMT_YUV420P
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
@ AV_PIX_FMT_YUVJ444P
planar YUV 4:4:4, 24bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV444P and setting col...
#define LIBAVUTIL_VERSION_INT
Describe the class of an AVClass context structure.
enum AVColorRange color_range
MPEG vs JPEG YUV range.
AVCodec ff_librav1e_encoder
@ AVCHROMA_LOC_LEFT
MPEG-2/4 4:2:0, H.264 default for 4:2:0.
AVDictionary * rav1e_opts
@ AVCHROMA_LOC_TOPLEFT
ITU-R 601, SMPTE 274M 296M S314M(DV 4:1:1), mpeg2 4:2:2.
struct AVCodecInternal * internal
Private context used for internal data.
@ AV_PIX_FMT_YUVJ420P
planar YUV 4:2:0, 12bpp, full scale (JPEG), deprecated in favor of AV_PIX_FMT_YUV420P and setting col...
int64_t bit_rate
the average bitrate
const char * av_default_item_name(void *ptr)
Return the context name.
#define AV_PIX_FMT_YUV422P10
int av_base64_decode(uint8_t *out, const char *in_str, int out_size)
Decode a base64-encoded string.
#define AV_EF_EXPLODE
abort decoding on minor error detection
static int librav1e_receive_packet(AVCodecContext *avctx, AVPacket *pkt)
AVRational time_base
This is the fundamental unit of time (in seconds) in terms of which frame timestamps are represented.
char * stats_out
pass1 encoding statistics output buffer
#define AV_CODEC_CAP_AUTO_THREADS
Codec supports avctx->thread_count == 0 (auto).
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification.
int gop_size
the number of pictures in a group of pictures, or 0 for intra_only
#define AV_PIX_FMT_YUV422P12
#define AV_PIX_FMT_YUV444P12
static RaPixelRange range_map(enum AVPixelFormat pix_fmt, enum AVColorRange range)
int64_t dts
Decompression timestamp in AVStream->time_base units; the time at which the packet is decompressed.
#define AV_CODEC_FLAG_PASS2
Use internal 2pass ratecontrol in second pass mode.
static RaChromaSampling pix_fmt_map(enum AVPixelFormat pix_fmt)
#define AVERROR_EXTERNAL
Generic error in an external library.
int flags
A combination of AV_PKT_FLAG values.
AVChromaLocation
Location of chroma samples.
#define AV_BASE64_SIZE(x)
Calculate the output size needed to base64-encode x bytes to a null-terminated string.
int av_bsf_receive_packet(AVBSFContext *ctx, AVPacket *pkt)
Retrieve a filtered packet.
#define i(width, name, range_min, range_max)
int64_t pts
Presentation timestamp in AVStream->time_base units; the time at which the decompressed packet will b...
#define FF_CODEC_CAP_INIT_CLEANUP
The codec allows calling the close function for deallocation even if the init function returned a fai...
const char * name
Name of the codec implementation.
enum AVChromaLocation chroma_sample_location
This defines the location of chroma samples.
enum AVPixelFormat pix_fmt
Pixel format, see AV_PIX_FMT_xxx.
@ AVCOL_RANGE_MPEG
the normal 219*2^(n-8) "MPEG" YUV ranges
static const AVOption options[]
const char * class_name
The name of the class; usually it is the same name as the context structure type to which the AVClass...
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
#define AV_PIX_FMT_YUV420P12
int av_bsf_send_packet(AVBSFContext *ctx, AVPacket *pkt)
Submit a packet for filtering.
main external API structure.
int qmin
minimum quantizer
char * av_base64_encode(char *out, int out_size, const uint8_t *in, int in_size)
Encode data to base64 and null-terminate.
#define AV_CODEC_CAP_DELAY
Encoder or decoder requires flushing with NULL input at the end in order to give the complete and cor...
static RaChromaSamplePosition chroma_loc_map(enum AVChromaLocation chroma_loc)
@ AV_PIX_FMT_YUV444P
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
int draining
checks API usage: after codec draining, flush is required to resume operation
static int shift(int a, int b)
@ AV_PIX_FMT_YUV422P
planar YUV 4:2:2, 16bpp, (1 Cr & Cb sample per 2x1 Y samples)
enum AVPixelFormat librav1e_pix_fmts[]
static int librav1e_send_frame(AVCodecContext *avctx, const AVFrame *frame)
Descriptor that unambiguously describes how the bits of a pixel are stored in the up to 4 data planes...
This structure stores compressed data.
void av_bsf_free(AVBSFContext **pctx)
Free a bitstream filter context and everything associated with it; write NULL into the supplied point...
int width
picture width / height.
#define AVERROR_BUG
Internal bug, also see AVERROR_BUG2.
#define AVERROR_INVALIDDATA
Invalid data found when processing input.
AVColorRange
MPEG vs JPEG YUV range.
int av_bsf_alloc(const AVBitStreamFilter *filter, AVBSFContext **pctx)
Allocate a context for a given bitstream filter.
static av_cold int librav1e_encode_close(AVCodecContext *avctx)
#define AV_CODEC_FLAG_PASS1
Use internal 2pass ratecontrol in first pass mode.
static int get_stats(AVCodecContext *avctx, int eos)