Go to the documentation of this file.
43 buf[1] =
s->raw_seq->seq_profile << 5 |
s->raw_seq->seq_level_idx[0];
44 buf[2] =
s->raw_seq->seq_tier[0] << 7 |
45 s->raw_seq->color_config.high_bitdepth << 6 |
46 s->raw_seq->color_config.twelve_bit << 5 |
47 s->raw_seq->color_config.mono_chrome << 4 |
48 s->raw_seq->color_config.subsampling_x << 3 |
49 s->raw_seq->color_config.subsampling_y << 2 |
50 s->raw_seq->color_config.chroma_sample_position;
52 if (
s->raw_seq->initial_display_delay_present_flag)
54 s->raw_seq->initial_display_delay_present_flag << 4 |
55 s->raw_seq->initial_display_delay_minus_1[0];
58 memcpy(buf + 4,
s->seq_data_ref->data,
s->seq_data_ref->size);
59 data = CFDataCreate(kCFAllocatorDefault, buf,
s->seq_data_ref->size + 4);
86 for (
int i =
s->start_unit; i < s->nb_unit;
i++)
88 s->current_obu.units[
i].data_size);
93 .
p.
name =
"av1_videotoolbox",
int ff_videotoolbox_common_end_frame(AVCodecContext *avctx, AVFrame *frame)
int ff_videotoolbox_uninit(AVCodecContext *avctx)
AVHWAccel p
The public AVHWAccel.
This structure describes decoded (raw) audio or video data.
int ff_videotoolbox_common_init(AVCodecContext *avctx)
struct AVCodecInternal * internal
Private context used for internal data.
int ff_videotoolbox_frame_params(AVCodecContext *avctx, AVBufferRef *hw_frames_ctx)
void * hwaccel_priv_data
hwaccel-specific private data
const char * name
Name of the hardware accelerated codec.
@ AV_PIX_FMT_VIDEOTOOLBOX
hardware decoding through Videotoolbox
#define i(width, name, range_min, range_max)
these buffered frames must be flushed immediately if a new input produces new the filter must not call request_frame to get more It must just process the frame or queue it The task of requesting more frames is left to the filter s request_frame method or the application If a filter has several the filter must be ready for frames arriving randomly on any input any filter with several inputs will most likely require some kind of queuing mechanism It is perfectly acceptable to have a limited queue and to drop frames when the inputs are too unbalanced request_frame For filters that do not use the this method is called when a frame is wanted on an output For a it should directly call filter_frame on the corresponding output For a if there are queued frames already one of these frames should be pushed If the filter should request a frame on one of its repeatedly until at least one frame has been pushed Return or at least make progress towards producing a frame
main external API structure.
the frame and frame reference mechanism is intended to as much as expensive copies of that data while still allowing the filters to produce correct results The data is stored in buffers represented by AVFrame structures Several references can point to the same frame buffer