// FFMPEG Video Encoder Integration for OBS Studio // Copyright (c) 2019 Michael Fabian Dirks // // Permission is hereby granted, free of charge, to any person obtaining a copy // of this software and associated documentation files (the "Software"), to deal // in the Software without restriction, including without limitation the rights // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell // copies of the Software, and to permit persons to whom the Software is // furnished to do so, subject to the following conditions: // // The above copyright notice and this permission notice shall be included in all // copies or substantial portions of the Software. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE // SOFTWARE. #include "ffmpeg-encoder.hpp" #include "strings.hpp" #include #include "codecs/hevc.hpp" #include "ffmpeg/tools.hpp" #include "handlers/debug_handler.hpp" #include "handlers/nvenc_h264_handler.hpp" #include "handlers/nvenc_hevc_handler.hpp" #include "handlers/prores_aw_handler.hpp" #include "obs/gs/gs-helper.hpp" #include "plugin.hpp" #include "utility.hpp" extern "C" { #pragma warning(push) #pragma warning(disable : 4244) #include #include #include #include #include #include #pragma warning(pop) } #ifdef WIN32 #define HARDWARE_ENCODING #include "ffmpeg/hwapi/d3d11.hpp" #endif // FFmpeg #define ST_FFMPEG "FFmpegEncoder" #define ST_FFMPEG_CUSTOMSETTINGS "FFmpegEncoder.CustomSettings" #define KEY_FFMPEG_CUSTOMSETTINGS "FFmpeg.CustomSettings" #define ST_FFMPEG_THREADS "FFmpegEncoder.Threads" #define KEY_FFMPEG_THREADS "FFmpeg.Threads" #define ST_FFMPEG_COLORFORMAT "FFmpegEncoder.ColorFormat" #define KEY_FFMPEG_COLORFORMAT "FFmpeg.ColorFormat" #define ST_FFMPEG_STANDARDCOMPLIANCE "FFmpegEncoder.StandardCompliance" #define KEY_FFMPEG_STANDARDCOMPLIANCE "FFmpeg.StandardCompliance" #define ST_FFMPEG_GPU "FFmpegEncoder.GPU" #define KEY_FFMPEG_GPU "FFmpeg.GPU" #define ST_KEYFRAMES "FFmpegEncoder.KeyFrames" #define ST_KEYFRAMES_INTERVALTYPE "FFmpegEncoder.KeyFrames.IntervalType" #define ST_KEYFRAMES_INTERVALTYPE_(x) "FFmpegEncoder.KeyFrames.IntervalType." D_VSTR(x) #define KEY_KEYFRAMES_INTERVALTYPE "KeyFrames.IntervalType" #define ST_KEYFRAMES_INTERVAL "FFmpegEncoder.KeyFrames.Interval" #define ST_KEYFRAMES_INTERVAL_SECONDS "FFmpegEncoder.KeyFrames.Interval.Seconds" #define KEY_KEYFRAMES_INTERVAL_SECONDS "KeyFrames.Interval.Seconds" #define ST_KEYFRAMES_INTERVAL_FRAMES "FFmpegEncoder.KeyFrames.Interval.Frames" #define KEY_KEYFRAMES_INTERVAL_FRAMES "KeyFrames.Interval.Frames" using namespace streamfx::encoder::ffmpeg; using namespace streamfx::encoder::codec; enum class keyframe_type { SECONDS, FRAMES }; static void* _create(obs_data_t* settings, obs_encoder_t* encoder) noexcept try { return reinterpret_cast(new ffmpeg_instance(settings, encoder)); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return nullptr; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return nullptr; } static void* _create_texture(obs_data_t* settings, obs_encoder_t* encoder) noexcept try { return reinterpret_cast(new ffmpeg_instance(settings, encoder, true)); } catch (const std::exception& ex) { ffmpeg_factory* fac = reinterpret_cast(obs_encoder_get_type_data(encoder)); LOG_ERROR("<%s> Failed to create GPU assisted encoder due to exception: %s.", fac->get_avcodec()->name, ex.what()); return obs_encoder_create_rerouted(encoder, fac->get_fallback().oei.id); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return nullptr; } static void _destroy(void* ptr) noexcept try { if (ptr) delete reinterpret_cast(ptr); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); } static const char* _get_name(void* type_data) noexcept try { return reinterpret_cast(type_data)->get_info().readable_name.c_str(); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return nullptr; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return nullptr; } static const char* _get_name_fallback(void* type_data) noexcept try { return reinterpret_cast(type_data)->get_fallback().readable_name.c_str(); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return nullptr; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return nullptr; } static void _get_defaults(obs_data_t* settings, void* type_data) noexcept try { reinterpret_cast(type_data)->get_defaults(settings); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); }; static void _get_defaults_texture(obs_data_t* settings, void* type_data) noexcept try { reinterpret_cast(type_data)->get_defaults(settings, true); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); }; static obs_properties_t* _get_properties(void* ptr, void* type_data) noexcept try { obs_properties_t* props = obs_properties_create(); if (type_data != nullptr) { reinterpret_cast(type_data)->get_properties(props); } if (ptr != nullptr) { reinterpret_cast(ptr)->get_properties(props); } return props; } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return reinterpret_cast(0); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return reinterpret_cast(0); } static obs_properties_t* _get_properties_texture(void* ptr, void* type_data) noexcept try { obs_properties_t* props = obs_properties_create(); if (type_data != nullptr) { reinterpret_cast(type_data)->get_properties(props, true); } if (ptr != nullptr) { reinterpret_cast(ptr)->get_properties(props, true); } return props; } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return reinterpret_cast(0); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return reinterpret_cast(0); } static bool _update(void* ptr, obs_data_t* settings) noexcept try { return reinterpret_cast(ptr)->update(settings); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return false; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return false; } static bool _get_sei_data(void* ptr, std::uint8_t** sei_data, size_t* size) noexcept try { return reinterpret_cast(ptr)->get_sei_data(sei_data, size); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return false; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return false; } static bool _get_extra_data(void* ptr, std::uint8_t** extra_data, size_t* size) noexcept try { return reinterpret_cast(ptr)->get_extra_data(extra_data, size); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return false; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return false; } static void _get_video_info(void* ptr, struct video_scale_info* info) noexcept try { reinterpret_cast(ptr)->get_video_info(info); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); } static bool _encode(void* ptr, struct encoder_frame* frame, struct encoder_packet* packet, bool* received_packet) noexcept try { return reinterpret_cast(ptr)->video_encode(frame, packet, received_packet); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return false; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return false; } static bool _encode_texture(void* ptr, std::uint32_t handle, std::int64_t pts, std::uint64_t lock_key, std::uint64_t* next_key, struct encoder_packet* packet, bool* received_packet) noexcept try { return reinterpret_cast(ptr)->video_encode_texture(handle, pts, lock_key, next_key, packet, received_packet); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return false; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return false; } static void _get_audio_info(void* ptr, struct audio_convert_info* info) noexcept try { reinterpret_cast(ptr)->get_audio_info(info); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); } static std::size_t _get_frame_size(void* ptr) noexcept try { return reinterpret_cast(ptr)->get_frame_size(); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return 0; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return 0; } static bool _encode_audio(void* ptr, struct encoder_frame* frame, struct encoder_packet* packet, bool* received_packet) noexcept try { return reinterpret_cast(ptr)->audio_encode(frame, packet, received_packet); } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return false; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return false; } ffmpeg_factory::ffmpeg_factory(const AVCodec* codec) : info(), info_fallback(), avcodec_ptr(codec) { // Find Codec UI handler. _handler = ffmpeg_manager::get()->get_handler(avcodec_ptr->name); // Unique Id is FFmpeg name. info.uid = std::string(PLUGIN_NAME "-") + avcodec_ptr->name; // Also generate a human readable name while we're at it. { std::stringstream sstr; if (!ffmpeg_manager::get()->has_handler(avcodec_ptr->name)) { sstr << "[UNSUPPORTED] "; } sstr << (avcodec_ptr->long_name ? avcodec_ptr->long_name : avcodec_ptr->name); if (avcodec_ptr->long_name) { sstr << " (" << avcodec_ptr->name << ")"; } info.readable_name = sstr.str(); } // Assign Ids. { const AVCodecDescriptor* desc = avcodec_descriptor_get(avcodec_ptr->id); if (desc) { info.codec = desc->name; } else { // Fall back to encoder name in the case that FFmpeg itself doesn't know // what codec this actually is. info.codec = avcodec_ptr->name; } } info.oei.id = info.uid.c_str(); info.oei.codec = info.codec.c_str(); #ifndef _DEBUG // Is this a deprecated encoder? if (!ffmpeg_manager::get()->has_handler(avcodec_ptr->name)) { info.oei.caps |= OBS_ENCODER_CAP_DEPRECATED; } #endif // Hardware encoder? #ifdef HARDWARE_ENCODING if (::ffmpeg::tools::can_hardware_encode(avcodec_ptr)) { info_fallback.uid = info.uid + "_sw"; info_fallback.codec = info.codec; info_fallback.readable_name = info.readable_name + " (Software)"; // Copy capabilities and hide from view. info_fallback.oei.id = info_fallback.uid.c_str(); info_fallback.oei.codec = info.oei.codec; info_fallback.oei.caps = info.oei.caps; //info_fallback.oei.caps |= OBS_ENCODER_CAP_DEPRECATED; info.oei.caps |= OBS_ENCODER_CAP_PASS_TEXTURE; } #endif } ffmpeg_factory::~ffmpeg_factory() {} void ffmpeg_factory::register_encoder() { // Detect encoder type (only Video and Audio supported) if (avcodec_ptr->type == AVMediaType::AVMEDIA_TYPE_VIDEO) { info.oei.type = obs_encoder_type::OBS_ENCODER_VIDEO; } else if (avcodec_ptr->type == AVMediaType::AVMEDIA_TYPE_AUDIO) { info.oei.type = obs_encoder_type::OBS_ENCODER_AUDIO; } else { throw std::invalid_argument("unsupported codec type"); } // Register functions. info.oei.destroy = _destroy; info.oei.get_name = _get_name; info.oei.get_defaults2 = _get_defaults; info.oei.get_properties2 = _get_properties; info.oei.update = _update; info.oei.get_sei_data = _get_sei_data; info.oei.get_extra_data = _get_extra_data; if (avcodec_ptr->type == AVMediaType::AVMEDIA_TYPE_VIDEO) { info.oei.get_video_info = _get_video_info; } else if (avcodec_ptr->type == AVMediaType::AVMEDIA_TYPE_AUDIO) { info.oei.get_audio_info = _get_audio_info; info.oei.get_frame_size = _get_frame_size; info.oei.encode = _encode_audio; } // Finally store ourself as type data. info.oei.type_data = this; if (::ffmpeg::tools::can_hardware_encode(avcodec_ptr)) { info.oei.create = _create_texture; info.oei.encode_texture = _encode_texture; info.oei.get_defaults2 = _get_defaults_texture; info.oei.get_properties2 = _get_properties_texture; info_fallback.oei.type = info.oei.type; info_fallback.oei.create = _create; info_fallback.oei.destroy = _destroy; info_fallback.oei.get_name = _get_name_fallback; info_fallback.oei.get_defaults2 = _get_defaults; info_fallback.oei.get_properties2 = _get_properties; info_fallback.oei.update = _update; info_fallback.oei.get_sei_data = _get_sei_data; info_fallback.oei.get_extra_data = _get_extra_data; info_fallback.oei.get_video_info = _get_video_info; info_fallback.oei.encode = _encode; info_fallback.oei.type_data = this; } else { // Is not a GPU Encoder, don't implement fallback. info.oei.create = _create; info.oei.encode = _encode; } if (_handler) _handler->adjust_encoder_info(this, &info, &info_fallback); obs_register_encoder(&info.oei); LOG_DEBUG("Registered encoder #%llX with name '%s' and long name '%s' and caps %llX", avcodec_ptr, avcodec_ptr->name, avcodec_ptr->long_name, avcodec_ptr->capabilities); if (info_fallback.uid.size() > 0) { obs_register_encoder(&info_fallback.oei); LOG_DEBUG("Registered software fallback for encoder #%llX", avcodec_ptr); } } void ffmpeg_factory::get_defaults(obs_data_t* settings, bool hw_encode) { if (_handler) _handler->get_defaults(settings, avcodec_ptr, nullptr, hw_encode); if ((avcodec_ptr->capabilities & AV_CODEC_CAP_INTRA_ONLY) == 0) { obs_data_set_default_int(settings, KEY_KEYFRAMES_INTERVALTYPE, 0); obs_data_set_default_double(settings, KEY_KEYFRAMES_INTERVAL_SECONDS, 2.0); obs_data_set_default_int(settings, KEY_KEYFRAMES_INTERVAL_FRAMES, 300); } { // Integrated Options // FFmpeg obs_data_set_default_string(settings, KEY_FFMPEG_CUSTOMSETTINGS, ""); obs_data_set_default_int(settings, KEY_FFMPEG_COLORFORMAT, static_cast(AV_PIX_FMT_NONE)); obs_data_set_default_int(settings, KEY_FFMPEG_THREADS, 0); obs_data_set_default_int(settings, KEY_FFMPEG_GPU, -1); obs_data_set_default_int(settings, KEY_FFMPEG_STANDARDCOMPLIANCE, FF_COMPLIANCE_STRICT); } } static bool modified_keyframes(obs_properties_t* props, obs_property_t*, obs_data_t* settings) noexcept try { bool is_seconds = obs_data_get_int(settings, KEY_KEYFRAMES_INTERVALTYPE) == 0; obs_property_set_visible(obs_properties_get(props, KEY_KEYFRAMES_INTERVAL_FRAMES), !is_seconds); obs_property_set_visible(obs_properties_get(props, KEY_KEYFRAMES_INTERVAL_SECONDS), is_seconds); return true; } catch (const std::exception& ex) { LOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what()); return false; } catch (...) { LOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__); return false; } void ffmpeg_factory::get_properties(obs_properties_t* props, bool hw_encode) { if (_handler) _handler->get_properties(props, avcodec_ptr, nullptr, hw_encode); if (_handler && _handler->has_keyframe_support(this)) { // Key-Frame Options obs_properties_t* grp = props; if (!util::are_property_groups_broken()) { grp = obs_properties_create(); obs_properties_add_group(props, ST_KEYFRAMES, D_TRANSLATE(ST_KEYFRAMES), OBS_GROUP_NORMAL, grp); } { // Key-Frame Interval Type auto p = obs_properties_add_list(grp, KEY_KEYFRAMES_INTERVALTYPE, D_TRANSLATE(ST_KEYFRAMES_INTERVALTYPE), OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_INT); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_KEYFRAMES_INTERVALTYPE))); obs_property_set_modified_callback(p, modified_keyframes); obs_property_list_add_int(p, D_TRANSLATE(ST_KEYFRAMES_INTERVALTYPE_(Seconds)), 0); obs_property_list_add_int(p, D_TRANSLATE(ST_KEYFRAMES_INTERVALTYPE_(Frames)), 1); } { // Key-Frame Interval Seconds auto p = obs_properties_add_float(grp, KEY_KEYFRAMES_INTERVAL_SECONDS, D_TRANSLATE(ST_KEYFRAMES_INTERVAL), 0.00, std::numeric_limits::max(), 0.01); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_KEYFRAMES_INTERVAL))); obs_property_float_set_suffix(p, " seconds"); } { // Key-Frame Interval Frames auto p = obs_properties_add_int(grp, KEY_KEYFRAMES_INTERVAL_FRAMES, D_TRANSLATE(ST_KEYFRAMES_INTERVAL), 0, std::numeric_limits::max(), 1); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_KEYFRAMES_INTERVAL))); obs_property_int_set_suffix(p, " frames"); } } { obs_properties_t* grp = props; if (!util::are_property_groups_broken()) { auto prs = obs_properties_create(); obs_properties_add_group(props, ST_FFMPEG, D_TRANSLATE(ST_FFMPEG), OBS_GROUP_NORMAL, prs); grp = prs; } { // Custom Settings auto p = obs_properties_add_text(grp, KEY_FFMPEG_CUSTOMSETTINGS, D_TRANSLATE(ST_FFMPEG_CUSTOMSETTINGS), obs_text_type::OBS_TEXT_DEFAULT); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_FFMPEG_CUSTOMSETTINGS))); } if (_handler && _handler->is_hardware_encoder(this)) { auto p = obs_properties_add_int(grp, KEY_FFMPEG_GPU, D_TRANSLATE(ST_FFMPEG_GPU), -1, std::numeric_limits::max(), 1); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_FFMPEG_GPU))); } if (_handler && _handler->has_threading_support(this)) { auto p = obs_properties_add_int_slider(grp, KEY_FFMPEG_THREADS, D_TRANSLATE(ST_FFMPEG_THREADS), 0, static_cast(std::thread::hardware_concurrency() * 2), 1); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_FFMPEG_THREADS))); } if (_handler && _handler->has_pixel_format_support(this)) { auto p = obs_properties_add_list(grp, KEY_FFMPEG_COLORFORMAT, D_TRANSLATE(ST_FFMPEG_COLORFORMAT), OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_INT); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_FFMPEG_COLORFORMAT))); obs_property_list_add_int(p, D_TRANSLATE(S_STATE_AUTOMATIC), static_cast(AV_PIX_FMT_NONE)); for (auto ptr = avcodec_ptr->pix_fmts; *ptr != AV_PIX_FMT_NONE; ptr++) { obs_property_list_add_int(p, ::ffmpeg::tools::get_pixel_format_name(*ptr), static_cast(*ptr)); } } { auto p = obs_properties_add_list(grp, KEY_FFMPEG_STANDARDCOMPLIANCE, D_TRANSLATE(ST_FFMPEG_STANDARDCOMPLIANCE), OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_INT); obs_property_set_long_description(p, D_TRANSLATE(D_DESC(ST_FFMPEG_STANDARDCOMPLIANCE))); obs_property_list_add_int(p, D_TRANSLATE(ST_FFMPEG_STANDARDCOMPLIANCE ".VeryStrict"), FF_COMPLIANCE_VERY_STRICT); obs_property_list_add_int(p, D_TRANSLATE(ST_FFMPEG_STANDARDCOMPLIANCE ".Strict"), FF_COMPLIANCE_STRICT); obs_property_list_add_int(p, D_TRANSLATE(ST_FFMPEG_STANDARDCOMPLIANCE ".Normal"), FF_COMPLIANCE_NORMAL); obs_property_list_add_int(p, D_TRANSLATE(ST_FFMPEG_STANDARDCOMPLIANCE ".Unofficial"), FF_COMPLIANCE_UNOFFICIAL); obs_property_list_add_int(p, D_TRANSLATE(ST_FFMPEG_STANDARDCOMPLIANCE ".Experimental"), FF_COMPLIANCE_EXPERIMENTAL); } }; } const AVCodec* ffmpeg_factory::get_avcodec() { return avcodec_ptr; } const ffmpeg_info& ffmpeg_factory::get_info() { return info; } const ffmpeg_info& ffmpeg_factory::get_fallback() { return info_fallback; } void ffmpeg_instance::initialize_sw(obs_data_t* settings) { if (_codec->type == AVMEDIA_TYPE_VIDEO) { // Initialize Video Encoding auto voi = video_output_get_info(obs_encoder_video(_self)); // Find a suitable Pixel Format. AVPixelFormat _pixfmt_source = ::ffmpeg::tools::obs_videoformat_to_avpixelformat(voi->format); AVPixelFormat _pixfmt_target = static_cast(obs_data_get_int(settings, KEY_FFMPEG_COLORFORMAT)); if (_pixfmt_target == AV_PIX_FMT_NONE) { // Find the best conversion format. _pixfmt_target = ::ffmpeg::tools::get_least_lossy_format(_codec->pix_fmts, _pixfmt_source); if (_handler) // Allow Handler to override the automatic color format for sanity reasons. _handler->override_colorformat(_pixfmt_target, settings, _codec, _context); } else { // Use user override, guaranteed to be supported. bool is_format_supported = false; for (auto ptr = _codec->pix_fmts; *ptr != AV_PIX_FMT_NONE; ptr++) { if (*ptr == _pixfmt_target) { is_format_supported = true; } } if (!is_format_supported) { std::stringstream sstr; sstr << "Color Format '" << ::ffmpeg::tools::get_pixel_format_name(_pixfmt_target) << "' is not supported by the encoder."; throw std::runtime_error(sstr.str().c_str()); } } _context->width = static_cast(obs_encoder_get_width(_self)); _context->height = static_cast(obs_encoder_get_height(_self)); ::ffmpeg::tools::setup_obs_color(voi->colorspace, voi->range, _context); _context->pix_fmt = _pixfmt_target; _context->field_order = AV_FIELD_PROGRESSIVE; _context->ticks_per_frame = 1; _context->sample_aspect_ratio.num = _context->sample_aspect_ratio.den = 1; _context->framerate.num = _context->time_base.den = static_cast(voi->fps_num); _context->framerate.den = _context->time_base.num = static_cast(voi->fps_den); _swscale.set_source_size(static_cast(_context->width), static_cast(_context->height)); _swscale.set_source_color(_context->color_range == AVCOL_RANGE_JPEG, _context->colorspace); _swscale.set_source_format(_pixfmt_source); _swscale.set_target_size(static_cast(_context->width), static_cast(_context->height)); _swscale.set_target_color(_context->color_range == AVCOL_RANGE_JPEG, _context->colorspace); _swscale.set_target_format(_pixfmt_target); // Create Scaler if (!_swscale.initialize(SWS_POINT)) { std::stringstream sstr; sstr << "Initializing scaler failed for conversion from '" << ::ffmpeg::tools::get_pixel_format_name(_swscale.get_source_format()) << "' to '" << ::ffmpeg::tools::get_pixel_format_name(_swscale.get_target_format()) << "' with color space '" << ::ffmpeg::tools::get_color_space_name(_swscale.get_source_colorspace()) << "' and " << (_swscale.is_source_full_range() ? "full" : "partial") << " range."; throw std::runtime_error(sstr.str()); } } } void ffmpeg_instance::initialize_hw(obs_data_t*) { // Initialize Video Encoding auto voi = video_output_get_info(obs_encoder_video(_self)); _context->width = static_cast(voi->width); _context->height = static_cast(voi->height); _context->field_order = AV_FIELD_PROGRESSIVE; _context->ticks_per_frame = 1; _context->sample_aspect_ratio.num = _context->sample_aspect_ratio.den = 1; _context->framerate.num = _context->time_base.den = static_cast(voi->fps_num); _context->framerate.den = _context->time_base.num = static_cast(voi->fps_den); ::ffmpeg::tools::setup_obs_color(voi->colorspace, voi->range, _context); _context->sw_pix_fmt = ::ffmpeg::tools::obs_videoformat_to_avpixelformat(voi->format); #ifdef WIN32 _context->pix_fmt = AV_PIX_FMT_D3D11; #endif _context->hw_device_ctx = _hwinst->create_device_context(); _context->hw_frames_ctx = av_hwframe_ctx_alloc(_context->hw_device_ctx); if (!_context->hw_frames_ctx) throw std::runtime_error("Allocating hardware context failed, chosen pixel format is likely not supported."); AVHWFramesContext* ctx = reinterpret_cast(_context->hw_frames_ctx->data); ctx->width = _context->width; ctx->height = _context->height; ctx->format = _context->pix_fmt; ctx->sw_format = _context->sw_pix_fmt; if (av_hwframe_ctx_init(_context->hw_frames_ctx) < 0) throw std::runtime_error("Initializing hardware context failed, chosen pixel format is likely not supported."); } void ffmpeg_instance::push_free_frame(std::shared_ptr frame) { auto now = std::chrono::high_resolution_clock::now(); if (_free_frames.size() > 0) { if ((now - _free_frames_last_used) < std::chrono::seconds(1)) { _free_frames.push(frame); } } else { _free_frames.push(frame); _free_frames_last_used = std::chrono::high_resolution_clock::now(); } } std::shared_ptr ffmpeg_instance::pop_free_frame() { std::shared_ptr frame; if (_free_frames.size() > 0) { // Re-use existing frames first. frame = _free_frames.top(); _free_frames.pop(); } else { if (_hwinst) { frame = _hwinst->allocate_frame(_context->hw_frames_ctx); } else { frame = std::shared_ptr(av_frame_alloc(), [](AVFrame* frame) { av_frame_unref(frame); av_frame_free(&frame); }); frame->width = _context->width; frame->height = _context->height; frame->format = _context->pix_fmt; int res = av_frame_get_buffer(frame.get(), 32); if (res < 0) { throw std::runtime_error(::ffmpeg::tools::get_error_description(res)); } } } return frame; } void ffmpeg_instance::push_used_frame(std::shared_ptr frame) { _used_frames.push(frame); } std::shared_ptr ffmpeg_instance::pop_used_frame() { auto frame = _used_frames.front(); _used_frames.pop(); return frame; } ffmpeg_instance::ffmpeg_instance(obs_data_t* settings, obs_encoder_t* encoder, bool is_texture_encode) : _self(encoder), _factory(reinterpret_cast(obs_encoder_get_type_data(_self))), _codec(_factory->get_avcodec()), _context(nullptr), _lag_in_frames(0), _count_send_frames(0), _have_first_frame(false) { // Find a handler _handler = ffmpeg_manager::get()->get_handler(_codec->name); // Initialize GPU Stuff if (is_texture_encode) { // Abort if user specified manual override. if ((static_cast(obs_data_get_int(settings, KEY_FFMPEG_COLORFORMAT)) != AV_PIX_FMT_NONE) || (obs_data_get_int(settings, KEY_FFMPEG_GPU) != -1) || (obs_encoder_scaling_enabled(_self))) { throw std::runtime_error("Unable to create accelerated encoder due to user settings."); } #ifdef WIN32 auto gctx = gs::context(); if (gs_get_device_type() == GS_DEVICE_DIRECT3D_11) { _hwapi = std::make_shared<::ffmpeg::hwapi::d3d11>(); } #endif if (!_hwapi) { throw std::runtime_error("Failed to create acceleration device."); } _hwinst = _hwapi->create_from_obs(); } // Initialize context. _context = avcodec_alloc_context3(_codec); if (!_context) { LOG_ERROR("Failed to create context for encoder '%s'.", _codec->name); throw std::runtime_error("Failed to create encoder context."); } // Create 8MB of precached Packet data for use later on. av_init_packet(&_current_packet); av_new_packet(&_current_packet, 8 * 1024 * 1024); // 8 MB precached Packet size. if (is_texture_encode) { initialize_hw(settings); } else { initialize_sw(settings); } // Update settings update(settings); // Initialize Encoder auto gctx = gs::context(); int res = avcodec_open2(_context, _codec, NULL); if (res < 0) { std::stringstream sstr; sstr << "Initializing encoder '" << _codec->name << "' failed with error: " << ::ffmpeg::tools::get_error_description(res) << " (code " << res << ")"; throw std::runtime_error(sstr.str()); } } ffmpeg_instance::~ffmpeg_instance() { auto gctx = gs::context(); if (_context) { // Flush encoders that require it. if ((_codec->capabilities & AV_CODEC_CAP_DELAY) != 0) { avcodec_send_frame(_context, nullptr); while (avcodec_receive_packet(_context, &_current_packet) >= 0) { avcodec_send_frame(_context, nullptr); std::this_thread::sleep_for(std::chrono::milliseconds(1)); } } // Close and free context. avcodec_close(_context); avcodec_free_context(&_context); } av_packet_unref(&_current_packet); _swscale.finalize(); } void ffmpeg_instance::get_properties(obs_properties_t* props, bool hw_encode) { if (_handler) _handler->get_properties(props, _codec, _context, hw_encode); obs_property_set_enabled(obs_properties_get(props, KEY_KEYFRAMES_INTERVALTYPE), false); obs_property_set_enabled(obs_properties_get(props, KEY_KEYFRAMES_INTERVAL_SECONDS), false); obs_property_set_enabled(obs_properties_get(props, KEY_KEYFRAMES_INTERVAL_FRAMES), false); obs_property_set_enabled(obs_properties_get(props, KEY_FFMPEG_COLORFORMAT), false); obs_property_set_enabled(obs_properties_get(props, KEY_FFMPEG_THREADS), false); obs_property_set_enabled(obs_properties_get(props, KEY_FFMPEG_STANDARDCOMPLIANCE), false); obs_property_set_enabled(obs_properties_get(props, KEY_FFMPEG_GPU), false); } bool ffmpeg_instance::update(obs_data_t* settings) { // FFmpeg Options _context->debug = 0; _context->strict_std_compliance = static_cast(obs_data_get_int(settings, KEY_FFMPEG_STANDARDCOMPLIANCE)); /// Threading if (!_hwinst) { _context->thread_type = 0; if (_codec->capabilities & AV_CODEC_CAP_FRAME_THREADS) { _context->thread_type |= FF_THREAD_FRAME; } if (_codec->capabilities & AV_CODEC_CAP_SLICE_THREADS) { _context->thread_type |= FF_THREAD_SLICE; } if (_context->thread_type != 0) { int64_t threads = obs_data_get_int(settings, ST_FFMPEG_THREADS); if (threads > 0) { _context->thread_count = static_cast(threads); } else { _context->thread_count = static_cast(std::thread::hardware_concurrency()); } } else { _context->thread_count = 1; } // Frame Delay (Lag In Frames) _context->delay = _context->thread_count; } else { _context->delay = 0; } // Apply GPU Selection if (!_hwinst && ::ffmpeg::tools::can_hardware_encode(_codec)) { av_opt_set_int(_context, "gpu", (int)obs_data_get_int(settings, KEY_FFMPEG_GPU), AV_OPT_SEARCH_CHILDREN); } // Keyframes if (_handler && _handler->has_keyframe_support(_factory)) { // Key-Frame Options obs_video_info ovi; if (!obs_get_video_info(&ovi)) { throw std::runtime_error("obs_get_video_info failed, restart OBS Studio to fix it (hopefully)."); } int64_t kf_type = obs_data_get_int(settings, KEY_KEYFRAMES_INTERVALTYPE); bool is_seconds = (kf_type == 0); if (is_seconds) { _context->gop_size = static_cast(obs_data_get_double(settings, KEY_KEYFRAMES_INTERVAL_SECONDS) * (ovi.fps_num / ovi.fps_den)); } else { _context->gop_size = static_cast(obs_data_get_int(settings, KEY_KEYFRAMES_INTERVAL_FRAMES)); } _context->keyint_min = _context->gop_size; } // Handler Options if (_handler) _handler->update(settings, _codec, _context); { // FFmpeg Custom Options const char* opts = obs_data_get_string(settings, KEY_FFMPEG_CUSTOMSETTINGS); std::size_t opts_len = strnlen(opts, 65535); parse_ffmpeg_commandline(std::string{opts, opts + opts_len}); } // Handler Overrides if (_handler) _handler->override_update(this, settings); // Handler Logging if (_handler) { LOG_INFO("[%s] Initializing...", _codec->name); LOG_INFO("[%s] FFmpeg:", _codec->name); LOG_INFO("[%s] Custom Settings: %s", _codec->name, obs_data_get_string(settings, KEY_FFMPEG_CUSTOMSETTINGS)); LOG_INFO("[%s] Standard Compliance: %s", _codec->name, ::ffmpeg::tools::get_std_compliance_name(_context->strict_std_compliance)); LOG_INFO("[%s] Threading: %s (with %i threads)", _codec->name, ::ffmpeg::tools::get_thread_type_name(_context->thread_type), _context->thread_count); LOG_INFO("[%s] Video:", _codec->name); if (_hwinst) { LOG_INFO("[%s] Texture: %ldx%ld %s %s %s", _codec->name, _context->width, _context->height, ::ffmpeg::tools::get_pixel_format_name(_context->sw_pix_fmt), ::ffmpeg::tools::get_color_space_name(_context->colorspace), av_color_range_name(_context->color_range)); } else { LOG_INFO("[%s] Input: %ldx%ld %s %s %s", _codec->name, _swscale.get_source_width(), _swscale.get_source_height(), ::ffmpeg::tools::get_pixel_format_name(_swscale.get_source_format()), ::ffmpeg::tools::get_color_space_name(_swscale.get_source_colorspace()), _swscale.is_source_full_range() ? "Full" : "Partial"); LOG_INFO("[%s] Output: %ldx%ld %s %s %s", _codec->name, _swscale.get_target_width(), _swscale.get_target_height(), ::ffmpeg::tools::get_pixel_format_name(_swscale.get_target_format()), ::ffmpeg::tools::get_color_space_name(_swscale.get_target_colorspace()), _swscale.is_target_full_range() ? "Full" : "Partial"); if (!_hwinst) LOG_INFO("[%s] On GPU Index: %lli", _codec->name, obs_data_get_int(settings, KEY_FFMPEG_GPU)); } LOG_INFO("[%s] Framerate: %ld/%ld (%f FPS)", _codec->name, _context->time_base.den, _context->time_base.num, static_cast(_context->time_base.den) / static_cast(_context->time_base.num)); LOG_INFO("[%s] Keyframes: ", _codec->name); if (_context->keyint_min != _context->gop_size) { LOG_INFO("[%s] Minimum: %i frames", _codec->name, _context->keyint_min); LOG_INFO("[%s] Maximum: %i frames", _codec->name, _context->gop_size); } else { LOG_INFO("[%s] Distance: %i frames", _codec->name, _context->gop_size); } _handler->log_options(settings, _codec, _context); } return true; } void ffmpeg_instance::get_audio_info(audio_convert_info*) {} std::size_t ffmpeg_instance::get_frame_size() { return size_t(); } bool ffmpeg_instance::audio_encode(encoder_frame*, encoder_packet*, bool*) { return false; } void ffmpeg_instance::get_video_info(video_scale_info* vsi) { vsi->width = _swscale.get_source_width(); vsi->height = _swscale.get_source_height(); vsi->format = ::ffmpeg::tools::avpixelformat_to_obs_videoformat(_swscale.get_source_format()); } bool ffmpeg_instance::get_sei_data(std::uint8_t** data, size_t* size) { if (_sei_data.size() == 0) return false; *data = _sei_data.data(); *size = _sei_data.size(); return true; } bool ffmpeg_instance::get_extra_data(std::uint8_t** data, size_t* size) { if (_extra_data.size() == 0) return false; *data = _extra_data.data(); *size = _extra_data.size(); return true; } static inline void copy_data(encoder_frame* frame, AVFrame* vframe) { int h_chroma_shift, v_chroma_shift; av_pix_fmt_get_chroma_sub_sample(static_cast(vframe->format), &h_chroma_shift, &v_chroma_shift); for (std::size_t idx = 0; idx < MAX_AV_PLANES; idx++) { if (!frame->data[idx] || !vframe->data[idx]) continue; std::size_t plane_height = static_cast(vframe->height) >> (idx ? v_chroma_shift : 0); if (static_cast(vframe->linesize[idx]) == frame->linesize[idx]) { std::memcpy(vframe->data[idx], frame->data[idx], frame->linesize[idx] * plane_height); } else { std::size_t ls_in = static_cast(frame->linesize[idx]); std::size_t ls_out = static_cast(vframe->linesize[idx]); std::size_t bytes = ls_in < ls_out ? ls_in : ls_out; std::uint8_t* to = vframe->data[idx]; std::uint8_t* from = frame->data[idx]; for (std::size_t y = 0; y < plane_height; y++) { std::memcpy(to, from, bytes); to += ls_out; from += ls_in; } } } } bool ffmpeg_instance::video_encode(encoder_frame* frame, encoder_packet* packet, bool* received_packet) { std::shared_ptr vframe = pop_free_frame(); // Retrieve an empty frame. // Convert frame. { vframe->height = _context->height; vframe->format = _context->pix_fmt; vframe->color_range = _context->color_range; vframe->colorspace = _context->colorspace; vframe->color_primaries = _context->color_primaries; vframe->color_trc = _context->color_trc; vframe->pts = frame->pts; if ((_swscale.is_source_full_range() == _swscale.is_target_full_range()) && (_swscale.get_source_colorspace() == _swscale.get_target_colorspace()) && (_swscale.get_source_format() == _swscale.get_target_format())) { copy_data(frame, vframe.get()); } else { int res = _swscale.convert(reinterpret_cast(frame->data), reinterpret_cast(frame->linesize), 0, _context->height, vframe->data, vframe->linesize); if (res <= 0) { LOG_ERROR("Failed to convert frame: %s (%ld).", ::ffmpeg::tools::get_error_description(res), res); return false; } } } if (!encode_avframe(vframe, packet, received_packet)) return false; return true; } bool ffmpeg_instance::video_encode_texture(std::uint32_t handle, std::int64_t pts, std::uint64_t lock_key, std::uint64_t* next_lock_key, encoder_packet* packet, bool* received_packet) { if (handle == GS_INVALID_HANDLE) { LOG_ERROR("Received invalid handle."); *next_lock_key = lock_key; return false; } std::shared_ptr vframe = pop_free_frame(); _hwinst->copy_from_obs(_context->hw_frames_ctx, handle, lock_key, next_lock_key, vframe); vframe->color_range = _context->color_range; vframe->colorspace = _context->colorspace; vframe->color_primaries = _context->color_primaries; vframe->color_trc = _context->color_trc; vframe->pts = pts; if (!encode_avframe(vframe, packet, received_packet)) return false; *next_lock_key = lock_key; return true; } int ffmpeg_instance::receive_packet(bool* received_packet, struct encoder_packet* packet) { int res = 0; av_packet_unref(&_current_packet); { auto gctx = gs::context(); res = avcodec_receive_packet(_context, &_current_packet); } if (res != 0) { return res; } if (!_have_first_frame) { if (_codec->id == AV_CODEC_ID_H264) { std::uint8_t* tmp_packet; std::uint8_t* tmp_header; std::uint8_t* tmp_sei; std::size_t sz_packet, sz_header, sz_sei; obs_extract_avc_headers(_current_packet.data, static_cast(_current_packet.size), &tmp_packet, &sz_packet, &tmp_header, &sz_header, &tmp_sei, &sz_sei); if (sz_header) { _extra_data.resize(sz_header); std::memcpy(_extra_data.data(), tmp_header, sz_header); } if (sz_sei) { _sei_data.resize(sz_sei); std::memcpy(_sei_data.data(), tmp_sei, sz_sei); } // Not required, we only need the Extra Data and SEI Data anyway. //std::memcpy(_current_packet.data, tmp_packet, sz_packet); //_current_packet.size = static_cast(sz_packet); bfree(tmp_packet); bfree(tmp_header); bfree(tmp_sei); } else if (_codec->id == AV_CODEC_ID_HEVC) { hevc::extract_header_sei(_current_packet.data, static_cast(_current_packet.size), _extra_data, _sei_data); } else if (_context->extradata != nullptr) { _extra_data.resize(static_cast(_context->extradata_size)); std::memcpy(_extra_data.data(), _context->extradata, static_cast(_context->extradata_size)); } _have_first_frame = true; } // Allow Handler Post-Processing if (_handler) _handler->process_avpacket(_current_packet, _codec, _context); packet->type = OBS_ENCODER_VIDEO; packet->pts = _current_packet.pts; packet->dts = _current_packet.dts; packet->data = _current_packet.data; packet->size = static_cast(_current_packet.size); packet->keyframe = !!(_current_packet.flags & AV_PKT_FLAG_KEY); packet->drop_priority = packet->keyframe ? 0 : 1; *received_packet = true; push_free_frame(pop_used_frame()); return res; } int ffmpeg_instance::send_frame(std::shared_ptr const frame) { int res = 0; { auto gctx = gs::context(); res = avcodec_send_frame(_context, frame.get()); } if (res == 0) { push_used_frame(frame); } return res; } bool ffmpeg_instance::encode_avframe(std::shared_ptr frame, encoder_packet* packet, bool* received_packet) { bool sent_frame = false; bool recv_packet = false; bool should_lag = (_count_send_frames >= _lag_in_frames); auto loop_begin = std::chrono::high_resolution_clock::now(); auto loop_end = loop_begin + std::chrono::milliseconds(50); while ((!sent_frame || (should_lag && !recv_packet)) && !(std::chrono::high_resolution_clock::now() > loop_end)) { bool eagain_is_stupid = false; if (!sent_frame) { int res = send_frame(frame); switch (res) { case 0: sent_frame = true; frame = nullptr; break; case AVERROR(EAGAIN): // This means we should call receive_packet again, but what do we do with that data? // Why can't we queue on both? Do I really have to implement threading for this stuff? if (*received_packet == true) { LOG_WARNING("Skipped frame due to EAGAIN when a packet was already returned."); sent_frame = true; } eagain_is_stupid = true; break; case AVERROR(EOF): LOG_ERROR("Skipped frame due to end of stream."); sent_frame = true; break; default: LOG_ERROR("Failed to encode frame: %s (%ld).", ::ffmpeg::tools::get_error_description(res), res); return false; } } if (!recv_packet) { int res = receive_packet(received_packet, packet); switch (res) { case 0: recv_packet = true; break; case AVERROR(EOF): LOG_ERROR("Received end of file."); recv_packet = true; break; case AVERROR(EAGAIN): if (sent_frame) { recv_packet = true; } if (eagain_is_stupid) { LOG_ERROR("Both send and recieve returned EAGAIN, encoder is broken."); return false; } break; default: LOG_ERROR("Failed to receive packet: %s (%ld).", ::ffmpeg::tools::get_error_description(res), res); return false; } } if (!sent_frame || !recv_packet) { std::this_thread::sleep_for(std::chrono::milliseconds(1)); } } if (!sent_frame) push_free_frame(frame); return true; } bool ffmpeg_instance::is_hardware_encode() { return _hwinst != nullptr; } const AVCodec* ffmpeg_instance::get_avcodec() { return _codec; } const AVCodecContext* ffmpeg_instance::get_avcodeccontext() { return _context; } void ffmpeg_instance::parse_ffmpeg_commandline(std::string text) { // Steps to properly parse a command line: // 1. Split by space and package by quotes. // 2. Parse each resulting option individually. // First, we split by space and of course respect quotes while doing so. // That means that "-foo= bar" is stored as std::string("-foo= bar"), // and things like -foo="bar" is stored as std::string("-foo=\"bar\""). // However "-foo"=bar" -foo2=bar" is stored as std::string("-foo=bar -foo2=bar") // because the quote was not escaped. std::list opts; std::stringstream opt_stream{std::ios_base::in | std::ios_base::out | std::ios_base::binary}; std::stack quote_stack; for (std::size_t p = 0; p <= text.size(); p++) { char here = p < text.size() ? text.at(p) : 0; if (here == '\\') { std::size_t p2 = p + 1; if (p2 < text.size()) { char here2 = text.at(p2); if (isdigit(here2)) { // Octal // Not supported yet. p++; } else if (here2 == 'x') { // Hexadecimal // Not supported yet. p += 3; } else if (here2 == 'u') { // 4 or 8 wide Unicode. // Not supported yet. } else if (here2 == 'a') { opt_stream << '\a'; p++; } else if (here2 == 'b') { opt_stream << '\b'; p++; } else if (here2 == 'f') { opt_stream << '\f'; p++; } else if (here2 == 'n') { opt_stream << '\n'; p++; } else if (here2 == 'r') { opt_stream << '\r'; p++; } else if (here2 == 't') { opt_stream << '\t'; p++; } else if (here2 == 'v') { opt_stream << '\v'; p++; } else if (here2 == '\\') { opt_stream << '\\'; p++; } else if (here2 == '\'') { opt_stream << '\''; p++; } else if (here2 == '"') { opt_stream << '"'; p++; } else if (here2 == '?') { opt_stream << '\?'; p++; } } } else if ((here == '\'') || (here == '"')) { if (quote_stack.size() > 1) { opt_stream << here; } if (quote_stack.size() == 0) { quote_stack.push(here); } else if (quote_stack.top() == here) { quote_stack.pop(); } else { quote_stack.push(here); } } else if ((here == 0) || ((here == ' ') && (quote_stack.size() == 0))) { std::string ropt = opt_stream.str(); if (ropt.size() > 0) { opts.push_back(ropt); opt_stream.str(std::string()); opt_stream.clear(); } } else { opt_stream << here; } } // Now that we have a list of parameters as neatly grouped strings, and // have also dealt with escaping for the most part. We want to parse // an FFmpeg commandline option set here, so the first character in // the string must be a '-'. for (std::string& opt : opts) { // Skip empty options. if (opt.size() == 0) continue; // Skip options that don't start with a '-'. if (opt.at(0) != '-') { LOG_WARNING("Option '%s' is malformed, must start with a '-'.", opt.c_str()); continue; } // Skip options that don't contain a '='. const char* cstr = opt.c_str(); const char* eq_at = strchr(cstr, '='); if (eq_at == nullptr) { LOG_WARNING("Option '%s' is malformed, must contain a '='.", opt.c_str()); continue; } try { std::string key = opt.substr(1, static_cast((eq_at - cstr) - 1)); std::string value = opt.substr(static_cast((eq_at - cstr) + 1)); int res = av_opt_set(_context, key.c_str(), value.c_str(), AV_OPT_SEARCH_CHILDREN); if (res < 0) { LOG_WARNING("Option '%s' (key: '%s', value: '%s') encountered error: %s", opt.c_str(), key.c_str(), value.c_str(), ::ffmpeg::tools::get_error_description(res)); } } catch (const std::exception& ex) { LOG_ERROR("Option '%s' encountered exception: %s", opt.c_str(), ex.what()); } } } ffmpeg_manager::ffmpeg_manager() : _factories(), _handlers(), _debug_handler() { // Handlers _debug_handler = ::std::make_shared(); register_handler("prores_aw", ::std::make_shared()); register_handler("h264_nvenc", ::std::make_shared()); register_handler("hevc_nvenc", ::std::make_shared()); } ffmpeg_manager::~ffmpeg_manager() { _factories.clear(); } void ffmpeg_manager::register_handler(std::string codec, std::shared_ptr handler) { _handlers.emplace(codec, handler); } std::shared_ptr ffmpeg_manager::get_handler(std::string codec) { auto fnd = _handlers.find(codec); if (fnd != _handlers.end()) return fnd->second; #ifdef _DEBUG return _debug_handler; #else return nullptr; #endif } bool ffmpeg_manager::has_handler(std::string codec) { return (_handlers.find(codec) != _handlers.end()); } void ffmpeg_manager::register_encoders() { // Encoders void* iterator = nullptr; const AVCodec* codec = nullptr; for (codec = av_codec_iterate(&iterator); codec != nullptr; codec = av_codec_iterate(&iterator)) { if (!av_codec_is_encoder(codec)) continue; if ((codec->type == AVMediaType::AVMEDIA_TYPE_AUDIO) || (codec->type == AVMediaType::AVMEDIA_TYPE_VIDEO)) { try { auto factory = std::make_shared(codec); factory->register_encoder(); _factories.emplace(codec, factory); } catch (...) { } } } } std::shared_ptr _ffmepg_encoder_factory_instance = nullptr; void streamfx::encoder::ffmpeg::ffmpeg_manager::initialize() { if (!_ffmepg_encoder_factory_instance) { _ffmepg_encoder_factory_instance = std::make_shared(); _ffmepg_encoder_factory_instance->register_encoders(); } } void streamfx::encoder::ffmpeg::ffmpeg_manager::finalize() { _ffmepg_encoder_factory_instance.reset(); } std::shared_ptr streamfx::encoder::ffmpeg::ffmpeg_manager::get() { return _ffmepg_encoder_factory_instance; }