obs-StreamFX/source/filters/filter-autoframing.cpp

1281 lines
42 KiB
C++
Raw Normal View History

// AUTOGENERATED COPYRIGHT HEADER START
// Copyright (C) 2021-2023 Michael Fabian 'Xaymar' Dirks <info@xaymar.com>
// Copyright (C) 2022 lainon <GermanAizek@yandex.ru>
// AUTOGENERATED COPYRIGHT HEADER END
#include "filter-autoframing.hpp"
#include "obs/gs/gs-helper.hpp"
#include "util/util-logging.hpp"
#ifdef _DEBUG
#define ST_PREFIX "<%s> "
#define D_LOG_ERROR(x, ...) P_LOG_ERROR(ST_PREFIX##x, __FUNCTION_SIG__, __VA_ARGS__)
#define D_LOG_WARNING(x, ...) P_LOG_WARN(ST_PREFIX##x, __FUNCTION_SIG__, __VA_ARGS__)
#define D_LOG_INFO(x, ...) P_LOG_INFO(ST_PREFIX##x, __FUNCTION_SIG__, __VA_ARGS__)
#define D_LOG_DEBUG(x, ...) P_LOG_DEBUG(ST_PREFIX##x, __FUNCTION_SIG__, __VA_ARGS__)
#else
#define ST_PREFIX "<filter::autoframing> "
#define D_LOG_ERROR(...) P_LOG_ERROR(ST_PREFIX __VA_ARGS__)
#define D_LOG_WARNING(...) P_LOG_WARN(ST_PREFIX __VA_ARGS__)
#define D_LOG_INFO(...) P_LOG_INFO(ST_PREFIX __VA_ARGS__)
#define D_LOG_DEBUG(...) P_LOG_DEBUG(ST_PREFIX __VA_ARGS__)
#endif
// Auto-Framing is the process of tracking important information inside of a group of video or
// audio samples, and then automatically cutting away all the unnecessary parts. In our case, we
// will focus on video only as the audio field is already covered by other solutions, like Noise
// Gate, Denoising, etc. The implementation will rely on the Provider system, so varying
// functionality should be expected from all providers. Some providers may only offer a way to
// track a single face, others will allow groups, yet others will allow even non-humans to be
// tracked.
//
// The goal is to provide Auto-Framing for single person streams ('Solo') as well as group streams
// ('Group'), though the latter will only be available if the provider supports it. In 'Solo' mode
// the filter will perfectly frame a single person, and no more than that. In 'Group' mode, it will
// combine all important elements into a single frame, and track that instead. In the future, we
// might want to offer a third mode to give each tracked face a separate frame however this may
// exceed the intended complexity of this feature entirely.
/** Settings
* Framing
* Mode: How should things be tracked?
* Solo: Frame only a single face.
* Group: Frame many faces, group all into single frame.
* Padding: How many pixels/much % of tracked are should be kept
* Aspect Ratio: What Aspect Ratio should the framed output have?
* Stability: How stable is the framing against changes of tracked elements?
*
* Motion
* Motion Prediction: How much should we attempt to predict where tracked elements move?
* Smoothing: How much should the position between tracking attempts
*
* Advanced
* Provider: What provider should be used?
* Frequency: How often should we track? Every frame, every 2nd frame, etc.
*/
#define ST_I18N "Filter.AutoFraming"
#define ST_I18N_TRACKING ST_I18N ".Tracking"
#define ST_KEY_TRACKING_MODE "Tracking.Mode"
#define ST_I18N_TRACKING_MODE ST_I18N_TRACKING ".Mode"
#define ST_I18N_FRAMING_MODE_SOLO ST_I18N_TRACKING_MODE ".Solo"
#define ST_I18N_FRAMING_MODE_GROUP ST_I18N_TRACKING_MODE ".Group"
#define ST_KEY_TRACKING_FREQUENCY "Tracking.Frequency"
#define ST_I18N_TRACKING_FREQUENCY ST_I18N_TRACKING ".Frequency"
#define ST_I18N_MOTION ST_I18N ".Motion"
#define ST_KEY_MOTION_PREDICTION "Motion.Prediction"
#define ST_I18N_MOTION_PREDICTION ST_I18N_MOTION ".Prediction"
#define ST_KEY_MOTION_SMOOTHING "Motion.Smoothing"
#define ST_I18N_MOTION_SMOOTHING ST_I18N_MOTION ".Smoothing"
#define ST_I18N_FRAMING ST_I18N ".Framing"
#define ST_KEY_FRAMING_STABILITY "Framing.Stability"
#define ST_I18N_FRAMING_STABILITY ST_I18N_FRAMING ".Stability"
#define ST_KEY_FRAMING_PADDING "Framing.Padding"
#define ST_I18N_FRAMING_PADDING ST_I18N_FRAMING ".Padding"
#define ST_KEY_FRAMING_OFFSET "Framing.Offset"
#define ST_I18N_FRAMING_OFFSET ST_I18N_FRAMING ".Offset"
#define ST_KEY_FRAMING_ASPECTRATIO "Framing.AspectRatio"
#define ST_I18N_FRAMING_ASPECTRATIO ST_I18N_FRAMING ".AspectRatio"
#define ST_KEY_ADVANCED_PROVIDER "Provider"
#define ST_I18N_ADVANCED_PROVIDER ST_I18N ".Provider"
#define ST_I18N_ADVANCED_PROVIDER_NVIDIA_FACEDETECTION ST_I18N_ADVANCED_PROVIDER ".NVIDIA.FaceDetection"
#define ST_KALMAN_EEC 1.0f
using streamfx::filter::autoframing::autoframing_factory;
using streamfx::filter::autoframing::autoframing_instance;
using streamfx::filter::autoframing::tracking_provider;
static constexpr std::string_view HELP_URL = "https://github.com/Xaymar/obs-StreamFX/wiki/Filter-Auto-Framing";
static tracking_provider provider_priority[] = {
tracking_provider::NVIDIA_FACEDETECTION,
};
inline std::pair<bool, double_t> parse_text_as_size(const char* text)
{
double_t v = 0;
if (sscanf(text, "%lf", &v) == 1) {
const char* prc_chr = strrchr(text, '%');
if (prc_chr && (*prc_chr == '%')) {
return {true, v / 100.0};
} else {
return {false, v};
}
} else {
return {true, 1.0};
}
}
const char* streamfx::filter::autoframing::cstring(tracking_provider provider)
{
switch (provider) {
case tracking_provider::INVALID:
return "N/A";
case tracking_provider::AUTOMATIC:
return D_TRANSLATE(S_STATE_AUTOMATIC);
case tracking_provider::NVIDIA_FACEDETECTION:
return D_TRANSLATE(ST_I18N_ADVANCED_PROVIDER_NVIDIA_FACEDETECTION);
default:
throw std::runtime_error("Missing Conversion Entry");
}
}
std::string streamfx::filter::autoframing::string(tracking_provider provider)
{
return cstring(provider);
}
autoframing_instance::~autoframing_instance()
{
D_LOG_DEBUG("Finalizing... (Addr: 0x%" PRIuPTR ")", this);
{ // Unload the underlying effect ASAP.
std::unique_lock<std::mutex> ul(_provider_lock);
// De-queue the underlying task.
if (_provider_task) {
streamfx::threadpool()->pop(_provider_task);
_provider_task->await_completion();
_provider_task.reset();
}
// TODO: Make this asynchronous.
switch (_provider) {
#ifdef ENABLE_FILTER_DENOISING_NVIDIA
case tracking_provider::NVIDIA_FACEDETECTION:
nvar_facedetection_unload();
break;
#endif
default:
break;
}
}
}
autoframing_instance::autoframing_instance(obs_data_t* data, obs_source_t* self)
: source_instance(data, self),
_dirty(true), _size(1, 1), _out_size(1, 1),
_gfx_debug(), _standard_effect(), _input(), _vb(),
_provider(tracking_provider::INVALID), _provider_ui(tracking_provider::INVALID), _provider_ready(false), _provider_lock(), _provider_task(),
_track_mode(tracking_mode::SOLO), _track_frequency(1),
_motion_smoothing(0.0), _motion_smoothing_kalman_pnc(1.), _motion_smoothing_kalman_mnc(1.), _motion_prediction(0.0),
_frame_stability(0.), _frame_stability_kalman(1.), _frame_padding_prc(), _frame_padding(), _frame_offset_prc(), _frame_offset(), _frame_aspect_ratio(0.0),
_track_frequency_counter(0), _tracked_elements(), _predicted_elements(),
_frame_pos_x({1., 1., 1., 1.}), _frame_pos_y({1., 1., 1., 1.}), _frame_pos({0, 0}), _frame_size({1, 1}),
_debug(false)
{
D_LOG_DEBUG("Initializating... (Addr: 0x%" PRIuPTR ")", this);
{
::streamfx::obs::gs::context gctx;
// Get debug renderer.
_gfx_debug = ::streamfx::gfx::util::get();
// Create the render target for the input buffering.
_input = std::make_shared<::streamfx::obs::gs::rendertarget>(GS_RGBA_UNORM, GS_ZS_NONE);
_input->render(1, 1); // Preallocate the RT on the driver and GPU.
// Load the required effect.
_standard_effect = std::make_shared<::streamfx::obs::gs::effect>(::streamfx::data_file_path("effects/standard.effect"));
// Create the Vertex Buffer for rendering.
_vb = std::make_shared<::streamfx::obs::gs::vertex_buffer>(uint32_t{4}, uint8_t{1});
vec3_set(_vb->at(0).position, 0, 0, 0);
vec3_set(_vb->at(1).position, 1, 0, 0);
vec3_set(_vb->at(2).position, 0, 1, 0);
vec3_set(_vb->at(3).position, 1, 1, 0);
_vb->update(true);
}
if (data) {
load(data);
}
}
void autoframing_instance::load(obs_data_t* data)
{
// Update from passed data.
update(data);
}
void autoframing_instance::migrate(obs_data_t* data, uint64_t version)
{
if (version < STREAMFX_MAKE_VERSION(0, 11, 0, 0)) {
obs_data_unset_user_value(data, "ROI.Zoom");
obs_data_unset_user_value(data, "ROI.Offset.X");
obs_data_unset_user_value(data, "ROI.Offset.Y");
obs_data_unset_user_value(data, "ROI.Stability");
}
}
void autoframing_instance::update(obs_data_t* data)
{
// Tracking
_track_mode = static_cast<tracking_mode>(obs_data_get_int(data, ST_KEY_TRACKING_MODE));
{
if (const char* text = obs_data_get_string(data, ST_KEY_TRACKING_FREQUENCY); text != nullptr) {
float value = 0.;
if (sscanf(text, "%f", &value) == 1) {
if (const char* seconds = strchr(text, 's'); seconds == nullptr) {
value = 1.f / value; // Hz -> seconds
} else {
// No-op
}
}
_track_frequency = value;
}
}
_track_frequency_counter = 0;
// Motion
_motion_prediction = static_cast<float>(obs_data_get_double(data, ST_KEY_MOTION_PREDICTION)) / 100.f;
_motion_smoothing = static_cast<float>(obs_data_get_double(data, ST_KEY_MOTION_SMOOTHING)) / 100.f;
_motion_smoothing_kalman_pnc = streamfx::util::math::lerp<float>(1.0f, 0.00001f, _motion_smoothing);
_motion_smoothing_kalman_mnc = streamfx::util::math::lerp<float>(0.001f, 1000.0f, _motion_smoothing);
for (auto kv : _predicted_elements) {
// Regenerate filters.
kv.second->filter_pos_x = {_frame_stability_kalman, _motion_smoothing_kalman_mnc, ST_KALMAN_EEC, kv.second->filter_pos_x.get()};
kv.second->filter_pos_y = {_frame_stability_kalman, _motion_smoothing_kalman_mnc, ST_KALMAN_EEC, kv.second->filter_pos_y.get()};
}
// Framing
{ // Smoothing
_frame_stability = static_cast<float>(obs_data_get_double(data, ST_KEY_FRAMING_STABILITY)) / 100.f;
_frame_stability_kalman = streamfx::util::math::lerp<float>(1.0f, 0.00001f, _frame_stability);
_frame_pos_x = {_frame_stability_kalman, 1.0f, ST_KALMAN_EEC, _frame_pos_x.get()};
_frame_pos_y = {_frame_stability_kalman, 1.0f, ST_KALMAN_EEC, _frame_pos_y.get()};
_frame_size_x = {_frame_stability_kalman, 1.0f, ST_KALMAN_EEC, _frame_size_x.get()};
_frame_size_y = {_frame_stability_kalman, 1.0f, ST_KALMAN_EEC, _frame_size_y.get()};
}
{ // Padding
if (const char* text = obs_data_get_string(data, ST_KEY_FRAMING_PADDING ".X"); text != nullptr) {
float value = 0.;
if (sscanf(text, "%f", &value) == 1) {
if (const char* percent = strchr(text, '%'); percent != nullptr) {
// Flip sign, percent is negative.
value = -(value / 100.f);
_frame_padding_prc[0] = true;
} else {
_frame_padding_prc[0] = false;
}
}
_frame_padding.x = value;
}
if (const char* text = obs_data_get_string(data, ST_KEY_FRAMING_PADDING ".Y"); text != nullptr) {
float value = 0.;
if (sscanf(text, "%f", &value) == 1) {
if (const char* percent = strchr(text, '%'); percent != nullptr) {
// Flip sign, percent is negative.
value = -(value / 100.f);
_frame_padding_prc[1] = true;
} else {
_frame_padding_prc[1] = false;
}
}
_frame_padding.y = value;
}
}
{ // Offset
if (const char* text = obs_data_get_string(data, ST_KEY_FRAMING_OFFSET ".X"); text != nullptr) {
float value = 0.;
if (sscanf(text, "%f", &value) == 1) {
if (const char* percent = strchr(text, '%'); percent != nullptr) {
// Flip sign, percent is negative.
value = -(value / 100.f);
_frame_offset_prc[0] = true;
} else {
_frame_offset_prc[0] = false;
}
}
_frame_offset.x = value;
}
if (const char* text = obs_data_get_string(data, ST_KEY_FRAMING_OFFSET ".Y"); text != nullptr) {
float value = 0.;
if (sscanf(text, "%f", &value) == 1) {
if (const char* percent = strchr(text, '%'); percent != nullptr) {
// Flip sign, percent is negative.
value = -(value / 100.f);
_frame_offset_prc[1] = true;
} else {
_frame_offset_prc[1] = false;
}
}
_frame_offset.y = value;
}
}
{ // Aspect Ratio
_frame_aspect_ratio = static_cast<float>(_size.first) / static_cast<float>(_size.second);
if (const char* text = obs_data_get_string(data, ST_KEY_FRAMING_ASPECTRATIO); text != nullptr) {
if (const char* percent = strchr(text, ':'); percent != nullptr) {
float left = 0.;
float right = 0.;
if ((sscanf(text, "%f", &left) == 1) && (sscanf(percent + 1, "%f", &right) == 1)) {
_frame_aspect_ratio = left / right;
} else {
_frame_aspect_ratio = 0.0;
}
} else {
float value = 0.;
if (sscanf(text, "%f", &value) == 1) {
_frame_aspect_ratio = value;
} else {
_frame_aspect_ratio = 0.0;
}
}
}
}
// Advanced / Provider
{ // Check if the user changed which Denoising provider we use.
auto provider = static_cast<tracking_provider>(obs_data_get_int(data, ST_KEY_ADVANCED_PROVIDER));
if (provider == tracking_provider::AUTOMATIC) {
provider = autoframing_factory::get()->find_ideal_provider();
}
// Check if the provider was changed, and if so switch.
if (provider != _provider) {
_provider_ui = provider;
switch_provider(provider);
}
if (_provider_ready) {
std::unique_lock<std::mutex> ul(_provider_lock);
switch (_provider) {
#ifdef ENABLE_FILTER_UPSCALING_NVIDIA
case tracking_provider::NVIDIA_FACEDETECTION:
nvar_facedetection_update(data);
break;
#endif
default:
break;
}
}
}
_debug = obs_data_get_bool(data, "Debug");
}
void streamfx::filter::autoframing::autoframing_instance::properties(obs_properties_t* properties)
{
switch (_provider_ui) {
#ifdef ENABLE_FILTER_AUTOFRAMING_NVIDIA
case tracking_provider::NVIDIA_FACEDETECTION:
nvar_facedetection_properties(properties);
break;
#endif
default:
break;
}
}
uint32_t autoframing_instance::get_width()
{
if (_debug) {
return std::max<uint32_t>(_size.first, 1);
}
return std::max<uint32_t>(_out_size.first, 1);
}
uint32_t autoframing_instance::get_height()
{
if (_debug) {
return std::max<uint32_t>(_size.second, 1);
}
return std::max<uint32_t>(_out_size.second, 1);
}
void autoframing_instance::video_tick(float_t seconds)
{
auto target = obs_filter_get_target(_self);
auto width = obs_source_get_base_width(target);
auto height = obs_source_get_base_height(target);
_size = {width, height};
{ // Calculate output size for aspect ratio.
_out_size = _size;
if (_frame_aspect_ratio > 0.0) {
if (width > height) {
_out_size.first = static_cast<uint32_t>(std::lroundf(static_cast<float>(_out_size.second) * _frame_aspect_ratio), 0, std::numeric_limits<uint32_t>::max());
} else {
_out_size.second = static_cast<uint32_t>(std::lroundf(static_cast<float>(_out_size.first) * _frame_aspect_ratio), 0, std::numeric_limits<uint32_t>::max());
}
}
}
// Update tracking.
tracking_tick(seconds);
// Mark the effect as dirty.
_dirty = true;
}
void autoframing_instance::video_render(gs_effect_t* effect)
{
auto parent = obs_filter_get_parent(_self);
auto target = obs_filter_get_target(_self);
auto width = obs_source_get_base_width(target);
auto height = obs_source_get_base_height(target);
vec4 blank = vec4{0, 0, 0, 0};
// Ensure we have the bare minimum of valid information.
target = target ? target : parent;
effect = effect ? effect : obs_get_base_effect(OBS_EFFECT_DEFAULT);
// Skip the filter if:
// - The Provider isn't ready yet.
// - We don't have a target.
// - The width/height of the next filter in the chain is empty.
if (!_provider_ready || !target || (width == 0) || (height == 0)) {
obs_source_skip_video_filter(_self);
return;
}
#if defined(ENABLE_PROFILING) && !defined(D_PLATFORM_MAC) && _DEBUG
::streamfx::obs::gs::debug_marker profiler0{::streamfx::obs::gs::debug_color_source, "StreamFX Auto-Framing"};
::streamfx::obs::gs::debug_marker profiler0_0{::streamfx::obs::gs::debug_color_gray, "'%s' on '%s'", obs_source_get_name(_self), obs_source_get_name(parent)};
#endif
if (_dirty) {
// Capture the input.
if (obs_source_process_filter_begin(_self, GS_RGBA, OBS_ALLOW_DIRECT_RENDERING)) {
auto op = _input->render(width, height);
// Set correct projection matrix.
gs_ortho(0, static_cast<float>(width), 0, static_cast<float>(height), 0, 1);
// Clear the buffer
gs_clear(GS_CLEAR_COLOR | GS_CLEAR_DEPTH, &blank, 0, 0);
// Set GPU state
gs_blend_state_push();
gs_enable_color(true, true, true, true);
gs_enable_blending(false);
gs_enable_depth_test(false);
gs_enable_stencil_test(false);
gs_set_cull_mode(GS_NEITHER);
// Render
bool srgb = gs_framebuffer_srgb_enabled();
gs_enable_framebuffer_srgb(gs_get_linear_srgb());
obs_source_process_filter_end(_self, obs_get_base_effect(OBS_EFFECT_DEFAULT), width, height);
gs_enable_framebuffer_srgb(srgb);
// Reset GPU state
gs_blend_state_pop();
} else {
obs_source_skip_video_filter(_self);
return;
}
// Lock & Process the captured input with the provider.
if (_track_frequency_counter >= _track_frequency) {
_track_frequency_counter = 0;
std::unique_lock<std::mutex> ul(_provider_lock);
switch (_provider) {
#ifdef ENABLE_FILTER_DENOISING_NVIDIA
case tracking_provider::NVIDIA_FACEDETECTION:
nvar_facedetection_process();
break;
#endif
default:
obs_source_skip_video_filter(_self);
return;
}
}
_dirty = false;
}
{ // Draw the result for the next filter to use.
#if defined(ENABLE_PROFILING) && !defined(D_PLATFORM_MAC) && _DEBUG
::streamfx::obs::gs::debug_marker profiler1{::streamfx::obs::gs::debug_color_render, "Render"};
#endif
if (_debug) { // Debug Mode
gs_effect_set_texture(gs_effect_get_param_by_name(effect, "image"), _input->get_object());
while (gs_effect_loop(effect, "Draw")) {
gs_draw_sprite(nullptr, 0, _size.first, _size.second);
}
for (auto kv : _predicted_elements) {
// Tracked Area (Red)
_gfx_debug->draw_rectangle(kv.first->pos.x - kv.first->size.x / 2.f, kv.first->pos.y - kv.first->size.y / 2.f, kv.first->size.x, kv.first->size.y, true, 0x7E0000FF);
// Velocity Arrow (Black)
_gfx_debug->draw_arrow(kv.first->pos.x, kv.first->pos.y, kv.first->pos.x + kv.first->vel.x, kv.first->pos.y + kv.first->vel.y, 0., 0x7E000000);
// Predicted Area (Orange)
_gfx_debug->draw_rectangle(kv.second->mp_pos.x - kv.first->size.x / 2.f, kv.second->mp_pos.y - kv.first->size.y / 2.f, kv.first->size.x, kv.first->size.y, true, 0x7E007EFF);
// Filtered Area (Yellow)
_gfx_debug->draw_rectangle(kv.second->filter_pos_x.get() - kv.first->size.x / 2.f, kv.second->filter_pos_y.get() - kv.first->size.y / 2.f, kv.first->size.x, kv.first->size.y, true, 0x7E00FFFF);
// Offset Filtered Area (Blue)
_gfx_debug->draw_rectangle(kv.second->offset_pos.x - kv.first->size.x / 2.f, kv.second->offset_pos.y - kv.first->size.y / 2.f, kv.first->size.x, kv.first->size.y, true, 0x7EFF0000);
// Padded Offset Filtered Area (Cyan)
_gfx_debug->draw_rectangle(kv.second->offset_pos.x - kv.second->pad_size.x / 2.f, kv.second->offset_pos.y - kv.second->pad_size.y / 2.f, kv.second->pad_size.x, kv.second->pad_size.y, true, 0x7EFFFF00);
// Aspect-Ratio-Corrected Padded Offset Filtered Area (Green)
_gfx_debug->draw_rectangle(kv.second->offset_pos.x - kv.second->aspected_size.x / 2.f, kv.second->offset_pos.y - kv.second->aspected_size.y / 2.f, kv.second->aspected_size.x, kv.second->aspected_size.y, true, 0x7E00FF00);
}
// Final Region (White)
_gfx_debug->draw_rectangle(_frame_pos.x - _frame_size.x / 2.f, _frame_pos.y - _frame_size.y / 2.f, _frame_size.x, _frame_size.y, true, 0x7EFFFFFF);
} else {
float x0 = (_frame_pos.x - _frame_size.x / 2.f) / static_cast<float>(_size.first);
float x1 = (_frame_pos.x + _frame_size.x / 2.f) / static_cast<float>(_size.first);
float y0 = (_frame_pos.y - _frame_size.y / 2.f) / static_cast<float>(_size.second);
float y1 = (_frame_pos.y + _frame_size.y / 2.f) / static_cast<float>(_size.second);
{
auto v = _vb->at(0);
vec3_set(v.position, 0., 0., 0.);
v.uv[0]->x = x0;
v.uv[0]->y = y0;
}
{
auto v = _vb->at(1);
vec3_set(v.position, static_cast<float>(_out_size.first), 0., 0.);
v.uv[0]->x = x1;
v.uv[0]->y = y0;
}
{
auto v = _vb->at(2);
vec3_set(v.position, 0., static_cast<float>(_out_size.second), 0.);
v.uv[0]->x = x0;
v.uv[0]->y = y1;
}
{
auto v = _vb->at(3);
vec3_set(v.position, static_cast<float>(_out_size.first), static_cast<float>(_out_size.second), 0.);
v.uv[0]->x = x1;
v.uv[0]->y = y1;
}
gs_load_vertexbuffer(_vb->update(true));
if (!effect) {
if (_standard_effect->has_parameter("InputA", ::streamfx::obs::gs::effect_parameter::type::Texture)) {
_standard_effect->get_parameter("InputA").set_texture(_input->get_texture());
}
while (gs_effect_loop(_standard_effect->get_object(), "Texture")) {
gs_draw(GS_TRISTRIP, 0, 4);
}
} else {
gs_effect_set_texture(gs_effect_get_param_by_name(effect, "image"), _input->get_texture()->get_object());
while (gs_effect_loop(effect, "Draw")) {
gs_draw(GS_TRISTRIP, 0, 4);
}
}
gs_load_vertexbuffer(nullptr);
}
}
}
void streamfx::filter::autoframing::autoframing_instance::tracking_tick(float seconds)
{
{ // Increase the age of all elements, and kill off any that are "too old".
float threshold = (0.5f * (1.f / (1.f - _track_frequency)));
auto iter = _tracked_elements.begin();
while (iter != _tracked_elements.end()) {
// Increment the age by the tick duration.
(*iter)->age += seconds;
// If the age exceeds the threshold, remove it.
if ((*iter)->age >= threshold) {
if (iter == _tracked_elements.begin()) {
// Erase iter, then reset to start.
_predicted_elements.erase(*iter);
_tracked_elements.erase(iter);
iter = _tracked_elements.begin();
} else {
// Copy, then advance before erasing.
auto iter2 = iter;
iter++;
_predicted_elements.erase(*iter2);
_tracked_elements.erase(iter2);
}
} else {
// Move ahead.
iter++;
}
}
}
for (auto trck : _tracked_elements) { // Updated predicted elements
std::shared_ptr<pred_el> pred;
// Find the corresponding prediction element.
auto iter = _predicted_elements.find(trck);
if (iter == _predicted_elements.end()) {
pred = std::make_shared<pred_el>();
_predicted_elements.insert_or_assign(trck, pred);
pred->filter_pos_x = {_motion_smoothing_kalman_pnc, _motion_smoothing_kalman_mnc, ST_KALMAN_EEC, trck->pos.x};
pred->filter_pos_y = {_motion_smoothing_kalman_pnc, _motion_smoothing_kalman_mnc, ST_KALMAN_EEC, trck->pos.y};
} else {
pred = iter->second;
}
// Calculate absolute velocity.
vec2 vel;
vec2_copy(&vel, &trck->vel);
vec2_mulf(&vel, &vel, _motion_prediction);
vec2_mulf(&vel, &vel, seconds);
// Calculate predicted position.
vec2 pos;
if (trck->age > seconds) {
vec2_copy(&pos, &pred->mp_pos);
} else {
vec2_copy(&pos, &trck->pos);
}
vec2_add(&pos, &pos, &vel);
vec2_copy(&pred->mp_pos, &pos);
// Update filtered position.
pred->filter_pos_x.filter(pred->mp_pos.x);
pred->filter_pos_y.filter(pred->mp_pos.y);
// Update offset position.
vec2_set(&pred->offset_pos, pred->filter_pos_x.get(), pred->filter_pos_y.get());
if (_frame_offset_prc[0]) { // %
pred->offset_pos.x += trck->size.x * (-_frame_offset.x);
} else { // Pixels
pred->offset_pos.x += _frame_offset.x;
}
if (_frame_offset_prc[1]) { // %
pred->offset_pos.y += trck->size.y * (-_frame_offset.y);
} else { // Pixels
pred->offset_pos.y += _frame_offset.y;
}
// Calculate padded area.
vec2_copy(&pred->pad_size, &trck->size);
if (_frame_padding_prc[0]) { // %
pred->pad_size.x += trck->size.x * (-_frame_padding.x) * 2.f;
} else { // Pixels
pred->pad_size.x += _frame_padding.x * 2.f;
}
if (_frame_padding_prc[1]) { // %
pred->pad_size.y += trck->size.y * (-_frame_padding.y) * 2.f;
} else { // Pixels
pred->pad_size.y += _frame_padding.y * 2.f;
}
// Adjust to match aspect ratio (width / height).
vec2_copy(&pred->aspected_size, &pred->pad_size);
if (_frame_aspect_ratio > 0.0) {
if ((pred->aspected_size.x / pred->aspected_size.y) >= _frame_aspect_ratio) { // Ours > Target
pred->aspected_size.y = pred->aspected_size.x / _frame_aspect_ratio;
} else { // Target > Ours
pred->aspected_size.x = pred->aspected_size.y * _frame_aspect_ratio;
}
}
}
{ // Find final frame.
bool need_filter = true;
if (_predicted_elements.size() > 0) {
if (_track_mode == tracking_mode::SOLO) {
auto kv = _predicted_elements.rbegin();
_frame_pos_x.filter(kv->second->offset_pos.x);
_frame_pos_y.filter(kv->second->offset_pos.y);
vec2_set(&_frame_pos, _frame_pos_x.get(), _frame_pos_y.get());
vec2_copy(&_frame_size, &kv->second->aspected_size);
need_filter = false;
} else {
vec2 min;
vec2 max;
vec2_set(&min, std::numeric_limits<float>::max(), std::numeric_limits<float>::max());
vec2_set(&max, 0., 0.);
for (auto kv : _predicted_elements) {
vec2 size;
vec2 low;
vec2 high;
vec2_copy(&size, &kv.second->aspected_size);
vec2_mulf(&size, &size, .5f);
vec2_copy(&low, &kv.second->offset_pos);
vec2_copy(&high, &kv.second->offset_pos);
vec2_sub(&low, &low, &size);
vec2_add(&high, &high, &size);
if (low.x < min.x) {
min.x = low.x;
}
if (low.y < min.y) {
min.y = low.y;
}
if (high.x > max.x) {
max.x = high.x;
}
if (high.y > max.y) {
max.y = high.y;
}
}
// Calculate center.
vec2 center;
vec2_add(&center, &min, &max);
vec2_divf(&center, &center, 2.f);
// Assign center.
_frame_pos_x.filter(center.x);
_frame_pos_y.filter(center.y);
// Calculate size.
vec2 size;
vec2_copy(&size, &max);
vec2_sub(&size, &size, &min);
_frame_size_x.filter(size.x);
_frame_size_y.filter(size.y);
}
} else {
_frame_pos_x.filter(static_cast<float>(_size.first) / 2.f);
_frame_pos_y.filter(static_cast<float>(_size.second) / 2.f);
_frame_size_x.filter(static_cast<float>(_size.first));
_frame_size_y.filter(static_cast<float>(_size.second));
}
// Grab filtered data if needed, otherwise stick with direct data.
if (need_filter) {
vec2_set(&_frame_pos, _frame_pos_x.get(), _frame_pos_y.get());
vec2_set(&_frame_size, _frame_size_x.get(), _frame_size_y.get());
}
{ // Aspect Ratio correction is a three step process:
float aspect = _frame_aspect_ratio > 0. ? _frame_aspect_ratio : (static_cast<float>(_size.first) / static_cast<float>(_size.second));
{ // 1. Adjust aspect ratio so that all elements end up contained.
float frame_aspect = _frame_size.x / _frame_size.y;
if (aspect < frame_aspect) {
_frame_size.y = _frame_size.x / aspect;
} else {
_frame_size.x = _frame_size.y * aspect;
}
}
// 2. Limit the size of the frame to the allowed region, and adjust it so it's inside the frame.
// This will move the center, which might not be a wanted side effect.
vec4 rect;
rect.x = std::clamp<float>(_frame_pos.x - _frame_size.x / 2.f, 0.f, static_cast<float>(_size.first));
rect.z = std::clamp<float>(_frame_pos.x + _frame_size.x / 2.f, 0.f, static_cast<float>(_size.first));
rect.y = std::clamp<float>(_frame_pos.y - _frame_size.y / 2.f, 0.f, static_cast<float>(_size.second));
rect.w = std::clamp<float>(_frame_pos.y + _frame_size.y / 2.f, 0.f, static_cast<float>(_size.second));
_frame_pos.x = (rect.x + rect.z) / 2.f;
_frame_pos.y = (rect.y + rect.w) / 2.f;
_frame_size.x = (rect.z - rect.x);
_frame_size.y = (rect.w - rect.y);
{ // 3. Adjust the aspect ratio so that it matches the expected output aspect ratio.
float frame_aspect = _frame_size.x / _frame_size.y;
if (aspect < frame_aspect) {
_frame_size.x = _frame_size.y * aspect;
} else {
_frame_size.y = _frame_size.x / aspect;
}
}
}
}
// Increment tracking counter.
_track_frequency_counter += seconds;
}
struct switch_provider_data_t {
tracking_provider provider;
};
void streamfx::filter::autoframing::autoframing_instance::switch_provider(tracking_provider provider)
{
std::unique_lock<std::mutex> ul(_provider_lock);
// Safeguard against calls made from unlocked memory.
if (provider == _provider) {
return;
}
// This doesn't work correctly.
// - Need to allow multiple switches at once because OBS is weird.
// - Doesn't guarantee that the task is properly killed off.
// Log information.
D_LOG_INFO("Instance '%s' is switching provider from '%s' to '%s'.", obs_source_get_name(_self), cstring(_provider), cstring(provider));
// If there is an ongoing task to switch provider, cancel it.
if (_provider_task) {
// De-queue it.
streamfx::threadpool()->pop(_provider_task);
// Await the death of the task itself.
_provider_task->await_completion();
// Clear any memory associated with it.
_provider_task.reset();
}
// Build data to pass into the task.
auto spd = std::make_shared<switch_provider_data_t>();
spd->provider = _provider;
_provider = provider;
// Then spawn a new task to switch provider.
_provider_task = streamfx::threadpool()->push(std::bind(&autoframing_instance::task_switch_provider, this, std::placeholders::_1), spd);
}
void streamfx::filter::autoframing::autoframing_instance::task_switch_provider(util::threadpool::task_data_t data)
{
std::shared_ptr<switch_provider_data_t> spd = std::static_pointer_cast<switch_provider_data_t>(data);
// Mark the provider as no longer ready.
_provider_ready = false;
// Lock the provider from being used.
std::unique_lock<std::mutex> ul(_provider_lock);
try {
// Unload the previous provider.
switch (spd->provider) {
#ifdef ENABLE_FILTER_AUTOFRAMING_NVIDIA
case tracking_provider::NVIDIA_FACEDETECTION:
nvar_facedetection_unload();
break;
#endif
default:
break;
}
// Load the new provider.
switch (_provider) {
#ifdef ENABLE_FILTER_AUTOFRAMING_NVIDIA
case tracking_provider::NVIDIA_FACEDETECTION:
nvar_facedetection_load();
break;
#endif
default:
break;
}
// Log information.
D_LOG_INFO("Instance '%s' switched provider from '%s' to '%s'.", obs_source_get_name(_self), cstring(spd->provider), cstring(_provider));
_provider_ready = true;
} catch (std::exception const& ex) {
// Log information.
D_LOG_ERROR("Instance '%s' failed switching provider with error: %s", obs_source_get_name(_self), ex.what());
}
}
#ifdef ENABLE_FILTER_AUTOFRAMING_NVIDIA
void streamfx::filter::autoframing::autoframing_instance::nvar_facedetection_load()
{
_nvidia_fx = std::make_shared<::streamfx::nvidia::ar::facedetection>();
}
void streamfx::filter::autoframing::autoframing_instance::nvar_facedetection_unload()
{
_nvidia_fx.reset();
}
void streamfx::filter::autoframing::autoframing_instance::nvar_facedetection_process()
{
if (!_nvidia_fx) {
return;
}
// Frames may not move more than this distance.
float max_dst = sqrtf(static_cast<float>(_size.first * _size.first) + static_cast<float>(_size.second * _size.second)) * 0.667f;
max_dst *= 1.f / (1.f - _track_frequency); // Fine-tune this?
// Process the current frame (if requested).
_nvidia_fx->process(_input->get_texture());
// If there are tracked faces, merge them with the tracked elements.
if (auto edx = _nvidia_fx->count(); edx > 0) {
for (size_t idx = 0; idx < edx; idx++) {
float confidence = 0.;
auto rect = _nvidia_fx->at(idx, confidence);
// Skip elements that have not enough confidence of being a face.
// TODO: Make the threshold configurable.
if (confidence < .5) {
continue;
}
// Calculate centered position.
vec2 pos;
pos.x = rect.x + (rect.z / 2.f);
pos.y = rect.y + (rect.w / 2.f);
// Try and find a match in the current list of tracked elements.
std::shared_ptr<track_el> match;
float match_dst = max_dst;
for (const auto& el : _tracked_elements) {
// Skip "fresh" elements.
if (el->age < 0.00001) {
continue;
}
// Check if the distance is within acceptable bounds.
float dst = vec2_dist(&pos, &el->pos);
if ((dst < match_dst) && (dst < max_dst)) {
match_dst = dst;
match = el;
}
}
// Do we have a match?
if (!match) {
// No, so create a new one.
match = std::make_shared<track_el>();
// Insert it.
_tracked_elements.push_back(match);
// Update information.
vec2_copy(&match->pos, &pos);
vec2_set(&match->size, rect.z, rect.w);
vec2_set(&match->vel, 0., 0.);
match->age = 0.;
} else {
// Reset the age to 0.
match->age = 0.;
// Calculate the velocity between changes.
vec2 vel;
vec2_sub(&vel, &pos, &match->pos);
// Update information.
vec2_copy(&match->pos, &pos);
vec2_set(&match->size, rect.z, rect.w);
vec2_copy(&match->vel, &vel);
match->age = 0.;
}
}
}
}
void streamfx::filter::autoframing::autoframing_instance::nvar_facedetection_properties(obs_properties_t* props) {}
void streamfx::filter::autoframing::autoframing_instance::nvar_facedetection_update(obs_data_t* data)
{
if (!_nvidia_fx) {
return;
}
switch (_track_mode) {
case tracking_mode::SOLO:
_nvidia_fx->set_tracking_limit(1);
break;
case tracking_mode::GROUP:
_nvidia_fx->set_tracking_limit(_nvidia_fx->tracking_limit_range().second);
break;
}
}
#endif
autoframing_factory::autoframing_factory()
{
bool any_available = false;
// 1. Try and load any configured providers.
#ifdef ENABLE_FILTER_AUTOFRAMING_NVIDIA
try {
// Load CVImage and Video Effects SDK.
_nvcuda = ::streamfx::nvidia::cuda::obs::get();
_nvcvi = ::streamfx::nvidia::cv::cv::get();
_nvar = ::streamfx::nvidia::ar::ar::get();
_nvidia_available = true;
any_available |= _nvidia_available;
} catch (const std::exception& ex) {
_nvidia_available = false;
_nvar.reset();
_nvcvi.reset();
_nvcuda.reset();
D_LOG_WARNING("Failed to make NVIDIA providers available due to error: %s", ex.what());
} catch (...) {
_nvidia_available = false;
_nvar.reset();
_nvcvi.reset();
_nvcuda.reset();
D_LOG_WARNING("Failed to make NVIDIA providers available with unknown error.", nullptr);
}
#endif
// 2. Check if any of them managed to load at all.
if (!any_available) {
D_LOG_ERROR("All supported providers failed to initialize, disabling effect.", 0);
return;
}
// Register initial source.
_info.id = S_PREFIX "filter-autoframing";
_info.type = OBS_SOURCE_TYPE_FILTER;
_info.output_flags = OBS_SOURCE_VIDEO;
support_size(true);
finish_setup();
// Register proxy identifiers.
register_proxy("streamfx-filter-nvidia-face-tracking");
register_proxy("streamfx-nvidia-face-tracking");
}
autoframing_factory::~autoframing_factory() {}
const char* autoframing_factory::get_name()
{
return D_TRANSLATE(ST_I18N);
}
void autoframing_factory::get_defaults2(obs_data_t* data)
{
// Tracking
obs_data_set_default_int(data, ST_KEY_TRACKING_MODE, static_cast<int64_t>(tracking_mode::SOLO));
obs_data_set_default_string(data, ST_KEY_TRACKING_FREQUENCY, "20 Hz");
// Motion
obs_data_set_default_double(data, ST_KEY_MOTION_SMOOTHING, 33.333);
obs_data_set_default_double(data, ST_KEY_MOTION_PREDICTION, 200.0);
// Framing
obs_data_set_default_double(data, ST_KEY_FRAMING_STABILITY, 10.0);
obs_data_set_default_string(data, ST_KEY_FRAMING_PADDING ".X", "33.333 %");
obs_data_set_default_string(data, ST_KEY_FRAMING_PADDING ".Y", "33.333 %");
obs_data_set_default_string(data, ST_KEY_FRAMING_OFFSET ".X", " 0.00 %");
obs_data_set_default_string(data, ST_KEY_FRAMING_OFFSET ".Y", "-7.50 %");
obs_data_set_default_string(data, ST_KEY_FRAMING_ASPECTRATIO, "");
// Advanced
obs_data_set_default_int(data, ST_KEY_ADVANCED_PROVIDER, static_cast<int64_t>(tracking_provider::AUTOMATIC));
obs_data_set_default_bool(data, "Debug", false);
}
static bool modified_provider(obs_properties_t* props, obs_property_t*, obs_data_t* settings) noexcept
{
try {
return true;
} catch (const std::exception& ex) {
DLOG_ERROR("Unexpected exception in function '%s': %s.", __FUNCTION_NAME__, ex.what());
return false;
} catch (...) {
DLOG_ERROR("Unexpected exception in function '%s'.", __FUNCTION_NAME__);
return false;
}
}
obs_properties_t* autoframing_factory::get_properties2(autoframing_instance* data)
{
obs_properties_t* pr = obs_properties_create();
#ifdef ENABLE_FRONTEND
{
obs_properties_add_button2(pr, S_MANUAL_OPEN, D_TRANSLATE(S_MANUAL_OPEN), autoframing_factory::on_manual_open, nullptr);
}
#endif
{
auto grp = obs_properties_create();
obs_properties_add_group(pr, ST_I18N_TRACKING, D_TRANSLATE(ST_I18N_TRACKING), OBS_GROUP_NORMAL, grp);
{
auto p = obs_properties_add_list(grp, ST_KEY_TRACKING_MODE, D_TRANSLATE(ST_I18N_TRACKING_MODE), OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_INT);
obs_property_set_modified_callback(p, modified_provider);
obs_property_list_add_int(p, D_TRANSLATE(ST_I18N_FRAMING_MODE_SOLO), static_cast<int64_t>(tracking_mode::SOLO));
obs_property_list_add_int(p, D_TRANSLATE(ST_I18N_FRAMING_MODE_GROUP), static_cast<int64_t>(tracking_mode::GROUP));
}
{
auto p = obs_properties_add_text(grp, ST_KEY_TRACKING_FREQUENCY, D_TRANSLATE(ST_I18N_TRACKING_FREQUENCY), OBS_TEXT_DEFAULT);
}
}
{
auto grp = obs_properties_create();
obs_properties_add_group(pr, ST_I18N_MOTION, D_TRANSLATE(ST_I18N_MOTION), OBS_GROUP_NORMAL, grp);
{
auto p = obs_properties_add_float_slider(grp, ST_KEY_MOTION_SMOOTHING, D_TRANSLATE(ST_I18N_MOTION_SMOOTHING), 0.0, 100.0, 0.01);
obs_property_float_set_suffix(p, " %");
}
{
auto p = obs_properties_add_float_slider(grp, ST_KEY_MOTION_PREDICTION, D_TRANSLATE(ST_I18N_MOTION_PREDICTION), 0.0, 500.0, 0.01);
obs_property_float_set_suffix(p, " %");
}
}
{
auto grp = obs_properties_create();
obs_properties_add_group(pr, ST_I18N_FRAMING, D_TRANSLATE(ST_I18N_FRAMING), OBS_GROUP_NORMAL, grp);
{
auto p = obs_properties_add_float_slider(grp, ST_KEY_FRAMING_STABILITY, D_TRANSLATE(ST_I18N_FRAMING_STABILITY), 0.0, 100.0, 0.01);
obs_property_float_set_suffix(p, " %");
}
{
auto grp2 = obs_properties_create();
obs_properties_add_group(grp, ST_KEY_FRAMING_PADDING, D_TRANSLATE(ST_I18N_FRAMING_PADDING), OBS_GROUP_NORMAL, grp2);
{
auto p = obs_properties_add_text(grp2, ST_KEY_FRAMING_PADDING ".X", "X", OBS_TEXT_DEFAULT);
}
{
auto p = obs_properties_add_text(grp2, ST_KEY_FRAMING_PADDING ".Y", "Y", OBS_TEXT_DEFAULT);
}
}
{
auto grp2 = obs_properties_create();
obs_properties_add_group(grp, ST_KEY_FRAMING_OFFSET, D_TRANSLATE(ST_I18N_FRAMING_OFFSET), OBS_GROUP_NORMAL, grp2);
{
auto p = obs_properties_add_text(grp2, ST_KEY_FRAMING_OFFSET ".X", "X", OBS_TEXT_DEFAULT);
}
{
auto p = obs_properties_add_text(grp2, ST_KEY_FRAMING_OFFSET ".Y", "Y", OBS_TEXT_DEFAULT);
}
}
{
auto p = obs_properties_add_list(grp, ST_KEY_FRAMING_ASPECTRATIO, D_TRANSLATE(ST_I18N_FRAMING_ASPECTRATIO), OBS_COMBO_TYPE_EDITABLE, OBS_COMBO_FORMAT_STRING);
obs_property_list_add_string(p, "None", "");
obs_property_list_add_string(p, "1:1", "1:1");
obs_property_list_add_string(p, "3:2", "3:2");
obs_property_list_add_string(p, "2:3", "2:3");
obs_property_list_add_string(p, "4:3", "4:3");
obs_property_list_add_string(p, "3:4", "3:4");
obs_property_list_add_string(p, "5:4", "5:4");
obs_property_list_add_string(p, "4:5", "4:5");
obs_property_list_add_string(p, "16:9", "16:9");
obs_property_list_add_string(p, "9:16", "9:16");
obs_property_list_add_string(p, "16:10", "16:10");
obs_property_list_add_string(p, "10:16", "10:16");
obs_property_list_add_string(p, "21:9", "21:9");
obs_property_list_add_string(p, "9:21", "9:21");
obs_property_list_add_string(p, "21:10", "21:10");
obs_property_list_add_string(p, "10:21", "10:21");
obs_property_list_add_string(p, "32:9", "32:9");
obs_property_list_add_string(p, "9:32", "9:32");
obs_property_list_add_string(p, "32:10", "32:10");
obs_property_list_add_string(p, "10:32", "10:32");
}
}
if (data) {
data->properties(pr);
}
{ // Advanced Settings
auto grp = obs_properties_create();
obs_properties_add_group(pr, S_ADVANCED, D_TRANSLATE(S_ADVANCED), OBS_GROUP_NORMAL, grp);
{
auto p = obs_properties_add_list(grp, ST_KEY_ADVANCED_PROVIDER, D_TRANSLATE(ST_I18N_ADVANCED_PROVIDER), OBS_COMBO_TYPE_LIST, OBS_COMBO_FORMAT_INT);
obs_property_set_modified_callback(p, modified_provider);
obs_property_list_add_int(p, D_TRANSLATE(S_STATE_AUTOMATIC), static_cast<int64_t>(tracking_provider::AUTOMATIC));
#ifdef ENABLE_FILTER_AUTOFRAMING_NVIDIA
obs_property_list_add_int(p, D_TRANSLATE(ST_I18N_ADVANCED_PROVIDER_NVIDIA_FACEDETECTION), static_cast<int64_t>(tracking_provider::NVIDIA_FACEDETECTION));
#endif
}
obs_properties_add_bool(grp, "Debug", "Debug");
}
return pr;
}
#ifdef ENABLE_FRONTEND
bool streamfx::filter::autoframing::autoframing_factory::on_manual_open(obs_properties_t* props, obs_property_t* property, void* data)
{
streamfx::open_url(HELP_URL);
return false;
}
#endif
bool streamfx::filter::autoframing::autoframing_factory::is_provider_available(tracking_provider provider)
{
switch (provider) {
#ifdef ENABLE_FILTER_AUTOFRAMING_NVIDIA
case tracking_provider::NVIDIA_FACEDETECTION:
return _nvidia_available;
#endif
default:
return false;
}
}
tracking_provider streamfx::filter::autoframing::autoframing_factory::find_ideal_provider()
{
for (auto v : provider_priority) {
if (is_provider_available(v)) {
return v;
break;
}
}
return tracking_provider::INVALID;
}
std::shared_ptr<autoframing_factory> _filter_autoframing_factory_instance = nullptr;
void autoframing_factory::initialize()
{
try {
if (!_filter_autoframing_factory_instance)
_filter_autoframing_factory_instance = std::make_shared<autoframing_factory>();
} catch (const std::exception& ex) {
D_LOG_ERROR("Failed to initialize due to error: %s", ex.what());
} catch (...) {
D_LOG_ERROR("Failed to initialize due to unknown error.", "");
}
}
void autoframing_factory::finalize()
{
_filter_autoframing_factory_instance.reset();
}
std::shared_ptr<autoframing_factory> autoframing_factory::get()
{
return _filter_autoframing_factory_instance;
}