Move hw->sw frame mapping into a separate class

This commit is contained in:
Cameron Gutman 2023-03-26 14:41:02 -05:00
parent 5639bcc0f5
commit ba507d8046
5 changed files with 181 additions and 152 deletions

View file

@ -212,12 +212,14 @@ ffmpeg {
SOURCES += \
streaming/video/ffmpeg.cpp \
streaming/video/ffmpeg-renderers/sdlvid.cpp \
streaming/video/ffmpeg-renderers/swframemapper.cpp \
streaming/video/ffmpeg-renderers/pacer/pacer.cpp
HEADERS += \
streaming/video/ffmpeg.h \
streaming/video/ffmpeg-renderers/renderer.h \
streaming/video/ffmpeg-renderers/sdlvid.h \
streaming/video/ffmpeg-renderers/swframemapper.h \
streaming/video/ffmpeg-renderers/pacer/pacer.h
}
libva {

View file

@ -9,9 +9,8 @@ SdlRenderer::SdlRenderer()
: m_VideoFormat(0),
m_Renderer(nullptr),
m_Texture(nullptr),
m_SwPixelFormat(AV_PIX_FMT_NONE),
m_ColorSpace(-1),
m_MapFrame(false)
m_SwFrameMapper(this)
{
SDL_zero(m_OverlayTextures);
@ -92,6 +91,7 @@ bool SdlRenderer::initialize(PDECODER_PARAMETERS params)
Uint32 rendererFlags = SDL_RENDERER_ACCELERATED;
m_VideoFormat = params->videoFormat;
m_SwFrameMapper.setVideoFormat(m_VideoFormat);
if (params->videoFormat & VIDEO_FORMAT_MASK_10BIT) {
// SDL doesn't support rendering YUV 10-bit textures yet
@ -211,137 +211,6 @@ void SdlRenderer::renderOverlay(Overlay::OverlayType type)
}
}
bool SdlRenderer::initializeReadBackFormat(AVBufferRef* hwFrameCtxRef, AVFrame* testFrame)
{
auto hwFrameCtx = (AVHWFramesContext*)hwFrameCtxRef->data;
int err;
enum AVPixelFormat *formats;
AVFrame* outputFrame;
// This function must only be called once per instance
SDL_assert(m_SwPixelFormat == AV_PIX_FMT_NONE);
SDL_assert(!m_MapFrame);
// Try direct mapping before resorting to copying the frame
outputFrame = av_frame_alloc();
if (outputFrame != nullptr) {
err = av_hwframe_map(outputFrame, testFrame, AV_HWFRAME_MAP_READ);
if (err == 0) {
if (isPixelFormatSupported(m_VideoFormat, (AVPixelFormat)outputFrame->format)) {
m_SwPixelFormat = (AVPixelFormat)outputFrame->format;
m_MapFrame = true;
goto Exit;
}
else {
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"Skipping unsupported hwframe mapping format: %d",
outputFrame->format);
}
}
else {
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_map() is unsupported (error: %d)",
err);
SDL_assert(err == AVERROR(ENOSYS));
}
}
// Direct mapping didn't work, so let's see what transfer formats we have
err = av_hwframe_transfer_get_formats(hwFrameCtxRef, AV_HWFRAME_TRANSFER_DIRECTION_FROM, &formats, 0);
if (err < 0) {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_transfer_get_formats() failed: %d",
err);
goto Exit;
}
// NB: In this algorithm, we prefer to get a preferred hardware readback format
// and non-preferred rendering format rather than the other way around. This is
// why we loop through the readback format list in order, rather than searching
// for the format from getPreferredPixelFormat() in the list.
for (int i = 0; formats[i] != AV_PIX_FMT_NONE; i++) {
SDL_assert(m_VideoFormat != 0);
if (!isPixelFormatSupported(m_VideoFormat, formats[i])) {
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"Skipping unsupported hwframe transfer format %d",
formats[i]);
continue;
}
m_SwPixelFormat = formats[i];
break;
}
av_freep(&formats);
Exit:
av_frame_free(&outputFrame);
// If we didn't find any supported formats, try hwFrameCtx->sw_format.
if (m_SwPixelFormat == AV_PIX_FMT_NONE) {
if (isPixelFormatSupported(m_VideoFormat, hwFrameCtx->sw_format)) {
m_SwPixelFormat = hwFrameCtx->sw_format;
}
else {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"Unable to find compatible hwframe transfer format (sw_format = %d)",
hwFrameCtx->sw_format);
return false;
}
}
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"Selected hwframe->swframe format: %d (mapping: %s)",
m_SwPixelFormat,
m_MapFrame ? "yes" : "no");
return true;
}
AVFrame* SdlRenderer::getSwFrameFromHwFrame(AVFrame* hwFrame)
{
int err;
SDL_assert(m_SwPixelFormat != AV_PIX_FMT_NONE);
AVFrame* swFrame = av_frame_alloc();
if (swFrame == nullptr) {
return nullptr;
}
swFrame->format = m_SwPixelFormat;
if (m_MapFrame) {
// We don't use AV_HWFRAME_MAP_DIRECT here because it can cause huge
// performance penalties on Intel hardware with VAAPI due to mappings
// being uncached memory.
err = av_hwframe_map(swFrame, hwFrame, AV_HWFRAME_MAP_READ);
if (err < 0) {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_map() failed: %d",
err);
av_frame_free(&swFrame);
return nullptr;
}
}
else {
err = av_hwframe_transfer_data(swFrame, hwFrame, 0);
if (err < 0) {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_transfer_data() failed: %d",
err);
av_frame_free(&swFrame);
return nullptr;
}
// av_hwframe_transfer_data() doesn't transfer metadata
// (and can even nuke existing metadata in dst), so we
// will propagate metadata manually afterwards.
av_frame_copy_props(swFrame, hwFrame);
}
return swFrame;
}
void SdlRenderer::renderFrame(AVFrame* frame)
{
int err;
@ -355,17 +224,8 @@ ReadbackRetry:
// accelerated decoder, we'll need to read the frame
// back to render it.
// Find the native read-back format
if (m_SwPixelFormat == AV_PIX_FMT_NONE) {
initializeReadBackFormat(frame->hw_frames_ctx, frame);
// If we don't support any of the hw transfer formats, we should
// have failed inside testRenderFrame() and not made it here.
SDL_assert(m_SwPixelFormat != AV_PIX_FMT_NONE);
}
// Map or copy this hwframe to a swframe that we can work with
frame = swFrame = getSwFrameFromHwFrame(frame);
frame = swFrame = m_SwFrameMapper.getSwFrameFromHwFrame(frame);
if (swFrame == nullptr) {
return;
}
@ -579,11 +439,7 @@ bool SdlRenderer::testRenderFrame(AVFrame* frame)
}
#endif
if (!initializeReadBackFormat(frame->hw_frames_ctx, frame)) {
return false;
}
AVFrame* swFrame = getSwFrameFromHwFrame(frame);
AVFrame* swFrame = m_SwFrameMapper.getSwFrameFromHwFrame(frame);
if (swFrame == nullptr) {
return false;
}

View file

@ -1,6 +1,7 @@
#pragma once
#include "renderer.h"
#include "swframemapper.h"
#ifdef HAVE_CUDA
#include "cuda.h"
@ -19,18 +20,16 @@ public:
private:
void renderOverlay(Overlay::OverlayType type);
bool initializeReadBackFormat(AVBufferRef* hwFrameCtxRef, AVFrame* testFrame);
AVFrame* getSwFrameFromHwFrame(AVFrame* hwFrame);
int m_VideoFormat;
SDL_Renderer* m_Renderer;
SDL_Texture* m_Texture;
enum AVPixelFormat m_SwPixelFormat;
int m_ColorSpace;
bool m_MapFrame;
SDL_Texture* m_OverlayTextures[Overlay::OverlayMax];
SDL_Rect m_OverlayRects[Overlay::OverlayMax];
SwFrameMapper m_SwFrameMapper;
#ifdef HAVE_CUDA
CUDAGLInteropHelper* m_CudaGLHelper;
#endif

View file

@ -0,0 +1,153 @@
#include "swframemapper.h"
SwFrameMapper::SwFrameMapper(IFFmpegRenderer* renderer)
: m_Renderer(renderer),
m_VideoFormat(0),
m_SwPixelFormat(AV_PIX_FMT_NONE),
m_MapFrame(false)
{
}
void SwFrameMapper::setVideoFormat(int videoFormat)
{
m_VideoFormat = videoFormat;
}
bool SwFrameMapper::initializeReadBackFormat(AVBufferRef* hwFrameCtxRef, AVFrame* testFrame)
{
auto hwFrameCtx = (AVHWFramesContext*)hwFrameCtxRef->data;
int err;
enum AVPixelFormat *formats;
AVFrame* outputFrame;
// This function must only be called once per instance
SDL_assert(m_SwPixelFormat == AV_PIX_FMT_NONE);
SDL_assert(!m_MapFrame);
SDL_assert(m_VideoFormat != 0);
// Try direct mapping before resorting to copying the frame
outputFrame = av_frame_alloc();
if (outputFrame != nullptr) {
err = av_hwframe_map(outputFrame, testFrame, AV_HWFRAME_MAP_READ);
if (err == 0) {
if (m_Renderer->isPixelFormatSupported(m_VideoFormat, (AVPixelFormat)outputFrame->format)) {
m_SwPixelFormat = (AVPixelFormat)outputFrame->format;
m_MapFrame = true;
goto Exit;
}
else {
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"Skipping unsupported hwframe mapping format: %d",
outputFrame->format);
}
}
else {
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_map() is unsupported (error: %d)",
err);
SDL_assert(err == AVERROR(ENOSYS));
}
}
// Direct mapping didn't work, so let's see what transfer formats we have
err = av_hwframe_transfer_get_formats(hwFrameCtxRef, AV_HWFRAME_TRANSFER_DIRECTION_FROM, &formats, 0);
if (err < 0) {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_transfer_get_formats() failed: %d",
err);
goto Exit;
}
// NB: In this algorithm, we prefer to get a preferred hardware readback format
// and non-preferred rendering format rather than the other way around. This is
// why we loop through the readback format list in order, rather than searching
// for the format from getPreferredPixelFormat() in the list.
for (int i = 0; formats[i] != AV_PIX_FMT_NONE; i++) {
if (!m_Renderer->isPixelFormatSupported(m_VideoFormat, formats[i])) {
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"Skipping unsupported hwframe transfer format %d",
formats[i]);
continue;
}
m_SwPixelFormat = formats[i];
break;
}
av_freep(&formats);
Exit:
av_frame_free(&outputFrame);
// If we didn't find any supported formats, try hwFrameCtx->sw_format.
if (m_SwPixelFormat == AV_PIX_FMT_NONE) {
if (m_Renderer->isPixelFormatSupported(m_VideoFormat, hwFrameCtx->sw_format)) {
m_SwPixelFormat = hwFrameCtx->sw_format;
}
else {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"Unable to find compatible hwframe transfer format (sw_format = %d)",
hwFrameCtx->sw_format);
return false;
}
}
SDL_LogInfo(SDL_LOG_CATEGORY_APPLICATION,
"Selected hwframe->swframe format: %d (mapping: %s)",
m_SwPixelFormat,
m_MapFrame ? "yes" : "no");
return true;
}
AVFrame* SwFrameMapper::getSwFrameFromHwFrame(AVFrame* hwFrame)
{
int err;
// setVideoFormat() must have been called before our first frame
SDL_assert(m_VideoFormat != 0);
if (m_SwPixelFormat == AV_PIX_FMT_NONE) {
SDL_assert(hwFrame->hw_frames_ctx != nullptr);
if (!initializeReadBackFormat(hwFrame->hw_frames_ctx, hwFrame)) {
return nullptr;
}
}
AVFrame* swFrame = av_frame_alloc();
if (swFrame == nullptr) {
return nullptr;
}
swFrame->format = m_SwPixelFormat;
if (m_MapFrame) {
// We don't use AV_HWFRAME_MAP_DIRECT here because it can cause huge
// performance penalties on Intel hardware with VAAPI due to mappings
// being uncached memory.
err = av_hwframe_map(swFrame, hwFrame, AV_HWFRAME_MAP_READ);
if (err < 0) {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_map() failed: %d",
err);
av_frame_free(&swFrame);
return nullptr;
}
}
else {
err = av_hwframe_transfer_data(swFrame, hwFrame, 0);
if (err < 0) {
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
"av_hwframe_transfer_data() failed: %d",
err);
av_frame_free(&swFrame);
return nullptr;
}
// av_hwframe_transfer_data() doesn't transfer metadata
// (and can even nuke existing metadata in dst), so we
// will propagate metadata manually afterwards.
av_frame_copy_props(swFrame, hwFrame);
}
return swFrame;
}

View file

@ -0,0 +1,19 @@
#pragma once
#include "renderer.h"
class SwFrameMapper
{
public:
explicit SwFrameMapper(IFFmpegRenderer* renderer);
void setVideoFormat(int videoFormat);
AVFrame* getSwFrameFromHwFrame(AVFrame* hwFrame);
private:
bool initializeReadBackFormat(AVBufferRef* hwFrameCtxRef, AVFrame* testFrame);
IFFmpegRenderer* m_Renderer;
int m_VideoFormat;
enum AVPixelFormat m_SwPixelFormat;
bool m_MapFrame;
};