mirror of
https://github.com/moonlight-stream/moonlight-qt
synced 2024-12-15 13:52:28 +00:00
291 lines
9.2 KiB
Text
291 lines
9.2 KiB
Text
// Nasty hack to avoid conflict between AVFoundation and
|
|
// libavutil both defining AVMediaType
|
|
#define AVMediaType AVMediaType_FFmpeg
|
|
#include "vt.h"
|
|
#undef AVMediaType
|
|
|
|
#include <SDL_syswm.h>
|
|
#include <Limelight.h>
|
|
|
|
#include <QQueue>
|
|
|
|
#import <Cocoa/Cocoa.h>
|
|
#import <VideoToolbox/VideoToolbox.h>
|
|
#import <AVFoundation/AVFoundation.h>
|
|
#import <CoreVideo/CoreVideo.h>
|
|
|
|
#define FRAME_HISTORY_ENTRIES 8
|
|
|
|
class VTRenderer : public IFFmpegRenderer
|
|
{
|
|
public:
|
|
VTRenderer()
|
|
: m_HwContext(nullptr),
|
|
m_DisplayLayer(nullptr),
|
|
m_FormatDesc(nullptr),
|
|
m_View(nullptr),
|
|
m_DisplayLink(nullptr),
|
|
m_FrameQueueLock(0)
|
|
{
|
|
}
|
|
|
|
virtual ~VTRenderer()
|
|
{
|
|
if (m_HwContext != nullptr) {
|
|
av_buffer_unref(&m_HwContext);
|
|
}
|
|
|
|
if (m_FormatDesc != nullptr) {
|
|
CFRelease(m_FormatDesc);
|
|
}
|
|
|
|
if (m_DisplayLink != nullptr) {
|
|
CVDisplayLinkStop(m_DisplayLink);
|
|
CVDisplayLinkRelease(m_DisplayLink);
|
|
}
|
|
|
|
while (!m_FrameQueue.isEmpty()) {
|
|
AVFrame* frame = m_FrameQueue.dequeue();
|
|
av_frame_free(&frame);
|
|
}
|
|
}
|
|
|
|
void drawFrame(uint64_t vsyncTime)
|
|
{
|
|
OSStatus status;
|
|
|
|
SDL_AtomicLock(&m_FrameQueueLock);
|
|
|
|
int frameDropTarget;
|
|
|
|
// If the queue length history entries are large, be strict
|
|
// about dropping excess frames.
|
|
frameDropTarget = 1;
|
|
for (int i = 0; i < m_FrameQueueHistory.count(); i++) {
|
|
if (m_FrameQueueHistory[i] <= 1) {
|
|
// Be lenient as long as the queue length
|
|
// resolves before the end of frame history
|
|
frameDropTarget = 3;
|
|
}
|
|
}
|
|
|
|
if (m_FrameQueueHistory.count() == FRAME_HISTORY_ENTRIES) {
|
|
m_FrameQueueHistory.dequeue();
|
|
}
|
|
|
|
m_FrameQueueHistory.enqueue(m_FrameQueue.count());
|
|
|
|
// Catch up if we're several frames ahead
|
|
while (m_FrameQueue.count() > frameDropTarget) {
|
|
AVFrame* frame = m_FrameQueue.dequeue();
|
|
av_frame_free(&frame);
|
|
}
|
|
|
|
if (m_FrameQueue.isEmpty()) {
|
|
SDL_AtomicUnlock(&m_FrameQueueLock);
|
|
return;
|
|
}
|
|
|
|
// Grab the first frame
|
|
AVFrame* frame = m_FrameQueue.dequeue();
|
|
SDL_AtomicUnlock(&m_FrameQueueLock);
|
|
|
|
CVPixelBufferRef pixBuf = reinterpret_cast<CVPixelBufferRef>(frame->data[3]);
|
|
|
|
// If the format has changed or doesn't exist yet, construct it with the
|
|
// pixel buffer data
|
|
if (!m_FormatDesc || !CMVideoFormatDescriptionMatchesImageBuffer(m_FormatDesc, pixBuf)) {
|
|
if (m_FormatDesc != nullptr) {
|
|
CFRelease(m_FormatDesc);
|
|
}
|
|
status = CMVideoFormatDescriptionCreateForImageBuffer(kCFAllocatorDefault,
|
|
pixBuf, &m_FormatDesc);
|
|
if (status != noErr) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
|
|
"CMVideoFormatDescriptionCreateForImageBuffer() failed: %d",
|
|
status);
|
|
av_frame_free(&frame);
|
|
return;
|
|
}
|
|
}
|
|
|
|
// Queue this sample for the next v-sync
|
|
CMSampleTimingInfo timingInfo = {
|
|
.duration = kCMTimeInvalid,
|
|
.decodeTimeStamp = kCMTimeInvalid,
|
|
.presentationTimeStamp = CMTimeMake(vsyncTime, 1000 * 1000 * 1000)
|
|
};
|
|
|
|
CMSampleBufferRef sampleBuffer;
|
|
status = CMSampleBufferCreateReadyWithImageBuffer(kCFAllocatorDefault,
|
|
pixBuf,
|
|
m_FormatDesc,
|
|
&timingInfo,
|
|
&sampleBuffer);
|
|
if (status != noErr) {
|
|
SDL_LogError(SDL_LOG_CATEGORY_APPLICATION,
|
|
"CMSampleBufferCreateReadyWithImageBuffer() failed: %d",
|
|
status);
|
|
av_frame_free(&frame);
|
|
return;
|
|
}
|
|
|
|
[m_DisplayLayer enqueueSampleBuffer:sampleBuffer];
|
|
|
|
CFRelease(sampleBuffer);
|
|
av_frame_free(&frame);
|
|
}
|
|
|
|
static
|
|
CVReturn
|
|
displayLinkOutputCallback(
|
|
CVDisplayLinkRef,
|
|
const CVTimeStamp*,
|
|
const CVTimeStamp* vsyncTime,
|
|
CVOptionFlags,
|
|
CVOptionFlags*,
|
|
void *displayLinkContext)
|
|
{
|
|
VTRenderer* me = reinterpret_cast<VTRenderer*>(displayLinkContext);
|
|
|
|
me->drawFrame(vsyncTime->hostTime);
|
|
|
|
return kCVReturnSuccess;
|
|
}
|
|
|
|
virtual bool initialize(SDL_Window* window,
|
|
int videoFormat,
|
|
int,
|
|
int) override
|
|
{
|
|
int err;
|
|
|
|
if (videoFormat & VIDEO_FORMAT_MASK_H264) {
|
|
// Prior to 10.13, we'll just assume everything has
|
|
// H.264 support and fail open to allow VT decode.
|
|
#if __MAC_OS_X_VERSION_MAX_ALLOWED >= 101300
|
|
if (__builtin_available(macOS 10.13, *)) {
|
|
if (!VTIsHardwareDecodeSupported(kCMVideoCodecType_H264)) {
|
|
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION,
|
|
"No HW accelerated H.264 decode via VT");
|
|
return false;
|
|
}
|
|
}
|
|
else
|
|
#endif
|
|
{
|
|
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION,
|
|
"Assuming H.264 HW decode on < macOS 10.13");
|
|
}
|
|
}
|
|
else if (videoFormat & VIDEO_FORMAT_MASK_H265) {
|
|
#if __MAC_OS_X_VERSION_MAX_ALLOWED >= 101300
|
|
if (__builtin_available(macOS 10.13, *)) {
|
|
if (!VTIsHardwareDecodeSupported(kCMVideoCodecType_HEVC)) {
|
|
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION,
|
|
"No HW accelerated HEVC decode via VT");
|
|
return false;
|
|
}
|
|
}
|
|
else
|
|
#endif
|
|
{
|
|
// Fail closed for HEVC if we're not on 10.13+
|
|
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION,
|
|
"No HEVC support on < macOS 10.13");
|
|
return false;
|
|
}
|
|
}
|
|
|
|
SDL_SysWMinfo info;
|
|
|
|
SDL_VERSION(&info.version);
|
|
|
|
if (!SDL_GetWindowWMInfo(window, &info)) {
|
|
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION,
|
|
"SDL_GetWindowWMInfo() failed: %s",
|
|
SDL_GetError());
|
|
return false;
|
|
}
|
|
|
|
SDL_assert(info.subsystem == SDL_SYSWM_COCOA);
|
|
|
|
// SDL adds its own content view to listen for events.
|
|
// We need to add a subview for our display layer.
|
|
NSView* contentView = info.info.cocoa.window.contentView;
|
|
m_View = [[NSView alloc] initWithFrame:contentView.bounds];
|
|
m_View.wantsLayer = YES;
|
|
[contentView addSubview: m_View];
|
|
|
|
setupDisplayLayer();
|
|
|
|
err = av_hwdevice_ctx_create(&m_HwContext,
|
|
AV_HWDEVICE_TYPE_VIDEOTOOLBOX,
|
|
nullptr,
|
|
nullptr,
|
|
0);
|
|
if (err < 0) {
|
|
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION,
|
|
"av_hwdevice_ctx_create() failed for VT decoder: %d",
|
|
err);
|
|
return false;
|
|
}
|
|
|
|
CVDisplayLinkCreateWithActiveCGDisplays(&m_DisplayLink);
|
|
CVDisplayLinkSetOutputCallback(m_DisplayLink, displayLinkOutputCallback, this);
|
|
CVDisplayLinkStart(m_DisplayLink);
|
|
|
|
return true;
|
|
}
|
|
|
|
virtual bool prepareDecoderContext(AVCodecContext* context) override
|
|
{
|
|
context->hw_device_ctx = av_buffer_ref(m_HwContext);
|
|
return true;
|
|
}
|
|
|
|
virtual void renderFrame(AVFrame* frame) override
|
|
{
|
|
if (m_DisplayLayer.status == AVQueuedSampleBufferRenderingStatusFailed) {
|
|
SDL_LogWarn(SDL_LOG_CATEGORY_APPLICATION,
|
|
"Resetting failed AVSampleBufferDisplay layer");
|
|
setupDisplayLayer();
|
|
}
|
|
|
|
SDL_AtomicLock(&m_FrameQueueLock);
|
|
m_FrameQueue.enqueue(frame);
|
|
SDL_AtomicUnlock(&m_FrameQueueLock);
|
|
}
|
|
|
|
private:
|
|
void setupDisplayLayer()
|
|
{
|
|
CALayer* oldLayer = m_DisplayLayer;
|
|
|
|
m_DisplayLayer = [[AVSampleBufferDisplayLayer alloc] init];
|
|
m_DisplayLayer.bounds = m_View.bounds;
|
|
m_DisplayLayer.position = CGPointMake(CGRectGetMidX(m_View.bounds), CGRectGetMidY(m_View.bounds));
|
|
m_DisplayLayer.videoGravity = AVLayerVideoGravityResizeAspect;
|
|
|
|
CALayer* viewLayer = m_View.layer;
|
|
if (oldLayer != nil) {
|
|
[viewLayer replaceSublayer:oldLayer with:m_DisplayLayer];
|
|
}
|
|
else {
|
|
[viewLayer addSublayer:m_DisplayLayer];
|
|
}
|
|
}
|
|
|
|
AVBufferRef* m_HwContext;
|
|
AVSampleBufferDisplayLayer* m_DisplayLayer;
|
|
CMVideoFormatDescriptionRef m_FormatDesc;
|
|
NSView* m_View;
|
|
CVDisplayLinkRef m_DisplayLink;
|
|
QQueue<AVFrame*> m_FrameQueue;
|
|
QQueue<int> m_FrameQueueHistory;
|
|
SDL_SpinLock m_FrameQueueLock;
|
|
};
|
|
|
|
IFFmpegRenderer* VTRendererFactory::createRenderer() {
|
|
return new VTRenderer();
|
|
}
|