using Ryujinx.Common.Memory; using Ryujinx.Graphics.Nvdec.Vp9.Common; using Ryujinx.Graphics.Nvdec.Vp9.Types; using Ryujinx.Graphics.Video; using System; namespace Ryujinx.Graphics.Nvdec.Vp9 { public sealed class Decoder : IVp9Decoder { public bool IsHardwareAccelerated => false; private readonly MemoryAllocator _allocator = new(); public ISurface CreateSurface(int width, int height) { return new Surface(width, height); } private static ReadOnlySpan LiteralToFilter => [ Constants.EightTapSmooth, Constants.EightTap, Constants.EightTapSharp, Constants.Bilinear ]; public unsafe bool Decode( ref Vp9PictureInfo pictureInfo, ISurface output, ReadOnlySpan bitstream, ReadOnlySpan mvsIn, Span mvsOut) { Vp9Common cm = new() { FrameType = pictureInfo.IsKeyFrame ? FrameType.KeyFrame : FrameType.InterFrame, IntraOnly = pictureInfo.IntraOnly, Width = output.Width, Height = output.Height, SubsamplingX = 1, SubsamplingY = 1, UsePrevFrameMvs = pictureInfo.UsePrevInFindMvRefs, RefFrameSignBias = pictureInfo.RefFrameSignBias, BaseQindex = pictureInfo.BaseQIndex, YDcDeltaQ = pictureInfo.YDcDeltaQ, UvAcDeltaQ = pictureInfo.UvAcDeltaQ, UvDcDeltaQ = pictureInfo.UvDcDeltaQ, TxMode = (TxMode)pictureInfo.TransformMode, AllowHighPrecisionMv = pictureInfo.AllowHighPrecisionMv, InterpFilter = (byte)pictureInfo.InterpFilter, ReferenceMode = (ReferenceMode)pictureInfo.ReferenceMode, CompFixedRef = pictureInfo.CompFixedRef, CompVarRef = pictureInfo.CompVarRef, BitDepth = BitDepth.Bits8, Log2TileCols = pictureInfo.Log2TileCols, Log2TileRows = pictureInfo.Log2TileRows, Fc = new Ptr(ref pictureInfo.Entropy), Counts = new Ptr(ref pictureInfo.BackwardUpdateCounts) }; cm.Mb.Lossless = pictureInfo.Lossless; cm.Mb.Bd = 8; if (cm.InterpFilter != Constants.Switchable) { cm.InterpFilter = LiteralToFilter[cm.InterpFilter]; } cm.Seg.Enabled = pictureInfo.SegmentEnabled; cm.Seg.UpdateMap = pictureInfo.SegmentMapUpdate; cm.Seg.TemporalUpdate = pictureInfo.SegmentMapTemporalUpdate; cm.Seg.AbsDelta = (byte)pictureInfo.SegmentAbsDelta; cm.Seg.FeatureMask = pictureInfo.SegmentFeatureEnable; cm.Seg.FeatureData = pictureInfo.SegmentFeatureData; cm.Lf.FilterLevel = pictureInfo.LoopFilterLevel; cm.Lf.SharpnessLevel = pictureInfo.LoopFilterSharpnessLevel; cm.Lf.ModeRefDeltaEnabled = pictureInfo.ModeRefDeltaEnabled; cm.Lf.RefDeltas = pictureInfo.RefDeltas; cm.Lf.ModeDeltas = pictureInfo.ModeDeltas; cm.FrameRefs[0].Buf = (Surface)pictureInfo.LastReference; cm.FrameRefs[1].Buf = (Surface)pictureInfo.GoldenReference; cm.FrameRefs[2].Buf = (Surface)pictureInfo.AltReference; cm.Mb.CurBuf = (Surface)output; cm.Mb.SetupBlockPlanes(1, 1); int tileCols = 1 << pictureInfo.Log2TileCols; int tileRows = 1 << pictureInfo.Log2TileRows; // Video usually have only 4 columns, so more threads won't make a difference for those. // Try to not take all CPU cores for video decoding. int maxThreads = Math.Min(4, Environment.ProcessorCount / 2); cm.AllocTileWorkerData(_allocator, tileCols, tileRows, maxThreads); cm.AllocContextBuffers(_allocator, output.Width, output.Height); cm.InitContextBuffers(); cm.SetupSegmentationDequant(); cm.SetupScaleFactors(); cm.SetMvs(mvsIn); if (cm.Lf.FilterLevel != 0 && cm.SkipLoopFilter == 0) { LoopFilter.LoopFilterFrameInit(ref cm, cm.Lf.FilterLevel); } fixed (byte* dataPtr = bitstream) { try { if (maxThreads > 1 && tileRows == 1 && tileCols > 1) { DecodeFrame.DecodeTilesMt(ref cm, new ArrayPtr(dataPtr, bitstream.Length), maxThreads); LoopFilter.LoopFilterFrameMt( ref cm.Mb.CurBuf, ref cm, ref cm.Mb, cm.Lf.FilterLevel, false, false, maxThreads); } else { DecodeFrame.DecodeTiles(ref cm, new ArrayPtr(dataPtr, bitstream.Length)); LoopFilter.LoopFilterFrame( ref cm.Mb.CurBuf, ref cm, ref cm.Mb, cm.Lf.FilterLevel, false, false); } } catch (InternalErrorException) { return false; } } cm.GetMvs(mvsOut); cm.FreeTileWorkerData(_allocator); cm.FreeContextBuffers(_allocator); return true; } public void Dispose() { _allocator.Dispose(); } } }