using Ryujinx.Horizon.Common; using Ryujinx.Memory; using Ryujinx.Memory.Range; using System; using System.Buffers; using System.Collections.Generic; using System.Diagnostics; namespace Ryujinx.HLE.HOS.Kernel.Memory { class KPageTable : KPageTableBase { private readonly IVirtualMemoryManager _cpuMemory; protected override bool UsesPrivateAllocations => _cpuMemory.UsesPrivateAllocations; public KPageTable(KernelContext context, IVirtualMemoryManager cpuMemory, ulong reservedAddressSpaceSize) : base(context, reservedAddressSpaceSize) { _cpuMemory = cpuMemory; } /// protected override IEnumerable GetHostRegions(ulong va, ulong size) { return _cpuMemory.GetHostRegions(va, size); } /// protected override void GetPhysicalRegions(ulong va, ulong size, KPageList pageList) { var ranges = _cpuMemory.GetPhysicalRegions(va, size); foreach (var range in ranges) { pageList.AddRange(range.Address + DramMemoryMap.DramBase, range.Size / PageSize); } } /// protected override ReadOnlySequence GetReadOnlySequence(ulong va, int size) { return _cpuMemory.GetReadOnlySequence(va, size); } /// protected override ReadOnlySpan GetSpan(ulong va, int size) { return _cpuMemory.GetSpan(va, size); } /// protected override Result MapMemory(ulong src, ulong dst, ulong pagesCount, KMemoryPermission oldSrcPermission, KMemoryPermission newDstPermission) { KPageList pageList = new(); GetPhysicalRegions(src, pagesCount * PageSize, pageList); Result result = Reprotect(src, pagesCount, KMemoryPermission.None); if (result != Result.Success) { return result; } result = MapPages(dst, pageList, newDstPermission, MemoryMapFlags.Private, false, 0); if (result != Result.Success) { Result reprotectResult = Reprotect(src, pagesCount, oldSrcPermission); Debug.Assert(reprotectResult == Result.Success); } return result; } /// protected override Result UnmapMemory(ulong dst, ulong src, ulong pagesCount, KMemoryPermission oldDstPermission, KMemoryPermission newSrcPermission) { ulong size = pagesCount * PageSize; KPageList srcPageList = new(); KPageList dstPageList = new(); GetPhysicalRegions(src, size, srcPageList); GetPhysicalRegions(dst, size, dstPageList); if (!dstPageList.IsEqual(srcPageList)) { return KernelResult.InvalidMemRange; } Result result = Unmap(dst, pagesCount); if (result != Result.Success) { return result; } result = Reprotect(src, pagesCount, newSrcPermission); if (result != Result.Success) { Result mapResult = MapPages(dst, dstPageList, oldDstPermission, MemoryMapFlags.Private, false, 0); Debug.Assert(mapResult == Result.Success); } return result; } /// protected override Result MapPages( ulong dstVa, ulong pagesCount, ulong srcPa, KMemoryPermission permission, MemoryMapFlags flags, bool shouldFillPages, byte fillValue) { ulong size = pagesCount * PageSize; Context.CommitMemory(srcPa - DramMemoryMap.DramBase, size); _cpuMemory.Map(dstVa, srcPa - DramMemoryMap.DramBase, size, flags); if (DramMemoryMap.IsHeapPhysicalAddress(srcPa)) { Context.MemoryManager.IncrementPagesReferenceCount(srcPa, pagesCount); } if (shouldFillPages) { _cpuMemory.Fill(dstVa, size, fillValue); } return Result.Success; } /// protected override Result MapPages( ulong address, KPageList pageList, KMemoryPermission permission, MemoryMapFlags flags, bool shouldFillPages, byte fillValue) { using var scopedPageList = new KScopedPageList(Context.MemoryManager, pageList); ulong currentVa = address; foreach (var pageNode in pageList) { ulong addr = pageNode.Address - DramMemoryMap.DramBase; ulong size = pageNode.PagesCount * PageSize; Context.CommitMemory(addr, size); _cpuMemory.Map(currentVa, addr, size, flags); if (shouldFillPages) { _cpuMemory.Fill(currentVa, size, fillValue); } currentVa += size; } scopedPageList.SignalSuccess(); return Result.Success; } /// protected override Result MapForeign(IEnumerable regions, ulong va, ulong size) { ulong backingStart = (ulong)Context.Memory.Pointer; ulong backingEnd = backingStart + Context.Memory.Size; KPageList pageList = new(); foreach (HostMemoryRange region in regions) { // If the range is inside the physical memory, it is shared and we should increment the page count, // otherwise it is private and we don't need to increment the page count. if (region.Address >= backingStart && region.Address < backingEnd) { pageList.AddRange(region.Address - backingStart + DramMemoryMap.DramBase, region.Size / PageSize); } } using var scopedPageList = new KScopedPageList(Context.MemoryManager, pageList); foreach (var pageNode in pageList) { Context.CommitMemory(pageNode.Address - DramMemoryMap.DramBase, pageNode.PagesCount * PageSize); } ulong offset = 0; foreach (var region in regions) { _cpuMemory.MapForeign(va + offset, region.Address, region.Size); offset += region.Size; } scopedPageList.SignalSuccess(); return Result.Success; } /// protected override Result Unmap(ulong address, ulong pagesCount) { KPageList pagesToClose = new(); var regions = _cpuMemory.GetPhysicalRegions(address, pagesCount * PageSize); foreach (var region in regions) { ulong pa = region.Address + DramMemoryMap.DramBase; if (DramMemoryMap.IsHeapPhysicalAddress(pa)) { pagesToClose.AddRange(pa, region.Size / PageSize); } } _cpuMemory.Unmap(address, pagesCount * PageSize); pagesToClose.DecrementPagesReferenceCount(Context.MemoryManager); return Result.Success; } /// protected override Result Reprotect(ulong address, ulong pagesCount, KMemoryPermission permission) { _cpuMemory.Reprotect(address, pagesCount * PageSize, permission.Convert()); return Result.Success; } /// protected override Result ReprotectAndFlush(ulong address, ulong pagesCount, KMemoryPermission permission) { // TODO: Flush JIT cache. return Reprotect(address, pagesCount, permission); } /// protected override void SignalMemoryTracking(ulong va, ulong size, bool write) { _cpuMemory.SignalMemoryTracking(va, size, write); } /// protected override void Write(ulong va, ReadOnlySequence data) { _cpuMemory.Write(va, data); } /// protected override void Write(ulong va, ReadOnlySpan data) { _cpuMemory.Write(va, data); } } }