1
0
mirror of synced 2024-12-01 02:27:22 +01:00
Switch-Toolbox/File_Format_Library/FileFormats/Archives/SDF.cs
KillzXGaming 0c126e4155 More improvements.
Rewrote the compression handling from scatch. It's way easier and cleaner to add new formats code wise as it's handled like file formats.
Added wip TVOL support (Touhou Azure Reflections)
Added XCI support. Note I plan to improve NSP, XCI, NCA, etc later for exefs exporting.
The compression rework now compresses via streams, so files get decompressed properly within archives as streams.
Added hyrule warriors bin.gz compression along with archive rebuilding. Note i do not have texture rebuilding done just yet.
2019-09-15 19:13:01 -04:00

586 lines
22 KiB
C#

using System;
using System.Collections.Generic;
using System.IO;
using System.Linq;
using System.Threading.Tasks;
using Toolbox;
using System.Windows.Forms;
using Toolbox.Library;
using Toolbox.Library.IO;
using Toolbox.Library.Forms;
namespace FirstPlugin
{
public class SDF : IArchiveFile, IFileFormat
{
public FileType FileType { get; set; } = FileType.Archive;
public bool CanSave { get; set; }
public string[] Description { get; set; } = new string[] { "Snowdrop Engine Data Table Of Contents" };
public string[] Extension { get; set; } = new string[] { "*.sdftoc" };
public string FileName { get; set; }
public string FilePath { get; set; }
public IFileInfo IFileInfo { get; set; }
public bool CanAddFiles { get; set; } = false;
public bool CanRenameFiles { get; set; } = false;
public bool CanReplaceFiles { get; set; } = false;
public bool CanDeleteFiles { get; set; } = false;
public bool Identify(System.IO.Stream stream)
{
using (var reader = new Toolbox.Library.IO.FileReader(stream, true))
{
return reader.CheckSignature(4, "WEST");
}
}
public Type[] Types
{
get
{
List<Type> types = new List<Type>();
return types.ToArray();
}
}
public List<FileEntry> files = new List<FileEntry>();
public IEnumerable<ArchiveFileInfo> Files => files;
public void ClearFiles() { files.Clear(); }
SDFTOC_Header Header;
public SDFTOC_Block2[] block2Array;
//Thanks to https://github.com/GoldFarmer/rouge_sdf/blob/master/main.cpp for docs/structs
public void Load(System.IO.Stream stream)
{
using (var reader = new FileReader(stream))
{
reader.ByteOrder = Syroot.BinaryData.ByteOrder.LittleEndian;
//Read header
Header = new SDFTOC_Header();
Header.Read(reader);
//Read first id
var startId = new SDFTOC_ID(reader);
//Check this flag
byte Flag1 = reader.ReadByte();
if (Flag1 != 0)
{
byte[] unk = reader.ReadBytes(0x140);
}
if (Header.DataOffset != 0)
{
reader.SeekBegin(Header.DataOffset + 0x51);
//Here is the compressed block. Check the magic first
uint magic = reader.ReadUInt32();
reader.Seek(-4, SeekOrigin.Current);
//Read and decompress the compressed block
//Contains file names and block info
DecompressNameBlock(magic, reader.ReadBytes((int)Header.CompressedSize), Header);
//Read last id
var endId = new SDFTOC_ID(reader);
}
else
{
//Read first block
var block1 = reader.ReadInt32s((int)Header.Block1Count);
//Read ID blocks
var blockIds = new SDFTOC_ID[Header.Block1Count];
for (int i = 0; i < Header.Block1Count; i++)
{
blockIds[i] = new SDFTOC_ID(reader);
}
//Read block 2 (DDS headers)
block2Array = new SDFTOC_Block2[Header.Block2Count];
for (int i = 0; i < Header.Block2Count; i++)
{
block2Array[i] = new SDFTOC_Block2(reader, Header);
}
//Here is the compressed block. Check the magic first
uint magic = reader.ReadUInt32();
reader.Seek(-4, SeekOrigin.Current);
//Read and decompress the compressed block
//Contains file names and block info
DecompressNameBlock(magic, reader.ReadBytes((int)Header.CompressedSize), Header);
//Read last id
var endId = new SDFTOC_ID(reader);
}
Dictionary<string, int> Extensions = new Dictionary<string, int>();
for (int i = 0; i < FileEntries.Count; i++)
{
string ext = Utils.GetExtension(FileEntries[i].FileName);
if (!Extensions.ContainsKey(ext))
Extensions.Add(ext, 1);
else
Extensions[ext]++;
}
for (int i = 0; i < FileEntries.Count; i++)
{
string ext = Utils.GetExtension(FileEntries[i].FileName);
if (Extensions[ext] > 10000 && ext != ".mmb")
FileEntries[i].CanLoadFile = false;
files.Add(FileEntries[i]);
}
List<string> FilteredExtensions = new List<string>();
foreach (var ext in Extensions)
{
if (ext.Value > 10000 && ext.Key != ".mmb")
FilteredExtensions.Add(ext.Key);
}
if (FilteredExtensions.Count > 0)
{
MessageBox.Show($"File extensions have a very large amount of nodes used." +
$" This will be filtered out to prevent slow booting. {ExtsToString(FilteredExtensions.ToArray())}");
}
//Remove unused data
startId = null;
FilteredExtensions.Clear();
Extensions.Clear();
// block1 = new int[0];
// blockIds = new SDFTOC_ID[0];
// endId = null;
}
}
private string ExtsToString(string[] value)
{
string output = "";
for (int i = 0; i < value.Length; i++)
output += $" ({value[i]})";
return output;
}
private List<string> FilteredExtensions = new List<string>()
{ };
private TreeNode GetNodes(TreeNode parent, string[] fileList)
{
// build a TreeNode collection from the file list
foreach (string strPath in fileList)
{
// Every time we parse a new file path, we start at the top level again
TreeNode thisParent = parent;
// split the file path into pieces at every backslash
foreach (string pathPart in strPath.Split('\\'))
{
// check if we already have a node for this
TreeNode[] tn = thisParent.Nodes.Find(pathPart, false);
if (tn.Length == 0)
{
// no node found, so add one
thisParent = thisParent.Nodes.Add(pathPart, pathPart);
}
else
{
// we already have this node, so use it as the parent of the next part of the path
thisParent = tn[0];
}
}
}
return parent;
}
public void DecompressNameBlock(uint magic, byte[] CompressedBlock, SDFTOC_Header header)
{
byte[] decomp = null;
if (magic == 0xDFF25B82 || magic == 0xFD2FB528)
decomp = Zstb.SDecompress(CompressedBlock);
else if (magic == 0x184D2204 || header.Version >= 0x17)
decomp = STLibraryCompression.Type_LZ4.Decompress(CompressedBlock, 0, CompressedBlock.Length,(int)header.DecompressedSize);
else
decomp = STLibraryCompression.ZLIB.Decompress(CompressedBlock);
//Now it's decompressed lets parse!
using (var reader = new FileReader(decomp))
{
ParseNames(reader);
}
CompressedBlock = new byte[0];
decomp = new byte[0];
decomp = null;
CompressedBlock = null;
}
private ulong readVariadicInteger(int Count, FileReader reader)
{
ulong result = 0;
for (int i = 0; i < Count; i++)
{
result |= (ulong)(reader.ReadByte()) << (i * 8);
}
return result;
}
public bool AddFile(ArchiveFileInfo archiveFileInfo)
{
return false;
}
public bool DeleteFile(ArchiveFileInfo archiveFileInfo)
{
return false;
}
public class FileEntry : ArchiveFileInfo
{
public SDF SDFParent;
public string FolderPath;
public string FileBlockPath;
public ulong PackageID;
public ulong Offset;
public ulong DecompressedSize;
public List<ulong> CompressedSizes;
public ulong DdsType;
public bool UseDDS;
public bool IsCompressed = false;
public bool IsSwizzled = false;
public override byte[] FileData
{
get { return GetFileBytes(); }
set { base.FileData = value; }
}
public override string FileSize { get { return STMath.GetFileSize((long)DecompressedSize, 4); } }
public override object DisplayProperties { get { return this; } }
public override IFileFormat OpenFile()
{
byte[] Data = FileData;
var FileFormat = STFileLoader.OpenFileFormat(new MemoryStream(Data),
IOExtensions.RemoveIllegaleFolderNameCharacters(FileName), true);
if (FileFormat is DDS)
((DDS)FileFormat).SwitchSwizzle = IsSwizzled;
return FileFormat;
}
public byte[] GetFileBytes()
{
List<byte[]> Data = new List<byte[]>();
if (File.Exists(FileBlockPath))
{
var block = File.Open(FileBlockPath, FileMode.Open);
using (var stream = new FileReader(block))
{
if (CompressedSizes.Count == 0)
{
//Decompressed File
// string FileNameBlock = Path.Combine(FolderPath, FileName);
// string FolerPath = Path.GetDirectoryName(FileNameBlock);
// if (!Directory.Exists(FolerPath))
// Directory.CreateDirectory(FolerPath);
Data.Add(stream.getSection((int)Offset, (int)DecompressedSize));
}
else
{
var PageSize = (double)0x10000;
var DecompOffset = 0;
var CompOffset = 0;
IsCompressed = true;
if (UseDDS)
{
bool IsDX10 = false;
using (var filereader = new FileReader(SDFParent.block2Array[DdsType].Data))
{
filereader.Position = 84;
IsDX10 = filereader.ReadString(4) == "DX10";
if (IsDX10)
{
if (SDFParent.block2Array[DdsType].UsedBytes > 0x94)
IsSwizzled = true;
Data.Add(SDFParent.block2Array[DdsType].Data.Take((int)0x94).ToArray());
}
else
{
if (SDFParent.block2Array[DdsType].UsedBytes > 0x80)
IsSwizzled = true;
Data.Add(SDFParent.block2Array[DdsType].Data.Take((int)0x80).ToArray());
}
}
Console.WriteLine("UsedBytes " + SDFParent.block2Array[DdsType].UsedBytes + " IsSwizzled " + IsSwizzled);
}
for (var i = 0; i < CompressedSizes.Count; i++)
{
var decompSize = (int)Math.Min((int)DecompressedSize - DecompOffset, PageSize);
if (CompressedSizes[i] == 0 || decompSize == (int)CompressedSizes[i])
{
stream.Seek((int)Offset + CompOffset, SeekOrigin.Begin);
CompressedSizes[i] = (ulong)decompSize;
Data.Add( stream.ReadBytes(decompSize));
}
else
{
stream.Seek((int)Offset + CompOffset, SeekOrigin.Begin);
Data.Add(Zstb.SDecompress(stream.ReadBytes((int)CompressedSizes[i])));
}
DecompOffset += (int)decompSize;
CompOffset += (int)CompressedSizes[i];
}
}
}
block.Dispose();
}
return Utils.CombineByteArray(Data.ToArray());
}
}
public List<FileEntry> FileEntries = new List<FileEntry>();
public void ParseNames(FileReader reader, string Name = "")
{
char ch = reader.ReadChar();
if (ch == 0)
throw new Exception("Unexcepted byte in file tree");
if (ch >= 1 && ch <= 0x1f) //string part
{
while (ch-- > 0)
{
Name += reader.ReadChar();
}
ParseNames(reader, Name);
}
else if (ch >= 'A' && ch <= 'Z') //file entry
{
int var = Convert.ToInt32(ch - 'A');
ch = Convert.ToChar(var);
int count1 = ch & 7;
int flag1 = (ch >> 3) & 1;
// int flag1 = ch & 8;
if (count1 > 0)
{
uint strangeId = reader.ReadUInt32();
byte chr2 = reader.ReadByte();
int byteCount = chr2 & 3;
int byteValue = chr2 >> 2;
ulong DdsType = readVariadicInteger(byteCount, reader);
for (int chunkIndex = 0; chunkIndex < count1; chunkIndex++)
{
byte ch3 = reader.ReadByte();
// if (ch3 == 0)
// {
// break;
// }
int compressedSizeByteCount = (ch3 & 3) + 1;
int packageOffsetByteCount = (ch3 >> 2) & 7;
bool hasCompression = ((ch3 >> 5) & 1) != 0;
ulong decompressedSize =0;
ulong compressedSize = 0;
ulong packageOffset = 0;
long fileId = -1;
if (compressedSizeByteCount > 0)
{
decompressedSize = readVariadicInteger(compressedSizeByteCount, reader);
}
if (hasCompression)
{
compressedSize = readVariadicInteger(compressedSizeByteCount, reader);
}
if (packageOffsetByteCount != 0)
{
packageOffset = readVariadicInteger(packageOffsetByteCount, reader);
}
ulong packageId = readVariadicInteger(2, reader);
if (packageId >= Header.Block1Count)
{
// throw new InvalidDataException($"SDF Package ID ({packageId}) outside of TOC range ({ Header.Block1Count})");
}
List<ulong> compSizeArray = new List<ulong>();
if (hasCompression)
{
ulong pageCount = (decompressedSize + 0xffff) >> 16;
// var pageCount = NextMultiple(decompressedSize, 0x10000) / 0x10000;
if (pageCount > 1)
{
for (ulong page = 0; page < pageCount; page++)
{
ulong compSize = readVariadicInteger(2, reader);
compSizeArray.Add(compSize);
}
}
}
if (Header.Version < 0x16) //Unsure. Rabbids doesn't use it, newer versions don't.
{
fileId = (long)readVariadicInteger(4, reader);
}
if (compSizeArray.Count == 0 && hasCompression)
compSizeArray.Add(compressedSize);
DumpFile(Name, packageId, packageOffset, decompressedSize, compSizeArray, DdsType, chunkIndex != 0, byteCount != 0 && chunkIndex == 0);
}
}
if ((ch & 8) != 0) //flag1
{
byte ch3 = reader.ReadByte();
while (ch3-- > 0)
{
reader.ReadByte();
reader.ReadByte();
}
}
}
else
{
uint offset = reader.ReadUInt32();
ParseNames(reader, Name);
reader.Seek(offset, SeekOrigin.Begin);
ParseNames(reader, Name);
}
}
public void DumpFile(string Name, ulong packageId, ulong packageOffset, ulong decompresedSize,
List<ulong> compressedSize, ulong ddsType, bool Append, bool UseDDS)
{
string PathFolder = Path.GetDirectoryName(FilePath);
string layer;
if (packageId < 1000) layer = "A";
else if (packageId < 2000) layer = "B";
else if (packageId < 3000) layer = "C";
else layer = "D";
string ID = packageId.ToString("D" + 4);
string BlockFilePath = Path.Combine(PathFolder, $"sdf-{layer}-{ID}.sdfdata");
bool IsFile = !Name.Contains("dummy") && decompresedSize > 5;
if (IsFile)
{
FileEntries.Add(new FileEntry()
{
SDFParent = this,
FileName = Name,
FileBlockPath = BlockFilePath,
FolderPath = PathFolder,
CompressedSizes = compressedSize,
DdsType = ddsType,
UseDDS = UseDDS,
DecompressedSize = decompresedSize,
PackageID = packageId,
Offset = packageOffset,
});
}
}
public void Unload()
{
}
public void Save(System.IO.Stream stream)
{
}
public class SDFTOC_Header
{
public uint Version { get; set; }
public uint DecompressedSize { get; set; }
public uint CompressedSize { get; set; }
public uint Zero { get; set; }
public uint Block1Count { get; set; }
public uint Block2Count { get; set; }
public uint DataOffset { get; set; }
public void Read(FileReader reader)
{
reader.CheckSignature(4, "WEST");
reader.Seek(4, System.IO.SeekOrigin.Begin);
Version = reader.ReadUInt32();
DecompressedSize = reader.ReadUInt32();
if (Version >= 0x17)
DataOffset = reader.ReadUInt32(); //Or might be size of some sorta block data?
CompressedSize = reader.ReadUInt32();
Zero = reader.ReadUInt32();
Block1Count = reader.ReadUInt32();
Block2Count = reader.ReadUInt32();
}
}
public class SDFTOC_ID
{
public ulong ubisoft { get; set; }
public byte[] Data { get; set; }
public ulong massive { get; set; }
public SDFTOC_ID(FileReader reader)
{
ubisoft = reader.ReadUInt64();
Data = reader.ReadBytes(0x20);
massive = reader.ReadUInt64();
}
}
public class SDFTOC_Block2 //Seems to be for DDS headers
{
public uint UsedBytes { get; set; }
public byte[] Data { get; set; }
public SDFTOC_Block2(FileReader reader, SDFTOC_Header header)
{
if (header.Version == 22)
{
UsedBytes = reader.ReadUInt32();
Data = reader.ReadBytes(0xC8);
}
else
{
UsedBytes = reader.ReadUInt32();
Data = reader.ReadBytes(0x94);
}
}
}
}
}