Files
Matt Nadareski 7689c6dd07 Libraries
This change looks dramatic, but it's just separating out the already-split namespaces into separate top-level folders. In theory, every single one could be built into their own Nuget package. `SabreTools.Serialization` still builds the normal Nuget package that is used by all other projects and includes all namespaces.
2026-03-21 16:26:56 -04:00

251 lines
8.3 KiB
C#

using System.Collections.Generic;
using System.IO;
using SabreTools.Data.Models.GCF;
namespace SabreTools.Wrappers
{
public partial class GCF : WrapperBase<Data.Models.GCF.File>
{
#region Descriptive Properties
/// <inheritdoc/>
public override string DescriptionString => "Half-Life Game Cache File (GCF)";
#endregion
#region Extension Properties
/// <inheritdoc cref="Models.GCF.DataBlockHeader.BlockSize"/>
public uint BlockSize => Model.DataBlockHeader.BlockSize;
/// <summary>
/// Set of all data block offsets
/// </summary>
public long[]? DataBlockOffsets
{
get
{
// Use the cached value, if it exists
if (field is not null)
return field;
// Otherwise, build the data block set
field = new long[Model.DataBlockHeader.BlockCount];
for (int i = 0; i < Model.DataBlockHeader.BlockCount; i++)
{
long dataBlockOffset = Model.DataBlockHeader.FirstBlockOffset + (i * Model.DataBlockHeader.BlockSize);
field[i] = dataBlockOffset;
}
// Return the set of data blocks
return field;
}
} = null;
/// <summary>
/// Set of all files and their information
/// </summary>
public FileInfo[]? Files
{
get
{
// Use the cached value, if it exists
if (field is not null)
return field;
// Otherwise, scan and build the files
var files = new List<FileInfo>();
for (int i = 0; i < Model.DirectoryEntries.Length; i++)
{
// Get the directory entry
var directoryEntry = Model.DirectoryEntries[i];
var directoryMapEntry = Model.DirectoryMapEntries[i];
// If we have a directory, skip for now
#if NET20 || NET35
if ((directoryEntry.DirectoryFlags & HL_GCF_FLAG.HL_GCF_FLAG_FILE) == 0)
#else
if (!directoryEntry.DirectoryFlags.HasFlag(HL_GCF_FLAG.HL_GCF_FLAG_FILE))
#endif
continue;
// Otherwise, start building the file info
var fileInfo = new FileInfo()
{
Size = directoryEntry.ItemSize,
#if NET20 || NET35
Encrypted = (directoryEntry.DirectoryFlags & HL_GCF_FLAG.HL_GCF_FLAG_ENCRYPTED) != 0,
#else
Encrypted = directoryEntry.DirectoryFlags.HasFlag(HL_GCF_FLAG.HL_GCF_FLAG_ENCRYPTED),
#endif
};
var pathParts = new List<string> { Model.DirectoryNames[directoryEntry.NameOffset] ?? string.Empty };
var blockEntries = new List<BlockEntry>();
// Traverse the parent tree
uint index = directoryEntry.ParentIndex;
while (index != 0xFFFFFFFF)
{
var parentDirectoryEntry = Model.DirectoryEntries[index];
pathParts.Add(Model.DirectoryNames[parentDirectoryEntry.NameOffset] ?? string.Empty);
index = parentDirectoryEntry.ParentIndex;
}
// Traverse the block entries
index = directoryMapEntry.FirstBlockIndex;
while (index != Model.DataBlockHeader.BlockCount)
{
var nextBlock = Model.BlockEntries[index];
blockEntries.Add(nextBlock);
index = nextBlock.NextBlockEntryIndex;
}
// Reverse the path parts because of traversal
pathParts.Reverse();
// Build the remaining file info
#if NET20 || NET35
string[] pathArray = [.. pathParts];
string tempPath = string.Empty;
if (pathArray.Length == 0 || pathArray.Length == 1)
{
tempPath = pathArray[0];
}
else
{
for (int j = 0; j < pathArray.Length; j++)
{
if (j == 0)
tempPath = pathArray[j];
else
tempPath = Path.Combine(tempPath, pathArray[j]);
}
}
fileInfo.Path = tempPath;
#else
fileInfo.Path = Path.Combine([.. pathParts]);
#endif
fileInfo.BlockEntries = [.. blockEntries];
// Add the file info and continue
files.Add(fileInfo);
}
// Set and return the file infos
field = [.. files];
return field;
}
} = null;
#endregion
#region Constructors
/// <inheritdoc/>
public GCF(Data.Models.GCF.File model, byte[] data) : base(model, data) { }
/// <inheritdoc/>
public GCF(Data.Models.GCF.File model, byte[] data, int offset) : base(model, data, offset) { }
/// <inheritdoc/>
public GCF(Data.Models.GCF.File model, byte[] data, int offset, int length) : base(model, data, offset, length) { }
/// <inheritdoc/>
public GCF(Data.Models.GCF.File model, Stream data) : base(model, data) { }
/// <inheritdoc/>
public GCF(Data.Models.GCF.File model, Stream data, long offset) : base(model, data, offset) { }
/// <inheritdoc/>
public GCF(Data.Models.GCF.File model, Stream data, long offset, long length) : base(model, data, offset, length) { }
#endregion
#region Static Constructors
/// <summary>
/// Create an GCF from a byte array and offset
/// </summary>
/// <param name="data">Byte array representing the GCF</param>
/// <param name="offset">Offset within the array to parse</param>
/// <returns>An GCF wrapper on success, null on failure</returns>
public static GCF? Create(byte[]? data, int offset)
{
// If the data is invalid
if (data is null || data.Length == 0)
return null;
// If the offset is out of bounds
if (offset < 0 || offset >= data.Length)
return null;
// Create a memory stream and use that
var dataStream = new MemoryStream(data, offset, data.Length - offset);
return Create(dataStream);
}
/// <summary>
/// Create a GCF from a Stream
/// </summary>
/// <param name="data">Stream representing the GCF</param>
/// <returns>An GCF wrapper on success, null on failure</returns>
public static GCF? Create(Stream? data)
{
// If the data is invalid
if (data is null || !data.CanRead)
return null;
try
{
// Cache the current offset
long currentOffset = data.Position;
var model = new Serialization.Readers.GCF().Deserialize(data);
if (model is null)
return null;
return new GCF(model, data, currentOffset);
}
catch
{
return null;
}
}
#endregion
#region Helper Classes
/// <summary>
/// Class to contain all necessary file information
/// </summary>
public sealed class FileInfo
{
/// <summary>
/// Full item path
/// </summary>
public string? Path;
/// <summary>
/// File size
/// </summary>
public uint Size;
/// <summary>
/// Indicates if the block is encrypted
/// </summary>
public bool Encrypted;
/// <summary>
/// Array of block entries
/// </summary>
public BlockEntry[] BlockEntries = [];
}
#endregion
}
}