Files
sharpcompress/tests/SharpCompress.Test/Zip/Zip64AsyncTests.cs
2026-01-15 15:20:52 +00:00

249 lines
7.8 KiB
C#

using System;
using System.IO;
using System.Linq;
using System.Threading.Tasks;
using SharpCompress.Archives;
using SharpCompress.Common;
using SharpCompress.Common.Zip;
using SharpCompress.Compressors.Deflate;
using SharpCompress.Readers;
using SharpCompress.Readers.Zip;
using SharpCompress.Test.Mocks;
using SharpCompress.Writers;
using SharpCompress.Writers.Zip;
using Xunit;
namespace SharpCompress.Test.Zip;
public class Zip64AsyncTests : WriterTests
{
public Zip64AsyncTests()
: base(ArchiveType.Zip) { }
// 4GiB + 1
private const long FOUR_GB_LIMIT = ((long)uint.MaxValue) + 1;
//[Fact]
[Trait("format", "zip64")]
public async ValueTask Zip64_Single_Large_File_Async() =>
await RunSingleTestAsync(1, FOUR_GB_LIMIT, setZip64: true, forwardOnly: false);
//[Fact]
[Trait("format", "zip64")]
public async ValueTask Zip64_Two_Large_Files_Async() =>
await RunSingleTestAsync(2, FOUR_GB_LIMIT, setZip64: true, forwardOnly: false);
//[Fact]
[Trait("format", "zip64")]
public async ValueTask Zip64_Two_Small_files_Async() =>
// Multiple files, does not require zip64
await RunSingleTestAsync(2, FOUR_GB_LIMIT / 2, setZip64: false, forwardOnly: false);
// [Fact]
[Trait("format", "zip64")]
public async ValueTask Zip64_Two_Small_files_stream_Async() =>
await RunSingleTestAsync(2, FOUR_GB_LIMIT / 2, setZip64: false, forwardOnly: true);
// [Fact]
[Trait("format", "zip64")]
public async ValueTask Zip64_Two_Small_Files_Zip64_Async() =>
// Multiple files, use zip64 even though it is not required
await RunSingleTestAsync(2, FOUR_GB_LIMIT / 2, setZip64: true, forwardOnly: false);
// [Fact]
[Trait("format", "zip64")]
public async ValueTask Zip64_Single_Large_File_Fail_Async()
{
try
{
// One single file, should fail
await RunSingleTestAsync(1, FOUR_GB_LIMIT, setZip64: false, forwardOnly: false);
throw new InvalidOperationException("Test did not fail?");
}
catch (NotSupportedException) { }
}
// [Fact]
[Trait("zip64", "true")]
public async ValueTask Zip64_Single_Large_File_Zip64_Streaming_Fail_Async()
{
try
{
// One single file, should fail (fast) with zip64
await RunSingleTestAsync(1, FOUR_GB_LIMIT, setZip64: true, forwardOnly: true);
throw new InvalidOperationException("Test did not fail?");
}
catch (NotSupportedException) { }
}
// [Fact]
[Trait("zip64", "true")]
public async ValueTask Zip64_Single_Large_File_Streaming_Fail_Async()
{
try
{
// One single file, should fail once the write discovers the problem
await RunSingleTestAsync(1, FOUR_GB_LIMIT, setZip64: false, forwardOnly: true);
throw new InvalidOperationException("Test did not fail?");
}
catch (NotSupportedException) { }
}
public async ValueTask RunSingleTestAsync(
long files,
long filesize,
bool setZip64,
bool forwardOnly,
long writeChunkSize = 1024 * 1024,
string filename = "zip64-test-async.zip"
)
{
filename = Path.Combine(SCRATCH2_FILES_PATH, filename);
try
{
if (File.Exists(filename))
{
File.Delete(filename);
}
if (!File.Exists(filename))
{
await CreateZipArchiveAsync(
filename,
files,
filesize,
writeChunkSize,
setZip64,
forwardOnly
);
}
var resForward = await ReadForwardOnlyAsync(filename);
if (resForward.Item1 != files)
{
throw new InvalidOperationException(
$"Incorrect number of items reported: {resForward.Item1}, should have been {files}"
);
}
if (resForward.Item2 != files * filesize)
{
throw new InvalidOperationException(
$"Incorrect combined size reported: {resForward.Item2}, should have been {files * filesize}"
);
}
var resArchive = ReadArchive(filename);
if (resArchive.Item1 != files)
{
throw new InvalidOperationException(
$"Incorrect number of items reported: {resArchive.Item1}, should have been {files}"
);
}
if (resArchive.Item2 != files * filesize)
{
throw new InvalidOperationException(
$"Incorrect number of items reported: {resArchive.Item2}, should have been {files * filesize}"
);
}
}
finally
{
if (File.Exists(filename))
{
File.Delete(filename);
}
}
}
public async ValueTask CreateZipArchiveAsync(
string filename,
long files,
long filesize,
long chunksize,
bool setZip64,
bool forwardOnly
)
{
var data = new byte[chunksize];
// Use deflate for speed
var opts = new ZipWriterOptions(CompressionType.Deflate) { UseZip64 = setZip64 };
// Use no compression to ensure we hit the limits (actually inflates a bit, but seems better than using method==Store)
var eo = new ZipWriterEntryOptions { DeflateCompressionLevel = CompressionLevel.None };
using var zip = File.OpenWrite(filename);
using var st = forwardOnly ? (Stream)new ForwardOnlyStream(zip) : zip;
using var zipWriter = (ZipWriter)WriterFactory.OpenWriter(st, ArchiveType.Zip, opts);
for (var i = 0; i < files; i++)
{
using var str = zipWriter.WriteToStream(i.ToString(), eo);
var left = filesize;
while (left > 0)
{
var b = (int)Math.Min(left, data.Length);
// Use synchronous Write to match the sync version and avoid ForwardOnlyStream issues
await str.WriteAsync(data, 0, b);
left -= b;
}
}
}
public async ValueTask<Tuple<long, long>> ReadForwardOnlyAsync(string filename)
{
long count = 0;
long size = 0;
ZipEntry? prev = null;
using (var fs = File.OpenRead(filename))
{
var rd = ReaderFactory.OpenAsyncReader(
new AsyncOnlyStream(fs),
new ReaderOptions { LookForHeader = false }
);
await using (rd)
{
while (await rd.MoveToNextEntryAsync())
{
#if NETFRAMEWORK || NETSTANDARD2_0
using (var entryStream = await rd.OpenEntryStreamAsync())
{
await entryStream.SkipEntryAsync();
}
#else
await using (var entryStream = await rd.OpenEntryStreamAsync())
{
await entryStream.SkipEntryAsync();
}
#endif
count++;
if (prev != null)
{
size += prev.Size;
}
prev = (ZipEntry)rd.Entry;
}
}
}
if (prev != null)
{
size += prev.Size;
}
return new Tuple<long, long>(count, size);
}
public Tuple<long, long> ReadArchive(string filename)
{
using var archive = ArchiveFactory.OpenArchive(filename);
return new Tuple<long, long>(
archive.Entries.Count(),
archive.Entries.Select(x => x.Size).Sum()
);
}
}